Skip to content

Vjay15/Finetuning_with_scraps

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Fine Tuning and Inferencing

This repo contains all the python notebooks that were created in order to learn about Fine Tuning, LoRA and vLLM inferencing when you don't have any compute and Google Colab is your only friend, except for the notebook where I demonstrate OpenAI Fine Tuning for which the company I interned at gave access to me (Thank you Greatify!)

This repo explores LoRA fine-tuning using freely available tools and libraries, excluding OpenAI. All notebooks are designed for easy execution in Google Colab(T4 instance), and Python scripts require dependency installation.

The Datasets were synthetically generated using Claude and GPT by feeding them a Science and Social Paper

All of this was done as a part of my internship at Greatify.

The following libraries and frameworks are demonstrated:

  • Huggingface Transformers: For model loading and training.
  • TRL: Reinforcement Learning utilities for language models.
  • PEFT: Efficient parameter fine-tuning.
  • LLaMAFactory: Streamlined LLaMA model fine-tuning.
  • Unsloth FastLanguageModel: Fast and memory-efficient training.

Additionally, the project covers:

  • Using vLLM to infer with saved LoRA adapters and respective models.
  • Hosting a server in Google Colab and accessing the model via API, utilizing cloudflared for tunneling.

Hands-on examples are provided in notebooks for experimentation.

Resources

I am also adding the resources that I used in order to learn it so that it can be beneficial for everyone!

If you guys need any help or you got stuck do reach out to me!

Finally I want to thank my mentor at Greatify Santhosh for mentoring me and my pookie senior Kashiful Haque for helping me out by guiding me to right sources and also the Google Overlords to have given me access to the T4 GPU in 4 accounts 😉

I will make a blog and explain about the concepts here soon!

About

Study Repo for PEFT LoRA

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors