Skip to content

sivakura/LLM-FineTuning-Large-Language-Models

This branch is 58 commits behind rohan-paul/LLM-FineTuning-Large-Language-Models:main.

Folders and files

NameName
Last commit message
Last commit date

Latest commit

e48fd7a Β· Jan 15, 2024
Jan 5, 2024
Jan 11, 2024
Dec 24, 2023
Nov 27, 2023
Jan 8, 2024
Jan 3, 2024
Jan 5, 2024
Dec 5, 2023
Nov 27, 2023
Nov 27, 2023
Jan 4, 2024
Jan 4, 2024
Jan 4, 2024
Nov 27, 2023
Nov 27, 2023
Nov 27, 2023
Jan 4, 2024
Jan 4, 2024
Dec 13, 2023
Jan 4, 2024
Jan 15, 2024
Jan 9, 2024
Jan 4, 2024
Jan 4, 2024
Jan 15, 2024
Jan 4, 2024

Repository files navigation

Multiple LLM (Large Language Models) FineTuning Projects

For almost all of these I have detailed video in my YouTube Channel

Youtube Link

Find me here..


Fine-tuning LLM (and YouTube Video Explanations)

Notebook YouTube_Video
CodeLLaMA-34B - Conversational Agent Youtube Link
Inference Yarn-Llama-2-13b-128k with KV Cache to answer quiz on very long textbook Youtube Link
Mistral 7B FineTuning with_PEFT and QLORA Youtube Link
Falcon finetuning on openassistant-guanaco Youtube Link
Fine Tuning Phi 1_5 with PEFT and QLoRA Youtube Link
Web scraping with Large Language Models (LLM)-AnthropicAI + LangChainAI Youtube Link

Fine-tuning LLM

Notebook Colab
πŸ“Œ Finetune codellama-34B with QLoRA Open In Colab
πŸ“Œ Mixtral Chatbot with Gradio
πŸ“Œ togetherai api to run Mixtral Open In Colab
πŸ“Œ Integrating TogetherAI with LangChain πŸ¦™ Open In Colab
πŸ“Œ Mistral-7B-Instruct_GPTQ - Finetune on finance-alpaca dataset πŸ¦™ Open In Colab
πŸ“Œ Mistral 7b FineTuning with DPO Direct_Preference_Optimization Open In Colab
πŸ“Œ Finetune llama_2_GPTQ
πŸ“Œ TinyLlama with Unsloth and_RoPE_Scaling dolly-15 dataset Open In Colab
πŸ“Œ Tinyllama fine-tuning with Taylor_Swift Song lyrics Open In Colab

LLM Techniques and utils - Explained

LLM Concepts
πŸ“Œ DPO (Direct Preference Optimization) training and its datasets
πŸ“Œ 4-bit LLM Quantization with GPTQ
πŸ“Œ Quantize with HF Transformers
πŸ“Œ Understanding rank r in LoRA and related Matrix_Math
πŸ“Œ Rotary Embeddings (RopE) is one of the Fundamental Building Blocks of LlaMA-2 Implementation

Other Smaller Language Models

About

LLM (Large Language Model) FineTuning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 98.7%
  • Python 1.3%