LLM-FineTuning-Large-Language-Models

LLM (Large Language Model) FineTuning

Stars
452
Committers
2

LLM (Large Language Models) FineTuning Projects and notes on common practical techniques

Find me here..


Fine-tuning LLM (and YouTube Video Explanations)

Notebook ๐ŸŸ  YouTube Video
Finetune Llama-3-8B with unsloth 4bit quantized with ORPO Youtube Link
Llama-3 Finetuning on custom dataset with unsloth Youtube Link
CodeLLaMA-34B - Conversational Agent Youtube Link
Inference Yarn-Llama-2-13b-128k with KV Cache to answer quiz on very long textbook Youtube Link
Mistral 7B FineTuning with_PEFT and QLORA Youtube Link
Falcon finetuning on openassistant-guanaco Youtube Link
Fine Tuning Phi 1_5 with PEFT and QLoRA Youtube Link
Web scraping with Large Language Models (LLM)-AnthropicAI + LangChainAI Youtube Link

Fine-tuning LLM

Notebook Colab
๐Ÿ“Œ Gemma_2b_finetuning_ORPO_full_precision
๐Ÿ“Œ Jamba_Finetuning_Colab-Pro
๐Ÿ“Œ Finetune codellama-34B with QLoRA
๐Ÿ“Œ Mixtral Chatbot with Gradio
๐Ÿ“Œ togetherai api to run Mixtral
๐Ÿ“Œ Integrating TogetherAI with LangChain ๐Ÿฆ™
๐Ÿ“Œ Mistral-7B-Instruct_GPTQ - Finetune on finance-alpaca dataset ๐Ÿฆ™
๐Ÿ“Œ Mistral 7b FineTuning with DPO Direct_Preference_Optimization
๐Ÿ“Œ Finetune llama_2_GPTQ
๐Ÿ“Œ TinyLlama with Unsloth and_RoPE_Scaling dolly-15 dataset
๐Ÿ“Œ Tinyllama fine-tuning with Taylor_Swift Song lyrics

LLM Techniques and utils - Explained

LLM Concepts
๐Ÿ“Œ DPO (Direct Preference Optimization) training and its datasets
๐Ÿ“Œ 4-bit LLM Quantization with GPTQ
๐Ÿ“Œ Quantize with HF Transformers
๐Ÿ“Œ Understanding rank r in LoRA and related Matrix_Math
๐Ÿ“Œ Rotary Embeddings (RopE) is one of the Fundamental Building Blocks of LlaMA-2 Implementation
๐Ÿ“Œ Chat Templates in HuggingFace
๐Ÿ“Œ How is Mixtral 8x7B is a dense 47Bn param model
๐Ÿ“Œ The concept of validation log perplexity in LLM training - a note on fundamentals.
๐Ÿ“Œ Why we need to identify target_layers for LoRA/QLoRA
๐Ÿ“Œ Evaluate Token per sec
๐Ÿ“Œ traversing through nested attributes (or sub-modules) of a PyTorch module
๐Ÿ“Œ Implementation of Sparse Mixtures-of-Experts layer in PyTorch from Mistral Official Repo
๐Ÿ“Œ Util method to extract a specific token's representation from the last hidden states of a transformer model.
๐Ÿ“Œ Convert PyTorch model's parameters and tensors to half-precision floating-point format
๐Ÿ“Œ Quantizing ๐Ÿค— Transformers models with the GPTQ method
๐Ÿ“Œ Quantize Mixtral-8x7B so it can run in 24GB GPU
๐Ÿ“Œ What is GGML or GGUF in the world of Large Language Models ?

Other Smaller Language Models

Badges
Extracted from project README
Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link Youtube Link
Related Projects