Giter Site home page Giter Site logo

rohan-paul / llm-finetuning-large-language-models Goto Github PK

View Code? Open in Web Editor NEW
417.0 9.0 97.0 11.63 MB

LLM (Large Language Model) FineTuning

Jupyter Notebook 98.71% Python 1.29%
gpt-3 gpt3-turbo large-language-models llama2 llm llm-finetuning llm-inference llm-serving llm-training mistral-7b

llm-finetuning-large-language-models's Introduction

LLM (Large Language Models) FineTuning Projects and notes on common practical techniques

Find me here..


Fine-tuning LLM (and YouTube Video Explanations)

Notebook ๐ŸŸ  YouTube Video
Finetune Llama-3-8B with unsloth 4bit quantized with ORPO Youtube Link
Llama-3 Finetuning on custom dataset with unsloth Youtube Link
CodeLLaMA-34B - Conversational Agent Youtube Link
Inference Yarn-Llama-2-13b-128k with KV Cache to answer quiz on very long textbook Youtube Link
Mistral 7B FineTuning with_PEFT and QLORA Youtube Link
Falcon finetuning on openassistant-guanaco Youtube Link
Fine Tuning Phi 1_5 with PEFT and QLoRA Youtube Link
Web scraping with Large Language Models (LLM)-AnthropicAI + LangChainAI Youtube Link

Fine-tuning LLM

Notebook Colab
๐Ÿ“Œ Gemma_2b_finetuning_ORPO_full_precision Open In Colab
๐Ÿ“Œ Jamba_Finetuning_Colab-Pro Open In Colab
๐Ÿ“Œ Finetune codellama-34B with QLoRA Open In Colab
๐Ÿ“Œ Mixtral Chatbot with Gradio
๐Ÿ“Œ togetherai api to run Mixtral Open In Colab
๐Ÿ“Œ Integrating TogetherAI with LangChain ๐Ÿฆ™ Open In Colab
๐Ÿ“Œ Mistral-7B-Instruct_GPTQ - Finetune on finance-alpaca dataset ๐Ÿฆ™ Open In Colab
๐Ÿ“Œ Mistral 7b FineTuning with DPO Direct_Preference_Optimization Open In Colab
๐Ÿ“Œ Finetune llama_2_GPTQ
๐Ÿ“Œ TinyLlama with Unsloth and_RoPE_Scaling dolly-15 dataset Open In Colab
๐Ÿ“Œ Tinyllama fine-tuning with Taylor_Swift Song lyrics Open In Colab

LLM Techniques and utils - Explained

LLM Concepts
๐Ÿ“Œ DPO (Direct Preference Optimization) training and its datasets
๐Ÿ“Œ 4-bit LLM Quantization with GPTQ
๐Ÿ“Œ Quantize with HF Transformers
๐Ÿ“Œ Understanding rank r in LoRA and related Matrix_Math
๐Ÿ“Œ Rotary Embeddings (RopE) is one of the Fundamental Building Blocks of LlaMA-2 Implementation
๐Ÿ“Œ Chat Templates in HuggingFace
๐Ÿ“Œ How is Mixtral 8x7B is a dense 47Bn param model
๐Ÿ“Œ The concept of validation log perplexity in LLM training - a note on fundamentals.
๐Ÿ“Œ Why we need to identify target_layers for LoRA/QLoRA
๐Ÿ“Œ Evaluate Token per sec
๐Ÿ“Œ traversing through nested attributes (or sub-modules) of a PyTorch module
๐Ÿ“Œ Implementation of Sparse Mixtures-of-Experts layer in PyTorch from Mistral Official Repo
๐Ÿ“Œ Util method to extract a specific token's representation from the last hidden states of a transformer model.
๐Ÿ“Œ Convert PyTorch model's parameters and tensors to half-precision floating-point format
๐Ÿ“Œ Quantizing ๐Ÿค— Transformers models with the GPTQ method
๐Ÿ“Œ Quantize Mixtral-8x7B so it can run in 24GB GPU
๐Ÿ“Œ What is GGML or GGUF in the world of Large Language Models ?

Other Smaller Language Models

llm-finetuning-large-language-models's People

Contributors

eltociear avatar rohan-paul avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

llm-finetuning-large-language-models's Issues

issue with model compilation

model = TFAutoModelForSequenceClassification.from_pretrained("bert-base-cased", num_labels=3)
model.compile(
optimizer=tf.keras.optimizers.Adam(learning_rate=5e-5),
loss=tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True),
metrics=tf.metrics.SparseCategoricalAccuracy(),
)
throwing this particular error:
ValueError: Could not interpret optimizer identifier: <keras.src.optimizers.adam.Adam object at 0x7f8a8336ae00>
in google colab

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.