Hackable implementation of state-of-the-art open-source LLMs based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
License: Apache License 2.0
Python 99.08%Jupyter Notebook 0.92%
lit-gpt's Introduction
Hi there, I'm Blake Mallory ๐
I'm a Machine Learning Engineer, Educator, Freelancer. Check out my YouTube!
๐ง I have a Bachelor of Science in Computer Engineering from the University of Cincinnati.
๐ญ Through my contract work and my Youtube channel I explore the latest in technology, typically with regards to large NLP models.
๐ฌ Ask me about : Machine learning, Deep learning, Computer vision, Natural language processing, and Cryptocurrency