Giter Site home page Giter Site logo

adithya-s-k / companionllm Goto Github PK

View Code? Open in Web Editor NEW
34.0 1.0 5.0 41.08 MB

CompanionLLM - A framework to finetune LLMs to be your own sentient conversational companion

License: MIT License

Jupyter Notebook 100.00%
fine-tuning huggingface llama llama2 llamacpp lora peft mit-license open-source finetuning

companionllm's Introduction

Hi ๐Ÿ‘‹, I'm Adithya S Kolavi

Typing SVG

GitHub Statistics ๐Ÿ“ƒ

My Skills

Proficient

Web Development

IoT

Arduino Firebase OpenCV

Machine Learning

Python Keras

Dev Tools and Software

Git GitLab Docker Nginx Linux
Blender Illustrator Figma Adobe XD Lightroom
WordPress Dreamweaver

Learning & Exploring

Web Development

Angular Vue.js Django Ruby on Rails Electron
.NET Redis Webpack Chart.js Invision
Gatsby Dreamweaver WooCommerce XAMPP Maria DB
Deno .Net Core Prisma

IoT

Raspberry Pi Azure AWS

Machine Learning

TensorFlow pytorch

Cloud Computing

GCP Terraform Linux Azure
Nginx XAMPP Power Bi Kubernetes

Websites I have Built

shunya website aikya website
samarpana website pcsports website
enable website courserf website
arithemania hackathon website Money mate

Badges and Certificates

An image of @savatar's Holopin badges, which is a link to view their full Holopin profile

dcc-2022-8 dcc-2022-8 dcc-2022-8 dcc-2022-8



companionllm's People

Contributors

adithya-s-k avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar

companionllm's Issues

Add Dataset Generation Notebook

Description

Create a Jupyter notebook for generating the dataset required for fine-tuning the CompanionLLama model. The dataset should be carefully curated, incorporating elements from the original Samantha dataset and additional contextual data to enhance the model's ability to emulate sentience.

Tasks

  • Develop a Jupyter notebook for dataset preparation.
  • Ensure the notebook includes instructions for generating the dataset with different approaches.
  • Document the steps and processes involved in dataset preparation.
  • Test the dataset generation notebook to verify its functionality.
  • Update the README to provide a link to the newly created notebook.

Implement Full Weights Fine-Tuning Using Axalotal

Description

Implement full weights fine-tuning of the CompanionLLama model using the Axalotal framework. This will help enhance the model's performance and adapt it to specific tasks or domains.

Tasks

  • Research and explore the Axalotal framework for fine-tuning language models.
  • Develop a workflow for implementing full weights fine-tuning using Axalotal.
  • Execute the fine-tuning process on the CompanionLLama model.
  • Assess the impact of full weights fine-tuning on the model's responses and capabilities.
  • Document the fine-tuning process and outcomes.
  • Update the README to provide information on Axalotal-based fine-tuning.

Optimize Fine-Tuning Hyperparameters

Description: Experiment with different hyperparameters and training strategies to optimize the fine-tuning process. Document your findings and suggest the best configuration for fine-tuning the model.

Skills Needed: Machine learning, Python, Git

Documentation Update

Description: Review and update the project's documentation. Ensure that it accurately reflects the current state of the project, including setup instructions, contribution guidelines, and code documentation.

Skills Needed: Technical writing, Markdown, Git

Question about Mistral_7B_qLora_Finetuning

Thanks for sharing this ๐Ÿ‘‰ https://github.com/adithya-s-k/CompanionLLM/blob/main/Mistral_7B_qLora_Finetuning.ipynb
I learn a lot from that, btw I'm not quite understand what I miss there

Issues

  1. Generated instruction, the 1-3 characters get cut off. e.g . CRE get cut off and appear ATE instead of CREATE
  2. Most of the time Generated instruction didn't match Ground truth.

Here's what I got.

The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
Setting `pad_token_id` to `eos_token_id`:2 for open-end generation.
Prompt:
<s>
Generate a SQL query to create a table containing Movie information (title, release year, genre).
Not applicable
[INST]




Generated instruction:
ATE TABLE Movies (
    id INT PRIMARY KEY,
    title VARCHAR(255),
    release_year INT,
    genre VARCHAR(255)
);  [/INST] \n CREATE TABLE Movies (
    id INT PRIMARY KEY,
    title VARCHAR(255),
    release_year INT,
    genre VARCHAR(255)
);  [/INST]

Ground truth:
CREATE TABLE Movies (
 title VARCHAR(50) NOT NULL,
 release_year INT NOT NULL,
 genre VARCHAR(20)
);

Is this normal? Any hint to improve this?

Thanks

Compare Between Avalon-Llama-7b and Avalo-Mitsral-7b

Description

Conduct a thorough comparison between the Avalon-Llama-7b model and the Avalo-Mitsral-7b model. Evaluate their performance, capabilities, and responses to assess the strengths and weaknesses of each model.

Tasks

  • Perform a detailed comparison of the Avalon-Llama-7b and Avalo-Mitsral-7b models.
  • Assess their performance in various conversational scenarios.
  • Evaluate their ability to provide meaningful responses and emulate sentience.
  • Document the findings and insights from the comparison.
  • Update the README to include a section summarizing the comparison results.

Implement Response Context Awareness

Description: Work on improving the model's responses by making it more context-aware. Enhance the response generation to consider the previous parts of the conversation for more natural and coherent replies.

Skills Needed: Natural language processing, Python, Git

Add Mitsral 7b LLM Fine-Tuning and Inference Notebooks

Description

Create Jupyter notebooks for fine-tuning the Mitsral 7b LLM using the CompanionLLama dataset and for performing inference with the fine-tuned model. These notebooks will be crucial for refining the model's responses and evaluating its performance.

Tasks

  • Develop Jupyter notebooks for fine-tuning Mitsral 7b LLM with the CompanionLLama dataset.
  • Create a notebook for performing inference with the fine-tuned Mitsral 7b LLM.
  • Ensure the notebooks are well-documented with clear instructions and explanations.
  • Test the notebooks to verify their functionality and performance.
  • Update the README to provide links to the newly created notebooks.

Add Gradio Interface for Inference

Implement a Gradio interface for the CompanionLLama model to allow users to interact with the model through a web-based interface. This can make it more accessible and user-friendly.

Skills Needed: Python, Gradio, Git

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.