Giter Site home page Giter Site logo

llm-mistral-invoice-cpu's People

Contributors

abaranovskis-redsamurai avatar ryzxxn avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

llm-mistral-invoice-cpu's Issues

How to use GPU / TPU?

Hey,

thanks so much for this project. Really amazing to see whats possible with mistral.

I happen to have a M1 Pro MacBook and was wondering if you could point me in the direction on how to use the GPU or "neural engine" to speed up the process?

Thanks so much

repo missing

the mistral-7b-instruct-v0.1.Q5_K_M.gguf repo is not awailable on huggingface

No of Tokens Exceeding Max Tokens

Hi,

First of all, thank you for this repo

right now I am trying to use Mistral Q8 on gpu and while running main.py I am getting the following error. I wanted to know how do I decide the optimal. I am attaching the config.yam for reference. Please help
Uploading Capture.PNG…

Using RAG

First, thank you so much for this repo.

I want to use my embeddings and local llm model.

What code spaces should i change? Can you help me?

How to speed up the app

Your app works well,
I would like to know how one may speed up the app.
I asked Bard for the same question, here is its response - https://g.co/gemini/share/caba9f27a56b

Asked Claude 3, got these response:

To improve the performance of CTransformers in Langchain, you can consider the following strategies:

Use Quantization: CTransformers supports quantization, which can significantly reduce the model size and improve inference speed. You can enable quantization when loading the model using the quantize parameter in CTransformerModel.
Optimize CTransformer Configuration: CTransformer provides various configuration options to optimize performance, such as setting the number of threads, using CPU or GPU, and enabling optimizations like Operator Fusion. You can adjust these settings in the CTransformerModel initialization.
Batching: If you're processing multiple inputs at once, you can leverage batching to improve performance. CTransformers supports batched inference, which can be more efficient than processing inputs individually.
Caching: If you're performing repetitive computations on the same input, you can cache the results to avoid redundant computations. Langchain provides caching utilities that can be used with CTransformers.
Model Pruning: If the model size is too large, you can consider pruning techniques to reduce the number of parameters while maintaining reasonable performance. Langchain doesn't directly support model pruning, but you can use external libraries like Hugging Face Transformers for this purpose.
Hardware Acceleration: If you have access to hardware accelerators like GPUs or specialized AI accelerators, you can leverage them to further improve performance. CTransformers supports GPU inference, which can provide significant speedups for certain models and tasks.

It seems that GPU is an easy solution: https://g.co/gemini/share/fc9472590c4f

What would you recommend?
Thanks a lot

how to add vectore store/faiss?

in your project config, need vectorestore/faiss. but in your repo its not include
how to download it. im done download faiss store and add to my env but still cant run

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.