Giter Site home page Giter Site logo

Comments (5)

stduhpf avatar stduhpf commented on May 25, 2024 1

Fixed!
(though you should update README)

from fastllama.

stduhpf avatar stduhpf commented on May 25, 2024

Also I get this error when I try to run it after quantize:

[Info]: Func('Model') loading model(='ALPACA-LORA-13B') from ./models/ALPACA-LORA-13B/alpaca-lora-q4_0.bin - please wait ...
[Info]: Func('KVCacheBuffer::init') kv self size  =  512.00 MB
[Error]: Func('Model') invalid model file ./models/ALPACA-LORA-13B/alpaca-lora-q4_0.bin (bad magic)
[Error]: Func('FastLlama::Params::build') Unable to load model
Traceback (most recent call last):
  File "examples/python/example-alpaca.py", line 11, in <module>
    model = Model(
  File "examples/python/build/fastllama.py", line 208, in __init__
    raise RuntimeError("Unable to load model")
RuntimeError: Unable to load model
Exception ignored in: <function Model.__del__ at 0x7f4e2d3cb310>
Traceback (most recent call last):
  File "examples/python/build/fastllama.py", line 404, in __del__
    signal.signal(signal.SIGINT, signal.SIG_DFL)
  File "~/miniconda3/envs/fastllama/lib/python3.8/signal.py", line 47, in signal
    handler = _signal.signal(_enum_to_int(signalnum), _enum_to_int(handler))
TypeError: signal handler must be signal.SIG_IGN, signal.SIG_DFL, or a callable object

from fastllama.

PotatoSpudowski avatar PotatoSpudowski commented on May 25, 2024

Ah yes,
Good catch. I was trying to move away from my hacky implementation. In our old method, for large models, creating just one file was not a good idea because of high memory usage during conversion and quantization steps.

I wanted to add a way to split the models into smaller parts during the exporting part.
Please give me some time I will get it done by tonight.

Update: Using the convert.py script should work for alpaca models now as well as the base llama models

from fastllama.

PotatoSpudowski avatar PotatoSpudowski commented on May 25, 2024

Meanwhile you can try the sync_llama branch for now.
All ggml models exported to huggingface work now. Been testing Vicuna and Alpaca 7B and 13B

from fastllama.

PotatoSpudowski avatar PotatoSpudowski commented on May 25, 2024

This should be fixed in the new PR, can you please check?

from fastllama.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.