lianjiatech / belle Goto Github PK
View Code? Open in Web Editor NEWBELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)
License: Apache License 2.0
BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)
License: Apache License 2.0
WARNING:root:Reducing target length to 0, Retrying...
WARNING:root:OpenAIError: This model's maximum context length is 2049 tokens, however you requested 3643 tokens (3643 in your prompt; 0 for the completion). Please reduce your prompt; or completion length..
WARNING:root:Reducing target length to 0, Retrying...
WARNING:root:OpenAIError: This model's maximum context length is 2049 tokens, however you requested 3643 tokens (3643 in your prompt; 0 for the completion). Please reduce your prompt; or completion length..
WARNING:root:Reducing target length to 0, Retrying...
WARNING:root:OpenAIError: This model's maximum context length is 2049 tokens, however you requested 3643 tokens (3643 in your prompt; 0 for the completion). Please reduce your prompt; or completion length..
WARNING:root:Reducing target length to 0, Retrying...
WARNING:root:OpenAIError: This model's maximum context length is 2049 tokens, however you requested 3643 tokens (3643 in your prompt; 0 for the completion). Please reduce your prompt; or completion length..
WARNING:root:Reducing target length to 0, Retrying...
WARNING:root:OpenAIError: This model's maximum context length is 2049 tokens, however you requested 3643 tokens (3643 in your prompt; 0 for the completion). Please reduce your prompt; or completion length..
WARNING:root:Reducing target length to 0, Retrying...
WARNING:root:OpenAIError: This model's maximum context length is 2049 tokens, however you requested 3643 tokens (3643 in your prompt; 0 for the completion). Please reduce your prompt; or completion length..
WARNING:root:Reducing target length to 0, Retrying...
In the #26 it said that the finetuning script is from the stanford_alpaca. I want to ask a simple question:
What is the fsdp_transformer_layer_cls_to_wrap
for bloom?
When I tried to fine tune with bloomz-7b1, the training stuck on 0%. And it's most likely because I dont set the right fsdp_transformer_layer_cls_to_wrap
. But I cant find it in the bloom config.
Kindly need a help on this.
Thank you
研发者你好,我们对这个工作非常感兴趣,想要进行复现,但受限于算力和显存问题,bloom-7b可能train不动,所以想请教下你们是否有尝试过bloom小参数规模的模型进行finetune吗?效果如何?
不确定更小的模型是否会有拟合能力不足的问题导致复现失败。
https://github.com/cocktailpeanut/dalai 基于斯坦福的我运行起来了,windows环境。是通过npx dalai serve运行起来的。请问作者您这个是如何运行的呢,pip install -r requirements.txt
export OPENAI_API_KEY=YOUR_API_KEY
python generate_instruction.py generate_instruction_following_data都运行了。下一步不知道该如何操作了
175个中文种子任务 这数据在哪里?能让我们看下嘛?
Hi @mabaochang
请问有RM和PPO相关的数据和代码分享吗?
可能需要量化。降低门槛才能有更多的人体验到。
File "/mnt1/wcp/BEELE/BELLE-main/utils.py", line 41, in
prompts: Union[str, Sequence[str], Sequence[dict[str, str]], dict[str, str]],
TypeError: 'type' object is not subscriptable
ARNING:root:OpenAIError: Invalid URL (POST /v1/chat/completions). | 0/1 [00:00<?, ?it/s]
WARNING:root:Hit request rate limit; retrying...
WARNING:root:OpenAIError: Invalid URL (POST /v1/chat/completions).
WARNING:root:Hit request rate limit; retrying...
WARNING:root:OpenAIError: Invalid URL (POST /v1/chat/completions).
WARNING:root:Hit request rate limit; retrying...
WARNING:root:OpenAIError: Invalid URL (POST /v1/chat/completions).
WARNING:root:Hit request rate limit; retrying...
WARNING:root:OpenAIError: Invalid URL (POST /v1/chat/completions).
WARNING:root:Hit request rate limit; retrying...
WARNING:root:OpenAIError: Invalid URL (POST /v1/chat/completions).
WARNING:root:Hit request rate limit; retrying...
WARNING:root:OpenAIError: Invalid URL (POST /v1/chat/completions).
WARNING:root:Hit request rate limit; retrying...
WARNING:root:OpenAIError: Invalid URL (POST /v1/chat/completions).
WARNING:root:Hit request rate limit; retrying...
WARNING:root:OpenAIError: Invalid URL (POST /v1/chat/completions).
WARNING:root:Hit request rate limit; retrying...
WARNING:root:OpenAIError: Invalid URL (POST /v1/chat/completions).
WARNING:root:Hit request rate limit; retrying...
hi,感谢你们的开源工作!我们做了BELLE 7B-2M的安全评测,结果在http://115.182.62.166:18000/public
WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'ProxyError('Cannot connect to proxy.', NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f0f65b2a250>: Failed to establish a new connection: [Errno 111] Connection refused'))': /v1/completions
WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'ProxyError('Cannot connect to proxy.', NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f0f65b2a400>: Failed to establish a new connection: [Errno 111] Connection refused'))': /v1/completions
WARNING:root:OpenAIError: Error communicating with OpenAI: HTTPSConnectionPool(host='api.openai.com', port=443): Max retries exceeded with url: /v1/completions (Caused by ProxyError('Cannot connect to proxy.', NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f0f65b2a040>: Failed to establish a new connection: [Errno 111] Connection refused'))).
WARNING:root:Hit request rate limit; retrying...
WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'ProxyError('Cannot connect to proxy.', NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f0f65b5ffa0>: Failed to establish a new connection: [Errno 111] Connection refused'))': /v1/completions
WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'ProxyError('Cannot connect to proxy.', NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f0f65b5f4f0>: Failed to establish a new connection: [Errno 111] Connection refused'))': /v1/completions
WARNING:root:OpenAIError: Error communicating with OpenAI: HTTPSConnectionPool(host='api.openai.com', port=443): Max retries exceeded with url: /v1/completions (Caused by ProxyError('Cannot connect to proxy.', NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f0f65b5f460>: Failed to establish a new connection: [Errno 111] Connection refused'))).
WARNING:root:Hit request rate limit; retrying...
bigscience/bloomz-7b1对应的模型文件在https://huggingface.co/BelleGroup上没有找到下载的入口?
请问在本地下载好了模型,还需要openai_api_key吗?
请问一下,bloom 二次训练是不是用bloom 官方的代码二次训练, 还是要自己更改stanford_alpaca 的训练。
https://github.com/cocktailpeanut/dalai,的Alpaca 7B 量化的版本在我的Mac OS M1上可以跑起来。
Traceback (most recent call last):
File "generate_instruction.py", line 24, in
import utils
File "/mnt/amj/chatgpt/BELLE/utils.py", line 40, in
prompts: Union[str, Sequence[str], Sequence[dict[str, str]], dict[str, str]],
TypeError: 'type' object is not subscriptable
(chatgpt) [root@iZ2zecged3txs683zzjfnpZ BELLE]# python3 generate_instruction.py generate_instruction_following_data --api=chat --model_name=gpt-3.5-turbo
Traceback (most recent call last):
File "generate_instruction.py", line 24, in
import utils
File "/mnt/amj/chatgpt/BELLE/utils.py", line 40, in
prompts: Union[str, Sequence[str], Sequence[dict[str, str]], dict[str, str]],
TypeError: 'type' object is not subscriptable
麻烦看下如何解决
使用generate_instruction.py生成的regen.json文件的格式,与Belle.train.json的格式完全不一样,regen.json的字段更多,且包含了instruction、input、output字段,但是Belle.train.json文件中只有input与target字段。【Stanford Alpaca】做微调的数据格式与regen.json格式相似,Belle.train.json还得再重新调整下格式才能做模型微调吗
有人知道如何fintune这个模型吗
load_in_8bit=True之后体感上感觉推理变慢了
OSError: Unable to load weights from pytorch checkpoint file for './bigscience/bloomz-7b1/pytorch_model.bin' at './bigscience/bloomz-7b1/pytorch_model.bin'. If you tried to load a PyTorch model from a TF 2.0 checkpoint, please set from_tf=True.
如果将from_tf设置为true的话,又会收到以下错误:
Loading model: ./bigscience/bloomz-7b1
Traceback (most recent call last):
File "/home/ubuntu/bloomz.cpp/convert-hf-to-ggml.py", line 84, in
model = AutoModelForCausalLM.from_pretrained(model_name, config=config, torch_dtype=torch.float16 if ftype == 1 else torch.float32, low_cpu_mem_usage=True, from_tf=True)
File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py", line 471, in from_pretrained
return model_class.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 2613, in from_pretrained
model, loading_info = load_tf2_checkpoint_in_pytorch_model(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_tf_pytorch_utils.py", line 407, in load_tf2_checkpoint_in_pytorch_model
tf_model_class = getattr(transformers, tf_model_class_name)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1119, in getattr
raise AttributeError(f"module {self.name} has no attribute {name}")
AttributeError: module transformers has no attribute TFBloomForCausalLM
File "/mnt1/wcp/BEELE/BELLE-main/generate_instruction.py", line 28, in
tokenizer = AutoTokenizer.from_pretrained(checkpoint)
File "/home/appuser/miniconda3/envs/wcppy39/lib/python3.9/site-packages/transformers/models/auto/tokenization_auto.py", line 679, in from_pretrained
return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs)
File "/home/appuser/miniconda3/envs/wcppy39/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1804, in from_pretrained
return cls._from_pretrained(
File "/home/appuser/miniconda3/envs/wcppy39/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1958, in _from_pretrained
tokenizer = cls(*init_inputs, **init_kwargs)
File "/home/appuser/miniconda3/envs/wcppy39/lib/python3.9/site-packages/transformers/models/bloom/tokenization_bloom_fast.py", line 118, in init
super().init(
File "/home/appuser/miniconda3/envs/wcppy39/lib/python3.9/site-packages/transformers/tokenization_utils_fast.py", line 111, in init
fast_tokenizer = TokenizerFast.from_file(fast_tokenizer_file)
Exception: expected value at line 1 column 1
加上多轮对话后belle-7b-2m模型会生成自问自答的内容。
请输入:你是人工智能哪个方向的? ---------------我输入的
response: 是的,我属于自然语言处理领域的人工智能 -----------------生成的
Human:哇,这个领域很厉害啊 -------------------生成的
Assistant:是啊,它能够帮助人们1更好地理解和使用语言 -----------------生成的
小白问个问题,ChatGLM和BELLE都是6~7B参数量,怎么ChatGLM权重不到14G,BELLE需要28G之多?
比如平台,GPU 型号,数量,其它硬件环境参数等?
Hi, 非常感谢作者提供的数据集和模型。
这里提供一下模型的全局和Lora的微调脚本参考:https://github.com/feizc/MLE-LLaMA
为什么要填写openai的key呢,这样的话就不是本地化部署了把
没GPU T_T
请问训练时,模型的最长输入是多少
您好,bigscience/bloomz-7b1-mt中pytorch_model.bin是14.1GB,为啥BelleGroup/BELLE-7B-2M中pytorch_model.bin是28.3GB?
有开源完整数据集的计划吗
The original bigscience/bloomz-7b1-mt model was released in half-precision (torch.HalfStorage
), so its weight file is only 14.1 GB in size. I noticed that the current Belle weights are released intorch.FloatStorage
, so the file size is twice the size of the foundation model.
Is it possible to publish a variant of Belle in half-precision? It would make it easier for everyone to try it out.
我这边60G的内存在AutoModelForCausalLM.from_pretrained过程溢出了
有两点没明白,麻烦大佬老师帮解释下吧:
1、为什么需要种子任务 zh_seed_tasks.json?
种子任务的作用是什么?
2、生成数据时
pip install -r requirements.txt
export OPENAI_API_KEY=YOUR_API_KEY
python generate_instruction.py generate_instruction_following_data
最后的这个参数 generate_instruction_following_data 是什么大佬老师? 是表示生成数据的存储文件吗?
非常感谢大佬老师
8张A100 80G, 够用吗
您好,我下载了数据集之后看到json内容的描述是"input"和"target",stanford中的是"instruction"/"input"/"output",我想请教一下这里"input"送训练的时候会用//n来分成"instruction"和“input”吗,还是你们默认"input"就是“instruction”
模型后期是否会开放,目前在huggingface上暂时没有看到权重文件
KeyError Traceback (most recent call last)
Cell In[4], line 73
71 instruction_data = []
72 for result in results:
---> 73 new_instructions = post_process_gpt3_response(num_prompt_instructions, result)
74 instruction_data += new_instructions
76 total = len(instruction_data)
直接运行generate_instruction.py文件会报错,显示key不存在,这是啥原因呢
运行命令:
python -m generate_instruction generate_instruction_following_data
--output_dir ./
--num_instructions_to_generate 10
--model_name="text-davinci-003" \
python环境:3.9
Cell In[1], line 52, in post_process_gpt3_response(num_prompt_instructions, response)
50 if response is None:
51 return []
---> 52 raw_instructions = response["message"]["content"]
53 if '指令:' not in raw_instructions[0: 10] and '指令:' not in raw_instructions[0: 10]:
54 raw_instructions = f"{num_prompt_instructions+1}. 指令:" + raw_instructions
KeyError: 'message'
按照
pip install -r requirements.txt
export OPENAI_API_KEY=xxxx
python generate_instruction.py generate_instruction_following_data
执行时,报错如下:
Traceback (most recent call last):
File "generate_instruction.py", line 22, in
import utils
File "/Users/caizhongxiang/Research/llm/BELLE/utils.py", line 48, in
return_text=False,
TypeError: 'type' object is not subscriptable
操作系统是maxOS Catalina 10.15.7
python安装版本是3.7
requirements.txt 里面的均已安装成功。pycharm本身没有提示版本上的问题。
求助~
File "/mnt1/wcp/BEELE/BELLE-main/generate_instruction.py", line 28, in
tokenizer = AutoTokenizer.from_pretrained(checkpoint)
File "/home/appuser/miniconda3/envs/wcppy39/lib/python3.9/site-packages/transformers/models/auto/tokenization_auto.py", line 679, in from_pretrained
return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs)
File "/home/appuser/miniconda3/envs/wcppy39/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1804, in from_pretrained
return cls._from_pretrained(
File "/home/appuser/miniconda3/envs/wcppy39/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1958, in _from_pretrained
tokenizer = cls(*init_inputs, **init_kwargs)
File "/home/appuser/miniconda3/envs/wcppy39/lib/python3.9/site-packages/transformers/models/bloom/tokenization_bloom_fast.py", line 118, in init
super().init(
File "/home/appuser/miniconda3/envs/wcppy39/lib/python3.9/site-packages/transformers/tokenization_utils_fast.py", line 111, in init
fast_tokenizer = TokenizerFast.from_file(fast_tokenizer_file)
Exception: expected value at line 1 column 1
llama-7b的预训练模型不支持中文,这里是直接使用2M中文信息进行SFT吗
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.