Comments (4)
Hi, such options are given in the shell command, which we have not documented yet. Roughly here is how the training is invoked:
accelerate launch -m src/magicoder/train.py \
--model_key $MODEL_KEY \
--model_name_or_path $MODEL_KEY \
--use_flash_attention True \
--datafile_paths $DATASET_PATH \
--output_dir $OUTPUT_DIR \
--bf16 True \
--num_train_epochs 2 \
--per_device_train_batch_size 2 \
--gradient_accumulation_steps 128 \
--group_by_length False \
--ddp_find_unused_parameters False \
--optim adafactor \
--max_grad_norm -1 \
--warmup_steps $WARMUP_STEP \
--learning_rate 5e-5 \
--lr_scheduler_type linear
We will give a more clear documentation later.
from magicoder.
Thank you for your reply, it worked! Looking forward to the clear documentation~
from magicoder.
Hey thx for the answer, looking forward to the whole scripts
from magicoder.
Hi, such options are given in the shell command, which we have not documented yet. Roughly here is how the training is invoked:
accelerate launch -m src/magicoder/train.py \ --model_key $MODEL_KEY \ --model_name_or_path $MODEL_KEY \ --use_flash_attention True \ --datafile_paths $DATASET_PATH \ --output_dir $OUTPUT_DIR \ --bf16 True \ --num_train_epochs 2 \ --per_device_train_batch_size 2 \ --gradient_accumulation_steps 128 \ --group_by_length False \ --ddp_find_unused_parameters False \ --optim adafactor \ --max_grad_norm -1 \ --warmup_steps $WARMUP_STEP \ --learning_rate 5e-5 \ --lr_scheduler_type linearWe will give a more clear documentation later.
Hey thx for the answer, Is clear documentation done?
from magicoder.
Related Issues (20)
- Training data format for Magicoder-OSS-Instruct-75K HOT 4
- So many impressive experiments ! Are there any experiments with neftune ? HOT 1
- The correctness of solution HOT 1
- used Dilated attenton instead of Vanilla Attention in Llama model and fine-tuen the model ,
- How do you set the 'stop_words' parameter
- Are the training loss and validation loss recorded? HOT 4
- Data collection and generation HOT 1
- Got same problem that model only return lots of '\n' HOT 5
- Achieved close performance of MagicoderS by finetuning only with `evol-codealpaca-v1`. HOT 8
- A scaling law of instruction-code-data would be very interesting... HOT 3
- catastrophic forgetting problem HOT 1
- The templates used in reproducing the eval results: why adding the instruction again after "### Response: "? HOT 1
- 8台A40机器上复现magicoder-S-DS-6.7B的结果
- Is it normal to take more than one hour to get the humanevalplus results?
- HuggingFace Playground has failed
- Quantised Finetuning on 22GB*4 GPUs
- A question of the generated data from the starcoderdata HOT 2
- Overlap between Magicoder-Evol-Instruct-110K and HumanEval HOT 2
- Code for the evaluations on APPS.
- Inquiry about Paper Details of Magicoder
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from magicoder.