Comments (8)
@alvarobartt hi, thx for your detailed explanation at another post, looks like I misuse device_map="auto"
for training, I will follow your instruction to try later. big thx again.
from trl.
Hi here too @blaze7451 can you share with me a Gist with the full fine-tuning script so that I can provide you the FSDP configuration and command to run after testing that myself? But as mentioned in the other issue, if you use accelerate
with FSDP and remove device_map="auto"
you shouldn't encounter this issue. Thanks in advance!
from trl.
@alvarobartt Hi buddy, thx. I will try again later.
from trl.
@blaze7451 Hey, I'm facing the same issue, do you know how to fix it?
from trl.
hi @alvarobartt, thanks for your response. I hit this issue too, from the same article. I have some experience with device_map="auto"
to fine-tune models like phi-2, Gemma with multiple GPUs. It is really useful when the single GPU does not have enough memory to load the model and training data. Do you have any suggestion here that we can use ORPO on multiple GPUs in notebook environment(like kaggle)? Thank you.
from trl.
But the issue for you @Aisuko is that you're not able to fine-tune Llama 3B with device_map="auto"
right? Or is it that you're just unable to use trl.ORPOTrainer
+ device_map="auto"
?
from trl.
But the issue for you @Aisuko is that you're not able to fine-tune Llama 3B with
device_map="auto"
right? Or is it that you're just unable to usetrl.ORPOTrainer
+device_map="auto"
?
Hi @alvarobartt thanks for your quick replying. Let's make it clearly, can we use trl.ORPOTrainer
to fine-tune model on multiple GPUs with device_map="auto"
parameter?
from trl.
Hi @alvarobartt thanks for your quick replying. Let's make it clearly, can we use
trl.ORPOTrainer
to fine-tune model on multiple GPUs withdevice_map="auto"
parameter?
Hi @Aisuko as mentioned before I don't have much experience on relying on device_map="auto"
to fine-tune models, indeed I thought it was not even supported, but it seems it is with some tweaks. After reading some issues I realised that @younesbelkada answered a similar question for the KTOTrainer
that can be easily solved apparently, please read #1560 (comment)
from trl.
Related Issues (20)
- KTOTrainer fails to compute loss when model is loaded across multiple GPUs HOT 4
- Support for training LLM's using RLAIF methods HOT 2
- SFT - Dataset Packing causes C10d timeout HOT 1
- TrlParser not respecting config
- Reproducing LLaVA with Mistral backened
- Bug in calling model.eval() in PPO
- vsft_llava.py:
- NameError: name 'PeftConfig' is not defined HOT 4
- Loss value returned by the SFT Trainer
- Using PEFT causes model to not predict EOS HOT 2
- Using SFTTrainer and the DeepSpeed Zero-3 Method training, the data map function runs, but the map function does not run in parallel GPU form. HOT 1
- IPO loss computation vs changes in DPO dataset format
- Anthropic HH new dataset format repeats the prompt HOT 4
- Group Relative Policy Optimization Trainer
- bugs in trained save and evaluation for DPO training with deepspeed_zero3 HOT 3
- config.json file is missing. HOT 4
- The data is consuming too much GPU memory In the SFTTrainer HOT 3
- [Question] Does TRL support DPO trainer with simulated environment for generating training data like PPO's step-based training? HOT 1
- [Question]How should I combine DPOTrainer and Accelerate for training? HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from trl.