-
Notifications
You must be signed in to change notification settings - Fork 26.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
cannot import name 'ShardedDDPOption' from 'transformers.trainer' #33242
Comments
cc @muellerzr and @SunMarc |
Hey @nishitanand, thanks for reporting ! Could you share your traceback ? This shouldn't happen as with your current version of transformers (4.43.4), |
Hi, I uninstalled and installed transformers again. I have tried with transformers version 4.44.2 as well. Same error. Here is the traceback: Traceback (most recent call last): |
That's indeed the case. It looks like the code in their repo needs to be updated to work with the current Trainer. Sorry for the breaking change. Do you know what replaced |
Hi @muellerzr, I'd really appreciate it if you could throw some light on the issue. I'm working on a priority project. |
Hi @SunMarc, any pointers on how to solve the issue? |
System Info
I am getting the following error, but this error should not be there -
cannot import name 'ShardedDDPOption' from 'transformers.trainer'
I have the following versions installed -
tokenizers-0.19.1
transformers-4.43.4
huggingface-hub-0.24.6
I have upgraded Vicuna -7v-v1.5 to llama 3.1 8B in this github repo - https://github.com/baaivision/EVE
This works with the vicuna-7b-v1.5, but not with llama3.1 8B. It should work as there isn't much change. I earlier got rope error, but solved it by upgrading transformers as guided in this issue -
https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct/discussions/15
Who can help?
https://github.com/amyeroberts
https://github.com/muellerzr
https://github.com/SunMarc
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
I run bash eve7b_prealign.sh 0 localhost
This works with the vicuna-7b-v1.5, but not with llama3.1 8B. It should work as there isn't much change. I earlier got rope error, but solved it by upgrading transformers as guided in this issue -
https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct/discussions/15
Expected behavior
The model should start training
The text was updated successfully, but these errors were encountered: