-
Notifications
You must be signed in to change notification settings - Fork 320
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
"ppo_micro_batch_size" missing in examples/grpo_trainer/run_qwen2-7b_seq_balance.sh #145
Comments
@deter3 This should be missing. When we use seq_balance, we don't need to tune Instead, we should tune Your quick fix will not affect the actual training but can pass this wrongly placed assertion. We'll merge them shortly. |
thanks for the prompt reply . I will give a try then . quite steep learning curve :) . |
It’s never easy to achieve optimal performance :) |
Keep trying all kinds of combination , and the instructions and error are quite confusing . below is the one working in 8*a100 80gb . @PeterSH6 is it good to go ? star text line is the one changed from examples config .
|
You don't need that line. Can you pull the latest main and try again? |
commit 695bdbb , 8*a100 80gb , working with the config below , which is same as examples/grpo_trainer/run_qwen2-7b_seq_balance.sh Great thanks to @PeterSH6 , still working at Chinese new year holiday . We all should blame deepseek R1 paper !!!!
|
I figure the config file in examples/grpo_trainer/run_qwen2-7b_seq_balance.sh , missed one line which is
actor_rollout_ref.actor.ppo_micro_batch_size=64 \
please update . And it can be run on 8*A100 .
currently , we have error of "unsupported operand type(s) for %: 'int' and 'NoneType'" when missing "ppo_micro_batch_size" .
Thanks .
The text was updated successfully, but these errors were encountered: