-
Notifications
You must be signed in to change notification settings - Fork 109
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Large memory usage on MATH #80
Comments
Hi, I can confirm that the generation size is unbounded, which you can see in the task description
When I'll check if the paper defines a maximum expected generation size, else will fix the bound to the maximum answer size + 10% maybe? |
Yes, alternatively we could set the max gen size to something like 1024 or 2048 tokens since if a model cannot answer in that span then it is likely incorrect. You can see here that the authors chose 1024 tokens for models that aren't |
Sounds perfect, will use this rn! |
… with a much longer context size. Should fix #80
Caps it at 2048 even for models with a much longer context size. Should fix #80
Is the MATH benchmark expected to run for anything beyond
batch_size=1
?Running the following command for a small model gives OOM on a single node of H100s which is a bit surprising to me:
Strangely enough, bumping up the batch size for Mistral 7B is fine:
Perhaps there's some sort of unbounded generation occurring which is causing the memory to explode for certain models like Qwen?
The text was updated successfully, but these errors were encountered: