You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When attempting to train a model using Nemo 2403, a ValueError is raised indicating a conflict between the precision argument and the plugins argument being passed to the Trainer. The error message states that both precision=bf16-mixed and plugins=PipelineMixedPrecisionPlugin were received and cannot be used together.
megatron_gpt_continue_training mixtral 8x22b with Nemo 2403 image.
Expected behavior
This issue seems to be related to Issue #8848, where a similar error was encountered when converting Mistral/Mixtral models to the Nemo format.
PR #8908 partially addressed this issue by removing precision args in the trainer due to a PyTorch Lightning (PTL) update. However, the changes were not applied to the training script.
The text was updated successfully, but these errors were encountered:
Describe the bug
When attempting to train a model using Nemo 2403, a
ValueError
is raised indicating a conflict between theprecision
argument and theplugins
argument being passed to the Trainer. The error message states that bothprecision=bf16-mixed
andplugins=PipelineMixedPrecisionPlugin
were received and cannot be used together.Steps/Code to reproduce bug
Expected behavior
The text was updated successfully, but these errors were encountered: