From 47aadb1872842a4244f2aaffeee254a60c767f6b Mon Sep 17 00:00:00 2001 From: Pablo Garay Date: Tue, 7 May 2024 14:34:13 -0700 Subject: [PATCH] Revert "Add assert for max_steps to be positive in MegatronGPTSFTModel (#9110)" (#9128) This reverts commit b92e24c505b1b44d4155558a785c49d7d36f3f3d. --- .../nlp/models/language_modeling/megatron_gpt_sft_model.py | 1 - 1 file changed, 1 deletion(-) diff --git a/nemo/collections/nlp/models/language_modeling/megatron_gpt_sft_model.py b/nemo/collections/nlp/models/language_modeling/megatron_gpt_sft_model.py index 3c3259beff923..17938beb449e4 100644 --- a/nemo/collections/nlp/models/language_modeling/megatron_gpt_sft_model.py +++ b/nemo/collections/nlp/models/language_modeling/megatron_gpt_sft_model.py @@ -74,7 +74,6 @@ def __init__(self, cfg: DictConfig, trainer: Trainer): raise ImportError( "Apex was not found. Please see the NeMo README for installation instructions: https://github.com/NVIDIA/NeMo#megatron-gpt." ) - assert trainer.max_steps > 0, "max_steps for SFT can't be negative as its required to build the dataset" super().__init__(cfg, trainer=trainer) self.sep_id = cfg.get('sep_id', 49704) if hasattr(self.cfg.data, "validation_ds"):