Skip to content

Commit

Permalink
Fix llama3 pretraining NeMo 2.0 script (NVIDIA#10307)
Browse files Browse the repository at this point in the history
Signed-off-by: Hemil Desai <hemild@nvidia.com>
  • Loading branch information
hemildesai authored Aug 31, 2024
1 parent 9a22005 commit b698ae5
Showing 1 changed file with 1 addition and 3 deletions.
4 changes: 1 addition & 3 deletions examples/llm/run/llama3_pretraining.py
Original file line number Diff line number Diff line change
Expand Up @@ -140,7 +140,7 @@ def main():
# Uses configs from NeMo directly
pretrain = MODEL_SIZE_MAPPING[args.size]["nemo"]["pretrain"](
name=exp_name,
ckpt_dir=f"/{exp_name}/checkpoints",
ckpt_dir="/nemo_run/checkpoints",
)

# Overwrite the dataloader in the recipe to use your custom dataloader.
Expand Down Expand Up @@ -170,8 +170,6 @@ def main():
executor = local_executor_torchrun(nodes=pretrain.trainer.num_nodes, devices=pretrain.trainer.devices)

with run.Experiment(f"{exp_name}{args.tag}") as exp:
pretrain.log.dir = f"/{exp_name}/checkpoints"

for i in range(1):
exp.add(
pretrain,
Expand Down

0 comments on commit b698ae5

Please sign in to comment.