How to continue training from a checkpoint with Trainer? #2656
Unanswered
Superskyyy
asked this question in
Q&A
Replies: 1 comment
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I've been using PPO/RLOOTrainer and it seems the resume_from_checkpoint doens't work, upon seeing the code I was surprised that nothing implemented the checkpointing loading mechanism not even using the one from huggingface transfomers, (the trainer.train method doesn't take a resume_from_checkpoint arg)
How can I load back the checkpoint and resume the training? I assume people have been using this feature and I somehow missed the guide to do so.
Beta Was this translation helpful? Give feedback.
All reactions