-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add new fairseq_pretraining function for starting from checkpoint #255
Conversation
I noticed that this, when using the The reason for that is that, when |
To tackle the problem described above, I proposed to change the behavior in the |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The function run_fairseq_pretraining_from_checkpoint
that you introduce has huge overlap with the existing run_fairseq_pretraining
. Instead of copying the whole function and only changing a small part of it, please add checkpoint: Optional[tk.Path] = None
as an argument to the existing function and do the modifications you need if it is given, e.g.
if checkpoint is not None:
fairseq_args["checkpoint"]["continue_once"] = checkpoint
Done. Instead of |
fairseq will never see a |
For doing scheduled hard_negatives training, I added
run_fairseq_pretraining_from_checkpoint
function that takes an additionalcheckpoint
parameter and runs the pretraining job from that checkpoint