-
Notifications
You must be signed in to change notification settings - Fork 27.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
BART for Pre-Training #6743
Comments
This should help: #5096 (comment) |
@sshleifer - think this is the 3rd issue about Bart pre-training -> maybe it would be a good idea to release a small notebook at some point. |
@patil-suraj you took a stab at this at some point? this may have been optimistic :( |
Yes, I was trying to port fairseq dataset here, same for t5, I'll try to focus more on it when I'm done with current PRs, should strat with a notebook as Patrick said, then try to include it in examples/ |
@patrickvonplaten Does that mean I can train with Masked-input, input(label) and Decoder-input? |
yes, this should be possible |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
@patil-suraj any news on the pretraining script for Bart? |
If anyone wants to train their MBART model then feel free to use this. Contributions are welcome! |
@patil-suraj excuse me, is there any news on the pretraining script for Bart? Thanks. |
@thomas-li-sjtu you can try my toolkit if you like. It's based on transformers and allows for Bart/mbart pretraining. https://github.com/prajdabre/yanmtt |
Hi there, here is my problem. I hope to pretrain a bart model based on my own dataset and fine tune it for another task (not nmt). I noticed that your toolkit designs for nmt so maybe it is not the one I need. Anyway, thanks for your reply! |
@thomas-li-sjtu ok I understand. It's not just designed for NMT (despite its name). I've used it for summarisation and general NLG without problems. Good luck with your search. |
Wow that is awesome. I will try it for my task! |
@thomas-li-sjtu cool. Feel free to raise issues as it helps me add new functionality that may be of use to people. If you want to know how to use it for summarisation (or generic nlg) then look here: https://github.com/AI4Bharat/indic-bart |
Sorry to only come back to this issue now. If anyone is interested in adding this example script in For BART pre-training we need the text-infilling + sentence-permutation data collator which you could find here https://github.com/morganmcg1/rotobart/blob/main/data_collator.py#L223 With this collator you could then modify and use Let me know if anyone is interested. :) cc @patrickvonplaten |
I think the BART pre-training script is very useful for my work and many others. It is generous of you to add this example script in 'Transfromers' !!! |
Thanks for your reply and I think your method is absolutely feasible. But when I try it , I faced some errors that I can't fix. And could you please give me some help?
Run the changed script and I get errors below. Traceback (most recent call last): Waiting for your generous reply! @patil-suraj |
@Eurus-W make sure you convert the numpy arrays in the batch returned by |
❓ Questions & Help
How can I run BART pre-training?
I have data to pre-training(Masked LM)
The text was updated successfully, but these errors were encountered: