Skip to content

The more I pretrain (SSL), the worse fine-tuned model gets? #9175

Answered by nithinraok
riqiang-dp asked this question in Q&A
Discussion options

You must be logged in to vote

I am not sure if you are using conformer or fastconformer and the model size. These factors affect the training speed. @pzelasko could you validate lhotse arguments.

I could confirm that for non causal models pretraining definitely helps. With causal models we haven;t experimented much and I am currently training some very large FastConformer models, so I would only know sooner when the training gets finished. But based on my experiments, pretraining will always help for stable training, improved performance and faster convergence.

Replies: 5 comments 15 replies

Comment options

You must be logged in to vote
3 replies
@riqiang-dp
Comment options

@MostafaAhmed98
Comment options

@riqiang-dp
Comment options

Comment options

You must be logged in to vote
2 replies
@nithinraok
Comment options

@riqiang-dp
Comment options

Comment options

You must be logged in to vote
2 replies
@nithinraok
Comment options

@riqiang-dp
Comment options

Comment options

You must be logged in to vote
7 replies
@nithinraok
Comment options

Answer selected by riqiang-dp
@itzsimpl
Comment options

@riqiang-dp
Comment options

@riqiang-dp
Comment options

@pzelasko
Comment options

pzelasko Jun 5, 2024
Collaborator

Comment options

You must be logged in to vote
1 reply
@nithinraok
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
6 participants