You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi,
I just wanted to know what would be the ideal tokenizer model_max_length/max_length during inference of the model.
Does max_length affect generation quality of questions? If yes, then can you briefly explain me why.
Thanksss
P.S - I've been using 2048 as my max length.
The text was updated successfully, but these errors were encountered:
Ideal max length doesn't depends in the parameter size of the model. it depends on the trained data. As long as your computation support you can increase the max length but, the better output would be generated based on the trained data max length.
So I prefer to use the trained data max length as inference max length.
Ideal max length doesn't depends in the parameter size of the model. it depends on the trained data. As long as your computation support you can increase the max length but, the better output would be generated based on the trained data max length.
So I prefer to use the trained data max length as inference max length.
Okay so then what was the max_length used during training?
Are any training scripts provided?
Hi,
I just wanted to know what would be the ideal tokenizer model_max_length/max_length during inference of the model.
Does max_length affect generation quality of questions? If yes, then can you briefly explain me why.
Thanksss
P.S - I've been using 2048 as my max length.
The text was updated successfully, but these errors were encountered: