-
Notifications
You must be signed in to change notification settings - Fork 310
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Exploring Weak to Strong Generalization from a pre-training standpoint #12
Comments
i am not that familiar with the literature but there's which uses training time for strength https://aclanthology.org/2023.acl-long.796/. overall seems like a reasonable direction and I suspect there are many under-explored things in this space! |
i would guess it's just randomness, could be that the second training split is better for idiosyncratic reasons |
Created a dataset of weak,strong and transfer accuracies for pythia 1b,1,4B,2.8B models at 5 different stages of their pretraining https://github.com/rokosbasilisk/weak-to-strong/blob/EDA/eda/results_df.csv. |
In the paper, a "stronger" model is defined as a model with the same architecture but a greater number of parameters. I am curious if any research has been conducted regarding weak to strong generalization, where the weak-supervisor model is less pretrained, and the stronger-student is more pretrained.
I am currently exploring the use of Pythia-models checkpoints to assess performance on BoolQ (https://github.com/rokosbasilisk/weak-to-strong where weaker student model is a checkpoint of the model which is few steps before the stronger student model).
Has any prior work been undertaken in this direction? If not, could you provide insights into why this area remains unexplored?
The text was updated successfully, but these errors were encountered: