Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Exploring Weak to Strong Generalization from a pre-training standpoint #12

Open
rokosbasilisk opened this issue Dec 17, 2023 · 4 comments

Comments

@rokosbasilisk
Copy link

rokosbasilisk commented Dec 17, 2023

In the paper, a "stronger" model is defined as a model with the same architecture but a greater number of parameters. I am curious if any research has been conducted regarding weak to strong generalization, where the weak-supervisor model is less pretrained, and the stronger-student is more pretrained.

I am currently exploring the use of Pythia-models checkpoints to assess performance on BoolQ (https://github.com/rokosbasilisk/weak-to-strong where weaker student model is a checkpoint of the model which is few steps before the stronger student model).

Has any prior work been undertaken in this direction? If not, could you provide insights into why this area remains unexplored?

@rokosbasilisk rokosbasilisk changed the title weak-2-strong generalization for less-pretrained vs more-pretrained models Exploring Weak to Strong Generalization in from a pre-training standpoint Dec 17, 2023
@rokosbasilisk rokosbasilisk changed the title Exploring Weak to Strong Generalization in from a pre-training standpoint Exploring Weak to Strong Generalization from a pre-training standpoint Dec 18, 2023
@WuTheFWasThat
Copy link
Contributor

i am not that familiar with the literature but there's which uses training time for strength https://aclanthology.org/2023.acl-long.796/. overall seems like a reasonable direction and I suspect there are many under-explored things in this space!

@rokosbasilisk
Copy link
Author

rokosbasilisk commented Dec 20, 2023

Screenshot from 2023-12-20 14-20-08
i am running the train_weak_to_strong over a range of parameter sizes at different checkpoint steps,
surprisingly when the weak model and the strong model are exactly the same (in terms of both params and checkpoint steps) there is a gain in the accuracy for the stronger model in most cases as seen in "acc_diff" column. i am currently trying to check this holds true for much larger (till ~12B params over multiple checkpoint steps).

Any idea why this might happen?

@WuTheFWasThat
Copy link
Contributor

i would guess it's just randomness, could be that the second training split is better for idiosyncratic reasons

@rokosbasilisk
Copy link
Author

rokosbasilisk commented Jan 2, 2024

Created a dataset of weak,strong and transfer accuracies for pythia 1b,1,4B,2.8B models at 5 different stages of their pretraining https://github.com/rokosbasilisk/weak-to-strong/blob/EDA/eda/results_df.csv.
Currently doing some EDA to check effect Pretraining vs Parameters on w2s generalization. any suggessions are welcome

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants