Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training SaProt with a smaller ESM model #45

Open
BSharmi opened this issue Jul 6, 2024 · 3 comments
Open

Training SaProt with a smaller ESM model #45

BSharmi opened this issue Jul 6, 2024 · 3 comments

Comments

@BSharmi
Copy link

BSharmi commented Jul 6, 2024

Hi there!

Awesome work! Do you have a script that I can use to train SaProt with a smaller ESM model? I think I can use the config https://github.com/westlake-repl/SaProt/blob/main/config/pretrain/saprot.yaml but from the name
https://github.com/westlake-repl/SaProt/blob/main/config/pretrain/saprot.yaml#L14 it looks like the model is already trained?

Would appreciate any pointers!

Thank you!

@LTEnjoy
Copy link
Contributor

LTEnjoy commented Jul 6, 2024

Hi!

You can definitely use the config to train a smaller SaProt. The line 14 doesn't mean the model is trained. It is just the name of the training log. You actually train it from scratch.

If you want to train a smaller version, you have to first create a folder containing the configuration of your model like SaProt_650M_AF2 or SaProt_35M_AF2, except that you do not have to inlclude pytorch_model.bin. Then you only need to change the config_path to your folder path so you can train your own model.
image

Hope this could resolve your problem:)

@Binyun-Z
Copy link

I have the same question, if I train Saport on a small-scale dataset, how should I design my configuration file?

@LTEnjoy
Copy link
Contributor

LTEnjoy commented Sep 26, 2024

I have the same question, if I train Saport on a small-scale dataset, how should I design my configuration file?

Hi. Do you mean by training on a smaller dataset or training a smaller model?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants