Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About checkpoints to be used by finetune #44

Open
trmzpi02 opened this issue Oct 18, 2024 · 1 comment
Open

About checkpoints to be used by finetune #44

trmzpi02 opened this issue Oct 18, 2024 · 1 comment

Comments

@trmzpi02
Copy link

Hello! I am very interested in your work and see that you release the weight of Show-o before fine-tuning on LLaVA instructional tuning datasets.

I have the following two questions:

  1. I see that you recommend in the README to go to finetune on the basis of the show-o-512x512-wo-llava-tuning checkpoint, so why don't go to finetune on the basis of the show-o-512x512. Is it because there is a performance degradation on certain downstream tasks after fine-tuning on LLaVA instructional tuning datasets?

  2. If I want to fine-tune on certain visual downstream tasks, which checkpoint should I use?

@Sierkinhane
Copy link
Collaborator

Sierkinhane commented Oct 18, 2024

Hi, thanks for your interest in our work. If you'd like to reproduce our results, you can try the pre-trained one. Besides, because the final checkpoint was fine-tuned on the llava data, further fine-tuning will degrade the performance (overfitting). If there is new training data, you can directly fine-tune the final checkpoint I think.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants