-
Notifications
You must be signed in to change notification settings - Fork 37
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LightingModule optimizer issue - Thermostability fine-tuning #71
Comments
Hi, I think it's due to the imcompatibility with the version of pytorch-lightning. Could you degrade your pytorch-lightning to 1.8.3? |
Thanks, that solved the previous issue, and training started just fine. Also, is there any way to automatically select to not visualize the results (option 3) from the interactive prompt? I’m submitting my job through slurm, and the error I’m getting I assume is because of that
I also tried to set WANDB_MODE: dryrun in the config file, and moreover, by setting Thanks |
If you don't want to record your training then set |
To run it normally, I requested the necessary resources through srun in the terminal, and then run The issue was when I submitted the training through
I decided to create an account on wandb, but the problem through |
The problem is more likely due to the |
Hi all, I have been trying to run the example to fine tune the 650M model with the provided thermostability data. Unfortunately, I’m getting the following error
[rank 0]: TypeError: LightingModule.optimizer_step() takes from 4 to 5 positional arguments but 9 were given
I’m using exactly the same script provided, with only changing the number of GPU from 4 to 1, and CUDA_VISIBLE_DEVICES to 0.
Any help is greatly appreciated. Thank you
The text was updated successfully, but these errors were encountered: