Having trouble reproducing Large Model Training #595
-
Hi Everyone, I'm trying to understand the effects of fine-tuning on my dataset (vs just training from scratch). For fine-tuning I have been using the --foundation_model="large". I want to now retrain on my dataset (from scratch) and am trying to use the same settings as the large foundation model. Here I'm a bit stuck. The source page: https://github.com/ACEsuit/mace-mp/releases/tag/mace_mp_0 has a training script, but it doesn't seem to correspond to the 'large model'. E.g. when I run the archive script I get the following number of parameters: INFO: Number of parameters: 894352 But when running 'large' I get: Number of parameters: 2203928 (substantially more) I can of course try to tune each of these settings by hand and verify against the generated architecture in the log in each case, but it would great if I could have some ability to just launch the 'large' architecture without using the weights so I can see 'from scratch performance'. Maybe the script used to produce the 'large' model is somewhere i haven't been able to find, if so that would be great to use as well. running mace-torch==0.3.6 PS I think the large model uses a 4.5 cutoff not 6.0. It would probably be good to mention that in the publication version of the 'foundational models' paper which currently makes it look like the cutoff is 6 and not 4.5: |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
Hey @daniel-sintef , I realized we uploaded the wrong script for L=2 model on the mace-mp repo. Here is the correct one:
|
Beta Was this translation helpful? Give feedback.
Hey @daniel-sintef ,
I realized we uploaded the wrong script for L=2 model on the mace-mp repo. Here is the correct one: