Replies: 2 comments
-
Hi! Also note, that MACE is primarily a GPU code, and the CPU evaluation speeds are much slower than the GPU speeds, so I definitely recommend fixing that if possible. Otherwise your initial fitting script looks sensible to me. |
Beta Was this translation helpful? Give feedback.
-
Ok, thanks for the response. I've gotten the GPU version of lammps working, and it seems to be running well. I'll try out L=0,1,2 models to check system size / memory requirements. |
Beta Was this translation helpful? Give feedback.
-
Hello,
I'm in the process of learning how to use MACE, and have the CPU version of lammps installed, using the dev branch of mace to train the model, and I was wondering what parameters in model training will impact performance the most, or at least some settings that should be worth experimenting with. I'm currently doing some initial performance testing using a very small (192 atom) unit cell, and a L=1 model, and am seeing 1.535 timesteps/s on 40 cores and one node (lammps is reporting 2719% CPU usage), which makes me suspect I can probably find some increase to speed. I've not checked the GPU version of mace-lammps yet, as there's an issue with the cmake version on the cluster I'm using, but will check it soon. I've attached the current shell script I'm using to call the mace training.
mace_train.txt
Also, with the current lammps interface, does the choice of default_dtype= float32 vs. float64 during model training affect inference speed?
Beta Was this translation helpful? Give feedback.
All reactions