Memory issues and Einsum #748
-
When using large configurations (900 atoms) I run out of memory on A100 GPU,
Could this be solved by doing the einsum in batches, perhaps using some opt-einsum module? I saw this being used in other MLIP codes (but maybe you already thought about it and there's a reason it's not possible). In any case it's possible to avoid running out of memory by using e.g. an H100 GPU with more memory or choosing a smaller cutoff, but it would be nice if it can be avoided if someone really needs a larger cutoff and run with a smaller GPU. edit: I already set |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 15 replies
-
Hey @lucasdekam, We already use optimised einsums. Did you try cueq, it should enable you to load more atoms on your GPU. Are you training or evaluating? Also what is the size of your model (channels, L, rmax)? |
Beta Was this translation helpful? Give feedback.
You can look at the doc https://mace-docs.readthedocs.io/en/latest/guide/cuda_acceleration.html, these are cuda kernels developed by Nvidia to accelerate MACE. You can train with them.
Two notes: