Replies: 2 comments
-
|
Hi @UnHwanLee The first thing I'd suggest doing is to use the isolated atom energies for your system computed with the same reference method used for doing your AIMD. For more details, see https://nequip.readthedocs.io/en/latest/guide/configuration/model.html#energy-shifts-scales. This suggestion is based on the consistent (though small) energy offset for the 3 by 3 by 3 and large, yet consistent looking offset for the 4 by 4 by 4 cell. Let's see if this change works first. If some issues persist, you might consider augmenting your dataset with 1 by 1 by 1, 2 by 2 by 2, etc supercell training data, which might be helpful to show more variety and avoid being overfit on a specific supercell size. NequIP is a message-passing GNN, and your config shows It might also be worth trying Side note that you can experiment with |
Beta Was this translation helpful? Give feedback.
-
|
(I am a user, not a NEquIP developer) Even on the 3x3x3-trained data, your energy errors are dominated by an offset. That could be taken away by setting |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Dear NequIP Team,
I hope this message finds you well.
First of all, thank you for your outstanding work on NequIP and Allegro. I’ve been using your framework to study Li metal systems and electrolyte in battery, and it has been incredibly useful.
I would like to ask for your opinion or suggestions regarding an issue I encountered with the scale-transferability of a trained NequIP model.
I generated 3×3×3 Li bulk structures using AIMD at 323K, 900K, and 1500K to ensure thermal diversity.
For computational efficiency and to reduce redundancy, I extracted 100 frames from each 10 ps AIMD trajectory (i.e., 100 × 3 = 300 frames total) and used them as my training dataset.
I trained a NequIP model on these 300 configurations. The model performs well when predicting energies and forces for 3×3×3 Li bulk structures, including unseen configurations.
However, when I apply this model to a 4×4×4 Li bulk structure (also generated via AIMD at 323K), the predicted energy/force accuracy significantly degrades.
I observed the same issue not only in Li bulk systems but also when attempting to generalize to larger slabs and electrolyte systems.
This is counterintuitive to me, because I expected that, as NequIP is based on local environments, the model should—at least to some extent—generalize to larger systems composed of similar local atomic motifs.
Moreover, a key reason for using MLIPs in general is to enable scaling up MD simulations beyond DFT-accessible system sizes, which is why I expected small-scale training to work for larger-scale predictions.
Do you have any insights or recommendations for this type of scale-transferability problem?
Is this a known limitation or failure mode in practice?
Would you suggest any specific training strategies, model architecture changes, or dataset augmentations to improve transferability across system sizes?
Could this be related to r_cut, neighbor statistics, or other NequIP-specific configurations (e.g., min_atoms, num_neighbors_statistics.yaml, etc.)?
I’ve already tried several variations in training hyperparameters and data preprocessing but haven’t yet resolved this issue. Any guidance would be greatly appreciated.
Thank you very much in advance for your time and help.
This is the config.yaml that I used to train model.
It is the energy and force error graph in 3x3x3 test data. (It was trained in my model)


It is the energy and force error graph in 4x4x4 test data. (It was not trained in my model)


Beta Was this translation helpful? Give feedback.
All reactions