Replies: 1 comment 4 replies
-
I don't see your data, but I guess the energies in your data are quite similar to each other, where the difference between these points is much less than the loss. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, I'm training models with distributed training using 6 GPUs. No matter what the batch size I use, I get a similar strange behavior with the lcurve.out results.
The energy/force validation and training loss seem to follow an almost identical pattern Here is an example for energy loss:
Should the results be that similar across the whole training? It makes me think that the same or very similar frames are being fed into the training and the validation steps.
The data has been shuffled, and placed in a separate directory and this is given in the input json.
Beta Was this translation helpful? Give feedback.
All reactions