-
Hello DeePMD-kit community, I'm currently using DP-GEN to manage an automated workflow with DeePMD-kit 2.2.10 (GPU version) installed from the conda channel. My setup involves running the training (DeePMD-kit), exploration (LAMMPS), and labeling (VASP-6.4 GPU version) processes on a single GPU node equipped with 4 Nvidia Tesla V100 GPU cards, 360 GB memory, and 40 AMD CPU cores. While testing the DP-GEN workflow with the official test data, I've encountered an issue specific to LAMMPS GPU utilization:
My question is: Does the pre-compiled LAMMPS version (lmp) provided with DeePMD-kit not support GPU parallelization? If it does support GPU parallelization, what might be causing this issue, and how can I resolve it to fully utilize all available GPUs for the LAMMPS exploration step in my DP-GEN workflow? Do I need to recompile the LAMMPS by myself? Any insights or suggestions would be greatly appreciated. Thank you in advance for your help! System details:
Here is the settings in machine.json;
|
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
When using Different MPI libraries may not be ABI compatible, as the MPI ABI standard hasn't been adopted. So, do not mix the MPIs from the cluster and from Conda. |
Beta Was this translation helpful? Give feedback.
When using
para_deg
with GPUs,if_cuda_multi_devices
shoud be set totrue
.Different MPI libraries may not be ABI compatible, as the MPI ABI standard hasn't been adopted. So, do not mix the MPIs from the cluster and from Conda.