I would like to run Gromacs-2016.4 on GPU, but I couldn't setup the appropriate parameters of mdrun option of Gromacs and resulted in immediate termination of simulation.

I given 2 nodes for calculation, each consist of 20 cores and 2 GPUs.

I have attached the job script file and error file for more clarity.

my command to run simulation is

>module load compiler/intel-2017 cuda/9.0 cudampi/mvapich2-2.2 applic/v100/gromacs-2016.4

>srun /applic/applications/gromacs/2016.4/bin/gmx_mpi mdrun -v -deffnm step7_production -pin on -nb gpu

Running on 2 nodes with total 40 cores, 40 logical cores, 4 compatible GPUs

Cores per node: 20

Logical cores per node: 20

Compatible GPUs per node: 2

All nodes have identical type(s) of GPUs

Using 2 MPI processes

Using 20 OpenMP threads per MPI process

On host tesla18 2 compatible GPUs are present, with IDs 0,1

On host tesla18 1 GPU auto-selected for this run

Mapping of the GPU ID to the 1 PP rank in this node: 0

Note: potentially sub-optimal launch configuration, gmx mdrun started with less PP MPI process per node than GPUs available.

Each PP MPI process can use only one GPU, 1 GPU per node will be used.

This resulted in error file saying something like below

Fatal error: Your choice of number of MPI ranks and amount of resources results in using 20 OpenMP threads per rank, which is most likely inefficient. the optimum is usually between 2 and 6 threads per rank. If you want to run with this setup, specify the -ntomp option. But we suggest to change the number of MPI ranks.

Could expert of GPU computing suggest a setting for Gromacs?

Thanking you in advance.

Similar questions and discussions