I would like to run Gromacs-2016.4 on GPU, but I couldn't setup the appropriate parameters of mdrun option of Gromacs and resulted in immediate termination of simulation.
I given 2 nodes for calculation, each consist of 20 cores and 2 GPUs.
I have attached the job script file and error file for more clarity.
my command to run simulation is
>module load compiler/intel-2017 cuda/9.0 cudampi/mvapich2-2.2 applic/v100/gromacs-2016.4
>srun /applic/applications/gromacs/2016.4/bin/gmx_mpi mdrun -v -deffnm step7_production -pin on -nb gpu
Running on 2 nodes with total 40 cores, 40 logical cores, 4 compatible GPUs
Cores per node: 20
Logical cores per node: 20
Compatible GPUs per node: 2
All nodes have identical type(s) of GPUs
Using 2 MPI processes
Using 20 OpenMP threads per MPI process
On host tesla18 2 compatible GPUs are present, with IDs 0,1
On host tesla18 1 GPU auto-selected for this run
Mapping of the GPU ID to the 1 PP rank in this node: 0
Note: potentially sub-optimal launch configuration, gmx mdrun started with less PP MPI process per node than GPUs available.
Each PP MPI process can use only one GPU, 1 GPU per node will be used.
This resulted in error file saying something like below
Fatal error: Your choice of number of MPI ranks and amount of resources results in using 20 OpenMP threads per rank, which is most likely inefficient. the optimum is usually between 2 and 6 threads per rank. If you want to run with this setup, specify the -ntomp option. But we suggest to change the number of MPI ranks.
Could expert of GPU computing suggest a setting for Gromacs?
Thanking you in advance.