Dear list,
I'm running 4 instances of pmemd.cuda on 24 core machine with 4 K80 boards
with a single SLURM job ( ntasks-per-core=1), setting CUDA_VISIBLE_DEVICES
to send each job to a separate board.
GPU works fine but all CPU pmemd.cuda runs on core "0".
Manually running "taskset -cp ... " fixes the issue, and performance is
back to normal. But isn't practical.
A workaround was to run 4 separate jobs asking for 6cores/job, but this way
I block the resources other people could be using.
Could you give a hint on what to do ?
A related MPI issue with pmemd.cuda,
http://archive.ambermd.org/201701/0081.html
--
Prof. Diego Enry B. Gomes, PhD
Laboratório de Biologia Computacional.
Diretoria de Metrologia Aplicada às Ciências da Vida.
Instituto Nacional de Metrologia, Qualidade e Tecnologia
+55 21 2145 3070 | dgomes.pq.cnpq.br
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Wed Mar 01 2017 - 08:30:02 PST