I will examine this situation, but there is no way to get AMBER to scale
PME past 4 GPUs.  It is also not at all unlikely this is an MPI bug of some
sort having to do with the relatively new RDMA code in MPI.
On Jun 9, 2014 5:45 AM, "Ross Walker" <ross.rosswalker.co.uk> wrote:
> Hi Keisuke,
>
> My first question is why on earth are you trying to run a calculation
> across 8 GPUs? This is madness with AMBER 12 and doubly madness with AMBER
> 14. Did you even check the performance when running on 8 GPUS? - it will
> be horendously slow compared to running on just a single GPU. AMBER 12
> used to scale for C2050 type GPUs but then the GPUs got much quicker and
> the interconnect and PCI-E bus stagnated hence no more scaling. If you
> have 8 GPUs I'd recommend running 8 individual calculations.
>
> AMBER 14 supports peer to peer if your motherboard supports it but really
> needs PCI-E Gen 3 to function well and at this time is typically limited
> to 2 GPUs per node on the same IOH controller. There is no way 8 K20s
> could be used in this configuration.
>
> So while the code itself should not crash and there is possibly a bug here
> 8 GPUs with such a small system as DFHR PME is something we never test
> since there simply isn't a valid use case so it's something that is
> unlikely to be investigated.
>
> I suggest sticking to more realistic GPU counts and actually monitoring
> the performance of your calculations.
>
> All the best
> Ross
>
>
> On 6/9/14, 4:17 AM, "Keisuke Aono" <kaono.o.cc.titech.ac.jp> wrote:
>
> >Dear Amber-users,
> >
> >I had problem about multi-GPU calculation when using Amber 14.
> >
> >Error occurs when specify eight GPUs with Tesla K20Xm (4 node x
> >2GPUs/node).
> >On the other hand, the case of specify four GPUs,
> >calculation succeed (4 node x 1GPU/node or 2 node x 2 GPU/node).
> >This problem didn't occur when using Amber 12.
> >
> >Could you give me some advise on this problem?
> >
> >
> >I checked some case and get the following stderr.
> >
> >- Build with MVAPICH2 1.8.1, Intel Compiler 2013.1.046, CUDA 5.0
> >    forrtl: severe (174): SIGSEGV, segmentation fault occurred
> >
> >- Build with mpich2 3,0,3, Intel Compiler 2013.1.046, CUDA 5.5
> >    *** glibc detected *** pmemd.cuda.MPI: double free or corruption
> >(out): 0x0000000007e8e830 ***
> >
> >- Build with OpenMPI 1.4.2, Intel Compiler 2013.1.046, CUDA 5.5
> >    [t2a004110:28186] *** Process received signal ***
> >    [t2a004110:28186] Signal: Bus error (7)
> >    [t2a004110:28186] Signal code:  (128)
> >    [t2a004110:28186] Failing at address: (nil)
> >
> >- Build with OpenMPI 1.4.2, GNU Compiler 4.3.4, CUDA 5.5
> >    [t2a005039:05045] *** Process received signal ***
> >    [t2a005039:05045] Signal: Segmentation fault (11)
> >    [t2a005039:05045] Signal code:  (128)
> >    [t2a005039:05045] Failing at address: (nil)
> >
> >
> >Each case, mdout file shows temp = NaN and EPtot = ***********.
> >
> >    check COM velocity, temp:             NaN      NaN(Removed)
> >
> >     NSTEP =     1000   TIME(PS) =       8.000  TEMP(K) =      NaN  PRESS
> >=     0.0
> >     Etot   =            NaN  EKtot   =            NaN  EPtot      =
> >974005909.5961
> >     BOND   = 617477679.4037  ANGLE   =    252874.7313  DIHED      =
> >7138.6530
> >     1-4 NB =         0.0000  1-4 EEL =         0.0709  VDWAALS    =
> >356801602.4831
> >     EELEC  =   -533385.7459  EHBOND  =         0.0000  RESTRAINT  =
> >   0.0000
> >
> >--------------------------------------------------------------------------
> >----
> >
> >    check COM velocity, temp:             NaN      NaN(Removed)
> >
> >     NSTEP =     2000   TIME(PS) =      10.000  TEMP(K) =      NaN  PRESS
> >=     0.0
> >     Etot   =            NaN  EKtot   =            NaN  EPtot      =
> >**************
> >     BOND   =         0.0000  ANGLE   =    297187.8075  DIHED      =
> >   0.0000
> >     1-4 NB =         0.0000  1-4 EEL =         0.0000  VDWAALS    =
> >**************
> >     EELEC  = **************  EHBOND  =         0.0000  RESTRAINT  =
> >   0.0000
> >
> >--------------------------------------------------------------------------
> >----
> >
> >
> >Our computer cluster and test case Information is the following.
> >
> >> ./update_amber -version
> >Version is reported as <version>.<patches applied>
> >
> >        AmberTools version 14.02
> >             Amber version 14.00
> >
> >- SUSE Linux Enterprise Server 11 SP1
> >- GPU: (Driver: 319.82, CUDA 5.5 or 5.0, K20Xm)
> >- Command: mpirun -np 8 -hostfile hostfile pmemd.cuda.MPI -O -i mdin -o
> >mdout -p prmtop -c inpcrd
> >- Test Case
> >    Explicit Solvent(PME) 3.DHFR NVE = 23,558 atoms
> >    http://ambermd.org/gpus/benchmarks.htm
> >
> >
> >Best Regards,
> >Keisuke Aono
> >
> >------------------------
> >Keisuke Aono
> >Global Scientific Information and Computing Center(GSIC)
> >Tokyo Institute of Technology
> >Mail:kaono.o.cc.titech.ac.jp
> >2-12-1, Ookayama, Meguro-ku, TOKYO
> >152-8550 JAPAN
> >TEL:+81-3-5754-1375
> >
> >
> >
> >_______________________________________________
> >AMBER mailing list
> >AMBER.ambermd.org
> >http://lists.ambermd.org/mailman/listinfo/amber
>
>
>
> _______________________________________________
> AMBER mailing list
> AMBER.ambermd.org
> http://lists.ambermd.org/mailman/listinfo/amber
>
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Mon Jun 09 2014 - 10:00:02 PDT