After the changes you suggested, I got this information on the output
| Dynamic Memory, Types Used:
| Reals 688690
| Integers 595564
| Nonbonded Pairs Initial Allocation: 146264
| Running AMBER/MPI version on 64 nodes
and still a performance of 02.ns/day.
There is now a log file that didn't appear before, with the following
information:
Initial FFT Block Distribution Based on Workload Estimate:
FFT blocks assigned to 12 tasks
First FFT Block Distribution Based on Actual Workload:
FFT blocks assigned to 56 tasks
Image Distribution at run step 344:
Count of images assigned to each task:
340 437 412 335 542 572 542 516
291 256 99 1 0 0 0 0
0 0 0 0 0 0 230 184
1 0 244 352 436 6 82 219
23 1 2 64 137 283 173 59
290 133 233 81 253 198 341 173
280 330 367 267 157 117 407 125
361 374 533 455 606 646 1003 905
What does it mean?
On Mon, Aug 29, 2011 at 5:34 PM, Jason Swails <jason.swails.gmail.com>wrote:
> Every PBS system is set up differently, so it's impossible for us to tell
> what may be happening for sure. However, I suspect that you're not getting
> 64 CPUs like you think you are.
>
> On Mon, Aug 29, 2011 at 4:05 PM, Bruno Rodrigues <bbrodrigues.gmail.com
> >wrote:
>
> > Dear All,
> >
> > I'm trying to run parallel Amber 11 on a cluster with PBS. I've checked
> the
> > parallel installation and it's quite fine (the log file attached).
> >
> > However, the performance is always between 0.1 and 0.5 ns/day, no matter
> > the
> > number of processors I choose. Is there something missing in my script?
> >
> > Here are the changes I made on my configure (for the parallel version):
> > mpicc --> icc -lmpi
> > mpif90 --> ifort -lmpi
> >
> > This generated the correct config.h needed for the fortran compiler.
> >
> > However, the problem persists with gnu installing, so I guess it has
> > nothing
> > to do with the installation, but it's pretty much a submission problem.
> > Here
> > is an example of my job:
> >
> > #!/bin/bash
> > #
> > #################################################
> > # THIS JOB IS TO EQUILIBRATE THE SYSTEM AT 300K #
> > # TO BE USED IN FUTURE SIMULATIONS. IT STARTS #
> > # FROM THE EQUILIBRATION ON CHACOBO, WHERE 1ns #
> > # WAS PERFORMED AFTER THE DNA WAS RELEASED. #
> > #################################################
> > #
> > #PBS -S /bin/sh
> > #
> > # Nome do job
> > #PBS -N prod_slow
> > #
> > #Erro na saida padrao
> > #PBS -j oe
> > #
> > # Chamada do ambiente paralelo e numero de slots
> > #PBS -l select=64:ncpus=1
> > #PBS -l walltime=200:00:00
> >
> > #
> > cd $PBS_O_WORKDIR
> >
> > export sander=/home/u/bbr/bin/amber11/bin/pmemd.MPI
> >
>
> In here, add the line
>
> CPUS=`cat $PBS_NODEFILE | wc -l`
>
>
> >
> > l=heat20
> > f=prod01
> > mpiexec -n 64 $sander -O -i $PWD/$f.in -o $PWD/$f.out -inf $PWD/$f.inf \
> > -c $PWD/1D20_wat_tip3pf.$l -ref $PWD/1D20_wat_tip3pf.$l -r
> > $PWD/1D20_wat_tip3pf.$f \
> > -p $PWD/1D20_wat_tip3pf.top -x $PWD/1D20_wat_tip3pf$f.x -e
> > $PWD/1D20_wat_tip3pf$f.ene
> >
>
> change the beginning to "mpiexec -n $CPUS" instead of "mpiexec -n 64".
> pmemd.MPI should report how many processors are being used, which should
> help you make sure that you're at least allocating all the processors you
> want to be. You could also consider passing mpiexec the PBS_NODEFILE if
> you
> find out how to pass your mpiexec a hostfile or nodefile or something (this
> makes sure that each thread is bound to the proper processor).
>
> HTH,
> Jason
>
> --
> Jason M. Swails
> Quantum Theory Project,
> University of Florida
> Ph.D. Candidate
> 352-392-4032
> _______________________________________________
> AMBER mailing list
> AMBER.ambermd.org
> http://lists.ambermd.org/mailman/listinfo/amber
>
--
--
Bruno Barbosa Rodrigues
PhD Student - Physics Department
Universidade Federal de Minas Gerais - UFMG
Belo Horizonte - Brazil
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Mon Aug 29 2011 - 14:00:03 PDT