Hwankyu -
Okay, I have no idea what the parallel run issues are; there is insufficient 
information here to figure that out.  You need to run the amber test suite 
in parallel, and get that working, and then you know that sander and pmemd 
parallel versions are properly installed.  Regarding the pmemd single 
processor run, I cannot see any cause for failure; looks to me like the run 
may have blown up without flushing buffers.  The  most common cause for 
pmemd not running when sander does run on linux systems is stacksize issues. 
Is this a linux system?  If so, be sure there is a "limit stacksize 
unlimited" in your .cshrc (just having it in .login is often not enough for 
parallel runs; if you invoke things with bourne shell, you must use ulimit 
instead).  Also, set ntpr to something like 1 to debug a run; lots of things 
can happen in 500 steps, and you are only dumping output every 500 steps. 
Other than that, I don't see anything. You of course do need the parallel 
builds of both sander and pmemd (you do have two executables, built 
differently, for each?). Also note that while the parallel build of sander 
can be run on one processor, parallel pmemd requires a minimum of two 
processors (but it should die with a very clear message).  Anyway, check 
stacksize, set ntpr really low, and see what you get.
Regards - Bob Duke
----- Original Message ----- 
From: "Hwankyu Lee" <leehk.umich.edu>
To: <amber.scripps.edu>
Sent: Tuesday, April 26, 2005 4:35 PM
Subject: AMBER: doesn't work in pmemd, and no output in parallel simulations
> Dear Amber-users,
>
> I've done energy minimization of my system (~70,000 atoms, including 
> water), and
> then run MD with restraints on the solute by using sander. And then, I was
> trying to simulate this system, testing with pmemd and sander with 
> parallel and
> no parallel. Sander and pmemd passed example test, so there may be no 
> problem
> with installation. When I ran the simulation with sander at single CPU, it
> nicely worked. But, in ohter cases, I got two problems.
>
> 1) When I ran sander in one CPU, it nicely worked. However, when I ran 
> sander in
> multiple CPUs, I saw CPUs run but coulnd't see any output files (.out and
> .mdcrd). I'm sure that I made right directory for output files. Could you 
> tell
> me why I can't see the output files in the directory
> (/home/leehk/run/g5ace90-par4)?  This is script file.
> ---------------------------
> #PBS -S /bin/bash
> #PBS -l nodes=2,walltime=500:00:00
> #PBS -q protein
> #PBS -N sand-cpu4
> #PBS -j oe
> #PBS -M leehk.umich.edu
> #PBS -m abe
> #
> echo "I ran on `hostname`"
> cat $PBS_NODEFILE
> #
> export GMPICONF=/home/leehk/.gmpi/$PBS_JOBID
> export PATH=/usr/cac/mpich.eth/bin:$PATH
> #
> cd /home/leehk/run/g5ace90-par4
> mpirun -np 4 -machinefile $PBS_NODEFILE /home/leehk/exe.eth/sander -O -i 
> md2.in
> -o md2.out -p g5ace90.prmtop -c md1.rst -r g5ace90-md0.rst -x 
> g5ace90-md0.mdcrd
> #
> ---------------------------
>
> 2) When I ran pmemd at one CPU or multiple CPUs, I saw CPUs run, but when 
> I
> checked .out file, I saw that .out file stopped like below. Since this 
> system
> work with sander, I cannot understand why it didn't work with pmemd.
> I also attach my script file.
> -----------------------------
>
>          -------------------------------------------------------
>          Amber 8  SANDER                 Scripps/UCSF 2004
>          -------------------------------------------------------
>
> | PMEMD implementation of SANDER, Release 8.0
>
> | Run on 04/26/2005 at 15:28:30
>
>  [-O]verwriting output
>
> File Assignments:
> |   MDIN: md2.in
> |  MDOUT: md2.out
> | INPCRD: md1.rst
> |   PARM: g5ace90.prmtop
> | RESTRT: g5ace90-md0.rst
> |   REFC: refc
> |  MDVEL: mdvel
> |   MDEN: mden
> |  MDCRD: g5ace90-md0.mdcrd
> | MDINFO: mdinfo
>
>
> Here is the input file:
>
> dendrimer : 2ns of MD
> &cntrl
>  imin = 0, irest = 0, ntx = 7,
>  ntb = 2, pres0 = 1.0, ntp = 1,
>  taup = 5.0,
>  cut = 10, ntr = 0,
>  ntc = 2, ntf = 2,
>  tempi = 298.0, temp0 = 298.0,
>  ntt = 3, gamma_ln = 1.0,
>  nstlim = 1000000, dt = 0.002,
>  ntpr = 500, ntwx = 500, ntwr = 20000
> /
>
>
>
>
> | Largest sphere to fit in unit cell has radius =    40.001
>
> | Duplicated    0 dihedrals
>
> | Duplicated    0 dihedrals
>
> --------------------------------------------------------------------------------
>   1.  RESOURCE   USE:
> --------------------------------------------------------------------------------
>
> getting new box info from bottom of inpcrd
>
> NATOM  =   63187 NTYPES =      11 NBONH =   60807 MBONA  =    2367
> NTHETH =    6554 MTHETA =    2860 NPHIH =   10308 MPHIA  =    3601
> NHPARM =       0 NPARM  =       0 NNB   =  105162 NRES   =   19366
> NBONA  =    2367 NTHETA =    2860 NPHIA =    3601 NUMBND =      15
> NUMANG =      26 NPTRA  =       7 NATYP =      14 NPHB   =       1
> IFBOX  =       2 NMXRS  =     635 IFCAP =       0 NEXTRA =       0
> NCOPY  =       0
>
> | Coordinate Index Table dimensions:    17   17   17
> | Direct force subcell size =     5.7636    5.7636    5.7636
>
>     BOX TYPE: TRUNCATED OCTAHEDRON
>
> --------------------------------------------------------------------------------
>   2.  CONTROL  DATA  FOR  THE  RUN
> --------------------------------------------------------------------------------
>
> DEU
>
> General flags:
>     imin    =       0, nmropt  =       0
>
> Nature and format of input:
>     ntx     =       7, irest   =       0, ntrx    =       1
>
> Nature and format of output:
>     ntxo    =       1, ntpr    =     500, ntrx    =       1, ntwr    = 
> 20000
>     iwrap   =       0, ntwx    =     500, ntwv    =       0, ntwe    = 
> 0
>     ioutfm  =       0, ntwprt  =       0, idecomp =       0, rbornstat= 
> 0
>
> Potential function:
>     ntf     =       2, ntb     =       2, igb     =       0, nsnb    = 
> 25
>     ipol    =       0, gbsa    =       0, iesp    =       0
>     dielc   =   1.00000, cut     =  10.00000, intdiel =   1.00000
>     scnb    =   2.00000, scee    =   1.20000
>
> Frozen or restrained atoms:
>     ibelly  =       0, ntr     =       0
> ------------------------------------
> Script file is below.
> -------------------------------------
> #PBS -S /bin/bash
> #PBS -l nodes=1,walltime=500:00:00
> #PBS -q protein
> #PBS -N sanpme1cpu
> #PBS -j oe
> #PBS -M leehk.umich.edu
> #PBS -m ae
>
> echo "I ran on `hostname`"
>
> cd /home/leehk/run/g5ace90-pme
> /home/leehk/exe.ser/pmemd -O -i md2.in -o md2.out -p g5ace90.prmtop -c 
> md1.rst
> -r g5ace90-md0.rst -x g5ace90-md0.mdcrd &
> wait
> -------------------------------------
>
> best,
> Hwankyu.
> -----------------------------------------------------------------------
> The AMBER Mail Reflector
> To post, send mail to amber.scripps.edu
> To unsubscribe, send "unsubscribe amber" to majordomo.scripps.edu
> 
-----------------------------------------------------------------------
The AMBER Mail Reflector
To post, send mail to amber.scripps.edu
To unsubscribe, send "unsubscribe amber" to majordomo.scripps.edu
Received on Tue Apr 26 2005 - 22:53:00 PDT