RE: [AMBER] pmemd.cuda segmentation fault with iwrap=1 on NVT/NVE

From: Scott Le Grand <SLeGrand.nvidia.com>
Date: Fri, 28 May 2010 13:00:42 -0700

Can you email me an example that shows the iwrap behavior?

That's likely an easy fix...


-----Original Message-----
From: amber-bounces.ambermd.org [mailto:amber-bounces.ambermd.org] On Behalf Of Sasha Buzko
Sent: Friday, May 28, 2010 09:40
To: AMBER Mailing List
Subject: [AMBER] pmemd.cuda segmentation fault with iwrap=1 on NVT/NVE

Hi all,
I've compiled pmemd.cuda on a 64-bit CentOS 5.3 for use with a GTX280.
The tests came out pretty good - one failure (proteosome, low memory),
and 5 more generated errors. I'm giving the deviceQuery output below.

I'm testing pmemd.cuda with a pre-equilibrated explicit solvent system
(about 60 000 atoms). The NPT version ran perfectly fine with no errors.
However, runs at NVT or NVE parameters lead to an immediate segmentation
fault whenever iwrap=1. There are no error messages, just instant
failure. At the same time, both sander and regular pmemd run normally
with the input files below. I had to remove the iwrap=1 line for
pmemd.cuda to run.
Ideally, I'd like to keep iwrap enabled for long runs, so I wonder if
it's a known issue and whether it can be fixed.

Here are the input files for each run of pmemd.cuda

The NPT input parameters (works ok):
  imin = 0, irest = 1, ntx = 7,
  ntb = 2, pres0 = 1.0, ntp = 1,
  taup = 2.0, iwrap = 1,
  cut = 10, ntr = 0,
  ntc = 2, ntf = 2,
  temp0 = 300.0,
  ntt = 3, gamma_ln = 2.0,
  nstlim = 20000, dt = 0.001,
  ntpr = 1000, ntwx = 1000, ntwr = 1000

The NVT input parameters (segmentation fault):
  imin = 0, irest = 1, ntx = 7,
  ntb = 1,
  iwrap = 1,
  cut = 10, ntr = 0,
  ntc = 2, ntf = 2,
  temp0 = 300.0,
  ntt = 3, gamma_ln = 2.0,
  nstlim = 20000, dt = 0.001,
  ntpr = 1000, ntwx = 1000, ntwr = 1000

The NVE input parameters (segmentation fault):

  imin = 0, irest = 1, ntx = 7,
  ntb = 1, ntp = 0,
  iwrap = 1,
  cut = 10, ntr = 0,
  ntc = 2, ntf = 2,
  temp0 = 300.0,
  ntt = 0,
  nstlim = 20000, dt = 0.001,
  ntpr = 1000, ntwx = 1000, ntwe = 1000, ntwr = 1000

Any suggestions would be appreciated.

Thanks in advance


Sasha
 



[sasha.redwood release]$ ./deviceQuery
./deviceQuery Starting...

 CUDA Device Query (Runtime API) version (CUDART static linking)

There is 1 device supporting CUDA

Device 0: "GeForce GTX 280"
  CUDA Driver Version: 3.0
  CUDA Runtime Version: 3.0
  CUDA Capability Major revision number: 1
  CUDA Capability Minor revision number: 3
  Total amount of global memory: 1073020928 bytes
  Number of multiprocessors: 30
  Number of cores: 240
  Total amount of constant memory: 65536 bytes
  Total amount of shared memory per block: 16384 bytes
  Total number of registers available per block: 16384
  Warp size: 32
  Maximum number of threads per block: 512
  Maximum sizes of each dimension of a block: 512 x 512 x 64
  Maximum sizes of each dimension of a grid: 65535 x 65535 x 1
  Maximum memory pitch: 2147483647 bytes
  Texture alignment: 256 bytes
  Clock rate: 1.30 GHz
  Concurrent copy and execution: Yes
  Run time limit on kernels: Yes
  Integrated: No
  Support host page-locked memory mapping: Yes
  Compute mode: Default (multiple host
threads can use this device simultaneously)

deviceQuery, CUDA Driver = CUDART, CUDA Driver Version = 4243455, CUDA
Runtime Version = 3.0, NumDevs = 1, Device = GeForce GTX 280


PASSED

Press <Enter> to Quit...
-----------------------------------------------------------

_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
-----------------------------------------------------------------------------------
This email message is for the sole use of the intended recipient(s) and may contain
confidential information. Any unauthorized review, use, disclosure or distribution
is prohibited. If you are not the intended recipient, please contact the sender by
reply email and destroy all copies of the original message.
-----------------------------------------------------------------------------------

_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Fri May 28 2010 - 13:30:03 PDT
Custom Search