Re: [AMBER] combination of CHAMBER prmtop and pmemd.cuda is causing serious instability

From: Joshua Adelman <jla65.pitt.edu>
Date: Thu, 01 Dec 2011 09:13:55 -0500

Hi Scott,

Could you quickly clarify which users will be effected by this bug? Is it all pmemd.cuda users with systems setup in Chamber using NPT?

Thanks,
Josh


On Dec 1, 2011, at 12:12 AM, Scott Le Grand wrote:

> It was a bug. Checking in fix shortly... Dumb bookkeeping error...
> Repros make fixes easy!
>
> Scott
>
>
> On Wed, Nov 30, 2011 at 9:51 AM, Marc van der Kamp <marcvanderkamp.gmail.com
>> wrote:
>
>> Ok, here is an archive, with just NVT and NPT input and pmemd and
>> pmemd.cuda_DPDP outputs to keep it small, as well as runtests.sh and
>> runtests_gpu.sh with the commands used to run.
>> If anyone would like the prmtop and restart file to run the tests, I'll
>> send them off list.
>>
>> Thanks,
>> Marc
>>
>> On 30 November 2011 13:36, Marc van der Kamp <marcvanderkamp.gmail.com
>>> wrote:
>>
>>> Thanks for the replies!
>>>
>>> I've been following the steps suggested by Ross.
>>>
>>> The results are essentially the same as I previously reported:
>>>
>>> - with NVT input, pmemd, sander and pmemd.cuda_DPDP give very similar
>>> results.
>>>
>>> - with NPT input, pmemd and sander give similar results, BUT
>>> pmemd.cuda_DPDP results are very different, especially towards the end of
>>> the 200 steps, with Etot about 850 kcal/mole lower and EPtot about 500
>>> kcal/mole lower.
>>>
>>> I tried to attach a bz2-zipped tar-file with the inputs and outputs
>>> (mdout.npt.cpu, mdout.nvt.cpu, mdout.npt.DPDP, mdout.nvt.DPDP etc.), as
>>> well as the prmtop and inpcrd used. The run commands are listed in
>>> runtests.sh (for pmemd and sander) and runtests_DPDP.sh (for
>>> pmemd.cuda_DPDP). But just got a reply that the message was too big.
>> Can't
>>> access files right now, but I will try to send a reduced archive when I
>> do.
>>>
>>> Thanks in advance for looking into this!
>>>
>>> Marc
>>>
>>> PS My initial problem was that when I tried to compile pmemd.cuda_DPDP, I
>>> got errors during compilation of bintraj.o / bintraj.f90 (a list of
>>> undefined references to __netcdf_MOD_nf90_*)
>>>
>>> I'm using cuda 4.0.17. With the same setup, I previously successfully
>>> compiled pmemd.cuda_SDPD (i.e. identical CUDA_HOME, in which no changes
>>> have been made since the pmemd.cuda_SPDP compile).
>>>
>>> The only difference, as far as I can tell, is that I have since compiled
>>> AmberTools1.5 in the same AMBERHOME tree, whereas I did the initial
>>> pmemd.cuda_SPDP compilation in a 'clean' tree.
>>>
>>> So, I ended up making a new AMBERHOME tree and doing the compilation
>> there.
>>>
>>>
>>> On 29 November 2011 18:23, Ross Walker <ross.rosswalker.co.uk> wrote:
>>>
>>>> Hi Marc,
>>>>
>>>> Yes PMEMD.cuda was tested with chamber prmtops and should work no
>>>> problems.
>>>> It is possible that something is funky with the prmtop being produced by
>>>> chamber that means it is not strictly kosher. The GPU version of the
>> code
>>>> is
>>>> more strict about prmtop standards (atoms per molecule being correct
>> etc)
>>>> than the CPU code is.
>>>>
>>>> I would suggest doing the following to help with debugging this.
>>>>
>>>> 1) Compile the DPDP version of pmemd.cuda (assuming you have applied all
>>>> the
>>>> latest bugfixes)
>>>>
>>>> cd $AMBERHOME/AmberTools/src/
>>>> ./configure -cuda_DPDP gnu
>>>> cd ../../src
>>>> make cuda
>>>>
>>>> 2) Run with your prmtop and inpcrd file that gives you the issue. Set
>>>> ntt=1
>>>> and ig=12345 in the mdin file to avoid complications from different
>> random
>>>> number generators. Set ntpr=1, nstlim=200. Then run with both:
>>>>
>>>> $AMBERHOME/bin/pmemd -O -o mdout.cpu -x mdcrd.cpu -r restrt.cpu
>>>> $AMBERHOME/bin/sander -O -o mdout.cpu.sander -x mdcrd.cpu.sander -r
>>>> restrt.cpu.sander
>>>> $AMBERHOME/bin/pmemd.cuda_DPDP -O -o mdout.DPDP -x mdcrd.DPDP -r
>>>> restrt.DPDP
>>>>
>>>> Try this with both NVT and NPT calculations. In both cases the CPU
>> PMEMD,
>>>> Sander and GPU DPDP code should match to the precision of the output on
>>>> all
>>>> steps (you get some variation in the last few decimal places) and a few
>>>> differences in the virial but they should match.
>>>>
>>>> Post the results (and the input files here).
>>>>
>>>> All the best
>>>> Ross
>>>>
>>>>> -----Original Message-----
>>>>> From: Marc van der Kamp [mailto:marcvanderkamp.gmail.com]
>>>>> Sent: Tuesday, November 29, 2011 7:08 AM
>>>>> To: AMBER Mailing List
>>>>> Subject: Re: [AMBER] combination of CHAMBER prmtop and pmemd.cuda is
>>>>> causing serious instability
>>>>>
>>>>> It would be great if the pmemd.cuda people (Ross, Scott?) could
>> confirm
>>>>> that pmemd.cuda has indeed not been tested with the combination of a
>>>>> CHAMBER prmtop and NPT.
>>>>> Is this assumption true?
>>>>> And, will there be efforts to test this and iron out potential
>>>>> bugs/problems?
>>>>>
>>>>> As I'm expecting my system to change conformation, I would really
>>>>> prefer to
>>>>> run the production MD with NPT (as opposed to NVT after NPT
>>>>> equilibration).
>>>>> It would be a pity if I need to do (expensive) multi-CPU pmemd.MPI
>> runs
>>>>> instead of using the superspeedy pmemd.cuda on a single GPU card...
>>>>>
>>>>> If this would be helpful, I'd be happy to try and set up some small
>>>>> test-systems (e.g. alanine dipeptide in waterbox) and see if I can
>>>>> replicate the problem.
>>>>>
>>>>> Thanks,
>>>>> Marc
>>>>>
>>>>>
>>>>>
>>>>> On 24 November 2011 17:16, Marc van der Kamp
>>>>> <marcvanderkamp.gmail.com>wrote:
>>>>>
>>>>>> Hi Mark,
>>>>>>
>>>>>> Thanks for your input!
>>>>>> Unfortunately, pmemd.cuda doesn't support the do_charmm_dump_gold
>>>>> option
>>>>>> of the debugf namelist. So I can't compare the forces this way.
>>>>>> This may indicate that pmemd.cuda has never really been tested
>>>>> (fully)
>>>>>> with CHAMBER prmtops...
>>>>>>
>>>>>> Thanks,
>>>>>> Marc
>>>>>>
>>>>>>
>>>>>> On 24 November 2011 13:30, Mark Williamson <mjw.mjw.name> wrote:
>>>>>>
>>>>>>> On 11/24/11 12:20, Marc van der Kamp wrote:
>>>>>>>> To provide more info:
>>>>>>>> I've just finished running 1ns of NVE and NVT MD with pmemd.cuda,
>>>>> and
>>>>>>> they
>>>>>>>> DON'T give the issue described, with CA RMSD < 1.8 in 1ns
>>>>> simulation.
>>>>>>>> The problems therefore appear to arise with a combination of
>>>>> pmemd.cuda,
>>>>>>>> NPT (ntb=2, ntp=1) and (my) CHAMBER prmtop.
>>>>>>>> I would prefer to run this system with NPT, as a conformational
>>>>> change
>>>>>>> may
>>>>>>>> occur. Nothing as large as unfolding though, just a small part of
>>>>> the
>>>>>>>> protein opening up. I'm using a fairly large water box around the
>>>>>>> protein,
>>>>>>>> so perhaps NVT would still be ok for this. Any comments
>>>>> appreciated!
>>>>>>>>
>>>>>>>> --Marc
>>>>>>>
>>>>>>>
>>>>>>> Dear Marc,
>>>>>>>
>>>>>>> I'm not sure where the source of this issue lies at the moment, but
>>>>> I
>>>>>>> have initial debug route to narrow this down.
>>>>>>>
>>>>>>> Have you tried checking that the potential energy and resultant per
>>>>> atom
>>>>>>> forces of the first step between two identical runs in pmemd and
>>>>>>> pmemd.cuda are the same?
>>>>>>>
>>>>>>> The "do_charmm_dump_gold" option can be used for this:
>>>>>>>
>>>>>>> &debugf
>>>>>>> do_charmm_dump_gold = 1
>>>>>>> /
>>>>>>>
>>>>>>> and will dump the following:
>>>>>>>
>>>>>>> NATOM 24
>>>>>>> ENERGY
>>>>>>> ENER 0.6656019668295578D+02
>>>>>>> BOND 0.1253078375923905D+01
>>>>>>> ANGL 0.3101502989274569D+01
>>>>>>> DIHE -0.2481576955859662D+02
>>>>>>> VDW 0.3170732223102823D+01
>>>>>>> ELEC 0.8385065265325110D+02
>>>>>>> FORCE
>>>>>>> 1 0.1774846256096088D+00 -0.7072502507211014D+00
>>>>>>> 0.6898056336525684D+00
>>>>>>> 2 -0.2664878707118652D+00 -0.2989897287348136D+00
>>>>>>> -0.4514535076187112D+00
>>>>>>> 3 0.1307432194682785D+00 0.1309127935015375D+01
>>>>>>> 0.1002524982820262D+01
>>>>>>> ...etc..
>>>>>>>
>>>>>>>
>>>>>>> There are more examples in
>>>>> $AMBERHOME/AmberTools/test/chamber/dev_tests
>>>>>>> and also have a look at p 41 of
>>>>> http://ambermd.org/doc11/AmberTools.pdf
>>>>>>> When I was implementing this, I was using these tests to ensure
>> that
>>>>> I
>>>>>>> was getting the same potential energy and per atom forces from the
>>>>> AMBER
>>>>>>> md engines as I was from the charmm MD engine for the same system.
>>>>>>>
>>>>>>> This test could be used to see if there is a difference in forces
>>>>>>> between pmemd and pmemd.cuda MD engines. If the issue is not here,
>>>>> one
>>>>>>> may need to look into the integration within this ensemble.
>>>>>>>
>>>>>>> Regards,
>>>>>>>
>>>>>>> Mark
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> AMBER mailing list
>>>>>>> AMBER.ambermd.org
>>>>>>> http://lists.ambermd.org/mailman/listinfo/amber
>>>>>>>
>>>>>>
>>>>>>
>>>>> _______________________________________________
>>>>> AMBER mailing list
>>>>> AMBER.ambermd.org
>>>>> http://lists.ambermd.org/mailman/listinfo/amber
>>>>
>>>>
>>>> _______________________________________________
>>>> AMBER mailing list
>>>> AMBER.ambermd.org
>>>> http://lists.ambermd.org/mailman/listinfo/amber
>>>>
>>>
>>>
>>
>> _______________________________________________
>> AMBER mailing list
>> AMBER.ambermd.org
>> http://lists.ambermd.org/mailman/listinfo/amber
>>
>>
> _______________________________________________
> AMBER mailing list
> AMBER.ambermd.org
> http://lists.ambermd.org/mailman/listinfo/amber



_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Thu Dec 01 2011 - 06:30:03 PST
Custom Search