Re: [AMBER] experiences with EVGA GTX TITAN Superclocked - memtestG80 - UNDERclocking in Linux ?

From: Scott Le Grand <varelse2005.gmail.com>
Date: Sun, 2 Jun 2013 09:57:08 -0700

Interestingly, I have seen the same numeric difference in the repiprocal
electrostatic energy in 2 out of the 3 failures I've seen so far... This
will all ultimately make sense but we're not there yet...


On Sun, Jun 2, 2013 at 9:49 AM, ET <sketchfoot.gmail.com> wrote:

> Ok. No worries. Sounds promising. :) I'll set up the long GB simulations
> and post the results back later this evening.
>
>
> br,
> g
>
>
> On 2 June 2013 17:42, Scott Le Grand <varelse2005.gmail.com> wrote:
>
> > PPS My testbed also repros crashing so at least I have things isolated...
> >
> >
> > On Sun, Jun 2, 2013 at 9:33 AM, Scott Le Grand <varelse2005.gmail.com
> > >wrote:
> >
> > > PS this *might* indicate a software bug in cuFFT, but it needs more
> > > characterization... And things are going to get a little stream of
> > > consciousness from here because you're getting unfiltered raw data, so
> > > please don't draw any conclusions towards anything yet - I'm just
> letting
> > > you guys know what I'm finding out as I find it...
> > >
> > >
> > >
> > > On Sun, Jun 2, 2013 at 9:31 AM, Scott Le Grand <varelse2005.gmail.com
> > >wrote:
> > >
> > >> And bingo...
> > >>
> > >> At the very least, the reciprocal sum is intermittently
> inconsistent...
> > >> This explains the irreproducible behavior...
> > >>
> > >> And here's the level of inconsistency:
> > >> 31989.38940628897399 vs
> > >> 31989.39168370794505
> > >>
> > >> That's error at the level of 1e-7 or a somehow missed single-precision
> > >> transaction somewhere...
> > >>
> > >> The next question is figuring out why... This may or may not
> ultimately
> > >> explain the crashes you guys are also seeing...
> > >>
> > >>
> > >>
> > >> On Sun, Jun 2, 2013 at 9:07 AM, Scott Le Grand <varelse2005.gmail.com
> > >wrote:
> > >>
> > >>>
> > >>> Observations:
> > >>> 1. The degree to which the reproducibility is broken *does* appear to
> > >>> vary between individual Titan GPUs. One of my Titans breaks within
> 10K
> > >>> steps on cellulose, the other one made it to 100K steps twice without
> > doing
> > >>> so leading me to believe it could be trusted (until yesterday where I
> > now
> > >>> see it dies between 50K and 100K steps most of the time).
> > >>>
> > >>> 2. GB hasn't broken (yet). So could you run myoglobin for 500K and
> > >>> TRPcage for 1,000,000 steps and let's see if that's universal.
> > >>>
> > >>> 3. Turning on double-precision mode makes my Titan crash rather than
> > run
> > >>> irreproducibly, sigh...
> > >>>
> > >>> So whatever is going on is triggered by something in PME but not GB.
> > So
> > >>> that's either the radix sort, the FFT, the Ewald grid interpolation,
> > or the
> > >>> neighbor list code. Fixing this involves isolating this and figuring
> > out
> > >>> what exactly goes haywire. It could *still* be software at some very
> > small
> > >>> probability but the combination of both 680 and K20c with ECC off
> > running
> > >>> reliably is really pointing towards the Titans just being clocked too
> > >>> fast.
> > >>>
> > >>> So how long with this take? Asking people how long it takes to fix a
> > >>> bug never really works out well. That said, I found the 480 bug
> > within a
> > >>> week and my usual turnaround for a bug with a solid repro is <24
> hours.
> > >>>
> > >>> Scott
> > >>>
> > >>> On Sun, Jun 2, 2013 at 7:58 AM, Marek Maly <marek.maly.ujep.cz>
> wrote:
> > >>>
> > >>>> Hi all,
> > >>>>
> > >>>> here are my results after bugfix 18 application (see attachment).
> > >>>>
> > >>>> In principle I don't see any "drastical" changes.
> > >>>>
> > >>>> FACTOR_IX still perfectly stable/reproducible on both cards,
> > >>>>
> > >>>> JAC tests - problems with finishing AND/OR reproducibility the
> > >>>> same CELLULOSE_NVE although here it seems that my TITAN_1
> > >>>> has no problems with this test (but the same same trend I saw also
> > >>>> before bugfix 18 - see my older 500K steps test).
> > >>>>
> > >>>> But anyway bugfix 18 brought here one change.
> > >>>>
> > >>>> The err
> > >>>>
> > >>>>
> > >>>> #1 ERR writtent in mdout:
> > >>>> ------
> > >>>> | ERROR: max pairlist cutoff must be less than unit cell max
> sphere
> > >>>> radius!
> > >>>> ------
> > >>>>
> > >>>> was substituted with err/warning ?
> > >>>>
> > >>>> #0 no ERR writtent in mdout, ERR written in standard output
> > (nohup.out)
> > >>>> -----
> > >>>> Nonbond cells need to be recalculated, restart simulation from
> > previous
> > >>>> checkpoint
> > >>>> with a higher value for skinnb.
> > >>>>
> > >>>> -----
> > >>>>
> > >>>> Another thing,
> > >>>>
> > >>>> recently I started on another machine and GTX 580 GPU simulation of
> > >>>> relatively
> > >>>> big system ( 364275 atoms/PME ). The system is composed also from
> the
> > >>>> "exotic" molecules like polymers. ff12SB, gaff, GLYCAM forcefields
> > used
> > >>>> here. I had problem even with minimization part here, having big
> > energy
> > >>>> on the start:
> > >>>>
> > >>>> -----
> > >>>> NSTEP ENERGY RMS GMAX NAME
> > >>>> NUMBER
> > >>>> 1 2.8442E+09 2.1339E+02 1.7311E+04 O
> > 32998
> > >>>>
> > >>>> BOND = 11051.7467 ANGLE = 17720.4706 DIHED =
> > >>>> 18977.7584
> > >>>> VDWAALS = ************* EEL = -1257709.6203 HBOND =
> > >>>> 0.0000
> > >>>> 1-4 VDW = 7253.7412 1-4 EEL = 149867.0207 RESTRAINT =
> > >>>> 0.0000
> > >>>>
> > >>>> ----
> > >>>>
> > >>>> with no chance to minimize the system even with 50 000 steps in both
> > >>>> min cycles (with constrained and unconstrained solute) and hence
> > >>>> heating NVT
> > >>>> crashed immediately even with very small dt. I patched Amber12 here
> > >>>> with the
> > >>>> bugfix 18 and the minimization was done without any problem with
> > common
> > >>>> 5000 steps
> > >>>> (obtaining target Energy -1.4505E+06 while that initial was that
> > >>>> written above).
> > >>>>
> > >>>> So indeed bugfix 18 solved some issues, but unfortunately not those
> > >>>> related to
> > >>>> Titans.
> > >>>>
> > >>>> Here I will try to install cuda 5.5, recompile GPU Amber part with
> > >>>> this new
> > >>>> cuda version and repeat the 100K tests.
> > >>>>
> > >>>> Scott, let us know how finished your experiment with downclocking of
> > >>>> Titan.
> > >>>> Maybe the best choice would be here to flash Titan directly with
> your
> > >>>> K20c bios :))
> > >>>>
> > >>>> M.
> > >>>>
> > >>>>
> > >>>>
> > >>>>
> > >>>>
> > >>>>
> > >>>>
> > >>>>
> > >>>>
> > >>>> Dne Sat, 01 Jun 2013 21:09:46 +0200 Marek Maly <marek.maly.ujep.cz>
> > >>>> napsal/-a:
> > >>>>
> > >>>>
> > >>>> Hi,
> > >>>>>
> > >>>>> first of all thanks for providing of your test results !
> > >>>>>
> > >>>>> It seems that your results are more or less similar to that of
> > >>>>> mine maybe with the exception of the results on FactorIX tests
> > >>>>> where I had perfect stability and 100% or close to 100%
> > >>>>> reproducibility.
> > >>>>>
> > >>>>> Anyway the type of errs which you reported are the same which I
> > >>>>> obtained.
> > >>>>>
> > >>>>> So let's see if the bugfix 18 will help here (or at least on NPT
> > tests)
> > >>>>> or not. As I wrote just before few minutes, it seems that it was
> not
> > >>>>> still
> > >>>>> loaded
> > >>>>> to the given server, although it's description is already present
> on
> > >>>>> the
> > >>>>> given
> > >>>>> web page ( see http://ambermd.org/bugfixes12.**html<
> > http://ambermd.org/bugfixes12.html>).
> > >>>>>
> > >>>>> As you can see, this bugfix contains also changes in CPU code
> > although
> > >>>>> the majority is devoted to GPU code, so perhaps the best will be to
> > >>>>> recompile
> > >>>>> whole amber with this patch although this patch would be perhaps
> > >>>>> applied
> > >>>>> even after just
> > >>>>> GPU configure command ( i.e. ./configure -cuda -noX11 gnu ) but
> after
> > >>>>> consequent
> > >>>>> building, just the GPU binaries will be updated. Anyway I would
> > rather
> > >>>>> recompile
> > >>>>> whole Amber after this patch.
> > >>>>>
> > >>>>> Regarding to GPU test under linux you may try memtestG80
> > >>>>> (please use the updated/patched version from here
> > >>>>> https://github.com/ihaque/**memtestG80<
> > https://github.com/ihaque/memtestG80>
> > >>>>> )
> > >>>>>
> > >>>>> just use git command like:
> > >>>>>
> > >>>>> git clone https://github.com/ihaque/**memtestG80.git<
> > https://github.com/ihaque/memtestG80.git>PATCHED_MEMTEST-G80
> > >>>>>
> > >>>>> to download all the files and save them into directory named
> > >>>>> PATCHED_MEMTEST-G80.
> > >>>>>
> > >>>>> another possibility is to try perhaps similar (but maybe more up to
> > >>>>> date)
> > >>>>> test
> > >>>>> cuda_memtest ( http://sourceforge.net/**projects/cudagpumemtest/<
> > http://sourceforge.net/projects/cudagpumemtest/>).
> > >>>>>
> > >>>>> regarding ig value: If ig is not present in mdin, the default value
> > is
> > >>>>> used
> > >>>>> (e.g. 71277) if ig=-1 the random seed will be based on the current
> > date
> > >>>>> and time, and hence will be different for every run (not a good
> > variant
> > >>>>> for our testts). I simply deleted eventual ig records from all
> mdins
> > >>>>> so I
> > >>>>> assume that in each run the default seed 71277 was automatically
> > used.
> > >>>>>
> > >>>>> M.
> > >>>>>
> > >>>>>
> > >>>>>
> > >>>>>
> > >>>>>
> > >>>>>
> > >>>>> Dne Sat, 01 Jun 2013 20:26:16 +0200 ET <sketchfoot.gmail.com>
> > >>>>> napsal/-a:
> > >>>>>
> > >>>>> Hi,
> > >>>>>>
> > >>>>>> I've put the graphics card into a machine with the working GTX
> titan
> > >>>>>> that I
> > >>>>>> mentioned earlier.
> > >>>>>>
> > >>>>>> The Nvidia driver version is: 133.30
> > >>>>>>
> > >>>>>> Amber version is:
> > >>>>>> AmberTools version 13.03
> > >>>>>> Amber version 12.16
> > >>>>>>
> > >>>>>> I ran 50k steps with the amber benchmark using ig=43689 on both
> > cards.
> > >>>>>> For
> > >>>>>> the purpose of discriminating between them, the card I believe
> > >>>>>> (fingers
> > >>>>>> crossed) is working is called GPU-00_TeaNCake, whilst the other
> one
> > is
> > >>>>>> called GPU-01_008.
> > >>>>>>
> > >>>>>> *When I run the tests on GPU-01_008:*
> > >>>>>>
> > >>>>>> 1) All the tests (across 2x repeats) finish apart from the
> > following
> > >>>>>> which
> > >>>>>> have the errors listed:
> > >>>>>>
> > >>>>>> ------------------------------**--------------
> > >>>>>> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
> > >>>>>> Error: unspecified launch failure launching kernel kNLSkinTest
> > >>>>>> cudaFree GpuBuffer::Deallocate failed unspecified launch failure
> > >>>>>>
> > >>>>>> ------------------------------**--------------
> > >>>>>> CELLULOSE_PRODUCTION_NPT - 408,609 atoms PME
> > >>>>>> cudaMemcpy GpuBuffer::Download failed unspecified launch failure
> > >>>>>>
> > >>>>>> ------------------------------**--------------
> > >>>>>> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
> > >>>>>> Error: unspecified launch failure launching kernel kNLSkinTest
> > >>>>>> cudaFree GpuBuffer::Deallocate failed unspecified launch failure
> > >>>>>>
> > >>>>>> ------------------------------**--------------
> > >>>>>> CELLULOSE_PRODUCTION_NPT - 408,609 atoms PME
> > >>>>>> cudaMemcpy GpuBuffer::Download failed unspecified launch failure
> > >>>>>> grep: mdinfo.1GTX680: No such file or directory
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>> 2) The sdiff logs indicate that reproducibility across the two
> > >>>>>> repeats
> > >>>>>> is
> > >>>>>> as follows:
> > >>>>>>
> > >>>>>> *GB_myoglobin: *Reproducible across 50k steps
> > >>>>>> *GB_nucleosome:* Reproducible till step 7400
> > >>>>>> *GB_TRPCage:* Reproducible across 50k steps
> > >>>>>>
> > >>>>>> *PME_JAC_production_NVE: *No reproducibility shown from step 1,000
> > >>>>>> onwards
> > >>>>>> *PME_JAC_production_NPT*: Reproducible till step 1,000. Also
> > outfile
> > >>>>>> is
> > >>>>>> not written properly - blank gaps appear where something should
> have
> > >>>>>> been
> > >>>>>> written
> > >>>>>>
> > >>>>>> *PME_FactorIX_production_NVE:* Reproducible across 50k steps
> > >>>>>> *PME_FactorIX_production_NPT:* Reproducible across 50k steps
> > >>>>>>
> > >>>>>> *PME_Cellulose_production_NVE:*** Failure means that both runs do
> > not
> > >>>>>> finish
> > >>>>>> (see point1)
> > >>>>>> *PME_Cellulose_production_NPT: *Failure means that both runs do
> not
> > >>>>>> finish
> > >>>>>> (see point1)
> > >>>>>>
> > >>>>>> ##############################**##############################**
> > >>>>>> ###########################
> > >>>>>>
> > >>>>>> *When I run the tests on * *GPU-00_TeaNCake:*
> > >>>>>> *
> > >>>>>> *
> > >>>>>> 1) All the tests (across 2x repeats) finish apart from the
> > following
> > >>>>>> which
> > >>>>>> have the errors listed:
> > >>>>>> ------------------------------**-------
> > >>>>>> JAC_PRODUCTION_NPT - 23,558 atoms PME
> > >>>>>> PMEMD Terminated Abnormally!
> > >>>>>> ------------------------------**-------
> > >>>>>>
> > >>>>>>
> > >>>>>> 2) The sdiff logs indicate that reproducibility across the two
> > >>>>>> repeats
> > >>>>>> is
> > >>>>>> as follows:
> > >>>>>>
> > >>>>>> *GB_myoglobin:* Reproducible across 50k steps
> > >>>>>> *GB_nucleosome:* Reproducible across 50k steps
> > >>>>>> *GB_TRPCage:* Reproducible across 50k steps
> > >>>>>>
> > >>>>>> *PME_JAC_production_NVE:* No reproducibility shown from step
> 10,000
> > >>>>>> onwards
> > >>>>>> *PME_JAC_production_NPT: * No reproducibility shown from step
> 10,000
> > >>>>>> onwards. Also outfile is not written properly - blank gaps appear
> > >>>>>> where
> > >>>>>> something should have been written. Repeat 2 Crashes with error
> > noted
> > >>>>>> in
> > >>>>>> 1.
> > >>>>>>
> > >>>>>> *PME_FactorIX_production_NVE:* No reproducibility shown from step
> > >>>>>> 9,000
> > >>>>>> onwards
> > >>>>>> *PME_FactorIX_production_NPT: *Reproducible across 50k steps
> > >>>>>>
> > >>>>>> *PME_Cellulose_production_NVE: *No reproducibility shown from step
> > >>>>>> 5,000
> > >>>>>> onwards
> > >>>>>> *PME_Cellulose_production_NPT: ** *No reproducibility shown from
> > step
> > >>>>>> 29,000 onwards. Also outfile is not written properly - blank gaps
> > >>>>>> appear
> > >>>>>> where something should have been written.
> > >>>>>>
> > >>>>>>
> > >>>>>> Out files and sdiff files are included as attatchments
> > >>>>>>
> > >>>>>> ##############################**###################
> > >>>>>>
> > >>>>>> So I'm going to update my nvidia driver to the latest version and
> > >>>>>> patch
> > >>>>>> amber to the latest version and rerun the tests to see if there is
> > any
> > >>>>>> improvement. Could someone let me know if it is necessary to
> > recompile
> > >>>>>> any
> > >>>>>> or all of AMBER after applying the bugfixes?
> > >>>>>>
> > >>>>>> Additionally, I'm going to run memory tests and heaven benchmarks
> on
> > >>>>>> the
> > >>>>>> cards to check whether they are faulty or not.
> > >>>>>>
> > >>>>>> I'm thinking that there is a mix of hardware error/configuration
> > (esp
> > >>>>>> in
> > >>>>>> the case of GPU-01_008) and amber software error in this
> situation.
> > >>>>>> What
> > >>>>>> do
> > >>>>>> you guys think?
> > >>>>>>
> > >>>>>> Also am I right in thinking (from what Scott was saying) that all
> > the
> > >>>>>> benchmarks should be reproducible across 50k steps but begin to
> > >>>>>> diverge
> > >>>>>> at
> > >>>>>> around 100K steps? Is there any difference from in setting *ig *to
> > an
> > >>>>>> explicit number to removing it from the mdin file?
> > >>>>>>
> > >>>>>> br,
> > >>>>>> g
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>> On 31 May 2013 23:45, ET <sketchfoot.gmail.com> wrote:
> > >>>>>>
> > >>>>>> I don't need sysadmins, but sysadmins need me as it gives purpose
> > to
> > >>>>>>> their
> > >>>>>>> bureaucratic existence. A encountered evil if working in an
> > >>>>>>> institution
> > >>>>>>> or
> > >>>>>>> comapny IMO. Good science and indiviguality being sacrificed for
> > >>>>>>> standardisation and mediocrity in the intrerests of maintaing a
> > >>>>>>> system
> > >>>>>>> that
> > >>>>>>> focusses on maintaining the system and not the objective.
> > >>>>>>>
> > >>>>>>> You need root to move fwd on these things, unfortunately. and ppl
> > >>>>>>> with
> > >>>>>>> root are kinda like your parents when you try to borrow money
> from
> > >>>>>>> them
> > >>>>>>> .
> > >>>>>>> age 12 :D
> > >>>>>>> On May 31, 2013 9:34 PM, "Marek Maly" <marek.maly.ujep.cz>
> wrote:
> > >>>>>>>
> > >>>>>>> Sorry why do you need sysadmins :)) ?
> > >>>>>>>>
> > >>>>>>>> BTW here is the most recent driver:
> > >>>>>>>>
> > >>>>>>>> http://www.nvidia.com/object/**linux-display-amd64-319.23-**
> > >>>>>>>> driver.html<
> > http://www.nvidia.com/object/linux-display-amd64-319.23-driver.html>
> > >>>>>>>>
> > >>>>>>>> I do not remember anything easier than is to install driver
> > >>>>>>>> (especially
> > >>>>>>>> in case of binary (*.run) installer) :))
> > >>>>>>>>
> > >>>>>>>> M.
> > >>>>>>>>
> > >>>>>>>>
> > >>>>>>>>
> > >>>>>>>> Dne Fri, 31 May 2013 22:02:34 +0200 ET <sketchfoot.gmail.com>
> > >>>>>>>> napsal/-a:
> > >>>>>>>>
> > >>>>>>>> > Yup. I know. I replaced a 680 and the everknowing sysadmins
> are
> > >>>>>>>> reluctant
> > >>>>>>>> > to install drivers not in the repositoery as they are lame. :(
> > >>>>>>>> > On May 31, 2013 7:14 PM, "Marek Maly" <marek.maly.ujep.cz>
> > wrote:
> > >>>>>>>> >>
> > >>>>>>>> >> As I already wrote you,
> > >>>>>>>> >>
> > >>>>>>>> >> the first driver which properly/officially supports Titans,
> > >>>>>>>> should
> > >>>>>>>> be
> > >>>>>>>> >> 313.26 .
> > >>>>>>>> >>
> > >>>>>>>> >> Anyway I am curious mainly about your 100K repetitive tests
> > with
> > >>>>>>>> >> your Titan SC card. Especially in case of these tests (
> > JAC_NVE,
> > >>>>>>>> JAC_NPT
> > >>>>>>>> >> and CELLULOSE_NVE ) where
> > >>>>>>>> >> my Titans SC randomly failed or succeeded. In FACTOR_IX_NVE,
> > >>>>>>>> >> FACTOR_IX_NPT
> > >>>>>>>> >> tests both
> > >>>>>>>> >> my cards are perfectly stable (independently from drv.
> version)
> > >>>>>>>> and
> > >>>>>>>> also
> > >>>>>>>> >> the runs
> > >>>>>>>> >> are perfectly or almost perfectly reproducible.
> > >>>>>>>> >>
> > >>>>>>>> >> Also if your test will crash please report the eventual errs.
> > >>>>>>>> >>
> > >>>>>>>> >> To this moment I have this actual library of errs on my
> Titans
> > SC
> > >>>>>>>> GPUs.
> > >>>>>>>> >>
> > >>>>>>>> >> #1 ERR writtent in mdout:
> > >>>>>>>> >> ------
> > >>>>>>>> >> | ERROR: max pairlist cutoff must be less than unit cell
> max
> > >>>>>>>> sphere
> > >>>>>>>> >> radius!
> > >>>>>>>> >> ------
> > >>>>>>>> >>
> > >>>>>>>> >>
> > >>>>>>>> >> #2 no ERR writtent in mdout, ERR written in standard output
> > >>>>>>>> (nohup.out)
> > >>>>>>>> >>
> > >>>>>>>> >> ----
> > >>>>>>>> >> Error: unspecified launch failure launching kernel
> kNLSkinTest
> > >>>>>>>> >> cudaFree GpuBuffer::Deallocate failed unspecified launch
> > failure
> > >>>>>>>> >> ----
> > >>>>>>>> >>
> > >>>>>>>> >>
> > >>>>>>>> >> #3 no ERR writtent in mdout, ERR written in standard output
> > >>>>>>>> (nohup.out)
> > >>>>>>>> >> ----
> > >>>>>>>> >> cudaMemcpy GpuBuffer::Download failed unspecified launch
> > failure
> > >>>>>>>> >> ----
> > >>>>>>>> >>
> > >>>>>>>> >> Another question, regarding your Titan SC, it is also EVGA as
> > in
> > >>>>>>>> my
> > >>>>>>>> case
> > >>>>>>>> >> or it is another producer ?
> > >>>>>>>> >>
> > >>>>>>>> >> Thanks,
> > >>>>>>>> >>
> > >>>>>>>> >> M.
> > >>>>>>>> >>
> > >>>>>>>> >>
> > >>>>>>>> >>
> > >>>>>>>> >> Dne Fri, 31 May 2013 19:17:03 +0200 ET <sketchfoot.gmail.com
> >
> > >>>>>>>> napsal/-a:
> > >>>>>>>> >>
> > >>>>>>>> >> > Well, this is interesting...
> > >>>>>>>> >> >
> > >>>>>>>> >> > I ran 50k steps on the Titan on the other machine with
> driver
> > >>>>>>>> 310.44
> > >>>>>>>> >> and
> > >>>>>>>> >> > it
> > >>>>>>>> >> > passed all the GB steps. i.e totally identical results over
> > two
> > >>>>>>>> >> repeats.
> > >>>>>>>> >> > However, it failed all the PME tests after step 1000. I'm
> > >>>>>>>> going to
> > >>>>>>>> > update
> > >>>>>>>> >> > the driver and test it again.
> > >>>>>>>> >> >
> > >>>>>>>> >> > Files included as attachments.
> > >>>>>>>> >> >
> > >>>>>>>> >> > br,
> > >>>>>>>> >> > g
> > >>>>>>>> >> >
> > >>>>>>>> >> >
> > >>>>>>>> >> > On 31 May 2013 16:40, Marek Maly <marek.maly.ujep.cz>
> wrote:
> > >>>>>>>> >> >
> > >>>>>>>> >> >> One more thing,
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> can you please check under which frequency is running that
> > >>>>>>>> your
> > >>>>>>>> >> titan ?
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> As the base frequency of normal Titans is 837MHz and the
> > Boost
> > >>>>>>>> one
> > >>>>>>>> is
> > >>>>>>>> >> >> 876MHz I
> > >>>>>>>> >> >> assume that yor GPU is running automatically also under
> it's
> > >>>>>>>> boot
> > >>>>>>>> >> >> frequency (876MHz).
> > >>>>>>>> >> >> You can find this information e.g. in Amber mdout file.
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> You also mentioned some crashes in your previous email.
> Your
> > >>>>>>>> ERRs
> > >>>>>>>> >> were
> > >>>>>>>> >> >> something like those here:
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> #1 ERR writtent in mdout:
> > >>>>>>>> >> >> ------
> > >>>>>>>> >> >> | ERROR: max pairlist cutoff must be less than unit cell
> > max
> > >>>>>>>> sphere
> > >>>>>>>> >> >> radius!
> > >>>>>>>> >> >> ------
> > >>>>>>>> >> >>
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> #2 no ERR writtent in mdout, ERR written in standard
> output
> > >>>>>>>> >> (nohup.out)
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> ----
> > >>>>>>>> >> >> Error: unspecified launch failure launching kernel
> > kNLSkinTest
> > >>>>>>>> >> >> cudaFree GpuBuffer::Deallocate failed unspecified launch
> > >>>>>>>> failure
> > >>>>>>>> >> >> ----
> > >>>>>>>> >> >>
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> #3 no ERR writtent in mdout, ERR written in standard
> output
> > >>>>>>>> >> (nohup.out)
> > >>>>>>>> >> >> ----
> > >>>>>>>> >> >> cudaMemcpy GpuBuffer::Download failed unspecified launch
> > >>>>>>>> failure
> > >>>>>>>> >> >> ----
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> or you obtained some new/additional errs ?
> > >>>>>>>> >> >>
> > >>>>>>>> >> >>
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> M.
> > >>>>>>>> >> >>
> > >>>>>>>> >> >>
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> Dne Fri, 31 May 2013 17:30:57 +0200 filip fratev
> > >>>>>>>> >> <filipfratev.yahoo.com
> > >>>>>>>> >>
> > >>>>>>>> >> >> napsal/-a:
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> > Hi,
> > >>>>>>>> >> >> > This is what I obtained for 50K tests and "normal"
> > GTXTitan:
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > run1:
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >>
> > >>>>>>>> >
> > >>>>>>>> ------------------------------**------------------------------**
> > >>>>>>>> ------------------
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > A V E R A G E S O V E R 50 S T E P S
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> > >>>>>>>> 299.87
> > >>>>>>>> >> PRESS
> > >>>>>>>> >> >> > = 0.0
> > >>>>>>>> >> >> > Etot = -443237.1079 EKtot = 257679.9750
> EPtot
> > >>>>>>>> =
> > >>>>>>>> >> >> > -700917.0829
> > >>>>>>>> >> >> > BOND = 20193.1856 ANGLE = 53517.5432
> DIHED
> > >>>>>>>> =
> > >>>>>>>> >> >> > 23575.4648
> > >>>>>>>> >> >> > 1-4 NB = 21759.5524 1-4 EEL = 742552.5939
> > VDWAALS
> > >>>>>>>> =
> > >>>>>>>> >> >> > 96286.7714
> > >>>>>>>> >> >> > EELEC = -1658802.1941 EHBOND = 0.0000
> > >>>>>>>> RESTRAINT
> > >>>>>>>> =
> > >>>>>>>> >> >> > 0.0000
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >>
> > >>>>>>>> >
> > >>>>>>>>
> ------------------------------**------------------------------**
> > >>>>>>>> ------------------
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > R M S F L U C T U A T I O N S
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> > >>>>>>>> 0.33
> > >>>>>>>> >> PRESS
> > >>>>>>>> >> >> > = 0.0
> > >>>>>>>> >> >> > Etot = 11.2784 EKtot = 284.8999
> EPtot
> > >>>>>>>> =
> > >>>>>>>> >> >> > 289.0773
> > >>>>>>>> >> >> > BOND = 136.3417 ANGLE = 214.0054
> DIHED
> > >>>>>>>> =
> > >>>>>>>> >> >> > 59.4893
> > >>>>>>>> >> >> > 1-4 NB = 58.5891 1-4 EEL = 330.5400
> > VDWAALS
> > >>>>>>>> =
> > >>>>>>>> >> >> > 559.2079
> > >>>>>>>> >> >> > EELEC = 743.8771 EHBOND = 0.0000
> > >>>>>>>> RESTRAINT
> > >>>>>>>> =
> > >>>>>>>> >> >> > 0.0000
> > >>>>>>>> >> >> > |E(PBS) = 21.8119
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >>
> > >>>>>>>> >
> > >>>>>>>>
> ------------------------------**------------------------------**
> > >>>>>>>> ------------------
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > run2:
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >>
> > >>>>>>>> >
> > >>>>>>>> ------------------------------**------------------------------**
> > >>>>>>>> ------------------
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > A V E R A G E S O V E R 50 S T E P S
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> > >>>>>>>> 299.89
> > >>>>>>>> >> PRESS
> > >>>>>>>> >> >> > = 0.0
> > >>>>>>>> >> >> > Etot = -443240.0999 EKtot = 257700.0950
> EPtot
> > >>>>>>>> =
> > >>>>>>>> >> >> > -700940.1949
> > >>>>>>>> >> >> > BOND = 20241.9174 ANGLE = 53644.6694
> DIHED
> > >>>>>>>> =
> > >>>>>>>> >> >> > 23541.3737
> > >>>>>>>> >> >> > 1-4 NB = 21803.1898 1-4 EEL = 742754.2254
> > VDWAALS
> > >>>>>>>> =
> > >>>>>>>> >> >> > 96298.8308
> > >>>>>>>> >> >> > EELEC = -1659224.4013 EHBOND = 0.0000
> > >>>>>>>> RESTRAINT
> > >>>>>>>> =
> > >>>>>>>> >> >> > 0.0000
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >>
> > >>>>>>>> >
> > >>>>>>>>
> ------------------------------**------------------------------**
> > >>>>>>>> ------------------
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > R M S F L U C T U A T I O N S
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> > >>>>>>>> 0.41
> > >>>>>>>> >> PRESS
> > >>>>>>>> >> >> > = 0.0
> > >>>>>>>> >> >> > Etot = 10.7633 EKtot = 348.2819
> EPtot
> > >>>>>>>> =
> > >>>>>>>> >> >> > 353.9918
> > >>>>>>>> >> >> > BOND = 106.5314 ANGLE = 196.7052
> DIHED
> > >>>>>>>> =
> > >>>>>>>> >> >> > 69.7476
> > >>>>>>>> >> >> > 1-4 NB = 60.3435 1-4 EEL = 400.7466
> > VDWAALS
> > >>>>>>>> =
> > >>>>>>>> >> >> > 462.7763
> > >>>>>>>> >> >> > EELEC = 651.9857 EHBOND = 0.0000
> > >>>>>>>> RESTRAINT
> > >>>>>>>> =
> > >>>>>>>> >> >> > 0.0000
> > >>>>>>>> >> >> > |E(PBS) = 17.0642
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >>
> > >>>>>>>> >
> > >>>>>>>>
> ------------------------------**------------------------------**
> > >>>>>>>> ------------------
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >>
> > >>>>>>>> >
> > >>>>>>>> ------------------------------**------------------------------**
> > >>>>>>>> --------------------
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > ______________________________**__
> > >>>>>>>> >> >> > From: Marek Maly <marek.maly.ujep.cz>
> > >>>>>>>> >> >> > To: AMBER Mailing List <amber.ambermd.org>
> > >>>>>>>> >> >> > Sent: Friday, May 31, 2013 3:34 PM
> > >>>>>>>> >> >> > Subject: Re: [AMBER] experiences with EVGA GTX TITAN
> > >>>>>>>> Superclocked
> > >>>>>>>> -
> > >>>>>>>> >> >> > memtestG80 - UNDERclocking in Linux ?
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > Hi here are my 100K results for driver 313.30 (and still
> > >>>>>>>> Cuda
> > >>>>>>>> 5.0).
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > The results are rather similar to those obtained
> > >>>>>>>> >> >> > under my original driver 319.17 (see the first table
> > >>>>>>>> >> >> > which I sent in this thread).
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > M.
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> > Dne Fri, 31 May 2013 12:29:59 +0200 Marek Maly <
> > >>>>>>>> marek.maly.ujep.cz>
> > >>>>>>>> >> >> > napsal/-a:
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >> Hi,
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >> please try to run at lest 100K tests twice to verify
> > exact
> > >>>>>>>> >> >> >> reproducibility
> > >>>>>>>> >> >> >> of the results on the given card. If you find in any
> mdin
> > >>>>>>>> file
> > >>>>>>>> >> ig=-1
> > >>>>>>>> >> >> >> just
> > >>>>>>>> >> >> >> delete it to ensure that you are using the identical
> > random
> > >>>>>>>> seed
> > >>>>>>>> >> for
> > >>>>>>>> >> >> >> both
> > >>>>>>>> >> >> >> runs. You can eventually omit NUCLEOSOME test
> > >>>>>>>> >> >> >> as it is too much time consuming.
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >> Driver 310.44 ?????
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >> As far as I know the proper support for titans is from
> > >>>>>>>> version
> > >>>>>>>> > 313.26
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >> see e.g. here :
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >>
> > >>>>>>>> >
> > >>>>>>>> http://www.geeks3d.com/**20130306/nvidia-releases-r313-**
> > >>>>>>>> 26-for-linux-with-gtx-titan-**support/<
> >
> http://www.geeks3d.com/20130306/nvidia-releases-r313-26-for-linux-with-gtx-titan-support/
> > >
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >> BTW: On my site downgrade to drv. 313.30 did not solved
> > the
> > >>>>>>>> >> >> situation, I
> > >>>>>>>> >> >> >> will post
> > >>>>>>>> >> >> >> my results soon here.
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >> M.
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >> Dne Fri, 31 May 2013 12:21:21 +0200 ET <
> > >>>>>>>> sketchfoot.gmail.com>
> > >>>>>>>> >> >> napsal/-a:
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >>> ps. I have another install of amber on another
> computer
> > >>>>>>>> with
> > >>>>>>>> a
> > >>>>>>>> >> >> >>> different
> > >>>>>>>> >> >> >>> Titan and different Driver Version: 310.44.
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>> In the interests of thrashing the proverbial horse,
> I'll
> > >>>>>>>> run
> > >>>>>>>> the
> > >>>>>>>> >> >> >>> benchmark
> > >>>>>>>> >> >> >>> for 50k steps. :P
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>> br,
> > >>>>>>>> >> >> >>> g
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>> On 31 May 2013 11:17, ET <sketchfoot.gmail.com>
> wrote:
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>>> Hi, I just ran the Amber benchmark for the default
> > (10000
> > >>>>>>>> steps)
> > >>>>>>>> >> >> on my
> > >>>>>>>> >> >> >>>> Titan.
> > >>>>>>>> >> >> >>>>
> > >>>>>>>> >> >> >>>> Using sdiff -sB showed that the two runs were
> > completely
> > >>>>>>>> > identical.
> > >>>>>>>> >> >> >>>> I've
> > >>>>>>>> >> >> >>>> attached compressed files of the mdout & diff files.
> > >>>>>>>> >> >> >>>>
> > >>>>>>>> >> >> >>>> br,
> > >>>>>>>> >> >> >>>> g
> > >>>>>>>> >> >> >>>>
> > >>>>>>>> >> >> >>>>
> > >>>>>>>> >> >> >>>> On 30 May 2013 23:41, Marek Maly <marek.maly.ujep.cz
> >
> > >>>>>>>> wrote:
> > >>>>>>>> >> >> >>>>
> > >>>>>>>> >> >> >>>>> OK, let's see. The eventual downclocking I see as
> the
> > >>>>>>>> very
> > >>>>>>>> last
> > >>>>>>>> >> >> >>>>> possibility
> > >>>>>>>> >> >> >>>>> (if I don't decide for RMAing). But now still some
> > other
> > >>>>>>>> >> >> experiments
> > >>>>>>>> >> >> >>>>> are
> > >>>>>>>> >> >> >>>>> available :))
> > >>>>>>>> >> >> >>>>> I just started 100K tests under 313.30 driver. For
> > today
> > >>>>>>>> good
> > >>>>>>>> >> >> night
> > >>>>>>>> >> >> >>>>> ...
> > >>>>>>>> >> >> >>>>>
> > >>>>>>>> >> >> >>>>> M.
> > >>>>>>>> >> >> >>>>>
> > >>>>>>>> >> >> >>>>> Dne Fri, 31 May 2013 00:45:49 +0200 Scott Le Grand
> > >>>>>>>> >> >> >>>>> <varelse2005.gmail.com
> > >>>>>>>> >> >> >>>>> >
> > >>>>>>>> >> >> >>>>> napsal/-a:
> > >>>>>>>> >> >> >>>>>
> > >>>>>>>> >> >> >>>>> > It will be very interesting if this behavior
> > persists
> > >>>>>>>> after
> > >>>>>>>> >> >> >>>>> downclocking.
> > >>>>>>>> >> >> >>>>> >
> > >>>>>>>> >> >> >>>>> > But right now, Titan 0 *looks* hosed and Titan 1
> > >>>>>>>> *looks*
> > >>>>>>>> like
> > >>>>>>>> > it
> > >>>>>>>> >> >> >>>>> needs
> > >>>>>>>> >> >> >>>>> > downclocking...
> > >>>>>>>> >> >> >>>>> > On May 30, 2013 3:20 PM, "Marek Maly"
> > >>>>>>>> <marek.maly.ujep.cz>
> > >>>>>>>> >> >> wrote:
> > >>>>>>>> >> >> >>>>> >
> > >>>>>>>> >> >> >>>>> >> Hi all,
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> here are my results from the 500K steps 2 x
> > repeated
> > >>>>>>>> > benchmarks
> > >>>>>>>> >> >> >>>>> >> under 319.23 driver and still Cuda 5.0 (see the
> > >>>>>>>> attached
> > >>>>>>>> >> table
> > >>>>>>>> >> >> ).
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> It is hard to say if the results are better or
> > worse
> > >>>>>>>> than
> > >>>>>>>> in
> > >>>>>>>> > my
> > >>>>>>>> >> >> >>>>> >> previous 100K test under driver 319.17.
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> While results from Cellulose test were improved
> and
> > >>>>>>>> the
> > >>>>>>>> > TITAN_1
> > >>>>>>>> >> >> >>>>> card
> > >>>>>>>> >> >> >>>>> >> even
> > >>>>>>>> >> >> >>>>> >> successfully finished all 500K steps moreover
> with
> > >>>>>>>> exactly
> > >>>>>>>> >> the
> > >>>>>>>> >> >> >>>>> same
> > >>>>>>>> >> >> >>>>> >> final
> > >>>>>>>> >> >> >>>>> >> energy !
> > >>>>>>>> >> >> >>>>> >> (TITAN_0 at least finished more than 100K steps
> and
> > >>>>>>>> in
> > >>>>>>>> >> RUN_01
> > >>>>>>>> >> >> even
> > >>>>>>>> >> >> >>>>> more
> > >>>>>>>> >> >> >>>>> >> than 400K steps)
> > >>>>>>>> >> >> >>>>> >> In JAC_NPT test no GPU was able to finish at
> least
> > >>>>>>>> 100K
> > >>>>>>>> >> steps
> > >>>>>>>> >> >> and
> > >>>>>>>> >> >> >>>>> the
> > >>>>>>>> >> >> >>>>> >> results from JAC_NVE
> > >>>>>>>> >> >> >>>>> >> test are also not too much convincing.
> > FACTOR_IX_NVE
> > >>>>>>>> and
> > >>>>>>>> >> >> >>>>> FACTOR_IX_NPT
> > >>>>>>>> >> >> >>>>> >> were successfully
> > >>>>>>>> >> >> >>>>> >> finished with 100% reproducibility in
> > FACTOR_IX_NPT
> > >>>>>>>> case
> > >>>>>>>> >> (on
> > >>>>>>>> >> >> both
> > >>>>>>>> >> >> >>>>> >> cards)
> > >>>>>>>> >> >> >>>>> >> and almost
> > >>>>>>>> >> >> >>>>> >> 100% reproducibility in case of FACTOR_IX_NVE
> > (again
> > >>>>>>>> 100%
> > >>>>>>>> in
> > >>>>>>>> >> >> case
> > >>>>>>>> >> >> >>>>> of
> > >>>>>>>> >> >> >>>>> >> TITAN_1). TRPCAGE, MYOGLOBIN
> > >>>>>>>> >> >> >>>>> >> again finished without any problem with 100%
> > >>>>>>>> >> reproducibility.
> > >>>>>>>> >> >> >>>>> NUCLEOSOME
> > >>>>>>>> >> >> >>>>> >> test was not done
> > >>>>>>>> >> >> >>>>> >> this time due to high time requirements. If you
> > find
> > >>>>>>>> in
> > >>>>>>>> the
> > >>>>>>>> >> >> table
> > >>>>>>>> >> >> >>>>> >> positive
> > >>>>>>>> >> >> >>>>> >> number finishing with
> > >>>>>>>> >> >> >>>>> >> K (which means "thousands") it means the last
> > number
> > >>>>>>>> of
> > >>>>>>>> step
> > >>>>>>>> >> >> >>>>> written in
> > >>>>>>>> >> >> >>>>> >> mdout before crash.
> > >>>>>>>> >> >> >>>>> >> Below are all the 3 types of detected errs with
> > >>>>>>>> relevant
> > >>>>>>>> >> >> >>>>> systems/rounds
> > >>>>>>>> >> >> >>>>> >> where the given err
> > >>>>>>>> >> >> >>>>> >> appeared.
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> Now I will try just 100K tests under ETs
> favourite
> > >>>>>>>> driver
> > >>>>>>>> >> >> version
> > >>>>>>>> >> >> >>>>> 313.30
> > >>>>>>>> >> >> >>>>> >> :)) and then
> > >>>>>>>> >> >> >>>>> >> I will eventually try to experiment with cuda 5.5
> > >>>>>>>> which
> > >>>>>>>> I
> > >>>>>>>> >> >> already
> > >>>>>>>> >> >> >>>>> >> downloaded from the
> > >>>>>>>> >> >> >>>>> >> cuda zone ( I had to become cuda developer for
> this
> > >>>>>>>> :))
> > >>>>>>>> )
> > >>>>>>>> >> BTW
> > >>>>>>>> >> >> ET
> > >>>>>>>> >> >> >>>>> thanks
> > >>>>>>>> >> >> >>>>> >> for the frequency info !
> > >>>>>>>> >> >> >>>>> >> and I am still ( perhaps not alone :)) ) very
> > curious
> > >>>>>>>> about
> > >>>>>>>> >> >> your 2
> > >>>>>>>> >> >> >>>>> x
> > >>>>>>>> >> >> >>>>> >> repeated Amber benchmark tests with superclocked
> > >>>>>>>> Titan.
> > >>>>>>>> >> Indeed
> > >>>>>>>> >> >> >>>>> that
> > >>>>>>>> >> >> >>>>> I
> > >>>>>>>> >> >> >>>>> am
> > >>>>>>>> >> >> >>>>> >> very curious also about that Ross "hot" patch.
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> M.
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> ERRORS DETECTED DURING THE 500K steps tests with
> > >>>>>>>> driver
> > >>>>>>>> >> 319.23
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> #1 ERR writtent in mdout:
> > >>>>>>>> >> >> >>>>> >> ------
> > >>>>>>>> >> >> >>>>> >> | ERROR: max pairlist cutoff must be less than
> > unit
> > >>>>>>>> cell
> > >>>>>>>> >> max
> > >>>>>>>> >> >> >>>>> sphere
> > >>>>>>>> >> >> >>>>> >> radius!
> > >>>>>>>> >> >> >>>>> >> ------
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_1 JAC_NPT (at least 5000 steps
> > >>>>>>>> successfully
> > >>>>>>>> > done
> > >>>>>>>> >> >> >>>>> before
> > >>>>>>>> >> >> >>>>> >> crash)
> > >>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 JAC_NPT (at least 8000 steps
> > >>>>>>>> successfully
> > >>>>>>>> > done
> > >>>>>>>> >> >> >>>>> before
> > >>>>>>>> >> >> >>>>> >> crash)
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> #2 no ERR writtent in mdout, ERR written in
> > standard
> > >>>>>>>> output
> > >>>>>>>> >> >> >>>>> (nohup.out)
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> ----
> > >>>>>>>> >> >> >>>>> >> Error: unspecified launch failure launching
> kernel
> > >>>>>>>> >> kNLSkinTest
> > >>>>>>>> >> >> >>>>> >> cudaFree GpuBuffer::Deallocate failed unspecified
> > >>>>>>>> launch
> > >>>>>>>> >> >> failure
> > >>>>>>>> >> >> >>>>> >> ----
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_1 CELLULOSE_NVE (at least 437 000
> > steps
> > >>>>>>>> >> >> successfully
> > >>>>>>>> >> >> >>>>> done
> > >>>>>>>> >> >> >>>>> >> before crash)
> > >>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 JAC_NVE (at least 162 000 steps
> > >>>>>>>> >> successfully
> > >>>>>>>> >> >> done
> > >>>>>>>> >> >> >>>>> >> before
> > >>>>>>>> >> >> >>>>> >> crash)
> > >>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 CELLULOSE_NVE (at least 117 000
> > steps
> > >>>>>>>> >> >> successfully
> > >>>>>>>> >> >> >>>>> done
> > >>>>>>>> >> >> >>>>> >> before crash)
> > >>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_1 JAC_NVE (at least 119 000 steps
> > >>>>>>>> >> successfully
> > >>>>>>>> >> >> done
> > >>>>>>>> >> >> >>>>> >> before
> > >>>>>>>> >> >> >>>>> >> crash)
> > >>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_2 JAC_NVE (at least 43 000 steps
> > >>>>>>>> successfully
> > >>>>>>>> >> >> done
> > >>>>>>>> >> >> >>>>> before
> > >>>>>>>> >> >> >>>>> >> crash)
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> #3 no ERR writtent in mdout, ERR written in
> > standard
> > >>>>>>>> output
> > >>>>>>>> >> >> >>>>> (nohup.out)
> > >>>>>>>> >> >> >>>>> >> ----
> > >>>>>>>> >> >> >>>>> >> cudaMemcpy GpuBuffer::Download failed unspecified
> > >>>>>>>> launch
> > >>>>>>>> >> >> failure
> > >>>>>>>> >> >> >>>>> >> ----
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_1 JAC_NPT (at least 77 000 steps
> > >>>>>>>> successfully
> > >>>>>>>> >> >> done
> > >>>>>>>> >> >> >>>>> before
> > >>>>>>>> >> >> >>>>> >> crash)
> > >>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_2 JAC_NPT (at least 58 000 steps
> > >>>>>>>> successfully
> > >>>>>>>> >> >> done
> > >>>>>>>> >> >> >>>>> before
> > >>>>>>>> >> >> >>>>> >> crash)
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> Dne Thu, 30 May 2013 21:27:17 +0200 Scott Le
> Grand
> > >>>>>>>> >> >> >>>>> >> <varelse2005.gmail.com>
> > >>>>>>>> >> >> >>>>> >> napsal/-a:
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> Oops meant to send that to Jason...
> > >>>>>>>> >> >> >>>>> >>>
> > >>>>>>>> >> >> >>>>> >>> Anyway, before we all panic, we need to get
> K20's
> > >>>>>>>> behavior
> > >>>>>>>> >> >> >>>>> analyzed
> > >>>>>>>> >> >> >>>>> >>> here.
> > >>>>>>>> >> >> >>>>> >>> If it's deterministic, this truly is a hardware
> > >>>>>>>> issue. If
> > >>>>>>>> >> >> not,
> > >>>>>>>> >> >> >>>>> then
> > >>>>>>>> >> >> >>>>> it
> > >>>>>>>> >> >> >>>>> >>> gets interesting because 680 is deterministic as
> > far
> > >>>>>>>> as I
> > >>>>>>>> >> can
> > >>>>>>>> >> >> >>>>> tell...
> > >>>>>>>> >> >> >>>>> >>> On May 30, 2013 12:24 PM, "Scott Le Grand"
> > >>>>>>>> >> >> >>>>> <varelse2005.gmail.com>
> > >>>>>>>> >> >> >>>>> >>> wrote:
> > >>>>>>>> >> >> >>>>> >>>
> > >>>>>>>> >> >> >>>>> >>> If the errors are not deterministically
> > triggered,
> > >>>>>>>> they
> > >>>>>>>> >> >> probably
> > >>>>>>>> >> >> >>>>> >>> won't be
> > >>>>>>>> >> >> >>>>> >>>> fixed by the patch, alas...
> > >>>>>>>> >> >> >>>>> >>>> On May 30, 2013 12:15 PM, "Jason Swails"
> > >>>>>>>> >> >> >>>>> <jason.swails.gmail.com>
> > >>>>>>>> >> >> >>>>> >>>> wrote:
> > >>>>>>>> >> >> >>>>> >>>>
> > >>>>>>>> >> >> >>>>> >>>> Just a reminder to everyone based on what Ross
> > >>>>>>>> said:
> > >>>>>>>> >> there
> > >>>>>>>> >> >> is a
> > >>>>>>>> >> >> >>>>> >>>> pending
> > >>>>>>>> >> >> >>>>> >>>>> patch to pmemd.cuda that will be coming out
> > >>>>>>>> shortly
> > >>>>>>>> >> (maybe
> > >>>>>>>> >> >> even
> > >>>>>>>> >> >> >>>>> >>>>> within
> > >>>>>>>> >> >> >>>>> >>>>> hours). It's entirely possible that several
> of
> > >>>>>>>> these
> > >>>>>>>> > errors
> > >>>>>>>> >> >> >>>>> are
> > >>>>>>>> >> >> >>>>> >>>>> fixed
> > >>>>>>>> >> >> >>>>> >>>>> by
> > >>>>>>>> >> >> >>>>> >>>>> this patch.
> > >>>>>>>> >> >> >>>>> >>>>>
> > >>>>>>>> >> >> >>>>> >>>>> All the best,
> > >>>>>>>> >> >> >>>>> >>>>> Jason
> > >>>>>>>> >> >> >>>>> >>>>>
> > >>>>>>>> >> >> >>>>> >>>>>
> > >>>>>>>> >> >> >>>>> >>>>> On Thu, May 30, 2013 at 2:46 PM, filip fratev
> <
> > >>>>>>>> >> >> >>>>> filipfratev.yahoo.com>
> > >>>>>>>> >> >> >>>>> >>>>> wrote:
> > >>>>>>>> >> >> >>>>> >>>>>
> > >>>>>>>> >> >> >>>>> >>>>> > I have observed the same crashes from time
> to
> > >>>>>>>> time. I
> > >>>>>>>> > will
> > >>>>>>>> >> >> >>>>> run
> > >>>>>>>> >> >> >>>>> >>>>> cellulose
> > >>>>>>>> >> >> >>>>> >>>>> > nve for 100k and will past results here.
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> > All the best,
> > >>>>>>>> >> >> >>>>> >>>>> > Filip
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> > ______________________________****__
> > >>>>>>>> >> >> >>>>> >>>>> > From: Scott Le Grand <
> varelse2005.gmail.com>
> > >>>>>>>> >> >> >>>>> >>>>> > To: AMBER Mailing List <amber.ambermd.org>
> > >>>>>>>> >> >> >>>>> >>>>> > Sent: Thursday, May 30, 2013 9:01 PM
> > >>>>>>>> >> >> >>>>> >>>>> > Subject: Re: [AMBER] experiences with EVGA
> GTX
> > >>>>>>>> TITAN
> > >>>>>>>> >> >> >>>>> Superclocked
> > >>>>>>>> >> >> >>>>> -
> > >>>>>>>> >> >> >>>>> >>>>> > memtestG80 - UNDERclocking in Linux ?
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> > Run cellulose nve for 100k iterations twice
> .
> > >>>>>>>> If
> > >>>>>>>> the
> > >>>>>>>> >> >> final
> > >>>>>>>> >> >> >>>>> >>>>> energies
> > >>>>>>>> >> >> >>>>> >>>>> don't
> > >>>>>>>> >> >> >>>>> >>>>> > match, you have a hardware issue. No need
> to
> > >>>>>>>> play
> > >>>>>>>> with
> > >>>>>>>> >> >> ntpr
> > >>>>>>>> >> >> >>>>> or
> > >>>>>>>> >> >> >>>>> any
> > >>>>>>>> >> >> >>>>> >>>>> other
> > >>>>>>>> >> >> >>>>> >>>>> > variable.
> > >>>>>>>> >> >> >>>>> >>>>> > On May 30, 2013 10:58 AM, <
> > pavel.banas.upol.cz>
> > >>>>>>>> wrote:
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > Dear all,
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > I would also like to share one of my
> > >>>>>>>> experience
> > >>>>>>>> with
> > >>>>>>>> >> >> titan
> > >>>>>>>> >> >> >>>>> >>>>> cards. We
> > >>>>>>>> >> >> >>>>> >>>>> have
> > >>>>>>>> >> >> >>>>> >>>>> > > one gtx titan card and with one system
> (~55k
> > >>>>>>>> atoms,
> > >>>>>>>> > NVT,
> > >>>>>>>> >> >> >>>>> >>>>> RNA+waters)
> > >>>>>>>> >> >> >>>>> >>>>> we
> > >>>>>>>> >> >> >>>>> >>>>> > run
> > >>>>>>>> >> >> >>>>> >>>>> > > into same troubles you are describing. I
> was
> > >>>>>>>> also
> > >>>>>>>> >> >> playing
> > >>>>>>>> >> >> >>>>> with
> > >>>>>>>> >> >> >>>>> >>>>> ntpr
> > >>>>>>>> >> >> >>>>> >>>>> to
> > >>>>>>>> >> >> >>>>> >>>>> > > figure out what is going on, step by
> step. I
> > >>>>>>>> >> understand
> > >>>>>>>> >> >> >>>>> that
> > >>>>>>>> >> >> >>>>> the
> > >>>>>>>> >> >> >>>>> >>>>> code
> > >>>>>>>> >> >> >>>>> >>>>> is
> > >>>>>>>> >> >> >>>>> >>>>> > > using different routines for calculation
> > >>>>>>>> >> >> energies+forces or
> > >>>>>>>> >> >> >>>>> only
> > >>>>>>>> >> >> >>>>> >>>>> forces.
> > >>>>>>>> >> >> >>>>> >>>>> > > The
> > >>>>>>>> >> >> >>>>> >>>>> > > simulations of other systems are perfectly
> > >>>>>>>> stable,
> > >>>>>>>> >> >> running
> > >>>>>>>> >> >> >>>>> for
> > >>>>>>>> >> >> >>>>> >>>>> days
> > >>>>>>>> >> >> >>>>> >>>>> and
> > >>>>>>>> >> >> >>>>> >>>>> > > weeks. Only that particular system
> > >>>>>>>> systematically
> > >>>>>>>> >> ends
> > >>>>>>>> >> >> up
> > >>>>>>>> >> >> >>>>> with
> > >>>>>>>> >> >> >>>>> >>>>> this
> > >>>>>>>> >> >> >>>>> >>>>> > error.
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > However, there was one interesting issue.
> > >>>>>>>> When I
> > >>>>>>>> set
> > >>>>>>>> >> >> >>>>> ntpr=1,
> > >>>>>>>> >> >> >>>>> the
> > >>>>>>>> >> >> >>>>> >>>>> error
> > >>>>>>>> >> >> >>>>> >>>>> > > vanished (systematically in multiple runs)
> > and
> > >>>>>>>> the
> > >>>>>>>> >> >> >>>>> simulation
> > >>>>>>>> >> >> >>>>> was
> > >>>>>>>> >> >> >>>>> >>>>> able to
> > >>>>>>>> >> >> >>>>> >>>>> > > run for more than millions of steps (I was
> > not
> > >>>>>>>> let
> > >>>>>>>> it
> > >>>>>>>> >> >> >>>>> running
> > >>>>>>>> >> >> >>>>> for
> > >>>>>>>> >> >> >>>>> >>>>> weeks
> > >>>>>>>> >> >> >>>>> >>>>> > as
> > >>>>>>>> >> >> >>>>> >>>>> > > in the meantime I shifted that simulation
> to
> > >>>>>>>> other
> > >>>>>>>> >> card
> > >>>>>>>> >> >> -
> > >>>>>>>> >> >> >>>>> need
> > >>>>>>>> >> >> >>>>> >>>>> data,
> > >>>>>>>> >> >> >>>>> >>>>> not
> > >>>>>>>> >> >> >>>>> >>>>> > > testing). All other setting of ntpr
> failed.
> > >>>>>>>> As I
> > >>>>>>>> read
> > >>>>>>>> >> >> this
> > >>>>>>>> >> >> >>>>> >>>>> discussion, I
> > >>>>>>>> >> >> >>>>> >>>>> > > tried to set ene_avg_sampling=1 with some
> > high
> > >>>>>>>> value
> > >>>>>>>> >> of
> > >>>>>>>> >> >> >>>>> ntpr
> > >>>>>>>> >> >> >>>>> (I
> > >>>>>>>> >> >> >>>>> >>>>> expected
> > >>>>>>>> >> >> >>>>> >>>>> > > that this will shift the code to
> permanently
> > >>>>>>>> use
> > >>>>>>>> the
> > >>>>>>>> >> >> >>>>> >>>>> force+energies
> > >>>>>>>> >> >> >>>>> >>>>> part
> > >>>>>>>> >> >> >>>>> >>>>> > of
> > >>>>>>>> >> >> >>>>> >>>>> > > the code, similarly to ntpr=1), but the
> > error
> > >>>>>>>> >> occurred
> > >>>>>>>> >> >> >>>>> again.
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > I know it is not very conclusive for
> finding
> > >>>>>>>> out
> > >>>>>>>> what
> > >>>>>>>> > is
> > >>>>>>>> >> >> >>>>> >>>>> happening,
> > >>>>>>>> >> >> >>>>> >>>>> at
> > >>>>>>>> >> >> >>>>> >>>>> > > least
> > >>>>>>>> >> >> >>>>> >>>>> > > not for me. Do you have any idea, why
> ntpr=1
> > >>>>>>>> might
> > >>>>>>>> > help?
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > best regards,
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > Pavel
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > --
> > >>>>>>>> >> >> >>>>> >>>>> > > Pavel Banáš
> > >>>>>>>> >> >> >>>>> >>>>> > > pavel.banas.upol.cz
> > >>>>>>>> >> >> >>>>> >>>>> > > Department of Physical Chemistry,
> > >>>>>>>> >> >> >>>>> >>>>> > > Palacky University Olomouc
> > >>>>>>>> >> >> >>>>> >>>>> > > Czech Republic
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > ---------- Původní zpráva ----------
> > >>>>>>>> >> >> >>>>> >>>>> > > Od: Jason Swails <jason.swails.gmail.com>
> > >>>>>>>> >> >> >>>>> >>>>> > > Datum: 29. 5. 2013
> > >>>>>>>> >> >> >>>>> >>>>> > > Předmět: Re: [AMBER] experiences with EVGA
> > GTX
> > >>>>>>>> TITAN
> > >>>>>>>> >> >> >>>>> >>>>> Superclocked -
> > >>>>>>>> >> >> >>>>> >>>>> > > memtestG
> > >>>>>>>> >> >> >>>>> >>>>> > > 80 - UNDERclocking in Linux ?
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > "I'll answer a little bit:
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > NTPR=10 Etot after 2000 steps
> > >>>>>>>> >> >> >>>>> >>>>> > > >
> > >>>>>>>> >> >> >>>>> >>>>> > > > -443256.6711
> > >>>>>>>> >> >> >>>>> >>>>> > > > -443256.6711
> > >>>>>>>> >> >> >>>>> >>>>> > > >
> > >>>>>>>> >> >> >>>>> >>>>> > > > NTPR=200 Etot after 2000 steps
> > >>>>>>>> >> >> >>>>> >>>>> > > >
> > >>>>>>>> >> >> >>>>> >>>>> > > > -443261.0705
> > >>>>>>>> >> >> >>>>> >>>>> > > > -443261.0705
> > >>>>>>>> >> >> >>>>> >>>>> > > >
> > >>>>>>>> >> >> >>>>> >>>>> > > > Any idea why energies should depend on
> > >>>>>>>> frequency
> > >>>>>>>> of
> > >>>>>>>> >> >> >>>>> energy
> > >>>>>>>> >> >> >>>>> >>>>> records
> > >>>>>>>> >> >> >>>>> >>>>> > (NTPR)
> > >>>>>>>> >> >> >>>>> >>>>> > > ?
> > >>>>>>>> >> >> >>>>> >>>>> > > >
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > It is a subtle point, but the answer is
> > >>>>>>>> 'different
> > >>>>>>>> >> code
> > >>>>>>>> >> >> >>>>> paths.'
> > >>>>>>>> >> >> >>>>> >>>>> In
> > >>>>>>>> >> >> >>>>> >>>>> > > general, it is NEVER necessary to compute
> > the
> > >>>>>>>> actual
> > >>>>>>>> >> >> energy
> > >>>>>>>> >> >> >>>>> of a
> > >>>>>>>> >> >> >>>>> >>>>> molecule
> > >>>>>>>> >> >> >>>>> >>>>> > > during the course of standard molecular
> > >>>>>>>> dynamics
> > >>>>>>>> (by
> > >>>>>>>> >> >> >>>>> analogy, it
> > >>>>>>>> >> >> >>>>> >>>>> is
> > >>>>>>>> >> >> >>>>> >>>>> NEVER
> > >>>>>>>> >> >> >>>>> >>>>> > > necessary to compute atomic forces during
> > the
> > >>>>>>>> course
> > >>>>>>>> >> of
> > >>>>>>>> >> >> >>>>> random
> > >>>>>>>> >> >> >>>>> >>>>> Monte
> > >>>>>>>> >> >> >>>>> >>>>> > Carlo
> > >>>>>>>> >> >> >>>>> >>>>> > > sampling).
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > For performance's sake, then, pmemd.cuda
> > >>>>>>>> computes
> > >>>>>>>> >> only
> > >>>>>>>> >> >> the
> > >>>>>>>> >> >> >>>>> force
> > >>>>>>>> >> >> >>>>> >>>>> when
> > >>>>>>>> >> >> >>>>> >>>>> > > energies are not requested, leading to a
> > >>>>>>>> different
> > >>>>>>>> >> >> order of
> > >>>>>>>> >> >> >>>>> >>>>> operations
> > >>>>>>>> >> >> >>>>> >>>>> > for
> > >>>>>>>> >> >> >>>>> >>>>> > > those runs. This difference ultimately
> > causes
> > >>>>>>>> >> >> divergence.
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > To test this, try setting the variable
> > >>>>>>>> >> >> ene_avg_sampling=10
> > >>>>>>>> >> >> >>>>> in
> > >>>>>>>> >> >> >>>>> the
> > >>>>>>>> >> >> >>>>> >>>>> &cntrl
> > >>>>>>>> >> >> >>>>> >>>>> > > section. This will force pmemd.cuda to
> > compute
> > >>>>>>>> >> energies
> > >>>>>>>> >> >> >>>>> every 10
> > >>>>>>>> >> >> >>>>> >>>>> steps
> > >>>>>>>> >> >> >>>>> >>>>> > > (for energy averaging), which will in turn
> > >>>>>>>> make
> > >>>>>>>> the
> > >>>>>>>> >> >> >>>>> followed
> > >>>>>>>> >> >> >>>>> code
> > >>>>>>>> >> >> >>>>> >>>>> path
> > >>>>>>>> >> >> >>>>> >>>>> > > identical for any multiple-of-10 value of
> > >>>>>>>> ntpr.
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > > --
> > >>>>>>>> >> >> >>>>> >>>>> > > Jason M. Swails
> > >>>>>>>> >> >> >>>>> >>>>> > > Quantum Theory Project,
> > >>>>>>>> >> >> >>>>> >>>>> > > University of Florida
> > >>>>>>>> >> >> >>>>> >>>>> > > Ph.D. Candidate
> > >>>>>>>> >> >> >>>>> >>>>> > > 352-392-4032
> > >>>>>>>> >> >> >>>>> >>>>> > > ______________________________**
> > >>>>>>>> **_________________
> > >>>>>>>> >> >> >>>>> >>>>> > > AMBER mailing list
> > >>>>>>>> >> >> >>>>> >>>>> > > AMBER.ambermd.org
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> > >>>>>>>> mailman/listinfo/amber<
> > http://lists.ambermd.org/**mailman/listinfo/amber>
> > >>>>>>>> <
> > >>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> "
> > >>>>>>>> >> >> >>>>> >>>>> > > ______________________________**
> > >>>>>>>> **_________________
> > >>>>>>>> >> >> >>>>> >>>>> > > AMBER mailing list
> > >>>>>>>> >> >> >>>>> >>>>> > > AMBER.ambermd.org
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> > >>>>>>>> mailman/listinfo/amber<
> > http://lists.ambermd.org/**mailman/listinfo/amber>
> > >>>>>>>> <
> > >>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> > >
> > >>>>>>>> >> >> >>>>> >>>>> > ______________________________**
> > >>>>>>>> **_________________
> > >>>>>>>> >> >> >>>>> >>>>> > AMBER mailing list
> > >>>>>>>> >> >> >>>>> >>>>> > AMBER.ambermd.org
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> > >>>>>>>> mailman/listinfo/amber<
> > http://lists.ambermd.org/**mailman/listinfo/amber>
> > >>>>>>>> <
> > >>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> > ______________________________**
> > >>>>>>>> **_________________
> > >>>>>>>> >> >> >>>>> >>>>> > AMBER mailing list
> > >>>>>>>> >> >> >>>>> >>>>> > AMBER.ambermd.org
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> > >>>>>>>> mailman/listinfo/amber<
> > http://lists.ambermd.org/**mailman/listinfo/amber>
> > >>>>>>>> <
> > >>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >
> > >>>>>>>> >> >> >>>>> >>>>> >
> > >>>>>>>> >> >> >>>>> >>>>>
> > >>>>>>>> >> >> >>>>> >>>>>
> > >>>>>>>> >> >> >>>>> >>>>>
> > >>>>>>>> >> >> >>>>> >>>>> --
> > >>>>>>>> >> >> >>>>> >>>>> Jason M. Swails
> > >>>>>>>> >> >> >>>>> >>>>> Quantum Theory Project,
> > >>>>>>>> >> >> >>>>> >>>>> University of Florida
> > >>>>>>>> >> >> >>>>> >>>>> Ph.D. Candidate
> > >>>>>>>> >> >> >>>>> >>>>> 352-392-4032
> > >>>>>>>> >> >> >>>>> >>>>> ______________________________**
> > >>>>>>>> **_________________
> > >>>>>>>> >> >> >>>>> >>>>> AMBER mailing list
> > >>>>>>>> >> >> >>>>> >>>>> AMBER.ambermd.org
> > >>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> > >>>>>>>> mailman/listinfo/amber<
> > http://lists.ambermd.org/**mailman/listinfo/amber>
> > >>>>>>>> <
> > >>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >
> > >>>>>>>> >> >> >>>>> >>>>>
> > >>>>>>>> >> >> >>>>> >>>>>
> > >>>>>>>> >> >> >>>>> >>>> ______________________________**
> > >>>>>>>> **_________________
> > >>>>>>>> >> >> >>>>> >>> AMBER mailing list
> > >>>>>>>> >> >> >>>>> >>> AMBER.ambermd.org
> > >>>>>>>> >> >> >>>>> >>>
> > http://lists.ambermd.org/****mailman/listinfo/amber<
> > http://lists.ambermd.org/**mailman/listinfo/amber>
> > >>>>>>>> <
> > >>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >
> > >>>>>>>> >> >> >>>>> >>>
> > >>>>>>>> >> >> >>>>> >>> __________ Informace od ESET NOD32 Antivirus,
> > verze
> > >>>>>>>> >> databaze
> > >>>>>>>> >> >> 8394
> > >>>>>>>> >> >> >>>>> >>> (20130530) __________
> > >>>>>>>> >> >> >>>>> >>>
> > >>>>>>>> >> >> >>>>> >>> Tuto zpravu proveril ESET NOD32 Antivirus.
> > >>>>>>>> >> >> >>>>> >>>
> > >>>>>>>> >> >> >>>>> >>> http://www.eset.cz
> > >>>>>>>> >> >> >>>>> >>>
> > >>>>>>>> >> >> >>>>> >>>
> > >>>>>>>> >> >> >>>>> >>>
> > >>>>>>>> >> >> >>>>> >>>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >> --
> > >>>>>>>> >> >> >>>>> >> Tato zpráva byla vytvořena převratným poštovním
> > >>>>>>>> klientem
> > >>>>>>>> > Opery:
> > >>>>>>>> >> >> >>>>> >> http://www.opera.com/mail/
> > >>>>>>>> >> >> >>>>> >> ______________________________**_________________
> > >>>>>>>> >> >> >>>>> >> AMBER mailing list
> > >>>>>>>> >> >> >>>>> >> AMBER.ambermd.org
> > >>>>>>>> >> >> >>>>> >>
> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> >>
> > >>>>>>>> >> >> >>>>> > ______________________________**_________________
> > >>>>>>>> >> >> >>>>> > AMBER mailing list
> > >>>>>>>> >> >> >>>>> > AMBER.ambermd.org
> > >>>>>>>> >> >> >>>>> > http://lists.ambermd.org/**mailman/listinfo/amber
> <
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >> >> >>>>> >
> > >>>>>>>> >> >> >>>>> > __________ Informace od ESET NOD32 Antivirus,
> verze
> > >>>>>>>> databaze
> > >>>>>>>> >> >> 8394
> > >>>>>>>> >> >> >>>>> > (20130530) __________
> > >>>>>>>> >> >> >>>>> >
> > >>>>>>>> >> >> >>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
> > >>>>>>>> >> >> >>>>> >
> > >>>>>>>> >> >> >>>>> > http://www.eset.cz
> > >>>>>>>> >> >> >>>>> >
> > >>>>>>>> >> >> >>>>> >
> > >>>>>>>> >> >> >>>>> >
> > >>>>>>>> >> >> >>>>>
> > >>>>>>>> >> >> >>>>>
> > >>>>>>>> >> >> >>>>> --
> > >>>>>>>> >> >> >>>>> Tato zpráva byla vytvořena převratným poštovním
> > klientem
> > >>>>>>>> Opery:
> > >>>>>>>> >> >> >>>>> http://www.opera.com/mail/
> > >>>>>>>> >> >> >>>>>
> > >>>>>>>> >> >> >>>>> ______________________________**_________________
> > >>>>>>>> >> >> >>>>> AMBER mailing list
> > >>>>>>>> >> >> >>>>> AMBER.ambermd.org
> > >>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >> >> >>>>>
> > >>>>>>>> >> >> >>>>
> > >>>>>>>> >> >> >>>>
> > >>>>>>>> >> >> >>> ______________________________**_________________
> > >>>>>>>> >> >> >>> AMBER mailing list
> > >>>>>>>> >> >> >>> AMBER.ambermd.org
> > >>>>>>>> >> >> >>> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>> __________ Informace od ESET NOD32 Antivirus, verze
> > >>>>>>>> databaze
> > >>>>>>>> 8395
> > >>>>>>>> >> >> >>> (20130531) __________
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>> Tuto zpravu proveril ESET NOD32 Antivirus.
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>> http://www.eset.cz
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>>
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >>
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >> >
> > >>>>>>>> >> >>
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> --
> > >>>>>>>> >> >> Tato zpráva byla vytvořena převratným poštovním klientem
> > >>>>>>>> Opery:
> > >>>>>>>> >> >> http://www.opera.com/mail/
> > >>>>>>>> >> >>
> > >>>>>>>> >> >> ______________________________**_________________
> > >>>>>>>> >> >> AMBER mailing list
> > >>>>>>>> >> >> AMBER.ambermd.org
> > >>>>>>>> >> >> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >> >>
> > >>>>>>>> >> >
> > >>>>>>>> >> >
> > >>>>>>>> >> >
> > >>>>>>>> >> > __________ Informace od ESET NOD32 Antivirus, verze
> databaze
> > >>>>>>>> 8397
> > >>>>>>>> >> > (20130531) __________
> > >>>>>>>> >> >
> > >>>>>>>> >> > Tuto zpravu proveril ESET NOD32 Antivirus.
> > >>>>>>>> >> >
> > >>>>>>>> >> > GB_out_plus_diff_Files.tar.gz - poskozeny archiv
> > >>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> > >>>>>>>> >> GB_out_plus_diff_Files.tar
> > >>>>>>>> >> > - poskozeny archiv
> > >>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> > >>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
> > >>>>>>>> GB_out_plus_diff_Files.tar.gz -
> > >>>>>>>> >> > poskozeny archiv
> > >>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> > >>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
> > >>>>>>>> GB_out_plus_diff_Files.tar.gz >
> > >>>>>>>> >> GZIP
> > >>>>>>>> >> > > GB_out_plus_diff_Files.tar - poskozeny archiv
> > >>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> > >>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
> > >>>>>>>> GB_out_plus_diff_Files.tar.gz >
> > >>>>>>>> >> GZIP
> > >>>>>>>> >> > > GB_out_plus_diff_Files.tar > TAR >
> > >>>>>>>> GB_nucleosome-sim3.mdout-full -
> > >>>>>>>> >> > vyskytl se problem pri cteni archivu
> > >>>>>>>> >> > PME_out_plus_diff_Files.tar.gz - poskozeny archiv
> > >>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> > >>>>>>>> >> > PME_out_plus_diff_Files.tar - poskozeny archiv
> > >>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> > >>>>>>>> >> > PME_out_plus_diff_Files.tar > TAR >
> > >>>>>>>> PME_out_plus_diff_Files.tar.gz -
> > >>>>>>>> >> > poskozeny archiv
> > >>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> > >>>>>>>> >> > PME_out_plus_diff_Files.tar > TAR >
> > >>>>>>>> PME_out_plus_diff_Files.tar.gz >
> > >>>>>>>> >> > GZIP > PME_out_plus_diff_Files.tar - poskozeny archiv
> > >>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> > >>>>>>>> >> > PME_out_plus_diff_Filestar > TAR >
> > >>>>>>>> PME_out_plus_diff_Files.tar.gz
> > >>>>>>>> >
> > >>>>>>>> >> GZIP
> > >>>>>>>> >> > > PME_out_plus_diff_Files.tar > TAR >
> > >>>>>>>> >> > PME_JAC_production_NPT-sim3.**mdout-full - vyskytl se
> problem
> > >>>>>>>> pri
> > >>>>>>>> cteni
> > >>>>>>>> >> > archivu
> > >>>>>>>> >> >
> > >>>>>>>> >> > http://www.eset.cz
> > >>>>>>>> >> >
> > >>>>>>>> >>
> > >>>>>>>> >>
> > >>>>>>>> >> --
> > >>>>>>>> >> Tato zpráva byla vytvořena převratným poštovním klientem
> Opery:
> > >>>>>>>> >> http://www.opera.com/mail/
> > >>>>>>>> >>
> > >>>>>>>> >> ______________________________**_________________
> > >>>>>>>> >> AMBER mailing list
> > >>>>>>>> >> AMBER.ambermd.org
> > >>>>>>>> >> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> > ______________________________**_________________
> > >>>>>>>> > AMBER mailing list
> > >>>>>>>> > AMBER.ambermd.org
> > >>>>>>>> > http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>> >
> > >>>>>>>> > __________ Informace od ESET NOD32 Antivirus, verze databaze
> > 8398
> > >>>>>>>> > (20130531) __________
> > >>>>>>>> >
> > >>>>>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
> > >>>>>>>> >
> > >>>>>>>> > http://www.eset.cz
> > >>>>>>>> >
> > >>>>>>>> >
> > >>>>>>>> >
> > >>>>>>>>
> > >>>>>>>>
> > >>>>>>>> --
> > >>>>>>>> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> > >>>>>>>> http://www.opera.com/mail/
> > >>>>>>>>
> > >>>>>>>> ______________________________**_________________
> > >>>>>>>> AMBER mailing list
> > >>>>>>>> AMBER.ambermd.org
> > >>>>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> > http://lists.ambermd.org/mailman/listinfo/amber>
> > >>>>>>>>
> > >>>>>>>>
> > >>>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>> __________ Informace od ESET NOD32 Antivirus, verze databaze 8401
> > >>>>>> (20130601) __________
> > >>>>>>
> > >>>>>> Tuto zpravu proveril ESET NOD32 Antivirus.
> > >>>>>>
> > >>>>>> http://www.eset.cz
> > >>>>>>
> > >>>>>>
> > >>>>>
> > >>>>>
> > >>>>
> > >>>> --
> > >>>> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> > >>>> http://www.opera.com/mail/
> > >>>>
> > >>>> _______________________________________________
> > >>>> AMBER mailing list
> > >>>> AMBER.ambermd.org
> > >>>> http://lists.ambermd.org/mailman/listinfo/amber
> > >>>>
> > >>>>
> > >>>
> > >>
> > >
> > _______________________________________________
> > AMBER mailing list
> > AMBER.ambermd.org
> > http://lists.ambermd.org/mailman/listinfo/amber
> >
> _______________________________________________
> AMBER mailing list
> AMBER.ambermd.org
> http://lists.ambermd.org/mailman/listinfo/amber
>
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Sun Jun 02 2013 - 10:00:04 PDT
Custom Search