Re: [AMBER] experiences with EVGA GTX TITAN Superclocked - memtestG80 - UNDERclocking in Linux ?

From: Jonathan Gough <jonathan.d.gough.gmail.com>
Date: Wed, 19 Jun 2013 12:34:36 -0400

FWIW I posted GTX 780 results

here http://archive.ambermd.org/201306/0207.html

and here

http://archive.ambermd.org/201306/0211.html


If you would like me to test anything else, let me know.

Would Nvidia be willing to trade me a GTX 780 for my Titan?



On Wed, Jun 19, 2013 at 11:50 AM, Scott Le Grand <varelse2005.gmail.com>wrote:

> Hey Marek,
> No updates per se. I had a theory about what was going on that proved to
> be wrong after testing, but I'm still waiting on NVIDIA to report something
> beyond having reproed the problem.
>
> Really really really interested in GTX 780 data right now...
>
>
>
> On Wed, Jun 19, 2013 at 8:20 AM, Marek Maly <marek.maly.ujep.cz> wrote:
>
> > Hi all,
> >
> > just a small update from my site.
> >
> > As I have yesterday obtained announcement that the CUDA 5.5 is
> > now available for public (not just for developers).
> >
> > I downloaded it from here:
> >
> > https://developer.nvidia.com/**cuda-pre-production<
> https://developer.nvidia.com/cuda-pre-production>
> >
> > It is still "just" release candidate ( as all Amber/Titan club members
> > perfectly know :)) ).
> >
> > So I installed this newest release and recompiled Amber cuda code.
> >
> > I was hoping that maybe there was "silently" incorporated some
> > improvement (e.g. in cuFFT) as the result e.g. of Scott's bug report.
> >
> > The results of my 100K tests are attached. It seems that comparing to my
> > latest
> > tests with CUDA 5.5. release candidate from June 3rd (when it was
> > accessible just for CUDA developers in the form of *.run binary
> installer)
> > there
> > is some slight improvement - e.g. my more stable TITAN was able to finish
> > successfully
> > all the 100K tests including Cellulose twice. But there is still an issue
> > with JAC NVE/NPT irreproducible results. On my "less stable" TITAN the
> > results are slightly better
> > then those older ones as well but still not err free (JAC/CELLULOSE) -
> see
> > attached file.
> >
> > FACTOR IX NVE/NPT finished again with 100% reproducibility on both GPUs
> as
> > usually.
> >
> > Scott, do you have any update regarding the "cuFFT"/TITAN issue which you
> > reported/described
> > to NVIDIA guys ? The latest info from you regarding this story was, that
> > they were able to
> > reproduce the "cuFFT"/TITAN error as well. Do you have any more recent
> > information ? How long
> > time it might take to NVIDIA developers to fully solve such problem in
> > your opinion ?
> >
> > Another thing. It seems that you successfully solved the "GB/TITAN"
> > problem in case of bigger molecular systems, here is your relevant
> message
> > form June 7th.
> >
> > ------------------------------**----------------
> >
> > Really really interesting...
> >
> > I seem to have found a fix for the GB issues on my Titan - not so
> > surprisingly, it's the same fix as on GTX4xx/GTX5xx...
> >
> > But this doesn't yet explain the weirdness with cuFFT so we're not done
> > here yet...
> > ------------------------------**---------------
> >
> > It was already after the latest Amber12 bugfix18 was released and there
> > was no additional
> > bugfix released from that moment. So the "GB/TITAN" patch will be
> > released later maybe as the part of some bigger bugfix ? Or you simply
> > additionally included it into bugfix 18 after it's release ?
> >
> >
> > My last question maybe deserves the new separate thread, but anyway would
> > be interesting
> > to have some information how "Amber-stable" are GTX780 comparing to
> TITANS
> > (of course based
> > on experience of more users or on testing more than 1 or 2 GTX780 GPUs).
> >
> > Best wishes,
> >
> > Marek
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> > Dne Mon, 03 Jun 2013 01:57:36 +0200 Marek Maly <marek.maly.ujep.cz>
> > napsal/-a:
> >
> >
> > Hi here are my results with CUDA 5.5
> >> (Total energy at step 100K(PME)/1000K(GB) (driver 319.23, Amber12 bugfix
> >> 18 applied, cuda 5.5))
> >>
> >>
> >> No significant differences comparing the previous test with CUDA 5.0
> >> (I also added those data to the attached table with CUDA 5.5 test).
> >>
> >> Still the same trend instability in JAC tests, perfect stability and
> >> reproducibility
> >> in FACTOR_IX tests (interesting isn't it ? especially if we consider 23K
> >> atoms
> >> in JAC case and 90K atoms in case of FACTOR_IX). Again the same crashes
> in
> >> CELLULOSE
> >> test now also in case of TITAN_1. Also in stable and reproducible
> >> FACTOR_IX slightly
> >> changed the final energy values comparing to CUDA 5.0 case.
> >>
> >> GB simulations (1M steps) again perfectly stable and reproducible.
> >>
> >> So to conclude, Scott we trust you :)) !
> >>
> >> If you have any idea what to try else (except GPU bios editing, perhaps
> >> too
> >> premature step at this moment) let me know. I got just last idea,
> >> which could be perhaps to try change rand seed and see if it has any
> >> influence in actual trends (e.g. JAC versus FACTOR_IX).
> >>
> >> TO ET : I am curious about your test in single GPU configuration.
> >> Regarding
> >> to your Win tests, in my opinion it is just wasting of time. They
> perhaps
> >> tells
> >> you just something about the GPU performance not about the eventual GPU
> >> "soft" errs.
> >>
> >> If intensive memtestG80 and/or cuda_memtest results were negative there
> is
> >> in my opinion
> >> very unlikely that Win performace testers will find any errs, but I am
> not
> >> an expert
> >> here ...
> >>
> >> Anyway If you learn which tests the ebuyer is using to confirm GPU errs,
> >> let us know.
> >>
> >> M.
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >> Dne Sun, 02 Jun 2013 19:22:54 +0200 Marek Maly <marek.maly.ujep.cz>
> >> napsal/-a:
> >>
> >> Hi so I finally succeeded to compile GPU Amber part under CUDA 5.5
> >>> (after "hacking" of the configure2 file) with common results in
> >>> consequent tests:
> >>>
> >>> ------
> >>> 80 file comparisons passed
> >>> 9 file comparisons failed
> >>> 0 tests experienced errors
> >>> ------
> >>>
> >>> So now I am running the 100K(PME)/1000K(GB) repetitive benchmark tests
> >>> under
> >>> this configuration: drv. 319.23, CUDA 5.5. , bugfix 18 installed
> >>>
> >>> When I finish it I will report results here.
> >>>
> >>> M.
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> Dne Sun, 02 Jun 2013 18:44:23 +0200 Marek Maly <marek.maly.ujep.cz>
> >>> napsal/-a:
> >>>
> >>> Hi Scott thanks for the update !
> >>>>
> >>>> Anyway any explanation regarding "cuFFT hypothesis" why there are no
> >>>> problems
> >>>> with GTX 580, GTX 680 or even K20c ???
> >>>>
> >>>>
> >>>> meanwhile I also tried to recompile GPU part of Amber with
> >>>> cuda 5.5 installed before, I have obtained these errs
> >>>> already in configure phase:
> >>>>
> >>>> --------
> >>>> [root.dyn-138-272 amber12]# ./configure -cuda -noX11 gnu
> >>>> Checking for updates...
> >>>> Checking for available patches online. This may take a few seconds...
> >>>>
> >>>> Available AmberTools 13 patches:
> >>>>
> >>>> No patches available
> >>>>
> >>>> Available Amber 12 patches:
> >>>>
> >>>> No patches available
> >>>> Searching for python2... Found python2.6: /usr/bin/python2.6
> >>>> Error: Unsupported CUDA version 5.5 detected.
> >>>> AMBER requires CUDA version == 4.2 .or. 5.0
> >>>> Configure failed due to the errors above!
> >>>> ---------
> >>>>
> >>>> so it seems that Amber is possible to compile only with CUDA 4.2 or
> 5.0
> >>>> at
> >>>> the moment:
> >>>>
> >>>> and this part of configure2 file has to be edited:
> >>>>
> >>>>
> >>>> -----------
> >>>> nvcc="$CUDA_HOME/bin/nvcc"
> >>>> sm35flags='-gencode arch=compute_35,code=sm_35'
> >>>> sm30flags='-gencode arch=compute_30,code=sm_30'
> >>>> sm20flags='-gencode arch=compute_20,code=sm_20'
> >>>> sm13flags='-gencode arch=compute_13,code=sm_13'
> >>>> nvccflags="$sm13flags $sm20flags"
> >>>> cudaversion=`$nvcc --version | grep 'release' | cut -d' ' -f5 |
> cut
> >>>> -d',' -f1`
> >>>> if [ "$cudaversion" == "5.0" ]; then
> >>>> echo "CUDA Version $cudaversion detected"
> >>>> nvccflags="$nvccflags $sm30flags $sm35flags"
> >>>> elif [ "$cudaversion" == "4.2" ]; then
> >>>> echo "CUDA Version $cudaversion detected"
> >>>> nvccflags="$nvccflags $sm30flags"
> >>>> else
> >>>> echo "Error: Unsupported CUDA version $cudaversion detected."
> >>>> echo "AMBER requires CUDA version == 4.2 .or. 5.0"
> >>>> exit 1
> >>>> fi
> >>>> nvcc="$nvcc $nvccflags"
> >>>>
> >>>> fi
> >>>>
> >>>> -----------
> >>>>
> >>>> would it be just OK to change
> >>>> "if [ "$cudaversion" == "5.0" ]; then"
> >>>>
> >>>> to
> >>>>
> >>>> "if [ "$cudaversion" == "5.5" ]; then"
> >>>>
> >>>>
> >>>> or some more flags etc. should be defined here to proceed
> successfully ?
> >>>>
> >>>>
> >>>> BTW it seems Scott, that you are on the way to isolate the problem
> soon
> >>>> so maybe it's better to wait and not to loose time with cuda 5.5
> >>>> experiments.
> >>>>
> >>>> I just thought that cuda 5.5 might be more "friendly" to Titans :))
> e.g.
> >>>> in terms of cuFFT function ....
> >>>>
> >>>>
> >>>> I will keep fingers crossed :))
> >>>>
> >>>> M.
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>> Dne Sun, 02 Jun 2013 18:33:52 +0200 Scott Le Grand
> >>>> <varelse2005.gmail.com>
> >>>> napsal/-a:
> >>>>
> >>>> PS this *might* indicate a software bug in cuFFT, but it needs more
> >>>>> characterization... And things are going to get a little stream of
> >>>>> consciousness from here because you're getting unfiltered raw data,
> so
> >>>>> please don't draw any conclusions towards anything yet - I'm just
> >>>>> letting
> >>>>> you guys know what I'm finding out as I find it...
> >>>>>
> >>>>>
> >>>>>
> >>>>> On Sun, Jun 2, 2013 at 9:31 AM, Scott Le Grand
> >>>>> <varelse2005.gmail.com>wrote:
> >>>>>
> >>>>> And bingo...
> >>>>>>
> >>>>>> At the very least, the reciprocal sum is intermittently
> >>>>>> inconsistent...
> >>>>>> This explains the irreproducible behavior...
> >>>>>>
> >>>>>> And here's the level of inconsistency:
> >>>>>> 31989.38940628897399 vs
> >>>>>> 31989.39168370794505
> >>>>>>
> >>>>>> That's error at the level of 1e-7 or a somehow missed
> single-precision
> >>>>>> transaction somewhere...
> >>>>>>
> >>>>>> The next question is figuring out why... This may or may not
> >>>>>> ultimately
> >>>>>> explain the crashes you guys are also seeing...
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>> On Sun, Jun 2, 2013 at 9:07 AM, Scott Le Grand
> >>>>>> <varelse2005.gmail.com>wrote:
> >>>>>>
> >>>>>>
> >>>>>>> Observations:
> >>>>>>> 1. The degree to which the reproducibility is broken *does* appear
> to
> >>>>>>> vary between individual Titan GPUs. One of my Titans breaks within
> >>>>>>> 10K
> >>>>>>> steps on cellulose, the other one made it to 100K steps twice
> without
> >>>>>>> doing
> >>>>>>> so leading me to believe it could be trusted (until yesterday
> where I
> >>>>>>> now
> >>>>>>> see it dies between 50K and 100K steps most of the time).
> >>>>>>>
> >>>>>>> 2. GB hasn't broken (yet). So could you run myoglobin for 500K and
> >>>>>>> TRPcage for 1,000,000 steps and let's see if that's universal.
> >>>>>>>
> >>>>>>> 3. Turning on double-precision mode makes my Titan crash rather
> than
> >>>>>>> run
> >>>>>>> irreproducibly, sigh...
> >>>>>>>
> >>>>>>> So whatever is going on is triggered by something in PME but not
> GB.
> >>>>>>> So
> >>>>>>> that's either the radix sort, the FFT, the Ewald grid
> interpolation,
> >>>>>>> or the
> >>>>>>> neighbor list code. Fixing this involves isolating this and
> figuring
> >>>>>>> out
> >>>>>>> what exactly goes haywire. It could *still* be software at some
> very
> >>>>>>> small
> >>>>>>> probability but the combination of both 680 and K20c with ECC off
> >>>>>>> running
> >>>>>>> reliably is really pointing towards the Titans just being clocked
> too
> >>>>>>> fast.
> >>>>>>>
> >>>>>>> So how long with this take? Asking people how long it takes to
> fix a
> >>>>>>> bug
> >>>>>>> never really works out well. That said, I found the 480 bug
> within a
> >>>>>>> week
> >>>>>>> and my usual turnaround for a bug with a solid repro is <24 hours.
> >>>>>>>
> >>>>>>> Scott
> >>>>>>>
> >>>>>>> On Sun, Jun 2, 2013 at 7:58 AM, Marek Maly <marek.maly.ujep.cz>
> >>>>>>> wrote:
> >>>>>>>
> >>>>>>> Hi all,
> >>>>>>>>
> >>>>>>>> here are my results after bugfix 18 application (see attachment).
> >>>>>>>>
> >>>>>>>> In principle I don't see any "drastical" changes.
> >>>>>>>>
> >>>>>>>> FACTOR_IX still perfectly stable/reproducible on both cards,
> >>>>>>>>
> >>>>>>>> JAC tests - problems with finishing AND/OR reproducibility the
> >>>>>>>> same CELLULOSE_NVE although here it seems that my TITAN_1
> >>>>>>>> has no problems with this test (but the same same trend I saw also
> >>>>>>>> before bugfix 18 - see my older 500K steps test).
> >>>>>>>>
> >>>>>>>> But anyway bugfix 18 brought here one change.
> >>>>>>>>
> >>>>>>>> The err
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> #1 ERR writtent in mdout:
> >>>>>>>> ------
> >>>>>>>> | ERROR: max pairlist cutoff must be less than unit cell max
> >>>>>>>> sphere
> >>>>>>>> radius!
> >>>>>>>> ------
> >>>>>>>>
> >>>>>>>> was substituted with err/warning ?
> >>>>>>>>
> >>>>>>>> #0 no ERR writtent in mdout, ERR written in standard output
> >>>>>>>> (nohup.out)
> >>>>>>>> -----
> >>>>>>>> Nonbond cells need to be recalculated, restart simulation from
> >>>>>>>> previous
> >>>>>>>> checkpoint
> >>>>>>>> with a higher value for skinnb.
> >>>>>>>>
> >>>>>>>> -----
> >>>>>>>>
> >>>>>>>> Another thing,
> >>>>>>>>
> >>>>>>>> recently I started on another machine and GTX 580 GPU simulation
> of
> >>>>>>>> relatively
> >>>>>>>> big system ( 364275 atoms/PME ). The system is composed also from
> >>>>>>>> the
> >>>>>>>> "exotic" molecules like polymers. ff12SB, gaff, GLYCAM forcefields
> >>>>>>>> used
> >>>>>>>> here. I had problem even with minimization part here, having big
> >>>>>>>> energy
> >>>>>>>> on the start:
> >>>>>>>>
> >>>>>>>> -----
> >>>>>>>> NSTEP ENERGY RMS GMAX NAME
> >>>>>>>> NUMBER
> >>>>>>>> 1 2.8442E+09 2.1339E+02 1.7311E+04 O
> >>>>>>>> 32998
> >>>>>>>>
> >>>>>>>> BOND = 11051.7467 ANGLE = 17720.4706 DIHED =
> >>>>>>>> 18977.7584
> >>>>>>>> VDWAALS = ************* EEL = -1257709.6203 HBOND =
> >>>>>>>> 0.0000
> >>>>>>>> 1-4 VDW = 7253.7412 1-4 EEL = 149867.0207 RESTRAINT =
> >>>>>>>> 0.0000
> >>>>>>>>
> >>>>>>>> ----
> >>>>>>>>
> >>>>>>>> with no chance to minimize the system even with 50 000 steps in
> both
> >>>>>>>> min cycles (with constrained and unconstrained solute) and hence
> >>>>>>>> heating
> >>>>>>>> NVT
> >>>>>>>> crashed immediately even with very small dt. I patched Amber12
> here
> >>>>>>>> with
> >>>>>>>> the
> >>>>>>>> bugfix 18 and the minimization was done without any problem with
> >>>>>>>> common
> >>>>>>>> 5000 steps
> >>>>>>>> (obtaining target Energy -1.4505E+06 while that initial was that
> >>>>>>>> written
> >>>>>>>> above).
> >>>>>>>>
> >>>>>>>> So indeed bugfix 18 solved some issues, but unfortunately not
> those
> >>>>>>>> related to
> >>>>>>>> Titans.
> >>>>>>>>
> >>>>>>>> Here I will try to install cuda 5.5, recompile GPU Amber part
> with
> >>>>>>>> this
> >>>>>>>> new
> >>>>>>>> cuda version and repeat the 100K tests.
> >>>>>>>>
> >>>>>>>> Scott, let us know how finished your experiment with downclocking
> of
> >>>>>>>> Titan.
> >>>>>>>> Maybe the best choice would be here to flash Titan directly with
> >>>>>>>> your
> >>>>>>>> K20c bios :))
> >>>>>>>>
> >>>>>>>> M.
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> Dne Sat, 01 Jun 2013 21:09:46 +0200 Marek Maly <
> marek.maly.ujep.cz>
> >>>>>>>> napsal/-a:
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> Hi,
> >>>>>>>>
> >>>>>>>>>
> >>>>>>>>> first of all thanks for providing of your test results !
> >>>>>>>>>
> >>>>>>>>> It seems that your results are more or less similar to that of
> >>>>>>>>> mine maybe with the exception of the results on FactorIX tests
> >>>>>>>>> where I had perfect stability and 100% or close to 100%
> >>>>>>>>> reproducibility.
> >>>>>>>>>
> >>>>>>>>> Anyway the type of errs which you reported are the same which I
> >>>>>>>>> obtained.
> >>>>>>>>>
> >>>>>>>>> So let's see if the bugfix 18 will help here (or at least on NPT
> >>>>>>>>> tests)
> >>>>>>>>> or not. As I wrote just before few minutes, it seems that it was
> >>>>>>>>> not
> >>>>>>>>> still
> >>>>>>>>> loaded
> >>>>>>>>> to the given server, although it's description is already present
> >>>>>>>>> on
> >>>>>>>>> the
> >>>>>>>>> given
> >>>>>>>>> web page ( see
> >>>>>>>>> http://ambermd.org/bugfixes12.****html<
> http://ambermd.org/bugfixes12.**html>
> >>>>>>>>> <http://ambermd.org/**bugfixes12.html<
> http://ambermd.org/bugfixes12.html>
> >>>>>>>>> >).
> >>>>>>>>>
> >>>>>>>>> As you can see, this bugfix contains also changes in CPU code
> >>>>>>>>> although
> >>>>>>>>> the majority is devoted to GPU code, so perhaps the best will be
> to
> >>>>>>>>> recompile
> >>>>>>>>> whole amber with this patch although this patch would be perhaps
> >>>>>>>>> applied
> >>>>>>>>> even after just
> >>>>>>>>> GPU configure command ( i.e. ./configure -cuda -noX11 gnu ) but
> >>>>>>>>> after
> >>>>>>>>> consequent
> >>>>>>>>> building, just the GPU binaries will be updated. Anyway I would
> >>>>>>>>> rather
> >>>>>>>>> recompile
> >>>>>>>>> whole Amber after this patch.
> >>>>>>>>>
> >>>>>>>>> Regarding to GPU test under linux you may try memtestG80
> >>>>>>>>> (please use the updated/patched version from here
> >>>>>>>>> https://github.com/ihaque/****memtestG80<
> https://github.com/ihaque/**memtestG80>
> >>>>>>>>> <https://github.com/**ihaque/memtestG80<
> https://github.com/ihaque/memtestG80>
> >>>>>>>>> >
> >>>>>>>>> )
> >>>>>>>>>
> >>>>>>>>> just use git command like:
> >>>>>>>>>
> >>>>>>>>> git clone
> >>>>>>>>> https://github.com/ihaque/****memtestG80.git<
> https://github.com/ihaque/**memtestG80.git>
> >>>>>>>>> <https://github.**com/ihaque/memtestG80.git<
> https://github.com/ihaque/memtestG80.git>
> >>>>>>>>> >**PATCHED_MEMTEST-G80
> >>>>>>>>>
> >>>>>>>>> to download all the files and save them into directory named
> >>>>>>>>> PATCHED_MEMTEST-G80.
> >>>>>>>>>
> >>>>>>>>> another possibility is to try perhaps similar (but maybe more up
> to
> >>>>>>>>> date)
> >>>>>>>>> test
> >>>>>>>>> cuda_memtest (
> >>>>>>>>> http://sourceforge.net/****projects/cudagpumemtest/<
> http://sourceforge.net/**projects/cudagpumemtest/>
> >>>>>>>>> <http:**//sourceforge.net/projects/**cudagpumemtest/<
> http://sourceforge.net/projects/cudagpumemtest/>
> >>>>>>>>> >).
> >>>>>>>>>
> >>>>>>>>> regarding ig value: If ig is not present in mdin, the default
> value
> >>>>>>>>> is
> >>>>>>>>> used
> >>>>>>>>> (e.g. 71277) if ig=-1 the random seed will be based on the
> current
> >>>>>>>>> date
> >>>>>>>>> and time, and hence will be different for every run (not a good
> >>>>>>>>> variant
> >>>>>>>>> for our testts). I simply deleted eventual ig records from all
> >>>>>>>>> mdins
> >>>>>>>>> so
> >>>>>>>>> I
> >>>>>>>>> assume that in each run the default seed 71277 was automatically
> >>>>>>>>> used.
> >>>>>>>>>
> >>>>>>>>> M.
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> Dne Sat, 01 Jun 2013 20:26:16 +0200 ET <sketchfoot.gmail.com>
> >>>>>>>>> napsal/-a:
> >>>>>>>>>
> >>>>>>>>> Hi,
> >>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> I've put the graphics card into a machine with the working GTX
> >>>>>>>>>> titan
> >>>>>>>>>> that I
> >>>>>>>>>> mentioned earlier.
> >>>>>>>>>>
> >>>>>>>>>> The Nvidia driver version is: 133.30
> >>>>>>>>>>
> >>>>>>>>>> Amber version is:
> >>>>>>>>>> AmberTools version 13.03
> >>>>>>>>>> Amber version 12.16
> >>>>>>>>>>
> >>>>>>>>>> I ran 50k steps with the amber benchmark using ig=43689 on both
> >>>>>>>>>> cards.
> >>>>>>>>>> For
> >>>>>>>>>> the purpose of discriminating between them, the card I believe
> >>>>>>>>>> (fingers
> >>>>>>>>>> crossed) is working is called GPU-00_TeaNCake, whilst the other
> >>>>>>>>>> one
> >>>>>>>>>> is
> >>>>>>>>>> called GPU-01_008.
> >>>>>>>>>>
> >>>>>>>>>> *When I run the tests on GPU-01_008:*
> >>>>>>>>>>
> >>>>>>>>>> 1) All the tests (across 2x repeats) finish apart from the
> >>>>>>>>>> following
> >>>>>>>>>> which
> >>>>>>>>>> have the errors listed:
> >>>>>>>>>>
> >>>>>>>>>> ------------------------------****--------------
> >>>>>>>>>> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
> >>>>>>>>>> Error: unspecified launch failure launching kernel kNLSkinTest
> >>>>>>>>>> cudaFree GpuBuffer::Deallocate failed unspecified launch failure
> >>>>>>>>>>
> >>>>>>>>>> ------------------------------****--------------
> >>>>>>>>>> CELLULOSE_PRODUCTION_NPT - 408,609 atoms PME
> >>>>>>>>>> cudaMemcpy GpuBuffer::Download failed unspecified launch
> failure
> >>>>>>>>>>
> >>>>>>>>>> ------------------------------****--------------
> >>>>>>>>>> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
> >>>>>>>>>> Error: unspecified launch failure launching kernel kNLSkinTest
> >>>>>>>>>> cudaFree GpuBuffer::Deallocate failed unspecified launch failure
> >>>>>>>>>>
> >>>>>>>>>> ------------------------------****--------------
> >>>>>>>>>> CELLULOSE_PRODUCTION_NPT - 408,609 atoms PME
> >>>>>>>>>> cudaMemcpy GpuBuffer::Download failed unspecified launch
> failure
> >>>>>>>>>> grep: mdinfo.1GTX680: No such file or directory
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> 2) The sdiff logs indicate that reproducibility across the two
> >>>>>>>>>> repeats
> >>>>>>>>>> is
> >>>>>>>>>> as follows:
> >>>>>>>>>>
> >>>>>>>>>> *GB_myoglobin: *Reproducible across 50k steps
> >>>>>>>>>> *GB_nucleosome:* Reproducible till step 7400
> >>>>>>>>>> *GB_TRPCage:* Reproducible across 50k steps
> >>>>>>>>>>
> >>>>>>>>>> *PME_JAC_production_NVE: *No reproducibility shown from step
> 1,000
> >>>>>>>>>> onwards
> >>>>>>>>>> *PME_JAC_production_NPT*: Reproducible till step 1,000. Also
> >>>>>>>>>> outfile
> >>>>>>>>>> is
> >>>>>>>>>> not written properly - blank gaps appear where something should
> >>>>>>>>>> have
> >>>>>>>>>> been
> >>>>>>>>>> written
> >>>>>>>>>>
> >>>>>>>>>> *PME_FactorIX_production_NVE:* Reproducible across 50k steps
> >>>>>>>>>> *PME_FactorIX_production_NPT:* Reproducible across 50k steps
> >>>>>>>>>>
> >>>>>>>>>> *PME_Cellulose_production_NVE:***** Failure means that both runs
> >>>>>>>>>> do
> >>>>>>>>>> not
> >>>>>>>>>> finish
> >>>>>>>>>> (see point1)
> >>>>>>>>>> *PME_Cellulose_production_NPT: *Failure means that both runs do
> >>>>>>>>>> not
> >>>>>>>>>> finish
> >>>>>>>>>> (see point1)
> >>>>>>>>>>
> >>>>>>>>>> ##############################****############################**
> >>>>>>>>>> ##**
> >>>>>>>>>> ###########################
> >>>>>>>>>>
> >>>>>>>>>> *When I run the tests on * *GPU-00_TeaNCake:*
> >>>>>>>>>> *
> >>>>>>>>>> *
> >>>>>>>>>> 1) All the tests (across 2x repeats) finish apart from the
> >>>>>>>>>> following
> >>>>>>>>>> which
> >>>>>>>>>> have the errors listed:
> >>>>>>>>>> ------------------------------****-------
> >>>>>>>>>> JAC_PRODUCTION_NPT - 23,558 atoms PME
> >>>>>>>>>> PMEMD Terminated Abnormally!
> >>>>>>>>>> ------------------------------****-------
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> 2) The sdiff logs indicate that reproducibility across the two
> >>>>>>>>>> repeats
> >>>>>>>>>> is
> >>>>>>>>>> as follows:
> >>>>>>>>>>
> >>>>>>>>>> *GB_myoglobin:* Reproducible across 50k steps
> >>>>>>>>>> *GB_nucleosome:* Reproducible across 50k steps
> >>>>>>>>>> *GB_TRPCage:* Reproducible across 50k steps
> >>>>>>>>>>
> >>>>>>>>>> *PME_JAC_production_NVE:* No reproducibility shown from step
> >>>>>>>>>> 10,000
> >>>>>>>>>> onwards
> >>>>>>>>>> *PME_JAC_production_NPT: * No reproducibility shown from step
> >>>>>>>>>> 10,000
> >>>>>>>>>> onwards. Also outfile is not written properly - blank gaps
> appear
> >>>>>>>>>> where
> >>>>>>>>>> something should have been written. Repeat 2 Crashes with error
> >>>>>>>>>> noted
> >>>>>>>>>> in
> >>>>>>>>>> 1.
> >>>>>>>>>>
> >>>>>>>>>> *PME_FactorIX_production_NVE:* No reproducibility shown from
> step
> >>>>>>>>>> 9,000
> >>>>>>>>>> onwards
> >>>>>>>>>> *PME_FactorIX_production_NPT: *Reproducible across 50k steps
> >>>>>>>>>>
> >>>>>>>>>> *PME_Cellulose_production_NVE: *No reproducibility shown from
> step
> >>>>>>>>>> 5,000
> >>>>>>>>>> onwards
> >>>>>>>>>> *PME_Cellulose_production_NPT: ** *No reproducibility shown from
> >>>>>>>>>> step
> >>>>>>>>>> 29,000 onwards. Also outfile is not written properly - blank
> gaps
> >>>>>>>>>> appear
> >>>>>>>>>> where something should have been written.
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> Out files and sdiff files are included as attatchments
> >>>>>>>>>>
> >>>>>>>>>> ##############################****###################
> >>>>>>>>>>
> >>>>>>>>>> So I'm going to update my nvidia driver to the latest version
> and
> >>>>>>>>>> patch
> >>>>>>>>>> amber to the latest version and rerun the tests to see if there
> is
> >>>>>>>>>> any
> >>>>>>>>>> improvement. Could someone let me know if it is necessary to
> >>>>>>>>>> recompile
> >>>>>>>>>> any
> >>>>>>>>>> or all of AMBER after applying the bugfixes?
> >>>>>>>>>>
> >>>>>>>>>> Additionally, I'm going to run memory tests and heaven
> benchmarks
> >>>>>>>>>> on
> >>>>>>>>>> the
> >>>>>>>>>> cards to check whether they are faulty or not.
> >>>>>>>>>>
> >>>>>>>>>> I'm thinking that there is a mix of hardware error/configuration
> >>>>>>>>>> (esp
> >>>>>>>>>> in
> >>>>>>>>>> the case of GPU-01_008) and amber software error in this
> >>>>>>>>>> situation.
> >>>>>>>>>> What
> >>>>>>>>>> do
> >>>>>>>>>> you guys think?
> >>>>>>>>>>
> >>>>>>>>>> Also am I right in thinking (from what Scott was saying) that
> all
> >>>>>>>>>> the
> >>>>>>>>>> benchmarks should be reproducible across 50k steps but begin to
> >>>>>>>>>> diverge
> >>>>>>>>>> at
> >>>>>>>>>> around 100K steps? Is there any difference from in setting *ig
> *to
> >>>>>>>>>> an
> >>>>>>>>>> explicit number to removing it from the mdin file?
> >>>>>>>>>>
> >>>>>>>>>> br,
> >>>>>>>>>> g
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> On 31 May 2013 23:45, ET <sketchfoot.gmail.com> wrote:
> >>>>>>>>>>
> >>>>>>>>>> I don't need sysadmins, but sysadmins need me as it gives
> purpose
> >>>>>>>>>> to
> >>>>>>>>>>
> >>>>>>>>>>> their
> >>>>>>>>>>> bureaucratic existence. A encountered evil if working in an
> >>>>>>>>>>> institution
> >>>>>>>>>>> or
> >>>>>>>>>>> comapny IMO. Good science and indiviguality being sacrificed
> for
> >>>>>>>>>>> standardisation and mediocrity in the intrerests of maintaing a
> >>>>>>>>>>> system
> >>>>>>>>>>> that
> >>>>>>>>>>> focusses on maintaining the system and not the objective.
> >>>>>>>>>>>
> >>>>>>>>>>> You need root to move fwd on these things, unfortunately. and
> ppl
> >>>>>>>>>>> with
> >>>>>>>>>>> root are kinda like your parents when you try to borrow money
> >>>>>>>>>>> from
> >>>>>>>>>>> them
> >>>>>>>>>>> .
> >>>>>>>>>>> age 12 :D
> >>>>>>>>>>> On May 31, 2013 9:34 PM, "Marek Maly" <marek.maly.ujep.cz>
> >>>>>>>>>>> wrote:
> >>>>>>>>>>>
> >>>>>>>>>>> Sorry why do you need sysadmins :)) ?
> >>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> BTW here is the most recent driver:
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> http://www.nvidia.com/object/****linux-display-amd64-319.23-**<
> http://www.nvidia.com/object/**linux-display-amd64-319.23-**>
> >>>>>>>>>>>> driver.html<http://www.nvidia.**com/object/linux-display-**
> >>>>>>>>>>>> amd64-319.23-driver.html<
> http://www.nvidia.com/object/linux-display-amd64-319.23-driver.html>
> >>>>>>>>>>>> >
> >>>>>>>>>>>>
> >>>>>>>>>>>> I do not remember anything easier than is to install driver
> >>>>>>>>>>>> (especially
> >>>>>>>>>>>> in case of binary (*.run) installer) :))
> >>>>>>>>>>>>
> >>>>>>>>>>>> M.
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> Dne Fri, 31 May 2013 22:02:34 +0200 ET <sketchfoot.gmail.com>
> >>>>>>>>>>>> napsal/-a:
> >>>>>>>>>>>>
> >>>>>>>>>>>> > Yup. I know. I replaced a 680 and the everknowing sysadmins
> >>>>>>>>>>>> are
> >>>>>>>>>>>> reluctant
> >>>>>>>>>>>> > to install drivers not in the repositoery as they are lame.
> :(
> >>>>>>>>>>>> > On May 31, 2013 7:14 PM, "Marek Maly" <marek.maly.ujep.cz>
> >>>>>>>>>>>> wrote:
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> As I already wrote you,
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> the first driver which properly/officially supports Titans,
> >>>>>>>>>>>> should
> >>>>>>>>>>>> be
> >>>>>>>>>>>> >> 313.26 .
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> Anyway I am curious mainly about your 100K repetitive tests
> >>>>>>>>>>>> with
> >>>>>>>>>>>> >> your Titan SC card. Especially in case of these tests (
> >>>>>>>>>>>> JAC_NVE,
> >>>>>>>>>>>> JAC_NPT
> >>>>>>>>>>>> >> and CELLULOSE_NVE ) where
> >>>>>>>>>>>> >> my Titans SC randomly failed or succeeded. In
> FACTOR_IX_NVE,
> >>>>>>>>>>>> >> FACTOR_IX_NPT
> >>>>>>>>>>>> >> tests both
> >>>>>>>>>>>> >> my cards are perfectly stable (independently from drv.
> >>>>>>>>>>>> version)
> >>>>>>>>>>>> and
> >>>>>>>>>>>> also
> >>>>>>>>>>>> >> the runs
> >>>>>>>>>>>> >> are perfectly or almost perfectly reproducible.
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> Also if your test will crash please report the eventual
> errs.
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> To this moment I have this actual library of errs on my
> >>>>>>>>>>>> Titans
> >>>>>>>>>>>> SC
> >>>>>>>>>>>> GPUs.
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> #1 ERR writtent in mdout:
> >>>>>>>>>>>> >> ------
> >>>>>>>>>>>> >> | ERROR: max pairlist cutoff must be less than unit cell
> >>>>>>>>>>>> max
> >>>>>>>>>>>> sphere
> >>>>>>>>>>>> >> radius!
> >>>>>>>>>>>> >> ------
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> #2 no ERR writtent in mdout, ERR written in standard output
> >>>>>>>>>>>> (nohup.out)
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> ----
> >>>>>>>>>>>> >> Error: unspecified launch failure launching kernel
> >>>>>>>>>>>> kNLSkinTest
> >>>>>>>>>>>> >> cudaFree GpuBuffer::Deallocate failed unspecified launch
> >>>>>>>>>>>> failure
> >>>>>>>>>>>> >> ----
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> #3 no ERR writtent in mdout, ERR written in standard output
> >>>>>>>>>>>> (nohup.out)
> >>>>>>>>>>>> >> ----
> >>>>>>>>>>>> >> cudaMemcpy GpuBuffer::Download failed unspecified launch
> >>>>>>>>>>>> failure
> >>>>>>>>>>>> >> ----
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> Another question, regarding your Titan SC, it is also EVGA
> as
> >>>>>>>>>>>> in
> >>>>>>>>>>>> my
> >>>>>>>>>>>> case
> >>>>>>>>>>>> >> or it is another producer ?
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> Thanks,
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> M.
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> Dne Fri, 31 May 2013 19:17:03 +0200 ET <
> sketchfoot.gmail.com
> >>>>>>>>>>>> >
> >>>>>>>>>>>> napsal/-a:
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> > Well, this is interesting...
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> > I ran 50k steps on the Titan on the other machine with
> >>>>>>>>>>>> driver
> >>>>>>>>>>>> 310.44
> >>>>>>>>>>>> >> and
> >>>>>>>>>>>> >> > it
> >>>>>>>>>>>> >> > passed all the GB steps. i.e totally identical results
> over
> >>>>>>>>>>>> two
> >>>>>>>>>>>> >> repeats.
> >>>>>>>>>>>> >> > However, it failed all the PME tests after step 1000. I'm
> >>>>>>>>>>>> going
> >>>>>>>>>>>> to
> >>>>>>>>>>>> > update
> >>>>>>>>>>>> >> > the driver and test it again.
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> > Files included as attachments.
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> > br,
> >>>>>>>>>>>> >> > g
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> > On 31 May 2013 16:40, Marek Maly <marek.maly.ujep.cz>
> >>>>>>>>>>>> wrote:
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> >> One more thing,
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> can you please check under which frequency is running
> that
> >>>>>>>>>>>> your
> >>>>>>>>>>>> >> titan ?
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> As the base frequency of normal Titans is 837MHz and the
> >>>>>>>>>>>> Boost
> >>>>>>>>>>>> one
> >>>>>>>>>>>> is
> >>>>>>>>>>>> >> >> 876MHz I
> >>>>>>>>>>>> >> >> assume that yor GPU is running automatically also under
> >>>>>>>>>>>> it's
> >>>>>>>>>>>> boot
> >>>>>>>>>>>> >> >> frequency (876MHz).
> >>>>>>>>>>>> >> >> You can find this information e.g. in Amber mdout file.
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> You also mentioned some crashes in your previous email.
> >>>>>>>>>>>> Your
> >>>>>>>>>>>> ERRs
> >>>>>>>>>>>> >> were
> >>>>>>>>>>>> >> >> something like those here:
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> #1 ERR writtent in mdout:
> >>>>>>>>>>>> >> >> ------
> >>>>>>>>>>>> >> >> | ERROR: max pairlist cutoff must be less than unit
> cell
> >>>>>>>>>>>> max
> >>>>>>>>>>>> sphere
> >>>>>>>>>>>> >> >> radius!
> >>>>>>>>>>>> >> >> ------
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> #2 no ERR writtent in mdout, ERR written in standard
> >>>>>>>>>>>> output
> >>>>>>>>>>>> >> (nohup.out)
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> ----
> >>>>>>>>>>>> >> >> Error: unspecified launch failure launching kernel
> >>>>>>>>>>>> kNLSkinTest
> >>>>>>>>>>>> >> >> cudaFree GpuBuffer::Deallocate failed unspecified launch
> >>>>>>>>>>>> failure
> >>>>>>>>>>>> >> >> ----
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> #3 no ERR writtent in mdout, ERR written in standard
> >>>>>>>>>>>> output
> >>>>>>>>>>>> >> (nohup.out)
> >>>>>>>>>>>> >> >> ----
> >>>>>>>>>>>> >> >> cudaMemcpy GpuBuffer::Download failed unspecified launch
> >>>>>>>>>>>> failure
> >>>>>>>>>>>> >> >> ----
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> or you obtained some new/additional errs ?
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> M.
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> Dne Fri, 31 May 2013 17:30:57 +0200 filip fratev
> >>>>>>>>>>>> >> <filipfratev.yahoo.com
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> >> napsal/-a:
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> > Hi,
> >>>>>>>>>>>> >> >> > This is what I obtained for 50K tests and "normal"
> >>>>>>>>>>>> GTXTitan:
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > run1:
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >
> >>>>>>>>>>>>
> ------------------------------****----------------------------*
> >>>>>>>>>>>> *--**
> >>>>>>>>>>>> ------------------
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > A V E R A G E S O V E R 50 S T E P S
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> >>>>>>>>>>>> 299.87
> >>>>>>>>>>>> >> PRESS
> >>>>>>>>>>>> >> >> > = 0.0
> >>>>>>>>>>>> >> >> > Etot = -443237.1079 EKtot = 257679.9750
> >>>>>>>>>>>> EPtot
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > -700917.0829
> >>>>>>>>>>>> >> >> > BOND = 20193.1856 ANGLE = 53517.5432
> >>>>>>>>>>>> DIHED
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 23575.4648
> >>>>>>>>>>>> >> >> > 1-4 NB = 21759.5524 1-4 EEL = 742552.5939
> >>>>>>>>>>>> VDWAALS
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 96286.7714
> >>>>>>>>>>>> >> >> > EELEC = -1658802.1941 EHBOND = 0.0000
> >>>>>>>>>>>> RESTRAINT
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 0.0000
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >
> >>>>>>>>>>>>
> ------------------------------****----------------------------
> >>>>>>>>>>>> **--**
> >>>>>>>>>>>> ------------------
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > R M S F L U C T U A T I O N S
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> >>>>>>>>>>>> 0.33
> >>>>>>>>>>>> >> PRESS
> >>>>>>>>>>>> >> >> > = 0.0
> >>>>>>>>>>>> >> >> > Etot = 11.2784 EKtot = 284.8999
> >>>>>>>>>>>> EPtot
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 289.0773
> >>>>>>>>>>>> >> >> > BOND = 136.3417 ANGLE = 214.0054
> >>>>>>>>>>>> DIHED
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 59.4893
> >>>>>>>>>>>> >> >> > 1-4 NB = 58.5891 1-4 EEL = 330.5400
> >>>>>>>>>>>> VDWAALS
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 559.2079
> >>>>>>>>>>>> >> >> > EELEC = 743.8771 EHBOND = 0.0000
> >>>>>>>>>>>> RESTRAINT
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 0.0000
> >>>>>>>>>>>> >> >> > |E(PBS) = 21.8119
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >
> >>>>>>>>>>>>
> ------------------------------****----------------------------
> >>>>>>>>>>>> **--**
> >>>>>>>>>>>> ------------------
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > run2:
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >
> >>>>>>>>>>>>
> ------------------------------****----------------------------*
> >>>>>>>>>>>> *--**
> >>>>>>>>>>>> ------------------
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > A V E R A G E S O V E R 50 S T E P S
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> >>>>>>>>>>>> 299.89
> >>>>>>>>>>>> >> PRESS
> >>>>>>>>>>>> >> >> > = 0.0
> >>>>>>>>>>>> >> >> > Etot = -443240.0999 EKtot = 257700.0950
> >>>>>>>>>>>> EPtot
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > -700940.1949
> >>>>>>>>>>>> >> >> > BOND = 20241.9174 ANGLE = 53644.6694
> >>>>>>>>>>>> DIHED
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 23541.3737
> >>>>>>>>>>>> >> >> > 1-4 NB = 21803.1898 1-4 EEL = 742754.2254
> >>>>>>>>>>>> VDWAALS
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 96298.8308
> >>>>>>>>>>>> >> >> > EELEC = -1659224.4013 EHBOND = 0.0000
> >>>>>>>>>>>> RESTRAINT
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 0.0000
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >
> >>>>>>>>>>>>
> ------------------------------****----------------------------
> >>>>>>>>>>>> **--**
> >>>>>>>>>>>> ------------------
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > R M S F L U C T U A T I O N S
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> >>>>>>>>>>>> 0.41
> >>>>>>>>>>>> >> PRESS
> >>>>>>>>>>>> >> >> > = 0.0
> >>>>>>>>>>>> >> >> > Etot = 10.7633 EKtot = 348.2819
> >>>>>>>>>>>> EPtot
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 353.9918
> >>>>>>>>>>>> >> >> > BOND = 106.5314 ANGLE = 196.7052
> >>>>>>>>>>>> DIHED
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 69.7476
> >>>>>>>>>>>> >> >> > 1-4 NB = 60.3435 1-4 EEL = 400.7466
> >>>>>>>>>>>> VDWAALS
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 462.7763
> >>>>>>>>>>>> >> >> > EELEC = 651.9857 EHBOND = 0.0000
> >>>>>>>>>>>> RESTRAINT
> >>>>>>>>>>>> =
> >>>>>>>>>>>> >> >> > 0.0000
> >>>>>>>>>>>> >> >> > |E(PBS) = 17.0642
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >
> >>>>>>>>>>>>
> ------------------------------****----------------------------
> >>>>>>>>>>>> **--**
> >>>>>>>>>>>> ------------------
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >
> >>>>>>>>>>>>
> ------------------------------****----------------------------*
> >>>>>>>>>>>> *--**
> >>>>>>>>>>>> --------------------
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > ______________________________****__
> >>>>>>>>>>>> >> >> > From: Marek Maly <marek.maly.ujep.cz>
> >>>>>>>>>>>> >> >> > To: AMBER Mailing List <amber.ambermd.org>
> >>>>>>>>>>>> >> >> > Sent: Friday, May 31, 2013 3:34 PM
> >>>>>>>>>>>> >> >> > Subject: Re: [AMBER] experiences with EVGA GTX TITAN
> >>>>>>>>>>>> Superclocked
> >>>>>>>>>>>> -
> >>>>>>>>>>>> >> >> > memtestG80 - UNDERclocking in Linux ?
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > Hi here are my 100K results for driver 313.30 (and
> still
> >>>>>>>>>>>> Cuda
> >>>>>>>>>>>> 5.0).
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > The results are rather similar to those obtained
> >>>>>>>>>>>> >> >> > under my original driver 319.17 (see the first table
> >>>>>>>>>>>> >> >> > which I sent in this thread).
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > M.
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> > Dne Fri, 31 May 2013 12:29:59 +0200 Marek Maly <
> >>>>>>>>>>>> marek.maly.ujep.cz>
> >>>>>>>>>>>> >> >> > napsal/-a:
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >> Hi,
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >> please try to run at lest 100K tests twice to verify
> >>>>>>>>>>>> exact
> >>>>>>>>>>>> >> >> >> reproducibility
> >>>>>>>>>>>> >> >> >> of the results on the given card. If you find in any
> >>>>>>>>>>>> mdin
> >>>>>>>>>>>> file
> >>>>>>>>>>>> >> ig=-1
> >>>>>>>>>>>> >> >> >> just
> >>>>>>>>>>>> >> >> >> delete it to ensure that you are using the identical
> >>>>>>>>>>>> random
> >>>>>>>>>>>> seed
> >>>>>>>>>>>> >> for
> >>>>>>>>>>>> >> >> >> both
> >>>>>>>>>>>> >> >> >> runs. You can eventually omit NUCLEOSOME test
> >>>>>>>>>>>> >> >> >> as it is too much time consuming.
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >> Driver 310.44 ?????
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >> As far as I know the proper support for titans is
> from
> >>>>>>>>>>>> version
> >>>>>>>>>>>> > 313.26
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >> see e.g. here :
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> http://www.geeks3d.com/****20130306/nvidia-releases-r313-****
> <http://www.geeks3d.com/**20130306/nvidia-releases-r313-**>
> >>>>>>>>>>>> 26-for-linux-with-gtx-titan-****support/<http://www.geeks3d.
> **
> >>>>>>>>>>>> com/20130306/nvidia-releases-**r313-26-for-linux-with-gtx-**
> >>>>>>>>>>>> titan-support/<
> http://www.geeks3d.com/20130306/nvidia-releases-r313-26-for-linux-with-gtx-titan-support/
> >
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >> BTW: On my site downgrade to drv. 313.30 did not
> solved
> >>>>>>>>>>>> the
> >>>>>>>>>>>> >> >> situation, I
> >>>>>>>>>>>> >> >> >> will post
> >>>>>>>>>>>> >> >> >> my results soon here.
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >> M.
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >> Dne Fri, 31 May 2013 12:21:21 +0200 ET <
> >>>>>>>>>>>> sketchfoot.gmail.com>
> >>>>>>>>>>>> >> >> napsal/-a:
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >>> ps. I have another install of amber on another
> >>>>>>>>>>>> computer
> >>>>>>>>>>>> with
> >>>>>>>>>>>> a
> >>>>>>>>>>>> >> >> >>> different
> >>>>>>>>>>>> >> >> >>> Titan and different Driver Version: 310.44.
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>> In the interests of thrashing the proverbial horse,
> >>>>>>>>>>>> I'll
> >>>>>>>>>>>> run
> >>>>>>>>>>>> the
> >>>>>>>>>>>> >> >> >>> benchmark
> >>>>>>>>>>>> >> >> >>> for 50k steps. :P
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>> br,
> >>>>>>>>>>>> >> >> >>> g
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>> On 31 May 2013 11:17, ET <sketchfoot.gmail.com>
> >>>>>>>>>>>> wrote:
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>>> Hi, I just ran the Amber benchmark for the default
> >>>>>>>>>>>> (10000
> >>>>>>>>>>>> steps)
> >>>>>>>>>>>> >> >> on my
> >>>>>>>>>>>> >> >> >>>> Titan.
> >>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>> >> >> >>>> Using sdiff -sB showed that the two runs were
> >>>>>>>>>>>> completely
> >>>>>>>>>>>> > identical.
> >>>>>>>>>>>> >> >> >>>> I've
> >>>>>>>>>>>> >> >> >>>> attached compressed files of the mdout & diff
> files.
> >>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>> >> >> >>>> br,
> >>>>>>>>>>>> >> >> >>>> g
> >>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>> >> >> >>>> On 30 May 2013 23:41, Marek Maly <
> marek.maly.ujep.cz
> >>>>>>>>>>>> >
> >>>>>>>>>>>> wrote:
> >>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>> >> >> >>>>> OK, let's see. The eventual downclocking I see as
> >>>>>>>>>>>> the
> >>>>>>>>>>>> very
> >>>>>>>>>>>> last
> >>>>>>>>>>>> >> >> >>>>> possibility
> >>>>>>>>>>>> >> >> >>>>> (if I don't decide for RMAing). But now still some
> >>>>>>>>>>>> other
> >>>>>>>>>>>> >> >> experiments
> >>>>>>>>>>>> >> >> >>>>> are
> >>>>>>>>>>>> >> >> >>>>> available :))
> >>>>>>>>>>>> >> >> >>>>> I just started 100K tests under 313.30 driver. For
> >>>>>>>>>>>> today
> >>>>>>>>>>>> good
> >>>>>>>>>>>> >> >> night
> >>>>>>>>>>>> >> >> >>>>> ...
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> >> >> >>>>> M.
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> >> >> >>>>> Dne Fri, 31 May 2013 00:45:49 +0200 Scott Le Grand
> >>>>>>>>>>>> >> >> >>>>> <varelse2005.gmail.com
> >>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> napsal/-a:
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> >> >> >>>>> > It will be very interesting if this behavior
> >>>>>>>>>>>> persists
> >>>>>>>>>>>> after
> >>>>>>>>>>>> >> >> >>>>> downclocking.
> >>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> > But right now, Titan 0 *looks* hosed and Titan 1
> >>>>>>>>>>>> *looks*
> >>>>>>>>>>>> like
> >>>>>>>>>>>> > it
> >>>>>>>>>>>> >> >> >>>>> needs
> >>>>>>>>>>>> >> >> >>>>> > downclocking...
> >>>>>>>>>>>> >> >> >>>>> > On May 30, 2013 3:20 PM, "Marek Maly"
> >>>>>>>>>>>> <marek.maly.ujep.cz>
> >>>>>>>>>>>> >> >> wrote:
> >>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >> Hi all,
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> here are my results from the 500K steps 2 x
> >>>>>>>>>>>> repeated
> >>>>>>>>>>>> > benchmarks
> >>>>>>>>>>>> >> >> >>>>> >> under 319.23 driver and still Cuda 5.0 (see the
> >>>>>>>>>>>> attached
> >>>>>>>>>>>> >> table
> >>>>>>>>>>>> >> >> ).
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> It is hard to say if the results are better or
> >>>>>>>>>>>> worse
> >>>>>>>>>>>> than
> >>>>>>>>>>>> in
> >>>>>>>>>>>> > my
> >>>>>>>>>>>> >> >> >>>>> >> previous 100K test under driver 319.17.
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> While results from Cellulose test were improved
> >>>>>>>>>>>> and
> >>>>>>>>>>>> the
> >>>>>>>>>>>> > TITAN_1
> >>>>>>>>>>>> >> >> >>>>> card
> >>>>>>>>>>>> >> >> >>>>> >> even
> >>>>>>>>>>>> >> >> >>>>> >> successfully finished all 500K steps moreover
> >>>>>>>>>>>> with
> >>>>>>>>>>>> exactly
> >>>>>>>>>>>> >> the
> >>>>>>>>>>>> >> >> >>>>> same
> >>>>>>>>>>>> >> >> >>>>> >> final
> >>>>>>>>>>>> >> >> >>>>> >> energy !
> >>>>>>>>>>>> >> >> >>>>> >> (TITAN_0 at least finished more than 100K steps
> >>>>>>>>>>>> and in
> >>>>>>>>>>>> >> RUN_01
> >>>>>>>>>>>> >> >> even
> >>>>>>>>>>>> >> >> >>>>> more
> >>>>>>>>>>>> >> >> >>>>> >> than 400K steps)
> >>>>>>>>>>>> >> >> >>>>> >> In JAC_NPT test no GPU was able to finish at
> >>>>>>>>>>>> least
> >>>>>>>>>>>> 100K
> >>>>>>>>>>>> >> steps
> >>>>>>>>>>>> >> >> and
> >>>>>>>>>>>> >> >> >>>>> the
> >>>>>>>>>>>> >> >> >>>>> >> results from JAC_NVE
> >>>>>>>>>>>> >> >> >>>>> >> test are also not too much convincing.
> >>>>>>>>>>>> FACTOR_IX_NVE
> >>>>>>>>>>>> and
> >>>>>>>>>>>> >> >> >>>>> FACTOR_IX_NPT
> >>>>>>>>>>>> >> >> >>>>> >> were successfully
> >>>>>>>>>>>> >> >> >>>>> >> finished with 100% reproducibility in
> >>>>>>>>>>>> FACTOR_IX_NPT
> >>>>>>>>>>>> case
> >>>>>>>>>>>> >> (on
> >>>>>>>>>>>> >> >> both
> >>>>>>>>>>>> >> >> >>>>> >> cards)
> >>>>>>>>>>>> >> >> >>>>> >> and almost
> >>>>>>>>>>>> >> >> >>>>> >> 100% reproducibility in case of FACTOR_IX_NVE
> >>>>>>>>>>>> (again
> >>>>>>>>>>>> 100%
> >>>>>>>>>>>> in
> >>>>>>>>>>>> >> >> case
> >>>>>>>>>>>> >> >> >>>>> of
> >>>>>>>>>>>> >> >> >>>>> >> TITAN_1). TRPCAGE, MYOGLOBIN
> >>>>>>>>>>>> >> >> >>>>> >> again finished without any problem with 100%
> >>>>>>>>>>>> >> reproducibility.
> >>>>>>>>>>>> >> >> >>>>> NUCLEOSOME
> >>>>>>>>>>>> >> >> >>>>> >> test was not done
> >>>>>>>>>>>> >> >> >>>>> >> this time due to high time requirements. If you
> >>>>>>>>>>>> find
> >>>>>>>>>>>> in
> >>>>>>>>>>>> the
> >>>>>>>>>>>> >> >> table
> >>>>>>>>>>>> >> >> >>>>> >> positive
> >>>>>>>>>>>> >> >> >>>>> >> number finishing with
> >>>>>>>>>>>> >> >> >>>>> >> K (which means "thousands") it means the last
> >>>>>>>>>>>> number
> >>>>>>>>>>>> of
> >>>>>>>>>>>> step
> >>>>>>>>>>>> >> >> >>>>> written in
> >>>>>>>>>>>> >> >> >>>>> >> mdout before crash.
> >>>>>>>>>>>> >> >> >>>>> >> Below are all the 3 types of detected errs with
> >>>>>>>>>>>> relevant
> >>>>>>>>>>>> >> >> >>>>> systems/rounds
> >>>>>>>>>>>> >> >> >>>>> >> where the given err
> >>>>>>>>>>>> >> >> >>>>> >> appeared.
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> Now I will try just 100K tests under ETs
> >>>>>>>>>>>> favourite
> >>>>>>>>>>>> driver
> >>>>>>>>>>>> >> >> version
> >>>>>>>>>>>> >> >> >>>>> 313.30
> >>>>>>>>>>>> >> >> >>>>> >> :)) and then
> >>>>>>>>>>>> >> >> >>>>> >> I will eventually try to experiment with cuda
> 5.5
> >>>>>>>>>>>> which
> >>>>>>>>>>>> I
> >>>>>>>>>>>> >> >> already
> >>>>>>>>>>>> >> >> >>>>> >> downloaded from the
> >>>>>>>>>>>> >> >> >>>>> >> cuda zone ( I had to become cuda developer for
> >>>>>>>>>>>> this
> >>>>>>>>>>>> :))
> >>>>>>>>>>>> )
> >>>>>>>>>>>> >> BTW
> >>>>>>>>>>>> >> >> ET
> >>>>>>>>>>>> >> >> >>>>> thanks
> >>>>>>>>>>>> >> >> >>>>> >> for the frequency info !
> >>>>>>>>>>>> >> >> >>>>> >> and I am still ( perhaps not alone :)) ) very
> >>>>>>>>>>>> curious
> >>>>>>>>>>>> about
> >>>>>>>>>>>> >> >> your 2
> >>>>>>>>>>>> >> >> >>>>> x
> >>>>>>>>>>>> >> >> >>>>> >> repeated Amber benchmark tests with
> superclocked
> >>>>>>>>>>>> Titan.
> >>>>>>>>>>>> >> Indeed
> >>>>>>>>>>>> >> >> >>>>> that
> >>>>>>>>>>>> >> >> >>>>> I
> >>>>>>>>>>>> >> >> >>>>> am
> >>>>>>>>>>>> >> >> >>>>> >> very curious also about that Ross "hot" patch.
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> M.
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> ERRORS DETECTED DURING THE 500K steps tests
> with
> >>>>>>>>>>>> driver
> >>>>>>>>>>>> >> 319.23
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> #1 ERR writtent in mdout:
> >>>>>>>>>>>> >> >> >>>>> >> ------
> >>>>>>>>>>>> >> >> >>>>> >> | ERROR: max pairlist cutoff must be less
> than
> >>>>>>>>>>>> unit
> >>>>>>>>>>>> cell
> >>>>>>>>>>>> >> max
> >>>>>>>>>>>> >> >> >>>>> sphere
> >>>>>>>>>>>> >> >> >>>>> >> radius!
> >>>>>>>>>>>> >> >> >>>>> >> ------
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_1 JAC_NPT (at least 5000 steps
> >>>>>>>>>>>> successfully
> >>>>>>>>>>>> > done
> >>>>>>>>>>>> >> >> >>>>> before
> >>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 JAC_NPT (at least 8000 steps
> >>>>>>>>>>>> successfully
> >>>>>>>>>>>> > done
> >>>>>>>>>>>> >> >> >>>>> before
> >>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> #2 no ERR writtent in mdout, ERR written in
> >>>>>>>>>>>> standard
> >>>>>>>>>>>> output
> >>>>>>>>>>>> >> >> >>>>> (nohup.out)
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> ----
> >>>>>>>>>>>> >> >> >>>>> >> Error: unspecified launch failure launching
> >>>>>>>>>>>> kernel
> >>>>>>>>>>>> >> kNLSkinTest
> >>>>>>>>>>>> >> >> >>>>> >> cudaFree GpuBuffer::Deallocate failed
> unspecified
> >>>>>>>>>>>> launch
> >>>>>>>>>>>> >> >> failure
> >>>>>>>>>>>> >> >> >>>>> >> ----
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_1 CELLULOSE_NVE (at least 437 000
> >>>>>>>>>>>> steps
> >>>>>>>>>>>> >> >> successfully
> >>>>>>>>>>>> >> >> >>>>> done
> >>>>>>>>>>>> >> >> >>>>> >> before crash)
> >>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 JAC_NVE (at least 162 000
> steps
> >>>>>>>>>>>> >> successfully
> >>>>>>>>>>>> >> >> done
> >>>>>>>>>>>> >> >> >>>>> >> before
> >>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 CELLULOSE_NVE (at least 117 000
> >>>>>>>>>>>> steps
> >>>>>>>>>>>> >> >> successfully
> >>>>>>>>>>>> >> >> >>>>> done
> >>>>>>>>>>>> >> >> >>>>> >> before crash)
> >>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_1 JAC_NVE (at least 119 000
> steps
> >>>>>>>>>>>> >> successfully
> >>>>>>>>>>>> >> >> done
> >>>>>>>>>>>> >> >> >>>>> >> before
> >>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_2 JAC_NVE (at least 43 000 steps
> >>>>>>>>>>>> successfully
> >>>>>>>>>>>> >> >> done
> >>>>>>>>>>>> >> >> >>>>> before
> >>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> #3 no ERR writtent in mdout, ERR written in
> >>>>>>>>>>>> standard
> >>>>>>>>>>>> output
> >>>>>>>>>>>> >> >> >>>>> (nohup.out)
> >>>>>>>>>>>> >> >> >>>>> >> ----
> >>>>>>>>>>>> >> >> >>>>> >> cudaMemcpy GpuBuffer::Download failed
> unspecified
> >>>>>>>>>>>> launch
> >>>>>>>>>>>> >> >> failure
> >>>>>>>>>>>> >> >> >>>>> >> ----
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_1 JAC_NPT (at least 77 000 steps
> >>>>>>>>>>>> successfully
> >>>>>>>>>>>> >> >> done
> >>>>>>>>>>>> >> >> >>>>> before
> >>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_2 JAC_NPT (at least 58 000 steps
> >>>>>>>>>>>> successfully
> >>>>>>>>>>>> >> >> done
> >>>>>>>>>>>> >> >> >>>>> before
> >>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> Dne Thu, 30 May 2013 21:27:17 +0200 Scott Le
> >>>>>>>>>>>> Grand
> >>>>>>>>>>>> >> >> >>>>> >> <varelse2005.gmail.com>
> >>>>>>>>>>>> >> >> >>>>> >> napsal/-a:
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> Oops meant to send that to Jason...
> >>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>> >> >> >>>>> >>> Anyway, before we all panic, we need to get
> >>>>>>>>>>>> K20's
> >>>>>>>>>>>> behavior
> >>>>>>>>>>>> >> >> >>>>> analyzed
> >>>>>>>>>>>> >> >> >>>>> >>> here.
> >>>>>>>>>>>> >> >> >>>>> >>> If it's deterministic, this truly is a
> hardware
> >>>>>>>>>>>> issue. If
> >>>>>>>>>>>> >> >> not,
> >>>>>>>>>>>> >> >> >>>>> then
> >>>>>>>>>>>> >> >> >>>>> it
> >>>>>>>>>>>> >> >> >>>>> >>> gets interesting because 680 is deterministic
> as
> >>>>>>>>>>>> far
> >>>>>>>>>>>> as I
> >>>>>>>>>>>> >> can
> >>>>>>>>>>>> >> >> >>>>> tell...
> >>>>>>>>>>>> >> >> >>>>> >>> On May 30, 2013 12:24 PM, "Scott Le Grand"
> >>>>>>>>>>>> >> >> >>>>> <varelse2005.gmail.com>
> >>>>>>>>>>>> >> >> >>>>> >>> wrote:
> >>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>> >> >> >>>>> >>> If the errors are not deterministically
> >>>>>>>>>>>> triggered,
> >>>>>>>>>>>> they
> >>>>>>>>>>>> >> >> probably
> >>>>>>>>>>>> >> >> >>>>> >>> won't be
> >>>>>>>>>>>> >> >> >>>>> >>>> fixed by the patch, alas...
> >>>>>>>>>>>> >> >> >>>>> >>>> On May 30, 2013 12:15 PM, "Jason Swails"
> >>>>>>>>>>>> >> >> >>>>> <jason.swails.gmail.com>
> >>>>>>>>>>>> >> >> >>>>> >>>> wrote:
> >>>>>>>>>>>> >> >> >>>>> >>>>
> >>>>>>>>>>>> >> >> >>>>> >>>> Just a reminder to everyone based on what
> Ross
> >>>>>>>>>>>> said:
> >>>>>>>>>>>> >> there
> >>>>>>>>>>>> >> >> is a
> >>>>>>>>>>>> >> >> >>>>> >>>> pending
> >>>>>>>>>>>> >> >> >>>>> >>>>> patch to pmemd.cuda that will be coming out
> >>>>>>>>>>>> shortly
> >>>>>>>>>>>> >> (maybe
> >>>>>>>>>>>> >> >> even
> >>>>>>>>>>>> >> >> >>>>> >>>>> within
> >>>>>>>>>>>> >> >> >>>>> >>>>> hours). It's entirely possible that several
> >>>>>>>>>>>> of
> >>>>>>>>>>>> these
> >>>>>>>>>>>> > errors
> >>>>>>>>>>>> >> >> >>>>> are
> >>>>>>>>>>>> >> >> >>>>> >>>>> fixed
> >>>>>>>>>>>> >> >> >>>>> >>>>> by
> >>>>>>>>>>>> >> >> >>>>> >>>>> this patch.
> >>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>> >> >> >>>>> >>>>> All the best,
> >>>>>>>>>>>> >> >> >>>>> >>>>> Jason
> >>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>> >> >> >>>>> >>>>> On Thu, May 30, 2013 at 2:46 PM, filip
> fratev
> >>>>>>>>>>>> <
> >>>>>>>>>>>> >> >> >>>>> filipfratev.yahoo.com>
> >>>>>>>>>>>> >> >> >>>>> >>>>> wrote:
> >>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>> >> >> >>>>> >>>>> > I have observed the same crashes from time
> >>>>>>>>>>>> to
> >>>>>>>>>>>> time. I
> >>>>>>>>>>>> > will
> >>>>>>>>>>>> >> >> >>>>> run
> >>>>>>>>>>>> >> >> >>>>> >>>>> cellulose
> >>>>>>>>>>>> >> >> >>>>> >>>>> > nve for 100k and will past results here.
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > All the best,
> >>>>>>>>>>>> >> >> >>>>> >>>>> > Filip
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > ______________________________******__
> >>>>>>>>>>>> >> >> >>>>> >>>>> > From: Scott Le Grand
> >>>>>>>>>>>> <varelse2005.gmail.com>
> >>>>>>>>>>>> >> >> >>>>> >>>>> > To: AMBER Mailing List <amber.ambermd.org
> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > Sent: Thursday, May 30, 2013 9:01 PM
> >>>>>>>>>>>> >> >> >>>>> >>>>> > Subject: Re: [AMBER] experiences with EVGA
> >>>>>>>>>>>> GTX
> >>>>>>>>>>>> TITAN
> >>>>>>>>>>>> >> >> >>>>> Superclocked
> >>>>>>>>>>>> >> >> >>>>> -
> >>>>>>>>>>>> >> >> >>>>> >>>>> > memtestG80 - UNDERclocking in Linux ?
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > Run cellulose nve for 100k iterations
> twice
> >>>>>>>>>>>> . If
> >>>>>>>>>>>> the
> >>>>>>>>>>>> >> >> final
> >>>>>>>>>>>> >> >> >>>>> >>>>> energies
> >>>>>>>>>>>> >> >> >>>>> >>>>> don't
> >>>>>>>>>>>> >> >> >>>>> >>>>> > match, you have a hardware issue. No need
> >>>>>>>>>>>> to
> >>>>>>>>>>>> play
> >>>>>>>>>>>> with
> >>>>>>>>>>>> >> >> ntpr
> >>>>>>>>>>>> >> >> >>>>> or
> >>>>>>>>>>>> >> >> >>>>> any
> >>>>>>>>>>>> >> >> >>>>> >>>>> other
> >>>>>>>>>>>> >> >> >>>>> >>>>> > variable.
> >>>>>>>>>>>> >> >> >>>>> >>>>> > On May 30, 2013 10:58 AM,
> >>>>>>>>>>>> <pavel.banas.upol.cz>
> >>>>>>>>>>>> wrote:
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Dear all,
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > I would also like to share one of my
> >>>>>>>>>>>> experience
> >>>>>>>>>>>> with
> >>>>>>>>>>>> >> >> titan
> >>>>>>>>>>>> >> >> >>>>> >>>>> cards. We
> >>>>>>>>>>>> >> >> >>>>> >>>>> have
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > one gtx titan card and with one system
> >>>>>>>>>>>> (~55k
> >>>>>>>>>>>> atoms,
> >>>>>>>>>>>> > NVT,
> >>>>>>>>>>>> >> >> >>>>> >>>>> RNA+waters)
> >>>>>>>>>>>> >> >> >>>>> >>>>> we
> >>>>>>>>>>>> >> >> >>>>> >>>>> > run
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > into same troubles you are describing. I
> >>>>>>>>>>>> was
> >>>>>>>>>>>> also
> >>>>>>>>>>>> >> >> playing
> >>>>>>>>>>>> >> >> >>>>> with
> >>>>>>>>>>>> >> >> >>>>> >>>>> ntpr
> >>>>>>>>>>>> >> >> >>>>> >>>>> to
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > figure out what is going on, step by
> step.
> >>>>>>>>>>>> I
> >>>>>>>>>>>> >> understand
> >>>>>>>>>>>> >> >> >>>>> that
> >>>>>>>>>>>> >> >> >>>>> the
> >>>>>>>>>>>> >> >> >>>>> >>>>> code
> >>>>>>>>>>>> >> >> >>>>> >>>>> is
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > using different routines for calculation
> >>>>>>>>>>>> >> >> energies+forces or
> >>>>>>>>>>>> >> >> >>>>> only
> >>>>>>>>>>>> >> >> >>>>> >>>>> forces.
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > The
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > simulations of other systems are
> perfectly
> >>>>>>>>>>>> stable,
> >>>>>>>>>>>> >> >> running
> >>>>>>>>>>>> >> >> >>>>> for
> >>>>>>>>>>>> >> >> >>>>> >>>>> days
> >>>>>>>>>>>> >> >> >>>>> >>>>> and
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > weeks. Only that particular system
> >>>>>>>>>>>> systematically
> >>>>>>>>>>>> >> ends
> >>>>>>>>>>>> >> >> up
> >>>>>>>>>>>> >> >> >>>>> with
> >>>>>>>>>>>> >> >> >>>>> >>>>> this
> >>>>>>>>>>>> >> >> >>>>> >>>>> > error.
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > However, there was one interesting
> issue.
> >>>>>>>>>>>> When
> >>>>>>>>>>>> I
> >>>>>>>>>>>> set
> >>>>>>>>>>>> >> >> >>>>> ntpr=1,
> >>>>>>>>>>>> >> >> >>>>> the
> >>>>>>>>>>>> >> >> >>>>> >>>>> error
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > vanished (systematically in multiple
> runs)
> >>>>>>>>>>>> and
> >>>>>>>>>>>> the
> >>>>>>>>>>>> >> >> >>>>> simulation
> >>>>>>>>>>>> >> >> >>>>> was
> >>>>>>>>>>>> >> >> >>>>> >>>>> able to
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > run for more than millions of steps (I
> was
> >>>>>>>>>>>> not
> >>>>>>>>>>>> let
> >>>>>>>>>>>> it
> >>>>>>>>>>>> >> >> >>>>> running
> >>>>>>>>>>>> >> >> >>>>> for
> >>>>>>>>>>>> >> >> >>>>> >>>>> weeks
> >>>>>>>>>>>> >> >> >>>>> >>>>> > as
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > in the meantime I shifted that
> simulation
> >>>>>>>>>>>> to
> >>>>>>>>>>>> other
> >>>>>>>>>>>> >> card
> >>>>>>>>>>>> >> >> -
> >>>>>>>>>>>> >> >> >>>>> need
> >>>>>>>>>>>> >> >> >>>>> >>>>> data,
> >>>>>>>>>>>> >> >> >>>>> >>>>> not
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > testing). All other setting of ntpr
> >>>>>>>>>>>> failed.
> >>>>>>>>>>>> As
> >>>>>>>>>>>> I
> >>>>>>>>>>>> read
> >>>>>>>>>>>> >> >> this
> >>>>>>>>>>>> >> >> >>>>> >>>>> discussion, I
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > tried to set ene_avg_sampling=1 with
> some
> >>>>>>>>>>>> high
> >>>>>>>>>>>> value
> >>>>>>>>>>>> >> of
> >>>>>>>>>>>> >> >> >>>>> ntpr
> >>>>>>>>>>>> >> >> >>>>> (I
> >>>>>>>>>>>> >> >> >>>>> >>>>> expected
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > that this will shift the code to
> >>>>>>>>>>>> permanently
> >>>>>>>>>>>> use
> >>>>>>>>>>>> the
> >>>>>>>>>>>> >> >> >>>>> >>>>> force+energies
> >>>>>>>>>>>> >> >> >>>>> >>>>> part
> >>>>>>>>>>>> >> >> >>>>> >>>>> > of
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > the code, similarly to ntpr=1), but the
> >>>>>>>>>>>> error
> >>>>>>>>>>>> >> occurred
> >>>>>>>>>>>> >> >> >>>>> again.
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > I know it is not very conclusive for
> >>>>>>>>>>>> finding
> >>>>>>>>>>>> out
> >>>>>>>>>>>> what
> >>>>>>>>>>>> > is
> >>>>>>>>>>>> >> >> >>>>> >>>>> happening,
> >>>>>>>>>>>> >> >> >>>>> >>>>> at
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > least
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > not for me. Do you have any idea, why
> >>>>>>>>>>>> ntpr=1
> >>>>>>>>>>>> might
> >>>>>>>>>>>> > help?
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > best regards,
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Pavel
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > --
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Pavel Banáš
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > pavel.banas.upol.cz
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Department of Physical Chemistry,
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Palacky University Olomouc
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Czech Republic
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > ---------- Původní zpráva ----------
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Od: Jason Swails <
> jason.swails.gmail.com>
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Datum: 29. 5. 2013
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Předmět: Re: [AMBER] experiences with
> EVGA
> >>>>>>>>>>>> GTX
> >>>>>>>>>>>> TITAN
> >>>>>>>>>>>> >> >> >>>>> >>>>> Superclocked -
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > memtestG
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > 80 - UNDERclocking in Linux ?
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > "I'll answer a little bit:
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > NTPR=10 Etot after 2000 steps
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443256.6711
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443256.6711
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > > NTPR=200 Etot after 2000 steps
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443261.0705
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443261.0705
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > > Any idea why energies should depend on
> >>>>>>>>>>>> frequency
> >>>>>>>>>>>> of
> >>>>>>>>>>>> >> >> >>>>> energy
> >>>>>>>>>>>> >> >> >>>>> >>>>> records
> >>>>>>>>>>>> >> >> >>>>> >>>>> > (NTPR)
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > ?
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > It is a subtle point, but the answer is
> >>>>>>>>>>>> 'different
> >>>>>>>>>>>> >> code
> >>>>>>>>>>>> >> >> >>>>> paths.'
> >>>>>>>>>>>> >> >> >>>>> >>>>> In
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > general, it is NEVER necessary to
> compute
> >>>>>>>>>>>> the
> >>>>>>>>>>>> actual
> >>>>>>>>>>>> >> >> energy
> >>>>>>>>>>>> >> >> >>>>> of a
> >>>>>>>>>>>> >> >> >>>>> >>>>> molecule
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > during the course of standard molecular
> >>>>>>>>>>>> dynamics
> >>>>>>>>>>>> (by
> >>>>>>>>>>>> >> >> >>>>> analogy, it
> >>>>>>>>>>>> >> >> >>>>> >>>>> is
> >>>>>>>>>>>> >> >> >>>>> >>>>> NEVER
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > necessary to compute atomic forces
> during
> >>>>>>>>>>>> the
> >>>>>>>>>>>> course
> >>>>>>>>>>>> >> of
> >>>>>>>>>>>> >> >> >>>>> random
> >>>>>>>>>>>> >> >> >>>>> >>>>> Monte
> >>>>>>>>>>>> >> >> >>>>> >>>>> > Carlo
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > sampling).
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > For performance's sake, then, pmemd.cuda
> >>>>>>>>>>>> computes
> >>>>>>>>>>>> >> only
> >>>>>>>>>>>> >> >> the
> >>>>>>>>>>>> >> >> >>>>> force
> >>>>>>>>>>>> >> >> >>>>> >>>>> when
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > energies are not requested, leading to a
> >>>>>>>>>>>> different
> >>>>>>>>>>>> >> >> order of
> >>>>>>>>>>>> >> >> >>>>> >>>>> operations
> >>>>>>>>>>>> >> >> >>>>> >>>>> > for
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > those runs. This difference ultimately
> >>>>>>>>>>>> causes
> >>>>>>>>>>>> >> >> divergence.
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > To test this, try setting the variable
> >>>>>>>>>>>> >> >> ene_avg_sampling=10
> >>>>>>>>>>>> >> >> >>>>> in
> >>>>>>>>>>>> >> >> >>>>> the
> >>>>>>>>>>>> >> >> >>>>> >>>>> &cntrl
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > section. This will force pmemd.cuda to
> >>>>>>>>>>>> compute
> >>>>>>>>>>>> >> energies
> >>>>>>>>>>>> >> >> >>>>> every 10
> >>>>>>>>>>>> >> >> >>>>> >>>>> steps
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > (for energy averaging), which will in
> turn
> >>>>>>>>>>>> make
> >>>>>>>>>>>> the
> >>>>>>>>>>>> >> >> >>>>> followed
> >>>>>>>>>>>> >> >> >>>>> code
> >>>>>>>>>>>> >> >> >>>>> >>>>> path
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > identical for any multiple-of-10 value
> of
> >>>>>>>>>>>> ntpr.
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > --
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Jason M. Swails
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Quantum Theory Project,
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > University of Florida
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > Ph.D. Candidate
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > 352-392-4032
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > ______________________________****
> >>>>>>>>>>>> **_________________
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER mailing list
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER.ambermd.org
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> >>>>>>>>>>>> mailman/listinfo/amber<http://**
> lists.ambermd.org/**mailman/**
> >>>>>>>>>>>> listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> <
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> "
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > ______________________________****
> >>>>>>>>>>>> **_________________
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER mailing list
> >>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER.ambermd.org
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> >>>>>>>>>>>> mailman/listinfo/amber<http://**
> lists.ambermd.org/**mailman/**
> >>>>>>>>>>>> listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> <
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > ______________________________****
> >>>>>>>>>>>> **_________________
> >>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER mailing list
> >>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER.ambermd.org
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> >>>>>>>>>>>> mailman/listinfo/amber<http://**
> lists.ambermd.org/**mailman/**
> >>>>>>>>>>>> listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> <
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> > ______________________________****
> >>>>>>>>>>>> **_________________
> >>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER mailing list
> >>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER.ambermd.org
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> >>>>>>>>>>>> mailman/listinfo/amber<http://**
> lists.ambermd.org/**mailman/**
> >>>>>>>>>>>> listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> <
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>> >> >> >>>>> >>>>> --
> >>>>>>>>>>>> >> >> >>>>> >>>>> Jason M. Swails
> >>>>>>>>>>>> >> >> >>>>> >>>>> Quantum Theory Project,
> >>>>>>>>>>>> >> >> >>>>> >>>>> University of Florida
> >>>>>>>>>>>> >> >> >>>>> >>>>> Ph.D. Candidate
> >>>>>>>>>>>> >> >> >>>>> >>>>> 352-392-4032
> >>>>>>>>>>>> >> >> >>>>> >>>>> ______________________________****
> >>>>>>>>>>>> **_________________
> >>>>>>>>>>>> >> >> >>>>> >>>>> AMBER mailing list
> >>>>>>>>>>>> >> >> >>>>> >>>>> AMBER.ambermd.org
> >>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> >>>>>>>>>>>> mailman/listinfo/amber<http://**
> lists.ambermd.org/**mailman/**
> >>>>>>>>>>>> listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> <
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>> >> >> >>>>> >>>> ______________________________****
> >>>>>>>>>>>> **_________________
> >>>>>>>>>>>> >> >> >>>>> >>> AMBER mailing list
> >>>>>>>>>>>> >> >> >>>>> >>> AMBER.ambermd.org
> >>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>> http://lists.ambermd.org/******mailman/listinfo/amber<
> http://lists.ambermd.org/****mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/**mailman/**listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> <
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>> >> >> >>>>> >>> __________ Informace od ESET NOD32 Antivirus,
> >>>>>>>>>>>> verze
> >>>>>>>>>>>> >> databaze
> >>>>>>>>>>>> >> >> 8394
> >>>>>>>>>>>> >> >> >>>>> >>> (20130530) __________
> >>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>> >> >> >>>>> >>> Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>> >> >> >>>>> >>> http://www.eset.cz
> >>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >> --
> >>>>>>>>>>>> >> >> >>>>> >> Tato zpráva byla vytvořena převratným poštovním
> >>>>>>>>>>>> klientem
> >>>>>>>>>>>> > Opery:
> >>>>>>>>>>>> >> >> >>>>> >> http://www.opera.com/mail/
> >>>>>>>>>>>> >> >> >>>>> >> ______________________________**
> >>>>>>>>>>>> **_________________
> >>>>>>>>>>>> >> >> >>>>> >> AMBER mailing list
> >>>>>>>>>>>> >> >> >>>>> >> AMBER.ambermd.org
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>> >> >> >>>>> > ______________________________**
> >>>>>>>>>>>> **_________________
> >>>>>>>>>>>> >> >> >>>>> > AMBER mailing list
> >>>>>>>>>>>> >> >> >>>>> > AMBER.ambermd.org
> >>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> > __________ Informace od ESET NOD32 Antivirus,
> >>>>>>>>>>>> verze
> >>>>>>>>>>>> databaze
> >>>>>>>>>>>> >> >> 8394
> >>>>>>>>>>>> >> >> >>>>> > (20130530) __________
> >>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> > http://www.eset.cz
> >>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> >> >> >>>>> --
> >>>>>>>>>>>> >> >> >>>>> Tato zpráva byla vytvořena převratným poštovním
> >>>>>>>>>>>> klientem
> >>>>>>>>>>>> Opery:
> >>>>>>>>>>>> >> >> >>>>> http://www.opera.com/mail/
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> >> >> >>>>>
> ______________________________****_________________
> >>>>>>>>>>>> >> >> >>>>> AMBER mailing list
> >>>>>>>>>>>> >> >> >>>>> AMBER.ambermd.org
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>> >> >> >>> ______________________________****_________________
> >>>>>>>>>>>> >> >> >>> AMBER mailing list
> >>>>>>>>>>>> >> >> >>> AMBER.ambermd.org
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>> __________ Informace od ESET NOD32 Antivirus, verze
> >>>>>>>>>>>> databaze
> >>>>>>>>>>>> 8395
> >>>>>>>>>>>> >> >> >>> (20130531) __________
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>> Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>> http://www.eset.cz
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >> >
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> --
> >>>>>>>>>>>> >> >> Tato zpráva byla vytvořena převratným poštovním klientem
> >>>>>>>>>>>> Opery:
> >>>>>>>>>>>> >> >> http://www.opera.com/mail/
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >> ______________________________****_________________
> >>>>>>>>>>>> >> >> AMBER mailing list
> >>>>>>>>>>>> >> >> AMBER.ambermd.org
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> >>
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> > __________ Informace od ESET NOD32 Antivirus, verze
> >>>>>>>>>>>> databaze
> >>>>>>>>>>>> 8397
> >>>>>>>>>>>> >> > (20130531) __________
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> > Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz - poskozeny archiv
> >>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>> >> GB_out_plus_diff_Files.tar
> >>>>>>>>>>>> >> > - poskozeny archiv
> >>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>> GB_out_plus_diff_Files.tar.gz -
> >>>>>>>>>>>> >> > poskozeny archiv
> >>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>> GB_out_plus_diff_Files.tar.gz >
> >>>>>>>>>>>> >> GZIP
> >>>>>>>>>>>> >> > > GB_out_plus_diff_Files.tar - poskozeny archiv
> >>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>> GB_out_plus_diff_Files.tar.gz >
> >>>>>>>>>>>> >> GZIP
> >>>>>>>>>>>> >> > > GB_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>> GB_nucleosome-sim3.mdout-full -
> >>>>>>>>>>>> >> > vyskytl se problem pri cteni archivu
> >>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz - poskozeny archiv
> >>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar - poskozeny archiv
> >>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>> PME_out_plus_diff_Files.tar.gz -
> >>>>>>>>>>>> >> > poskozeny archiv
> >>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>> PME_out_plus_diff_Files.tar.gz >
> >>>>>>>>>>>> >> > GZIP > PME_out_plus_diff_Files.tar - poskozeny archiv
> >>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>> >> > PME_out_plus_diff_Filestar > TAR >
> >>>>>>>>>>>> PME_out_plus_diff_Files.tar.gz
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >> GZIP
> >>>>>>>>>>>> >> > > PME_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>> >> > PME_JAC_production_NPT-sim3.****mdout-full - vyskytl se
> >>>>>>>>>>>> problem
> >>>>>>>>>>>> pri
> >>>>>>>>>>>> cteni
> >>>>>>>>>>>> >> > archivu
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >> > http://www.eset.cz
> >>>>>>>>>>>> >> >
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> --
> >>>>>>>>>>>> >> Tato zpráva byla vytvořena převratným poštovním klientem
> >>>>>>>>>>>> Opery:
> >>>>>>>>>>>> >> http://www.opera.com/mail/
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> >> ______________________________****_________________
> >>>>>>>>>>>> >> AMBER mailing list
> >>>>>>>>>>>> >> AMBER.ambermd.org
> >>>>>>>>>>>> >>
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> > ______________________________****_________________
> >>>>>>>>>>>> > AMBER mailing list
> >>>>>>>>>>>> > AMBER.ambermd.org
> >>>>>>>>>>>> >
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >
> >>>>>>>>>>>> > __________ Informace od ESET NOD32 Antivirus, verze databaze
> >>>>>>>>>>>> 8398
> >>>>>>>>>>>> > (20130531) __________
> >>>>>>>>>>>> >
> >>>>>>>>>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>> >
> >>>>>>>>>>>> > http://www.eset.cz
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >
> >>>>>>>>>>>> >
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> --
> >>>>>>>>>>>> Tato zpráva byla vytvořena převratným poštovním klientem
> Opery:
> >>>>>>>>>>>> http://www.opera.com/mail/
> >>>>>>>>>>>>
> >>>>>>>>>>>> ______________________________****_________________
> >>>>>>>>>>>> AMBER mailing list
> >>>>>>>>>>>> AMBER.ambermd.org
> >>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>> >
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> __________ Informace od ESET NOD32 Antivirus, verze databaze
> 8401
> >>>>>>>>>> (20130601) __________
> >>>>>>>>>>
> >>>>>>>>>> Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>
> >>>>>>>>>> http://www.eset.cz
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>> --
> >>>>>>>> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> >>>>>>>> http://www.opera.com/mail/
> >>>>>>>>
> >>>>>>>> ______________________________**_________________
> >>>>>>>> AMBER mailing list
> >>>>>>>> AMBER.ambermd.org
> >>>>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>
> >>>>>> ______________________________**_________________
> >>>>> AMBER mailing list
> >>>>> AMBER.ambermd.org
> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>
> >>>>> __________ Informace od ESET NOD32 Antivirus, verze databaze 8403
> >>>>> (20130602) __________
> >>>>>
> >>>>> Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>
> >>>>> http://www.eset.cz
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>
> >>>>
> >>>
> >>>
> >>
> >>
> >
> > --
> > Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> > http://www.opera.com/mail/
> >
> > _______________________________________________
> > AMBER mailing list
> > AMBER.ambermd.org
> > http://lists.ambermd.org/mailman/listinfo/amber
> >
> >
> _______________________________________________
> AMBER mailing list
> AMBER.ambermd.org
> http://lists.ambermd.org/mailman/listinfo/amber
>
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Wed Jun 19 2013 - 10:00:02 PDT
Custom Search