Re: [AMBER] experiences with EVGA GTX TITAN Superclocked - memtestG80 - UNDERclocking in Linux ?

From: Scott Le Grand <varelse2005.gmail.com>
Date: Wed, 19 Jun 2013 10:21:53 -0700

I have fixed the GB weirdness (I think). But now I have found a new issue
with IPS simulations where I can repro bad behavior in 5 minutes or less
(either Shake or the Local Forces misfire somehow in a way I can't quite
figure out).

I'm pretty dependent on NVIDIA at this point.

Scott



On Wed, Jun 19, 2013 at 9:55 AM, Marek Maly <marek.maly.ujep.cz> wrote:

> Hi Scott,
> thanks for update ! Regarding "GB/Titan" issue it is perhaps still
> in debugging phase too (like cuFFT/Titan problem) if I understood well.
>
> Best,
>
> Marek
>
>
>
>
> Dne Wed, 19 Jun 2013 17:50:51 +0200 Scott Le Grand <varelse2005.gmail.com>
> napsal/-a:
>
> > Hey Marek,
> > No updates per se. I had a theory about what was going on that proved to
> > be wrong after testing, but I'm still waiting on NVIDIA to report
> > something
> > beyond having reproed the problem.
> >
> > Really really really interested in GTX 780 data right now...
> >
> >
> >
> > On Wed, Jun 19, 2013 at 8:20 AM, Marek Maly <marek.maly.ujep.cz> wrote:
> >
> >> Hi all,
> >>
> >> just a small update from my site.
> >>
> >> As I have yesterday obtained announcement that the CUDA 5.5 is
> >> now available for public (not just for developers).
> >>
> >> I downloaded it from here:
> >>
> >> https://developer.nvidia.com/**cuda-pre-production<
> https://developer.nvidia.com/cuda-pre-production>
> >>
> >> It is still "just" release candidate ( as all Amber/Titan club members
> >> perfectly know :)) ).
> >>
> >> So I installed this newest release and recompiled Amber cuda code.
> >>
> >> I was hoping that maybe there was "silently" incorporated some
> >> improvement (e.g. in cuFFT) as the result e.g. of Scott's bug report.
> >>
> >> The results of my 100K tests are attached. It seems that comparing to my
> >> latest
> >> tests with CUDA 5.5. release candidate from June 3rd (when it was
> >> accessible just for CUDA developers in the form of *.run binary
> >> installer)
> >> there
> >> is some slight improvement - e.g. my more stable TITAN was able to
> >> finish
> >> successfully
> >> all the 100K tests including Cellulose twice. But there is still an
> >> issue
> >> with JAC NVE/NPT irreproducible results. On my "less stable" TITAN the
> >> results are slightly better
> >> then those older ones as well but still not err free (JAC/CELLULOSE) -
> >> see
> >> attached file.
> >>
> >> FACTOR IX NVE/NPT finished again with 100% reproducibility on both GPUs
> >> as
> >> usually.
> >>
> >> Scott, do you have any update regarding the "cuFFT"/TITAN issue which
> >> you
> >> reported/described
> >> to NVIDIA guys ? The latest info from you regarding this story was, that
> >> they were able to
> >> reproduce the "cuFFT"/TITAN error as well. Do you have any more recent
> >> information ? How long
> >> time it might take to NVIDIA developers to fully solve such problem in
> >> your opinion ?
> >>
> >> Another thing. It seems that you successfully solved the "GB/TITAN"
> >> problem in case of bigger molecular systems, here is your relevant
> >> message
> >> form June 7th.
> >>
> >> ------------------------------**----------------
> >>
> >> Really really interesting...
> >>
> >> I seem to have found a fix for the GB issues on my Titan - not so
> >> surprisingly, it's the same fix as on GTX4xx/GTX5xx...
> >>
> >> But this doesn't yet explain the weirdness with cuFFT so we're not done
> >> here yet...
> >> ------------------------------**---------------
> >>
> >> It was already after the latest Amber12 bugfix18 was released and there
> >> was no additional
> >> bugfix released from that moment. So the "GB/TITAN" patch will be
> >> released later maybe as the part of some bigger bugfix ? Or you simply
> >> additionally included it into bugfix 18 after it's release ?
> >>
> >>
> >> My last question maybe deserves the new separate thread, but anyway
> >> would
> >> be interesting
> >> to have some information how "Amber-stable" are GTX780 comparing to
> >> TITANS
> >> (of course based
> >> on experience of more users or on testing more than 1 or 2 GTX780 GPUs).
> >>
> >> Best wishes,
> >>
> >> Marek
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >> Dne Mon, 03 Jun 2013 01:57:36 +0200 Marek Maly <marek.maly.ujep.cz>
> >> napsal/-a:
> >>
> >>
> >> Hi here are my results with CUDA 5.5
> >>> (Total energy at step 100K(PME)/1000K(GB) (driver 319.23, Amber12
> >>> bugfix
> >>> 18 applied, cuda 5.5))
> >>>
> >>>
> >>> No significant differences comparing the previous test with CUDA 5.0
> >>> (I also added those data to the attached table with CUDA 5.5 test).
> >>>
> >>> Still the same trend instability in JAC tests, perfect stability and
> >>> reproducibility
> >>> in FACTOR_IX tests (interesting isn't it ? especially if we consider
> >>> 23K
> >>> atoms
> >>> in JAC case and 90K atoms in case of FACTOR_IX). Again the same
> >>> crashes in
> >>> CELLULOSE
> >>> test now also in case of TITAN_1. Also in stable and reproducible
> >>> FACTOR_IX slightly
> >>> changed the final energy values comparing to CUDA 5.0 case.
> >>>
> >>> GB simulations (1M steps) again perfectly stable and reproducible.
> >>>
> >>> So to conclude, Scott we trust you :)) !
> >>>
> >>> If you have any idea what to try else (except GPU bios editing, perhaps
> >>> too
> >>> premature step at this moment) let me know. I got just last idea,
> >>> which could be perhaps to try change rand seed and see if it has any
> >>> influence in actual trends (e.g. JAC versus FACTOR_IX).
> >>>
> >>> TO ET : I am curious about your test in single GPU configuration.
> >>> Regarding
> >>> to your Win tests, in my opinion it is just wasting of time. They
> >>> perhaps
> >>> tells
> >>> you just something about the GPU performance not about the eventual GPU
> >>> "soft" errs.
> >>>
> >>> If intensive memtestG80 and/or cuda_memtest results were negative
> >>> there is
> >>> in my opinion
> >>> very unlikely that Win performace testers will find any errs, but I am
> >>> not
> >>> an expert
> >>> here ...
> >>>
> >>> Anyway If you learn which tests the ebuyer is using to confirm GPU
> >>> errs,
> >>> let us know.
> >>>
> >>> M.
> >>>
> >>>
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> Dne Sun, 02 Jun 2013 19:22:54 +0200 Marek Maly <marek.maly.ujep.cz>
> >>> napsal/-a:
> >>>
> >>> Hi so I finally succeeded to compile GPU Amber part under CUDA 5.5
> >>>> (after "hacking" of the configure2 file) with common results in
> >>>> consequent tests:
> >>>>
> >>>> ------
> >>>> 80 file comparisons passed
> >>>> 9 file comparisons failed
> >>>> 0 tests experienced errors
> >>>> ------
> >>>>
> >>>> So now I am running the 100K(PME)/1000K(GB) repetitive benchmark tests
> >>>> under
> >>>> this configuration: drv. 319.23, CUDA 5.5. , bugfix 18 installed
> >>>>
> >>>> When I finish it I will report results here.
> >>>>
> >>>> M.
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>> Dne Sun, 02 Jun 2013 18:44:23 +0200 Marek Maly <marek.maly.ujep.cz>
> >>>> napsal/-a:
> >>>>
> >>>> Hi Scott thanks for the update !
> >>>>>
> >>>>> Anyway any explanation regarding "cuFFT hypothesis" why there are no
> >>>>> problems
> >>>>> with GTX 580, GTX 680 or even K20c ???
> >>>>>
> >>>>>
> >>>>> meanwhile I also tried to recompile GPU part of Amber with
> >>>>> cuda 5.5 installed before, I have obtained these errs
> >>>>> already in configure phase:
> >>>>>
> >>>>> --------
> >>>>> [root.dyn-138-272 amber12]# ./configure -cuda -noX11 gnu
> >>>>> Checking for updates...
> >>>>> Checking for available patches online. This may take a few seconds...
> >>>>>
> >>>>> Available AmberTools 13 patches:
> >>>>>
> >>>>> No patches available
> >>>>>
> >>>>> Available Amber 12 patches:
> >>>>>
> >>>>> No patches available
> >>>>> Searching for python2... Found python2.6: /usr/bin/python2.6
> >>>>> Error: Unsupported CUDA version 5.5 detected.
> >>>>> AMBER requires CUDA version == 4.2 .or. 5.0
> >>>>> Configure failed due to the errors above!
> >>>>> ---------
> >>>>>
> >>>>> so it seems that Amber is possible to compile only with CUDA 4.2 or
> >>>>> 5.0
> >>>>> at
> >>>>> the moment:
> >>>>>
> >>>>> and this part of configure2 file has to be edited:
> >>>>>
> >>>>>
> >>>>> -----------
> >>>>> nvcc="$CUDA_HOME/bin/nvcc"
> >>>>> sm35flags='-gencode arch=compute_35,code=sm_35'
> >>>>> sm30flags='-gencode arch=compute_30,code=sm_30'
> >>>>> sm20flags='-gencode arch=compute_20,code=sm_20'
> >>>>> sm13flags='-gencode arch=compute_13,code=sm_13'
> >>>>> nvccflags="$sm13flags $sm20flags"
> >>>>> cudaversion=`$nvcc --version | grep 'release' | cut -d' ' -f5 |
> >>>>> cut
> >>>>> -d',' -f1`
> >>>>> if [ "$cudaversion" == "5.0" ]; then
> >>>>> echo "CUDA Version $cudaversion detected"
> >>>>> nvccflags="$nvccflags $sm30flags $sm35flags"
> >>>>> elif [ "$cudaversion" == "4.2" ]; then
> >>>>> echo "CUDA Version $cudaversion detected"
> >>>>> nvccflags="$nvccflags $sm30flags"
> >>>>> else
> >>>>> echo "Error: Unsupported CUDA version $cudaversion detected."
> >>>>> echo "AMBER requires CUDA version == 4.2 .or. 5.0"
> >>>>> exit 1
> >>>>> fi
> >>>>> nvcc="$nvcc $nvccflags"
> >>>>>
> >>>>> fi
> >>>>>
> >>>>> -----------
> >>>>>
> >>>>> would it be just OK to change
> >>>>> "if [ "$cudaversion" == "5.0" ]; then"
> >>>>>
> >>>>> to
> >>>>>
> >>>>> "if [ "$cudaversion" == "5.5" ]; then"
> >>>>>
> >>>>>
> >>>>> or some more flags etc. should be defined here to proceed
> >>>>> successfully ?
> >>>>>
> >>>>>
> >>>>> BTW it seems Scott, that you are on the way to isolate the problem
> >>>>> soon
> >>>>> so maybe it's better to wait and not to loose time with cuda 5.5
> >>>>> experiments.
> >>>>>
> >>>>> I just thought that cuda 5.5 might be more "friendly" to Titans :))
> >>>>> e.g.
> >>>>> in terms of cuFFT function ....
> >>>>>
> >>>>>
> >>>>> I will keep fingers crossed :))
> >>>>>
> >>>>> M.
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>> Dne Sun, 02 Jun 2013 18:33:52 +0200 Scott Le Grand
> >>>>> <varelse2005.gmail.com>
> >>>>> napsal/-a:
> >>>>>
> >>>>> PS this *might* indicate a software bug in cuFFT, but it needs more
> >>>>>> characterization... And things are going to get a little stream of
> >>>>>> consciousness from here because you're getting unfiltered raw data,
> >>>>>> so
> >>>>>> please don't draw any conclusions towards anything yet - I'm just
> >>>>>> letting
> >>>>>> you guys know what I'm finding out as I find it...
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>> On Sun, Jun 2, 2013 at 9:31 AM, Scott Le Grand
> >>>>>> <varelse2005.gmail.com>wrote:
> >>>>>>
> >>>>>> And bingo...
> >>>>>>>
> >>>>>>> At the very least, the reciprocal sum is intermittently
> >>>>>>> inconsistent...
> >>>>>>> This explains the irreproducible behavior...
> >>>>>>>
> >>>>>>> And here's the level of inconsistency:
> >>>>>>> 31989.38940628897399 vs
> >>>>>>> 31989.39168370794505
> >>>>>>>
> >>>>>>> That's error at the level of 1e-7 or a somehow missed
> >>>>>>> single-precision
> >>>>>>> transaction somewhere...
> >>>>>>>
> >>>>>>> The next question is figuring out why... This may or may not
> >>>>>>> ultimately
> >>>>>>> explain the crashes you guys are also seeing...
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>> On Sun, Jun 2, 2013 at 9:07 AM, Scott Le Grand
> >>>>>>> <varelse2005.gmail.com>wrote:
> >>>>>>>
> >>>>>>>
> >>>>>>>> Observations:
> >>>>>>>> 1. The degree to which the reproducibility is broken *does*
> >>>>>>>> appear to
> >>>>>>>> vary between individual Titan GPUs. One of my Titans breaks
> >>>>>>>> within
> >>>>>>>> 10K
> >>>>>>>> steps on cellulose, the other one made it to 100K steps twice
> >>>>>>>> without
> >>>>>>>> doing
> >>>>>>>> so leading me to believe it could be trusted (until yesterday
> >>>>>>>> where I
> >>>>>>>> now
> >>>>>>>> see it dies between 50K and 100K steps most of the time).
> >>>>>>>>
> >>>>>>>> 2. GB hasn't broken (yet). So could you run myoglobin for 500K
> >>>>>>>> and
> >>>>>>>> TRPcage for 1,000,000 steps and let's see if that's universal.
> >>>>>>>>
> >>>>>>>> 3. Turning on double-precision mode makes my Titan crash rather
> >>>>>>>> than
> >>>>>>>> run
> >>>>>>>> irreproducibly, sigh...
> >>>>>>>>
> >>>>>>>> So whatever is going on is triggered by something in PME but not
> >>>>>>>> GB.
> >>>>>>>> So
> >>>>>>>> that's either the radix sort, the FFT, the Ewald grid
> >>>>>>>> interpolation,
> >>>>>>>> or the
> >>>>>>>> neighbor list code. Fixing this involves isolating this and
> >>>>>>>> figuring
> >>>>>>>> out
> >>>>>>>> what exactly goes haywire. It could *still* be software at some
> >>>>>>>> very
> >>>>>>>> small
> >>>>>>>> probability but the combination of both 680 and K20c with ECC off
> >>>>>>>> running
> >>>>>>>> reliably is really pointing towards the Titans just being clocked
> >>>>>>>> too
> >>>>>>>> fast.
> >>>>>>>>
> >>>>>>>> So how long with this take? Asking people how long it takes to
> >>>>>>>> fix a
> >>>>>>>> bug
> >>>>>>>> never really works out well. That said, I found the 480 bug
> >>>>>>>> within a
> >>>>>>>> week
> >>>>>>>> and my usual turnaround for a bug with a solid repro is <24 hours.
> >>>>>>>>
> >>>>>>>> Scott
> >>>>>>>>
> >>>>>>>> On Sun, Jun 2, 2013 at 7:58 AM, Marek Maly <marek.maly.ujep.cz>
> >>>>>>>> wrote:
> >>>>>>>>
> >>>>>>>> Hi all,
> >>>>>>>>>
> >>>>>>>>> here are my results after bugfix 18 application (see attachment).
> >>>>>>>>>
> >>>>>>>>> In principle I don't see any "drastical" changes.
> >>>>>>>>>
> >>>>>>>>> FACTOR_IX still perfectly stable/reproducible on both cards,
> >>>>>>>>>
> >>>>>>>>> JAC tests - problems with finishing AND/OR reproducibility the
> >>>>>>>>> same CELLULOSE_NVE although here it seems that my TITAN_1
> >>>>>>>>> has no problems with this test (but the same same trend I saw
> >>>>>>>>> also
> >>>>>>>>> before bugfix 18 - see my older 500K steps test).
> >>>>>>>>>
> >>>>>>>>> But anyway bugfix 18 brought here one change.
> >>>>>>>>>
> >>>>>>>>> The err
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> #1 ERR writtent in mdout:
> >>>>>>>>> ------
> >>>>>>>>> | ERROR: max pairlist cutoff must be less than unit cell max
> >>>>>>>>> sphere
> >>>>>>>>> radius!
> >>>>>>>>> ------
> >>>>>>>>>
> >>>>>>>>> was substituted with err/warning ?
> >>>>>>>>>
> >>>>>>>>> #0 no ERR writtent in mdout, ERR written in standard output
> >>>>>>>>> (nohup.out)
> >>>>>>>>> -----
> >>>>>>>>> Nonbond cells need to be recalculated, restart simulation from
> >>>>>>>>> previous
> >>>>>>>>> checkpoint
> >>>>>>>>> with a higher value for skinnb.
> >>>>>>>>>
> >>>>>>>>> -----
> >>>>>>>>>
> >>>>>>>>> Another thing,
> >>>>>>>>>
> >>>>>>>>> recently I started on another machine and GTX 580 GPU simulation
> >>>>>>>>> of
> >>>>>>>>> relatively
> >>>>>>>>> big system ( 364275 atoms/PME ). The system is composed also from
> >>>>>>>>> the
> >>>>>>>>> "exotic" molecules like polymers. ff12SB, gaff, GLYCAM
> >>>>>>>>> forcefields
> >>>>>>>>> used
> >>>>>>>>> here. I had problem even with minimization part here, having big
> >>>>>>>>> energy
> >>>>>>>>> on the start:
> >>>>>>>>>
> >>>>>>>>> -----
> >>>>>>>>> NSTEP ENERGY RMS GMAX NAME
> >>>>>>>>> NUMBER
> >>>>>>>>> 1 2.8442E+09 2.1339E+02 1.7311E+04 O
> >>>>>>>>> 32998
> >>>>>>>>>
> >>>>>>>>> BOND = 11051.7467 ANGLE = 17720.4706 DIHED =
> >>>>>>>>> 18977.7584
> >>>>>>>>> VDWAALS = ************* EEL = -1257709.6203 HBOND =
> >>>>>>>>> 0.0000
> >>>>>>>>> 1-4 VDW = 7253.7412 1-4 EEL = 149867.0207 RESTRAINT =
> >>>>>>>>> 0.0000
> >>>>>>>>>
> >>>>>>>>> ----
> >>>>>>>>>
> >>>>>>>>> with no chance to minimize the system even with 50 000 steps in
> >>>>>>>>> both
> >>>>>>>>> min cycles (with constrained and unconstrained solute) and hence
> >>>>>>>>> heating
> >>>>>>>>> NVT
> >>>>>>>>> crashed immediately even with very small dt. I patched Amber12
> >>>>>>>>> here
> >>>>>>>>> with
> >>>>>>>>> the
> >>>>>>>>> bugfix 18 and the minimization was done without any problem with
> >>>>>>>>> common
> >>>>>>>>> 5000 steps
> >>>>>>>>> (obtaining target Energy -1.4505E+06 while that initial was that
> >>>>>>>>> written
> >>>>>>>>> above).
> >>>>>>>>>
> >>>>>>>>> So indeed bugfix 18 solved some issues, but unfortunately not
> >>>>>>>>> those
> >>>>>>>>> related to
> >>>>>>>>> Titans.
> >>>>>>>>>
> >>>>>>>>> Here I will try to install cuda 5.5, recompile GPU Amber part
> >>>>>>>>> with
> >>>>>>>>> this
> >>>>>>>>> new
> >>>>>>>>> cuda version and repeat the 100K tests.
> >>>>>>>>>
> >>>>>>>>> Scott, let us know how finished your experiment with
> >>>>>>>>> downclocking of
> >>>>>>>>> Titan.
> >>>>>>>>> Maybe the best choice would be here to flash Titan directly with
> >>>>>>>>> your
> >>>>>>>>> K20c bios :))
> >>>>>>>>>
> >>>>>>>>> M.
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> Dne Sat, 01 Jun 2013 21:09:46 +0200 Marek Maly
> >>>>>>>>> <marek.maly.ujep.cz>
> >>>>>>>>> napsal/-a:
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> Hi,
> >>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> first of all thanks for providing of your test results !
> >>>>>>>>>>
> >>>>>>>>>> It seems that your results are more or less similar to that of
> >>>>>>>>>> mine maybe with the exception of the results on FactorIX tests
> >>>>>>>>>> where I had perfect stability and 100% or close to 100%
> >>>>>>>>>> reproducibility.
> >>>>>>>>>>
> >>>>>>>>>> Anyway the type of errs which you reported are the same which I
> >>>>>>>>>> obtained.
> >>>>>>>>>>
> >>>>>>>>>> So let's see if the bugfix 18 will help here (or at least on NPT
> >>>>>>>>>> tests)
> >>>>>>>>>> or not. As I wrote just before few minutes, it seems that it was
> >>>>>>>>>> not
> >>>>>>>>>> still
> >>>>>>>>>> loaded
> >>>>>>>>>> to the given server, although it's description is already
> >>>>>>>>>> present
> >>>>>>>>>> on
> >>>>>>>>>> the
> >>>>>>>>>> given
> >>>>>>>>>> web page ( see
> >>>>>>>>>> http://ambermd.org/bugfixes12.****html<
> http://ambermd.org/bugfixes12.**html>
> >>>>>>>>>> <http://ambermd.org/**bugfixes12.html<
> http://ambermd.org/bugfixes12.html>
> >>>>>>>>>> >).
> >>>>>>>>>>
> >>>>>>>>>> As you can see, this bugfix contains also changes in CPU code
> >>>>>>>>>> although
> >>>>>>>>>> the majority is devoted to GPU code, so perhaps the best will
> >>>>>>>>>> be to
> >>>>>>>>>> recompile
> >>>>>>>>>> whole amber with this patch although this patch would be perhaps
> >>>>>>>>>> applied
> >>>>>>>>>> even after just
> >>>>>>>>>> GPU configure command ( i.e. ./configure -cuda -noX11 gnu ) but
> >>>>>>>>>> after
> >>>>>>>>>> consequent
> >>>>>>>>>> building, just the GPU binaries will be updated. Anyway I would
> >>>>>>>>>> rather
> >>>>>>>>>> recompile
> >>>>>>>>>> whole Amber after this patch.
> >>>>>>>>>>
> >>>>>>>>>> Regarding to GPU test under linux you may try memtestG80
> >>>>>>>>>> (please use the updated/patched version from here
> >>>>>>>>>> https://github.com/ihaque/****memtestG80<
> https://github.com/ihaque/**memtestG80>
> >>>>>>>>>> <https://github.com/**ihaque/memtestG80<
> https://github.com/ihaque/memtestG80>
> >>>>>>>>>> >
> >>>>>>>>>> )
> >>>>>>>>>>
> >>>>>>>>>> just use git command like:
> >>>>>>>>>>
> >>>>>>>>>> git clone
> >>>>>>>>>> https://github.com/ihaque/****memtestG80.git<
> https://github.com/ihaque/**memtestG80.git>
> >>>>>>>>>> <https://github.**com/ihaque/memtestG80.git<
> https://github.com/ihaque/memtestG80.git>
> >>>>>>>>>> >**PATCHED_MEMTEST-G80
> >>>>>>>>>>
> >>>>>>>>>> to download all the files and save them into directory named
> >>>>>>>>>> PATCHED_MEMTEST-G80.
> >>>>>>>>>>
> >>>>>>>>>> another possibility is to try perhaps similar (but maybe more
> >>>>>>>>>> up to
> >>>>>>>>>> date)
> >>>>>>>>>> test
> >>>>>>>>>> cuda_memtest (
> >>>>>>>>>> http://sourceforge.net/****projects/cudagpumemtest/<
> http://sourceforge.net/**projects/cudagpumemtest/>
> >>>>>>>>>> <http:**//sourceforge.net/projects/**cudagpumemtest/<
> http://sourceforge.net/projects/cudagpumemtest/>
> >>>>>>>>>> >).
> >>>>>>>>>>
> >>>>>>>>>> regarding ig value: If ig is not present in mdin, the default
> >>>>>>>>>> value
> >>>>>>>>>> is
> >>>>>>>>>> used
> >>>>>>>>>> (e.g. 71277) if ig=-1 the random seed will be based on the
> >>>>>>>>>> current
> >>>>>>>>>> date
> >>>>>>>>>> and time, and hence will be different for every run (not a good
> >>>>>>>>>> variant
> >>>>>>>>>> for our testts). I simply deleted eventual ig records from all
> >>>>>>>>>> mdins
> >>>>>>>>>> so
> >>>>>>>>>> I
> >>>>>>>>>> assume that in each run the default seed 71277 was automatically
> >>>>>>>>>> used.
> >>>>>>>>>>
> >>>>>>>>>> M.
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> Dne Sat, 01 Jun 2013 20:26:16 +0200 ET <sketchfoot.gmail.com>
> >>>>>>>>>> napsal/-a:
> >>>>>>>>>>
> >>>>>>>>>> Hi,
> >>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> I've put the graphics card into a machine with the working GTX
> >>>>>>>>>>> titan
> >>>>>>>>>>> that I
> >>>>>>>>>>> mentioned earlier.
> >>>>>>>>>>>
> >>>>>>>>>>> The Nvidia driver version is: 133.30
> >>>>>>>>>>>
> >>>>>>>>>>> Amber version is:
> >>>>>>>>>>> AmberTools version 13.03
> >>>>>>>>>>> Amber version 12.16
> >>>>>>>>>>>
> >>>>>>>>>>> I ran 50k steps with the amber benchmark using ig=43689 on both
> >>>>>>>>>>> cards.
> >>>>>>>>>>> For
> >>>>>>>>>>> the purpose of discriminating between them, the card I believe
> >>>>>>>>>>> (fingers
> >>>>>>>>>>> crossed) is working is called GPU-00_TeaNCake, whilst the other
> >>>>>>>>>>> one
> >>>>>>>>>>> is
> >>>>>>>>>>> called GPU-01_008.
> >>>>>>>>>>>
> >>>>>>>>>>> *When I run the tests on GPU-01_008:*
> >>>>>>>>>>>
> >>>>>>>>>>> 1) All the tests (across 2x repeats) finish apart from the
> >>>>>>>>>>> following
> >>>>>>>>>>> which
> >>>>>>>>>>> have the errors listed:
> >>>>>>>>>>>
> >>>>>>>>>>> ------------------------------****--------------
> >>>>>>>>>>> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
> >>>>>>>>>>> Error: unspecified launch failure launching kernel kNLSkinTest
> >>>>>>>>>>> cudaFree GpuBuffer::Deallocate failed unspecified launch
> >>>>>>>>>>> failure
> >>>>>>>>>>>
> >>>>>>>>>>> ------------------------------****--------------
> >>>>>>>>>>> CELLULOSE_PRODUCTION_NPT - 408,609 atoms PME
> >>>>>>>>>>> cudaMemcpy GpuBuffer::Download failed unspecified launch
> >>>>>>>>>>> failure
> >>>>>>>>>>>
> >>>>>>>>>>> ------------------------------****--------------
> >>>>>>>>>>> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
> >>>>>>>>>>> Error: unspecified launch failure launching kernel kNLSkinTest
> >>>>>>>>>>> cudaFree GpuBuffer::Deallocate failed unspecified launch
> >>>>>>>>>>> failure
> >>>>>>>>>>>
> >>>>>>>>>>> ------------------------------****--------------
> >>>>>>>>>>> CELLULOSE_PRODUCTION_NPT - 408,609 atoms PME
> >>>>>>>>>>> cudaMemcpy GpuBuffer::Download failed unspecified launch
> >>>>>>>>>>> failure
> >>>>>>>>>>> grep: mdinfo.1GTX680: No such file or directory
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> 2) The sdiff logs indicate that reproducibility across the two
> >>>>>>>>>>> repeats
> >>>>>>>>>>> is
> >>>>>>>>>>> as follows:
> >>>>>>>>>>>
> >>>>>>>>>>> *GB_myoglobin: *Reproducible across 50k steps
> >>>>>>>>>>> *GB_nucleosome:* Reproducible till step 7400
> >>>>>>>>>>> *GB_TRPCage:* Reproducible across 50k steps
> >>>>>>>>>>>
> >>>>>>>>>>> *PME_JAC_production_NVE: *No reproducibility shown from step
> >>>>>>>>>>> 1,000
> >>>>>>>>>>> onwards
> >>>>>>>>>>> *PME_JAC_production_NPT*: Reproducible till step 1,000. Also
> >>>>>>>>>>> outfile
> >>>>>>>>>>> is
> >>>>>>>>>>> not written properly - blank gaps appear where something should
> >>>>>>>>>>> have
> >>>>>>>>>>> been
> >>>>>>>>>>> written
> >>>>>>>>>>>
> >>>>>>>>>>> *PME_FactorIX_production_NVE:* Reproducible across 50k steps
> >>>>>>>>>>> *PME_FactorIX_production_NPT:* Reproducible across 50k steps
> >>>>>>>>>>>
> >>>>>>>>>>> *PME_Cellulose_production_NVE:***** Failure means that both
> >>>>>>>>>>> runs
> >>>>>>>>>>> do
> >>>>>>>>>>> not
> >>>>>>>>>>> finish
> >>>>>>>>>>> (see point1)
> >>>>>>>>>>> *PME_Cellulose_production_NPT: *Failure means that both runs do
> >>>>>>>>>>> not
> >>>>>>>>>>> finish
> >>>>>>>>>>> (see point1)
> >>>>>>>>>>>
> >>>>>>>>>>>
> ##############################****############################**
> >>>>>>>>>>> ##**
> >>>>>>>>>>> ###########################
> >>>>>>>>>>>
> >>>>>>>>>>> *When I run the tests on * *GPU-00_TeaNCake:*
> >>>>>>>>>>> *
> >>>>>>>>>>> *
> >>>>>>>>>>> 1) All the tests (across 2x repeats) finish apart from the
> >>>>>>>>>>> following
> >>>>>>>>>>> which
> >>>>>>>>>>> have the errors listed:
> >>>>>>>>>>> ------------------------------****-------
> >>>>>>>>>>> JAC_PRODUCTION_NPT - 23,558 atoms PME
> >>>>>>>>>>> PMEMD Terminated Abnormally!
> >>>>>>>>>>> ------------------------------****-------
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> 2) The sdiff logs indicate that reproducibility across the two
> >>>>>>>>>>> repeats
> >>>>>>>>>>> is
> >>>>>>>>>>> as follows:
> >>>>>>>>>>>
> >>>>>>>>>>> *GB_myoglobin:* Reproducible across 50k steps
> >>>>>>>>>>> *GB_nucleosome:* Reproducible across 50k steps
> >>>>>>>>>>> *GB_TRPCage:* Reproducible across 50k steps
> >>>>>>>>>>>
> >>>>>>>>>>> *PME_JAC_production_NVE:* No reproducibility shown from step
> >>>>>>>>>>> 10,000
> >>>>>>>>>>> onwards
> >>>>>>>>>>> *PME_JAC_production_NPT: * No reproducibility shown from step
> >>>>>>>>>>> 10,000
> >>>>>>>>>>> onwards. Also outfile is not written properly - blank gaps
> >>>>>>>>>>> appear
> >>>>>>>>>>> where
> >>>>>>>>>>> something should have been written. Repeat 2 Crashes with error
> >>>>>>>>>>> noted
> >>>>>>>>>>> in
> >>>>>>>>>>> 1.
> >>>>>>>>>>>
> >>>>>>>>>>> *PME_FactorIX_production_NVE:* No reproducibility shown from
> >>>>>>>>>>> step
> >>>>>>>>>>> 9,000
> >>>>>>>>>>> onwards
> >>>>>>>>>>> *PME_FactorIX_production_NPT: *Reproducible across 50k steps
> >>>>>>>>>>>
> >>>>>>>>>>> *PME_Cellulose_production_NVE: *No reproducibility shown from
> >>>>>>>>>>> step
> >>>>>>>>>>> 5,000
> >>>>>>>>>>> onwards
> >>>>>>>>>>> *PME_Cellulose_production_NPT: ** *No reproducibility shown
> >>>>>>>>>>> from
> >>>>>>>>>>> step
> >>>>>>>>>>> 29,000 onwards. Also outfile is not written properly - blank
> >>>>>>>>>>> gaps
> >>>>>>>>>>> appear
> >>>>>>>>>>> where something should have been written.
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> Out files and sdiff files are included as attatchments
> >>>>>>>>>>>
> >>>>>>>>>>> ##############################****###################
> >>>>>>>>>>>
> >>>>>>>>>>> So I'm going to update my nvidia driver to the latest version
> >>>>>>>>>>> and
> >>>>>>>>>>> patch
> >>>>>>>>>>> amber to the latest version and rerun the tests to see if
> >>>>>>>>>>> there is
> >>>>>>>>>>> any
> >>>>>>>>>>> improvement. Could someone let me know if it is necessary to
> >>>>>>>>>>> recompile
> >>>>>>>>>>> any
> >>>>>>>>>>> or all of AMBER after applying the bugfixes?
> >>>>>>>>>>>
> >>>>>>>>>>> Additionally, I'm going to run memory tests and heaven
> >>>>>>>>>>> benchmarks
> >>>>>>>>>>> on
> >>>>>>>>>>> the
> >>>>>>>>>>> cards to check whether they are faulty or not.
> >>>>>>>>>>>
> >>>>>>>>>>> I'm thinking that there is a mix of hardware
> >>>>>>>>>>> error/configuration
> >>>>>>>>>>> (esp
> >>>>>>>>>>> in
> >>>>>>>>>>> the case of GPU-01_008) and amber software error in this
> >>>>>>>>>>> situation.
> >>>>>>>>>>> What
> >>>>>>>>>>> do
> >>>>>>>>>>> you guys think?
> >>>>>>>>>>>
> >>>>>>>>>>> Also am I right in thinking (from what Scott was saying) that
> >>>>>>>>>>> all
> >>>>>>>>>>> the
> >>>>>>>>>>> benchmarks should be reproducible across 50k steps but begin to
> >>>>>>>>>>> diverge
> >>>>>>>>>>> at
> >>>>>>>>>>> around 100K steps? Is there any difference from in setting *ig
> >>>>>>>>>>> *to
> >>>>>>>>>>> an
> >>>>>>>>>>> explicit number to removing it from the mdin file?
> >>>>>>>>>>>
> >>>>>>>>>>> br,
> >>>>>>>>>>> g
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> On 31 May 2013 23:45, ET <sketchfoot.gmail.com> wrote:
> >>>>>>>>>>>
> >>>>>>>>>>> I don't need sysadmins, but sysadmins need me as it gives
> >>>>>>>>>>> purpose
> >>>>>>>>>>> to
> >>>>>>>>>>>
> >>>>>>>>>>>> their
> >>>>>>>>>>>> bureaucratic existence. A encountered evil if working in an
> >>>>>>>>>>>> institution
> >>>>>>>>>>>> or
> >>>>>>>>>>>> comapny IMO. Good science and indiviguality being sacrificed
> >>>>>>>>>>>> for
> >>>>>>>>>>>> standardisation and mediocrity in the intrerests of maintaing
> >>>>>>>>>>>> a
> >>>>>>>>>>>> system
> >>>>>>>>>>>> that
> >>>>>>>>>>>> focusses on maintaining the system and not the objective.
> >>>>>>>>>>>>
> >>>>>>>>>>>> You need root to move fwd on these things, unfortunately. and
> >>>>>>>>>>>> ppl
> >>>>>>>>>>>> with
> >>>>>>>>>>>> root are kinda like your parents when you try to borrow money
> >>>>>>>>>>>> from
> >>>>>>>>>>>> them
> >>>>>>>>>>>> .
> >>>>>>>>>>>> age 12 :D
> >>>>>>>>>>>> On May 31, 2013 9:34 PM, "Marek Maly" <marek.maly.ujep.cz>
> >>>>>>>>>>>> wrote:
> >>>>>>>>>>>>
> >>>>>>>>>>>> Sorry why do you need sysadmins :)) ?
> >>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> BTW here is the most recent driver:
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> http://www.nvidia.com/object/****linux-display-amd64-319.23-**<
> http://www.nvidia.com/object/**linux-display-amd64-319.23-**>
> >>>>>>>>>>>>> driver.html<http://www.nvidia.**com/object/linux-display-**
> >>>>>>>>>>>>> amd64-319.23-driver.html<
> http://www.nvidia.com/object/linux-display-amd64-319.23-driver.html>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> I do not remember anything easier than is to install driver
> >>>>>>>>>>>>> (especially
> >>>>>>>>>>>>> in case of binary (*.run) installer) :))
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> M.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Dne Fri, 31 May 2013 22:02:34 +0200 ET <sketchfoot.gmail.com
> >
> >>>>>>>>>>>>> napsal/-a:
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> > Yup. I know. I replaced a 680 and the everknowing sysadmins
> >>>>>>>>>>>>> are
> >>>>>>>>>>>>> reluctant
> >>>>>>>>>>>>> > to install drivers not in the repositoery as they are
> >>>>>>>>>>>>> lame. :(
> >>>>>>>>>>>>> > On May 31, 2013 7:14 PM, "Marek Maly" <marek.maly.ujep.cz>
> >>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> As I already wrote you,
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> the first driver which properly/officially supports
> >>>>>>>>>>>>> Titans,
> >>>>>>>>>>>>> should
> >>>>>>>>>>>>> be
> >>>>>>>>>>>>> >> 313.26 .
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> Anyway I am curious mainly about your 100K repetitive
> >>>>>>>>>>>>> tests
> >>>>>>>>>>>>> with
> >>>>>>>>>>>>> >> your Titan SC card. Especially in case of these tests (
> >>>>>>>>>>>>> JAC_NVE,
> >>>>>>>>>>>>> JAC_NPT
> >>>>>>>>>>>>> >> and CELLULOSE_NVE ) where
> >>>>>>>>>>>>> >> my Titans SC randomly failed or succeeded. In
> >>>>>>>>>>>>> FACTOR_IX_NVE,
> >>>>>>>>>>>>> >> FACTOR_IX_NPT
> >>>>>>>>>>>>> >> tests both
> >>>>>>>>>>>>> >> my cards are perfectly stable (independently from drv.
> >>>>>>>>>>>>> version)
> >>>>>>>>>>>>> and
> >>>>>>>>>>>>> also
> >>>>>>>>>>>>> >> the runs
> >>>>>>>>>>>>> >> are perfectly or almost perfectly reproducible.
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> Also if your test will crash please report the eventual
> >>>>>>>>>>>>> errs.
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> To this moment I have this actual library of errs on my
> >>>>>>>>>>>>> Titans
> >>>>>>>>>>>>> SC
> >>>>>>>>>>>>> GPUs.
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> #1 ERR writtent in mdout:
> >>>>>>>>>>>>> >> ------
> >>>>>>>>>>>>> >> | ERROR: max pairlist cutoff must be less than unit cell
> >>>>>>>>>>>>> max
> >>>>>>>>>>>>> sphere
> >>>>>>>>>>>>> >> radius!
> >>>>>>>>>>>>> >> ------
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> #2 no ERR writtent in mdout, ERR written in standard
> >>>>>>>>>>>>> output
> >>>>>>>>>>>>> (nohup.out)
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> ----
> >>>>>>>>>>>>> >> Error: unspecified launch failure launching kernel
> >>>>>>>>>>>>> kNLSkinTest
> >>>>>>>>>>>>> >> cudaFree GpuBuffer::Deallocate failed unspecified launch
> >>>>>>>>>>>>> failure
> >>>>>>>>>>>>> >> ----
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> #3 no ERR writtent in mdout, ERR written in standard
> >>>>>>>>>>>>> output
> >>>>>>>>>>>>> (nohup.out)
> >>>>>>>>>>>>> >> ----
> >>>>>>>>>>>>> >> cudaMemcpy GpuBuffer::Download failed unspecified launch
> >>>>>>>>>>>>> failure
> >>>>>>>>>>>>> >> ----
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> Another question, regarding your Titan SC, it is also
> >>>>>>>>>>>>> EVGA as
> >>>>>>>>>>>>> in
> >>>>>>>>>>>>> my
> >>>>>>>>>>>>> case
> >>>>>>>>>>>>> >> or it is another producer ?
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> Thanks,
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> M.
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> Dne Fri, 31 May 2013 19:17:03 +0200 ET
> >>>>>>>>>>>>> <sketchfoot.gmail.com
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> napsal/-a:
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> > Well, this is interesting...
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> > I ran 50k steps on the Titan on the other machine with
> >>>>>>>>>>>>> driver
> >>>>>>>>>>>>> 310.44
> >>>>>>>>>>>>> >> and
> >>>>>>>>>>>>> >> > it
> >>>>>>>>>>>>> >> > passed all the GB steps. i.e totally identical results
> >>>>>>>>>>>>> over
> >>>>>>>>>>>>> two
> >>>>>>>>>>>>> >> repeats.
> >>>>>>>>>>>>> >> > However, it failed all the PME tests after step 1000.
> >>>>>>>>>>>>> I'm
> >>>>>>>>>>>>> going
> >>>>>>>>>>>>> to
> >>>>>>>>>>>>> > update
> >>>>>>>>>>>>> >> > the driver and test it again.
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> > Files included as attachments.
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> > br,
> >>>>>>>>>>>>> >> > g
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> > On 31 May 2013 16:40, Marek Maly <marek.maly.ujep.cz>
> >>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> >> One more thing,
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> can you please check under which frequency is running
> >>>>>>>>>>>>> that
> >>>>>>>>>>>>> your
> >>>>>>>>>>>>> >> titan ?
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> As the base frequency of normal Titans is 837MHz and
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> Boost
> >>>>>>>>>>>>> one
> >>>>>>>>>>>>> is
> >>>>>>>>>>>>> >> >> 876MHz I
> >>>>>>>>>>>>> >> >> assume that yor GPU is running automatically also under
> >>>>>>>>>>>>> it's
> >>>>>>>>>>>>> boot
> >>>>>>>>>>>>> >> >> frequency (876MHz).
> >>>>>>>>>>>>> >> >> You can find this information e.g. in Amber mdout file.
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> You also mentioned some crashes in your previous email.
> >>>>>>>>>>>>> Your
> >>>>>>>>>>>>> ERRs
> >>>>>>>>>>>>> >> were
> >>>>>>>>>>>>> >> >> something like those here:
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> #1 ERR writtent in mdout:
> >>>>>>>>>>>>> >> >> ------
> >>>>>>>>>>>>> >> >> | ERROR: max pairlist cutoff must be less than unit
> >>>>>>>>>>>>> cell
> >>>>>>>>>>>>> max
> >>>>>>>>>>>>> sphere
> >>>>>>>>>>>>> >> >> radius!
> >>>>>>>>>>>>> >> >> ------
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> #2 no ERR writtent in mdout, ERR written in standard
> >>>>>>>>>>>>> output
> >>>>>>>>>>>>> >> (nohup.out)
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> ----
> >>>>>>>>>>>>> >> >> Error: unspecified launch failure launching kernel
> >>>>>>>>>>>>> kNLSkinTest
> >>>>>>>>>>>>> >> >> cudaFree GpuBuffer::Deallocate failed unspecified
> >>>>>>>>>>>>> launch
> >>>>>>>>>>>>> failure
> >>>>>>>>>>>>> >> >> ----
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> #3 no ERR writtent in mdout, ERR written in standard
> >>>>>>>>>>>>> output
> >>>>>>>>>>>>> >> (nohup.out)
> >>>>>>>>>>>>> >> >> ----
> >>>>>>>>>>>>> >> >> cudaMemcpy GpuBuffer::Download failed unspecified
> >>>>>>>>>>>>> launch
> >>>>>>>>>>>>> failure
> >>>>>>>>>>>>> >> >> ----
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> or you obtained some new/additional errs ?
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> M.
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> Dne Fri, 31 May 2013 17:30:57 +0200 filip fratev
> >>>>>>>>>>>>> >> <filipfratev.yahoo.com
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> >> napsal/-a:
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> > Hi,
> >>>>>>>>>>>>> >> >> > This is what I obtained for 50K tests and "normal"
> >>>>>>>>>>>>> GTXTitan:
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > run1:
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> ------------------------------****----------------------------*
> >>>>>>>>>>>>> *--**
> >>>>>>>>>>>>> ------------------
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > A V E R A G E S O V E R 50 S T E P S
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> >>>>>>>>>>>>> 299.87
> >>>>>>>>>>>>> >> PRESS
> >>>>>>>>>>>>> >> >> > = 0.0
> >>>>>>>>>>>>> >> >> > Etot = -443237.1079 EKtot = 257679.9750
> >>>>>>>>>>>>> EPtot
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > -700917.0829
> >>>>>>>>>>>>> >> >> > BOND = 20193.1856 ANGLE = 53517.5432
> >>>>>>>>>>>>> DIHED
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 23575.4648
> >>>>>>>>>>>>> >> >> > 1-4 NB = 21759.5524 1-4 EEL = 742552.5939
> >>>>>>>>>>>>> VDWAALS
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 96286.7714
> >>>>>>>>>>>>> >> >> > EELEC = -1658802.1941 EHBOND = 0.0000
> >>>>>>>>>>>>> RESTRAINT
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 0.0000
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> ------------------------------****----------------------------
> >>>>>>>>>>>>> **--**
> >>>>>>>>>>>>> ------------------
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > R M S F L U C T U A T I O N S
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> >>>>>>>>>>>>> 0.33
> >>>>>>>>>>>>> >> PRESS
> >>>>>>>>>>>>> >> >> > = 0.0
> >>>>>>>>>>>>> >> >> > Etot = 11.2784 EKtot = 284.8999
> >>>>>>>>>>>>> EPtot
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 289.0773
> >>>>>>>>>>>>> >> >> > BOND = 136.3417 ANGLE = 214.0054
> >>>>>>>>>>>>> DIHED
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 59.4893
> >>>>>>>>>>>>> >> >> > 1-4 NB = 58.5891 1-4 EEL = 330.5400
> >>>>>>>>>>>>> VDWAALS
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 559.2079
> >>>>>>>>>>>>> >> >> > EELEC = 743.8771 EHBOND = 0.0000
> >>>>>>>>>>>>> RESTRAINT
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 0.0000
> >>>>>>>>>>>>> >> >> > |E(PBS) = 21.8119
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> ------------------------------****----------------------------
> >>>>>>>>>>>>> **--**
> >>>>>>>>>>>>> ------------------
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > run2:
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> ------------------------------****----------------------------*
> >>>>>>>>>>>>> *--**
> >>>>>>>>>>>>> ------------------
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > A V E R A G E S O V E R 50 S T E P S
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> >>>>>>>>>>>>> 299.89
> >>>>>>>>>>>>> >> PRESS
> >>>>>>>>>>>>> >> >> > = 0.0
> >>>>>>>>>>>>> >> >> > Etot = -443240.0999 EKtot = 257700.0950
> >>>>>>>>>>>>> EPtot
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > -700940.1949
> >>>>>>>>>>>>> >> >> > BOND = 20241.9174 ANGLE = 53644.6694
> >>>>>>>>>>>>> DIHED
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 23541.3737
> >>>>>>>>>>>>> >> >> > 1-4 NB = 21803.1898 1-4 EEL = 742754.2254
> >>>>>>>>>>>>> VDWAALS
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 96298.8308
> >>>>>>>>>>>>> >> >> > EELEC = -1659224.4013 EHBOND = 0.0000
> >>>>>>>>>>>>> RESTRAINT
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 0.0000
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> ------------------------------****----------------------------
> >>>>>>>>>>>>> **--**
> >>>>>>>>>>>>> ------------------
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > R M S F L U C T U A T I O N S
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
> >>>>>>>>>>>>> 0.41
> >>>>>>>>>>>>> >> PRESS
> >>>>>>>>>>>>> >> >> > = 0.0
> >>>>>>>>>>>>> >> >> > Etot = 10.7633 EKtot = 348.2819
> >>>>>>>>>>>>> EPtot
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 353.9918
> >>>>>>>>>>>>> >> >> > BOND = 106.5314 ANGLE = 196.7052
> >>>>>>>>>>>>> DIHED
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 69.7476
> >>>>>>>>>>>>> >> >> > 1-4 NB = 60.3435 1-4 EEL = 400.7466
> >>>>>>>>>>>>> VDWAALS
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 462.7763
> >>>>>>>>>>>>> >> >> > EELEC = 651.9857 EHBOND = 0.0000
> >>>>>>>>>>>>> RESTRAINT
> >>>>>>>>>>>>> =
> >>>>>>>>>>>>> >> >> > 0.0000
> >>>>>>>>>>>>> >> >> > |E(PBS) = 17.0642
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> ------------------------------****----------------------------
> >>>>>>>>>>>>> **--**
> >>>>>>>>>>>>> ------------------
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> ------------------------------****----------------------------*
> >>>>>>>>>>>>> *--**
> >>>>>>>>>>>>> --------------------
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > ______________________________****__
> >>>>>>>>>>>>> >> >> > From: Marek Maly <marek.maly.ujep.cz>
> >>>>>>>>>>>>> >> >> > To: AMBER Mailing List <amber.ambermd.org>
> >>>>>>>>>>>>> >> >> > Sent: Friday, May 31, 2013 3:34 PM
> >>>>>>>>>>>>> >> >> > Subject: Re: [AMBER] experiences with EVGA GTX TITAN
> >>>>>>>>>>>>> Superclocked
> >>>>>>>>>>>>> -
> >>>>>>>>>>>>> >> >> > memtestG80 - UNDERclocking in Linux ?
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > Hi here are my 100K results for driver 313.30 (and
> >>>>>>>>>>>>> still
> >>>>>>>>>>>>> Cuda
> >>>>>>>>>>>>> 5.0).
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > The results are rather similar to those obtained
> >>>>>>>>>>>>> >> >> > under my original driver 319.17 (see the first table
> >>>>>>>>>>>>> >> >> > which I sent in this thread).
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > M.
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> > Dne Fri, 31 May 2013 12:29:59 +0200 Marek Maly <
> >>>>>>>>>>>>> marek.maly.ujep.cz>
> >>>>>>>>>>>>> >> >> > napsal/-a:
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >> Hi,
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >> please try to run at lest 100K tests twice to verify
> >>>>>>>>>>>>> exact
> >>>>>>>>>>>>> >> >> >> reproducibility
> >>>>>>>>>>>>> >> >> >> of the results on the given card. If you find in any
> >>>>>>>>>>>>> mdin
> >>>>>>>>>>>>> file
> >>>>>>>>>>>>> >> ig=-1
> >>>>>>>>>>>>> >> >> >> just
> >>>>>>>>>>>>> >> >> >> delete it to ensure that you are using the identical
> >>>>>>>>>>>>> random
> >>>>>>>>>>>>> seed
> >>>>>>>>>>>>> >> for
> >>>>>>>>>>>>> >> >> >> both
> >>>>>>>>>>>>> >> >> >> runs. You can eventually omit NUCLEOSOME test
> >>>>>>>>>>>>> >> >> >> as it is too much time consuming.
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >> Driver 310.44 ?????
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >> As far as I know the proper support for titans is
> >>>>>>>>>>>>> from
> >>>>>>>>>>>>> version
> >>>>>>>>>>>>> > 313.26
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >> see e.g. here :
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> http://www.geeks3d.com/****20130306/nvidia-releases-r313-****<
> http://www.geeks3d.com/**20130306/nvidia-releases-r313-**>
> >>>>>>>>>>>>> 26-for-linux-with-gtx-titan-****support/<http://www.geeks3d.
> **
> >>>>>>>>>>>>> com/20130306/nvidia-releases-**r313-26-for-linux-with-gtx-**
> >>>>>>>>>>>>> titan-support/<
> http://www.geeks3d.com/20130306/nvidia-releases-r313-26-for-linux-with-gtx-titan-support/
> >
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >> BTW: On my site downgrade to drv. 313.30 did not
> >>>>>>>>>>>>> solved
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> >> >> situation, I
> >>>>>>>>>>>>> >> >> >> will post
> >>>>>>>>>>>>> >> >> >> my results soon here.
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >> M.
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >> Dne Fri, 31 May 2013 12:21:21 +0200 ET <
> >>>>>>>>>>>>> sketchfoot.gmail.com>
> >>>>>>>>>>>>> >> >> napsal/-a:
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >>> ps. I have another install of amber on another
> >>>>>>>>>>>>> computer
> >>>>>>>>>>>>> with
> >>>>>>>>>>>>> a
> >>>>>>>>>>>>> >> >> >>> different
> >>>>>>>>>>>>> >> >> >>> Titan and different Driver Version: 310.44.
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>> In the interests of thrashing the proverbial horse,
> >>>>>>>>>>>>> I'll
> >>>>>>>>>>>>> run
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> >> >> >>> benchmark
> >>>>>>>>>>>>> >> >> >>> for 50k steps. :P
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>> br,
> >>>>>>>>>>>>> >> >> >>> g
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>> On 31 May 2013 11:17, ET <sketchfoot.gmail.com>
> >>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>>> Hi, I just ran the Amber benchmark for the default
> >>>>>>>>>>>>> (10000
> >>>>>>>>>>>>> steps)
> >>>>>>>>>>>>> >> >> on my
> >>>>>>>>>>>>> >> >> >>>> Titan.
> >>>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>>> >> >> >>>> Using sdiff -sB showed that the two runs were
> >>>>>>>>>>>>> completely
> >>>>>>>>>>>>> > identical.
> >>>>>>>>>>>>> >> >> >>>> I've
> >>>>>>>>>>>>> >> >> >>>> attached compressed files of the mdout & diff
> >>>>>>>>>>>>> files.
> >>>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>>> >> >> >>>> br,
> >>>>>>>>>>>>> >> >> >>>> g
> >>>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>>> >> >> >>>> On 30 May 2013 23:41, Marek Maly
> >>>>>>>>>>>>> <marek.maly.ujep.cz
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>>> >> >> >>>>> OK, let's see. The eventual downclocking I see as
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> very
> >>>>>>>>>>>>> last
> >>>>>>>>>>>>> >> >> >>>>> possibility
> >>>>>>>>>>>>> >> >> >>>>> (if I don't decide for RMAing). But now still
> >>>>>>>>>>>>> some
> >>>>>>>>>>>>> other
> >>>>>>>>>>>>> >> >> experiments
> >>>>>>>>>>>>> >> >> >>>>> are
> >>>>>>>>>>>>> >> >> >>>>> available :))
> >>>>>>>>>>>>> >> >> >>>>> I just started 100K tests under 313.30 driver.
> >>>>>>>>>>>>> For
> >>>>>>>>>>>>> today
> >>>>>>>>>>>>> good
> >>>>>>>>>>>>> >> >> night
> >>>>>>>>>>>>> >> >> >>>>> ...
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> M.
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> Dne Fri, 31 May 2013 00:45:49 +0200 Scott Le
> >>>>>>>>>>>>> Grand
> >>>>>>>>>>>>> >> >> >>>>> <varelse2005.gmail.com
> >>>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> napsal/-a:
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> > It will be very interesting if this behavior
> >>>>>>>>>>>>> persists
> >>>>>>>>>>>>> after
> >>>>>>>>>>>>> >> >> >>>>> downclocking.
> >>>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> > But right now, Titan 0 *looks* hosed and Titan
> >>>>>>>>>>>>> 1
> >>>>>>>>>>>>> *looks*
> >>>>>>>>>>>>> like
> >>>>>>>>>>>>> > it
> >>>>>>>>>>>>> >> >> >>>>> needs
> >>>>>>>>>>>>> >> >> >>>>> > downclocking...
> >>>>>>>>>>>>> >> >> >>>>> > On May 30, 2013 3:20 PM, "Marek Maly"
> >>>>>>>>>>>>> <marek.maly.ujep.cz>
> >>>>>>>>>>>>> >> >> wrote:
> >>>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >> Hi all,
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> here are my results from the 500K steps 2 x
> >>>>>>>>>>>>> repeated
> >>>>>>>>>>>>> > benchmarks
> >>>>>>>>>>>>> >> >> >>>>> >> under 319.23 driver and still Cuda 5.0 (see
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> attached
> >>>>>>>>>>>>> >> table
> >>>>>>>>>>>>> >> >> ).
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> It is hard to say if the results are better or
> >>>>>>>>>>>>> worse
> >>>>>>>>>>>>> than
> >>>>>>>>>>>>> in
> >>>>>>>>>>>>> > my
> >>>>>>>>>>>>> >> >> >>>>> >> previous 100K test under driver 319.17.
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> While results from Cellulose test were
> >>>>>>>>>>>>> improved
> >>>>>>>>>>>>> and
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> > TITAN_1
> >>>>>>>>>>>>> >> >> >>>>> card
> >>>>>>>>>>>>> >> >> >>>>> >> even
> >>>>>>>>>>>>> >> >> >>>>> >> successfully finished all 500K steps moreover
> >>>>>>>>>>>>> with
> >>>>>>>>>>>>> exactly
> >>>>>>>>>>>>> >> the
> >>>>>>>>>>>>> >> >> >>>>> same
> >>>>>>>>>>>>> >> >> >>>>> >> final
> >>>>>>>>>>>>> >> >> >>>>> >> energy !
> >>>>>>>>>>>>> >> >> >>>>> >> (TITAN_0 at least finished more than 100K
> >>>>>>>>>>>>> steps
> >>>>>>>>>>>>> and in
> >>>>>>>>>>>>> >> RUN_01
> >>>>>>>>>>>>> >> >> even
> >>>>>>>>>>>>> >> >> >>>>> more
> >>>>>>>>>>>>> >> >> >>>>> >> than 400K steps)
> >>>>>>>>>>>>> >> >> >>>>> >> In JAC_NPT test no GPU was able to finish at
> >>>>>>>>>>>>> least
> >>>>>>>>>>>>> 100K
> >>>>>>>>>>>>> >> steps
> >>>>>>>>>>>>> >> >> and
> >>>>>>>>>>>>> >> >> >>>>> the
> >>>>>>>>>>>>> >> >> >>>>> >> results from JAC_NVE
> >>>>>>>>>>>>> >> >> >>>>> >> test are also not too much convincing.
> >>>>>>>>>>>>> FACTOR_IX_NVE
> >>>>>>>>>>>>> and
> >>>>>>>>>>>>> >> >> >>>>> FACTOR_IX_NPT
> >>>>>>>>>>>>> >> >> >>>>> >> were successfully
> >>>>>>>>>>>>> >> >> >>>>> >> finished with 100% reproducibility in
> >>>>>>>>>>>>> FACTOR_IX_NPT
> >>>>>>>>>>>>> case
> >>>>>>>>>>>>> >> (on
> >>>>>>>>>>>>> >> >> both
> >>>>>>>>>>>>> >> >> >>>>> >> cards)
> >>>>>>>>>>>>> >> >> >>>>> >> and almost
> >>>>>>>>>>>>> >> >> >>>>> >> 100% reproducibility in case of FACTOR_IX_NVE
> >>>>>>>>>>>>> (again
> >>>>>>>>>>>>> 100%
> >>>>>>>>>>>>> in
> >>>>>>>>>>>>> >> >> case
> >>>>>>>>>>>>> >> >> >>>>> of
> >>>>>>>>>>>>> >> >> >>>>> >> TITAN_1). TRPCAGE, MYOGLOBIN
> >>>>>>>>>>>>> >> >> >>>>> >> again finished without any problem with 100%
> >>>>>>>>>>>>> >> reproducibility.
> >>>>>>>>>>>>> >> >> >>>>> NUCLEOSOME
> >>>>>>>>>>>>> >> >> >>>>> >> test was not done
> >>>>>>>>>>>>> >> >> >>>>> >> this time due to high time requirements. If
> >>>>>>>>>>>>> you
> >>>>>>>>>>>>> find
> >>>>>>>>>>>>> in
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> >> >> table
> >>>>>>>>>>>>> >> >> >>>>> >> positive
> >>>>>>>>>>>>> >> >> >>>>> >> number finishing with
> >>>>>>>>>>>>> >> >> >>>>> >> K (which means "thousands") it means the last
> >>>>>>>>>>>>> number
> >>>>>>>>>>>>> of
> >>>>>>>>>>>>> step
> >>>>>>>>>>>>> >> >> >>>>> written in
> >>>>>>>>>>>>> >> >> >>>>> >> mdout before crash.
> >>>>>>>>>>>>> >> >> >>>>> >> Below are all the 3 types of detected errs
> >>>>>>>>>>>>> with
> >>>>>>>>>>>>> relevant
> >>>>>>>>>>>>> >> >> >>>>> systems/rounds
> >>>>>>>>>>>>> >> >> >>>>> >> where the given err
> >>>>>>>>>>>>> >> >> >>>>> >> appeared.
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> Now I will try just 100K tests under ETs
> >>>>>>>>>>>>> favourite
> >>>>>>>>>>>>> driver
> >>>>>>>>>>>>> >> >> version
> >>>>>>>>>>>>> >> >> >>>>> 313.30
> >>>>>>>>>>>>> >> >> >>>>> >> :)) and then
> >>>>>>>>>>>>> >> >> >>>>> >> I will eventually try to experiment with cuda
> >>>>>>>>>>>>> 5.5
> >>>>>>>>>>>>> which
> >>>>>>>>>>>>> I
> >>>>>>>>>>>>> >> >> already
> >>>>>>>>>>>>> >> >> >>>>> >> downloaded from the
> >>>>>>>>>>>>> >> >> >>>>> >> cuda zone ( I had to become cuda developer for
> >>>>>>>>>>>>> this
> >>>>>>>>>>>>> :))
> >>>>>>>>>>>>> )
> >>>>>>>>>>>>> >> BTW
> >>>>>>>>>>>>> >> >> ET
> >>>>>>>>>>>>> >> >> >>>>> thanks
> >>>>>>>>>>>>> >> >> >>>>> >> for the frequency info !
> >>>>>>>>>>>>> >> >> >>>>> >> and I am still ( perhaps not alone :)) ) very
> >>>>>>>>>>>>> curious
> >>>>>>>>>>>>> about
> >>>>>>>>>>>>> >> >> your 2
> >>>>>>>>>>>>> >> >> >>>>> x
> >>>>>>>>>>>>> >> >> >>>>> >> repeated Amber benchmark tests with
> >>>>>>>>>>>>> superclocked
> >>>>>>>>>>>>> Titan.
> >>>>>>>>>>>>> >> Indeed
> >>>>>>>>>>>>> >> >> >>>>> that
> >>>>>>>>>>>>> >> >> >>>>> I
> >>>>>>>>>>>>> >> >> >>>>> am
> >>>>>>>>>>>>> >> >> >>>>> >> very curious also about that Ross "hot" patch.
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> M.
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> ERRORS DETECTED DURING THE 500K steps tests
> >>>>>>>>>>>>> with
> >>>>>>>>>>>>> driver
> >>>>>>>>>>>>> >> 319.23
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> #1 ERR writtent in mdout:
> >>>>>>>>>>>>> >> >> >>>>> >> ------
> >>>>>>>>>>>>> >> >> >>>>> >> | ERROR: max pairlist cutoff must be less
> >>>>>>>>>>>>> than
> >>>>>>>>>>>>> unit
> >>>>>>>>>>>>> cell
> >>>>>>>>>>>>> >> max
> >>>>>>>>>>>>> >> >> >>>>> sphere
> >>>>>>>>>>>>> >> >> >>>>> >> radius!
> >>>>>>>>>>>>> >> >> >>>>> >> ------
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_1 JAC_NPT (at least 5000 steps
> >>>>>>>>>>>>> successfully
> >>>>>>>>>>>>> > done
> >>>>>>>>>>>>> >> >> >>>>> before
> >>>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 JAC_NPT (at least 8000 steps
> >>>>>>>>>>>>> successfully
> >>>>>>>>>>>>> > done
> >>>>>>>>>>>>> >> >> >>>>> before
> >>>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> #2 no ERR writtent in mdout, ERR written in
> >>>>>>>>>>>>> standard
> >>>>>>>>>>>>> output
> >>>>>>>>>>>>> >> >> >>>>> (nohup.out)
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> ----
> >>>>>>>>>>>>> >> >> >>>>> >> Error: unspecified launch failure launching
> >>>>>>>>>>>>> kernel
> >>>>>>>>>>>>> >> kNLSkinTest
> >>>>>>>>>>>>> >> >> >>>>> >> cudaFree GpuBuffer::Deallocate failed
> >>>>>>>>>>>>> unspecified
> >>>>>>>>>>>>> launch
> >>>>>>>>>>>>> >> >> failure
> >>>>>>>>>>>>> >> >> >>>>> >> ----
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_1 CELLULOSE_NVE (at least 437
> >>>>>>>>>>>>> 000
> >>>>>>>>>>>>> steps
> >>>>>>>>>>>>> >> >> successfully
> >>>>>>>>>>>>> >> >> >>>>> done
> >>>>>>>>>>>>> >> >> >>>>> >> before crash)
> >>>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 JAC_NVE (at least 162 000
> >>>>>>>>>>>>> steps
> >>>>>>>>>>>>> >> successfully
> >>>>>>>>>>>>> >> >> done
> >>>>>>>>>>>>> >> >> >>>>> >> before
> >>>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 CELLULOSE_NVE (at least 117
> >>>>>>>>>>>>> 000
> >>>>>>>>>>>>> steps
> >>>>>>>>>>>>> >> >> successfully
> >>>>>>>>>>>>> >> >> >>>>> done
> >>>>>>>>>>>>> >> >> >>>>> >> before crash)
> >>>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_1 JAC_NVE (at least 119 000
> >>>>>>>>>>>>> steps
> >>>>>>>>>>>>> >> successfully
> >>>>>>>>>>>>> >> >> done
> >>>>>>>>>>>>> >> >> >>>>> >> before
> >>>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_2 JAC_NVE (at least 43 000
> >>>>>>>>>>>>> steps
> >>>>>>>>>>>>> successfully
> >>>>>>>>>>>>> >> >> done
> >>>>>>>>>>>>> >> >> >>>>> before
> >>>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> #3 no ERR writtent in mdout, ERR written in
> >>>>>>>>>>>>> standard
> >>>>>>>>>>>>> output
> >>>>>>>>>>>>> >> >> >>>>> (nohup.out)
> >>>>>>>>>>>>> >> >> >>>>> >> ----
> >>>>>>>>>>>>> >> >> >>>>> >> cudaMemcpy GpuBuffer::Download failed
> >>>>>>>>>>>>> unspecified
> >>>>>>>>>>>>> launch
> >>>>>>>>>>>>> >> >> failure
> >>>>>>>>>>>>> >> >> >>>>> >> ----
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_1 JAC_NPT (at least 77 000
> >>>>>>>>>>>>> steps
> >>>>>>>>>>>>> successfully
> >>>>>>>>>>>>> >> >> done
> >>>>>>>>>>>>> >> >> >>>>> before
> >>>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_2 JAC_NPT (at least 58 000
> >>>>>>>>>>>>> steps
> >>>>>>>>>>>>> successfully
> >>>>>>>>>>>>> >> >> done
> >>>>>>>>>>>>> >> >> >>>>> before
> >>>>>>>>>>>>> >> >> >>>>> >> crash)
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> Dne Thu, 30 May 2013 21:27:17 +0200 Scott Le
> >>>>>>>>>>>>> Grand
> >>>>>>>>>>>>> >> >> >>>>> >> <varelse2005.gmail.com>
> >>>>>>>>>>>>> >> >> >>>>> >> napsal/-a:
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> Oops meant to send that to Jason...
> >>>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>>> >> >> >>>>> >>> Anyway, before we all panic, we need to get
> >>>>>>>>>>>>> K20's
> >>>>>>>>>>>>> behavior
> >>>>>>>>>>>>> >> >> >>>>> analyzed
> >>>>>>>>>>>>> >> >> >>>>> >>> here.
> >>>>>>>>>>>>> >> >> >>>>> >>> If it's deterministic, this truly is a
> >>>>>>>>>>>>> hardware
> >>>>>>>>>>>>> issue. If
> >>>>>>>>>>>>> >> >> not,
> >>>>>>>>>>>>> >> >> >>>>> then
> >>>>>>>>>>>>> >> >> >>>>> it
> >>>>>>>>>>>>> >> >> >>>>> >>> gets interesting because 680 is
> >>>>>>>>>>>>> deterministic as
> >>>>>>>>>>>>> far
> >>>>>>>>>>>>> as I
> >>>>>>>>>>>>> >> can
> >>>>>>>>>>>>> >> >> >>>>> tell...
> >>>>>>>>>>>>> >> >> >>>>> >>> On May 30, 2013 12:24 PM, "Scott Le Grand"
> >>>>>>>>>>>>> >> >> >>>>> <varelse2005.gmail.com>
> >>>>>>>>>>>>> >> >> >>>>> >>> wrote:
> >>>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>>> >> >> >>>>> >>> If the errors are not deterministically
> >>>>>>>>>>>>> triggered,
> >>>>>>>>>>>>> they
> >>>>>>>>>>>>> >> >> probably
> >>>>>>>>>>>>> >> >> >>>>> >>> won't be
> >>>>>>>>>>>>> >> >> >>>>> >>>> fixed by the patch, alas...
> >>>>>>>>>>>>> >> >> >>>>> >>>> On May 30, 2013 12:15 PM, "Jason Swails"
> >>>>>>>>>>>>> >> >> >>>>> <jason.swails.gmail.com>
> >>>>>>>>>>>>> >> >> >>>>> >>>> wrote:
> >>>>>>>>>>>>> >> >> >>>>> >>>>
> >>>>>>>>>>>>> >> >> >>>>> >>>> Just a reminder to everyone based on what
> >>>>>>>>>>>>> Ross
> >>>>>>>>>>>>> said:
> >>>>>>>>>>>>> >> there
> >>>>>>>>>>>>> >> >> is a
> >>>>>>>>>>>>> >> >> >>>>> >>>> pending
> >>>>>>>>>>>>> >> >> >>>>> >>>>> patch to pmemd.cuda that will be coming out
> >>>>>>>>>>>>> shortly
> >>>>>>>>>>>>> >> (maybe
> >>>>>>>>>>>>> >> >> even
> >>>>>>>>>>>>> >> >> >>>>> >>>>> within
> >>>>>>>>>>>>> >> >> >>>>> >>>>> hours). It's entirely possible that
> >>>>>>>>>>>>> several
> >>>>>>>>>>>>> of
> >>>>>>>>>>>>> these
> >>>>>>>>>>>>> > errors
> >>>>>>>>>>>>> >> >> >>>>> are
> >>>>>>>>>>>>> >> >> >>>>> >>>>> fixed
> >>>>>>>>>>>>> >> >> >>>>> >>>>> by
> >>>>>>>>>>>>> >> >> >>>>> >>>>> this patch.
> >>>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> >>>>> All the best,
> >>>>>>>>>>>>> >> >> >>>>> >>>>> Jason
> >>>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> >>>>> On Thu, May 30, 2013 at 2:46 PM, filip
> >>>>>>>>>>>>> fratev
> >>>>>>>>>>>>> <
> >>>>>>>>>>>>> >> >> >>>>> filipfratev.yahoo.com>
> >>>>>>>>>>>>> >> >> >>>>> >>>>> wrote:
> >>>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > I have observed the same crashes from
> >>>>>>>>>>>>> time
> >>>>>>>>>>>>> to
> >>>>>>>>>>>>> time. I
> >>>>>>>>>>>>> > will
> >>>>>>>>>>>>> >> >> >>>>> run
> >>>>>>>>>>>>> >> >> >>>>> >>>>> cellulose
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > nve for 100k and will past results here.
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > All the best,
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > Filip
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > ______________________________******__
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > From: Scott Le Grand
> >>>>>>>>>>>>> <varelse2005.gmail.com>
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > To: AMBER Mailing List
> >>>>>>>>>>>>> <amber.ambermd.org>
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > Sent: Thursday, May 30, 2013 9:01 PM
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > Subject: Re: [AMBER] experiences with
> >>>>>>>>>>>>> EVGA
> >>>>>>>>>>>>> GTX
> >>>>>>>>>>>>> TITAN
> >>>>>>>>>>>>> >> >> >>>>> Superclocked
> >>>>>>>>>>>>> >> >> >>>>> -
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > memtestG80 - UNDERclocking in Linux ?
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > Run cellulose nve for 100k iterations
> >>>>>>>>>>>>> twice
> >>>>>>>>>>>>> . If
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> >> >> final
> >>>>>>>>>>>>> >> >> >>>>> >>>>> energies
> >>>>>>>>>>>>> >> >> >>>>> >>>>> don't
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > match, you have a hardware issue. No
> >>>>>>>>>>>>> need
> >>>>>>>>>>>>> to
> >>>>>>>>>>>>> play
> >>>>>>>>>>>>> with
> >>>>>>>>>>>>> >> >> ntpr
> >>>>>>>>>>>>> >> >> >>>>> or
> >>>>>>>>>>>>> >> >> >>>>> any
> >>>>>>>>>>>>> >> >> >>>>> >>>>> other
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > variable.
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > On May 30, 2013 10:58 AM,
> >>>>>>>>>>>>> <pavel.banas.upol.cz>
> >>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Dear all,
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > I would also like to share one of my
> >>>>>>>>>>>>> experience
> >>>>>>>>>>>>> with
> >>>>>>>>>>>>> >> >> titan
> >>>>>>>>>>>>> >> >> >>>>> >>>>> cards. We
> >>>>>>>>>>>>> >> >> >>>>> >>>>> have
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > one gtx titan card and with one system
> >>>>>>>>>>>>> (~55k
> >>>>>>>>>>>>> atoms,
> >>>>>>>>>>>>> > NVT,
> >>>>>>>>>>>>> >> >> >>>>> >>>>> RNA+waters)
> >>>>>>>>>>>>> >> >> >>>>> >>>>> we
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > run
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > into same troubles you are describing.
> >>>>>>>>>>>>> I
> >>>>>>>>>>>>> was
> >>>>>>>>>>>>> also
> >>>>>>>>>>>>> >> >> playing
> >>>>>>>>>>>>> >> >> >>>>> with
> >>>>>>>>>>>>> >> >> >>>>> >>>>> ntpr
> >>>>>>>>>>>>> >> >> >>>>> >>>>> to
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > figure out what is going on, step by
> >>>>>>>>>>>>> step.
> >>>>>>>>>>>>> I
> >>>>>>>>>>>>> >> understand
> >>>>>>>>>>>>> >> >> >>>>> that
> >>>>>>>>>>>>> >> >> >>>>> the
> >>>>>>>>>>>>> >> >> >>>>> >>>>> code
> >>>>>>>>>>>>> >> >> >>>>> >>>>> is
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > using different routines for
> >>>>>>>>>>>>> calculation
> >>>>>>>>>>>>> >> >> energies+forces or
> >>>>>>>>>>>>> >> >> >>>>> only
> >>>>>>>>>>>>> >> >> >>>>> >>>>> forces.
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > The
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > simulations of other systems are
> >>>>>>>>>>>>> perfectly
> >>>>>>>>>>>>> stable,
> >>>>>>>>>>>>> >> >> running
> >>>>>>>>>>>>> >> >> >>>>> for
> >>>>>>>>>>>>> >> >> >>>>> >>>>> days
> >>>>>>>>>>>>> >> >> >>>>> >>>>> and
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > weeks. Only that particular system
> >>>>>>>>>>>>> systematically
> >>>>>>>>>>>>> >> ends
> >>>>>>>>>>>>> >> >> up
> >>>>>>>>>>>>> >> >> >>>>> with
> >>>>>>>>>>>>> >> >> >>>>> >>>>> this
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > error.
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > However, there was one interesting
> >>>>>>>>>>>>> issue.
> >>>>>>>>>>>>> When
> >>>>>>>>>>>>> I
> >>>>>>>>>>>>> set
> >>>>>>>>>>>>> >> >> >>>>> ntpr=1,
> >>>>>>>>>>>>> >> >> >>>>> the
> >>>>>>>>>>>>> >> >> >>>>> >>>>> error
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > vanished (systematically in multiple
> >>>>>>>>>>>>> runs)
> >>>>>>>>>>>>> and
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> >> >> >>>>> simulation
> >>>>>>>>>>>>> >> >> >>>>> was
> >>>>>>>>>>>>> >> >> >>>>> >>>>> able to
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > run for more than millions of steps (I
> >>>>>>>>>>>>> was
> >>>>>>>>>>>>> not
> >>>>>>>>>>>>> let
> >>>>>>>>>>>>> it
> >>>>>>>>>>>>> >> >> >>>>> running
> >>>>>>>>>>>>> >> >> >>>>> for
> >>>>>>>>>>>>> >> >> >>>>> >>>>> weeks
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > as
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > in the meantime I shifted that
> >>>>>>>>>>>>> simulation
> >>>>>>>>>>>>> to
> >>>>>>>>>>>>> other
> >>>>>>>>>>>>> >> card
> >>>>>>>>>>>>> >> >> -
> >>>>>>>>>>>>> >> >> >>>>> need
> >>>>>>>>>>>>> >> >> >>>>> >>>>> data,
> >>>>>>>>>>>>> >> >> >>>>> >>>>> not
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > testing). All other setting of ntpr
> >>>>>>>>>>>>> failed.
> >>>>>>>>>>>>> As
> >>>>>>>>>>>>> I
> >>>>>>>>>>>>> read
> >>>>>>>>>>>>> >> >> this
> >>>>>>>>>>>>> >> >> >>>>> >>>>> discussion, I
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > tried to set ene_avg_sampling=1 with
> >>>>>>>>>>>>> some
> >>>>>>>>>>>>> high
> >>>>>>>>>>>>> value
> >>>>>>>>>>>>> >> of
> >>>>>>>>>>>>> >> >> >>>>> ntpr
> >>>>>>>>>>>>> >> >> >>>>> (I
> >>>>>>>>>>>>> >> >> >>>>> >>>>> expected
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > that this will shift the code to
> >>>>>>>>>>>>> permanently
> >>>>>>>>>>>>> use
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> >> >> >>>>> >>>>> force+energies
> >>>>>>>>>>>>> >> >> >>>>> >>>>> part
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > of
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > the code, similarly to ntpr=1), but the
> >>>>>>>>>>>>> error
> >>>>>>>>>>>>> >> occurred
> >>>>>>>>>>>>> >> >> >>>>> again.
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > I know it is not very conclusive for
> >>>>>>>>>>>>> finding
> >>>>>>>>>>>>> out
> >>>>>>>>>>>>> what
> >>>>>>>>>>>>> > is
> >>>>>>>>>>>>> >> >> >>>>> >>>>> happening,
> >>>>>>>>>>>>> >> >> >>>>> >>>>> at
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > least
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > not for me. Do you have any idea, why
> >>>>>>>>>>>>> ntpr=1
> >>>>>>>>>>>>> might
> >>>>>>>>>>>>> > help?
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > best regards,
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Pavel
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > --
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Pavel Banáš
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > pavel.banas.upol.cz
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Department of Physical Chemistry,
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Palacky University Olomouc
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Czech Republic
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > ---------- Původní zpráva ----------
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Od: Jason Swails
> >>>>>>>>>>>>> <jason.swails.gmail.com>
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Datum: 29. 5. 2013
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Předmět: Re: [AMBER] experiences with
> >>>>>>>>>>>>> EVGA
> >>>>>>>>>>>>> GTX
> >>>>>>>>>>>>> TITAN
> >>>>>>>>>>>>> >> >> >>>>> >>>>> Superclocked -
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > memtestG
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > 80 - UNDERclocking in Linux ?
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > "I'll answer a little bit:
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > NTPR=10 Etot after 2000 steps
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443256.6711
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443256.6711
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > > NTPR=200 Etot after 2000 steps
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443261.0705
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443261.0705
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > > Any idea why energies should depend
> >>>>>>>>>>>>> on
> >>>>>>>>>>>>> frequency
> >>>>>>>>>>>>> of
> >>>>>>>>>>>>> >> >> >>>>> energy
> >>>>>>>>>>>>> >> >> >>>>> >>>>> records
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > (NTPR)
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > ?
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > It is a subtle point, but the answer is
> >>>>>>>>>>>>> 'different
> >>>>>>>>>>>>> >> code
> >>>>>>>>>>>>> >> >> >>>>> paths.'
> >>>>>>>>>>>>> >> >> >>>>> >>>>> In
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > general, it is NEVER necessary to
> >>>>>>>>>>>>> compute
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> actual
> >>>>>>>>>>>>> >> >> energy
> >>>>>>>>>>>>> >> >> >>>>> of a
> >>>>>>>>>>>>> >> >> >>>>> >>>>> molecule
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > during the course of standard molecular
> >>>>>>>>>>>>> dynamics
> >>>>>>>>>>>>> (by
> >>>>>>>>>>>>> >> >> >>>>> analogy, it
> >>>>>>>>>>>>> >> >> >>>>> >>>>> is
> >>>>>>>>>>>>> >> >> >>>>> >>>>> NEVER
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > necessary to compute atomic forces
> >>>>>>>>>>>>> during
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> course
> >>>>>>>>>>>>> >> of
> >>>>>>>>>>>>> >> >> >>>>> random
> >>>>>>>>>>>>> >> >> >>>>> >>>>> Monte
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > Carlo
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > sampling).
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > For performance's sake, then,
> >>>>>>>>>>>>> pmemd.cuda
> >>>>>>>>>>>>> computes
> >>>>>>>>>>>>> >> only
> >>>>>>>>>>>>> >> >> the
> >>>>>>>>>>>>> >> >> >>>>> force
> >>>>>>>>>>>>> >> >> >>>>> >>>>> when
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > energies are not requested, leading to
> >>>>>>>>>>>>> a
> >>>>>>>>>>>>> different
> >>>>>>>>>>>>> >> >> order of
> >>>>>>>>>>>>> >> >> >>>>> >>>>> operations
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > for
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > those runs. This difference ultimately
> >>>>>>>>>>>>> causes
> >>>>>>>>>>>>> >> >> divergence.
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > To test this, try setting the variable
> >>>>>>>>>>>>> >> >> ene_avg_sampling=10
> >>>>>>>>>>>>> >> >> >>>>> in
> >>>>>>>>>>>>> >> >> >>>>> the
> >>>>>>>>>>>>> >> >> >>>>> >>>>> &cntrl
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > section. This will force pmemd.cuda to
> >>>>>>>>>>>>> compute
> >>>>>>>>>>>>> >> energies
> >>>>>>>>>>>>> >> >> >>>>> every 10
> >>>>>>>>>>>>> >> >> >>>>> >>>>> steps
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > (for energy averaging), which will in
> >>>>>>>>>>>>> turn
> >>>>>>>>>>>>> make
> >>>>>>>>>>>>> the
> >>>>>>>>>>>>> >> >> >>>>> followed
> >>>>>>>>>>>>> >> >> >>>>> code
> >>>>>>>>>>>>> >> >> >>>>> >>>>> path
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > identical for any multiple-of-10 value
> >>>>>>>>>>>>> of
> >>>>>>>>>>>>> ntpr.
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > --
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Jason M. Swails
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Quantum Theory Project,
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > University of Florida
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > Ph.D. Candidate
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > 352-392-4032
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > ______________________________****
> >>>>>>>>>>>>> **_________________
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER mailing list
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER.ambermd.org
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> >>>>>>>>>>>>> mailman/listinfo/amber<http://**
> lists.ambermd.org/**mailman/**
> >>>>>>>>>>>>> listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> <
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> "
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > ______________________________****
> >>>>>>>>>>>>> **_________________
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER mailing list
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER.ambermd.org
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> >>>>>>>>>>>>> mailman/listinfo/amber<http://**
> lists.ambermd.org/**mailman/**
> >>>>>>>>>>>>> listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> <
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > ______________________________****
> >>>>>>>>>>>>> **_________________
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER mailing list
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER.ambermd.org
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> >>>>>>>>>>>>> mailman/listinfo/amber<http://**
> lists.ambermd.org/**mailman/**
> >>>>>>>>>>>>> listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> <
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > ______________________________****
> >>>>>>>>>>>>> **_________________
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER mailing list
> >>>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER.ambermd.org
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> >>>>>>>>>>>>> mailman/listinfo/amber<http://**
> lists.ambermd.org/**mailman/**
> >>>>>>>>>>>>> listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> <
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> >>>>> --
> >>>>>>>>>>>>> >> >> >>>>> >>>>> Jason M. Swails
> >>>>>>>>>>>>> >> >> >>>>> >>>>> Quantum Theory Project,
> >>>>>>>>>>>>> >> >> >>>>> >>>>> University of Florida
> >>>>>>>>>>>>> >> >> >>>>> >>>>> Ph.D. Candidate
> >>>>>>>>>>>>> >> >> >>>>> >>>>> 352-392-4032
> >>>>>>>>>>>>> >> >> >>>>> >>>>> ______________________________****
> >>>>>>>>>>>>> **_________________
> >>>>>>>>>>>>> >> >> >>>>> >>>>> AMBER mailing list
> >>>>>>>>>>>>> >> >> >>>>> >>>>> AMBER.ambermd.org
> >>>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
> >>>>>>>>>>>>> mailman/listinfo/amber<http://**
> lists.ambermd.org/**mailman/**
> >>>>>>>>>>>>> listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> <
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> >>>> ______________________________****
> >>>>>>>>>>>>> **_________________
> >>>>>>>>>>>>> >> >> >>>>> >>> AMBER mailing list
> >>>>>>>>>>>>> >> >> >>>>> >>> AMBER.ambermd.org
> >>>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>>> http://lists.ambermd.org/******mailman/listinfo/amber<
> http://lists.ambermd.org/****mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/**mailman/**listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> <
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>>> >> >> >>>>> >>> __________ Informace od ESET NOD32 Antivirus,
> >>>>>>>>>>>>> verze
> >>>>>>>>>>>>> >> databaze
> >>>>>>>>>>>>> >> >> 8394
> >>>>>>>>>>>>> >> >> >>>>> >>> (20130530) __________
> >>>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>>> >> >> >>>>> >>> Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>>> >> >> >>>>> >>> http://www.eset.cz
> >>>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>>> >> >> >>>>> >>>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >> --
> >>>>>>>>>>>>> >> >> >>>>> >> Tato zpráva byla vytvořena převratným
> >>>>>>>>>>>>> poštovním
> >>>>>>>>>>>>> klientem
> >>>>>>>>>>>>> > Opery:
> >>>>>>>>>>>>> >> >> >>>>> >> http://www.opera.com/mail/
> >>>>>>>>>>>>> >> >> >>>>> >> ______________________________**
> >>>>>>>>>>>>> **_________________
> >>>>>>>>>>>>> >> >> >>>>> >> AMBER mailing list
> >>>>>>>>>>>>> >> >> >>>>> >> AMBER.ambermd.org
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> >>
> >>>>>>>>>>>>> >> >> >>>>> > ______________________________**
> >>>>>>>>>>>>> **_________________
> >>>>>>>>>>>>> >> >> >>>>> > AMBER mailing list
> >>>>>>>>>>>>> >> >> >>>>> > AMBER.ambermd.org
> >>>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> > __________ Informace od ESET NOD32 Antivirus,
> >>>>>>>>>>>>> verze
> >>>>>>>>>>>>> databaze
> >>>>>>>>>>>>> >> >> 8394
> >>>>>>>>>>>>> >> >> >>>>> > (20130530) __________
> >>>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> > http://www.eset.cz
> >>>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>> >
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> >> >> >>>>> --
> >>>>>>>>>>>>> >> >> >>>>> Tato zpráva byla vytvořena převratným poštovním
> >>>>>>>>>>>>> klientem
> >>>>>>>>>>>>> Opery:
> >>>>>>>>>>>>> >> >> >>>>> http://www.opera.com/mail/
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> ______________________________****_________________
> >>>>>>>>>>>>> >> >> >>>>> AMBER mailing list
> >>>>>>>>>>>>> >> >> >>>>> AMBER.ambermd.org
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>>>>
> >>>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>>> >> >> >>>>
> >>>>>>>>>>>>> >> >> >>> ______________________________****_________________
> >>>>>>>>>>>>> >> >> >>> AMBER mailing list
> >>>>>>>>>>>>> >> >> >>> AMBER.ambermd.org
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>> __________ Informace od ESET NOD32 Antivirus, verze
> >>>>>>>>>>>>> databaze
> >>>>>>>>>>>>> 8395
> >>>>>>>>>>>>> >> >> >>> (20130531) __________
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>> Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>> http://www.eset.cz
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>>
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >>
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >> >
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> --
> >>>>>>>>>>>>> >> >> Tato zpráva byla vytvořena převratným poštovním
> >>>>>>>>>>>>> klientem
> >>>>>>>>>>>>> Opery:
> >>>>>>>>>>>>> >> >> http://www.opera.com/mail/
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >> ______________________________****_________________
> >>>>>>>>>>>>> >> >> AMBER mailing list
> >>>>>>>>>>>>> >> >> AMBER.ambermd.org
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> >>
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> > __________ Informace od ESET NOD32 Antivirus, verze
> >>>>>>>>>>>>> databaze
> >>>>>>>>>>>>> 8397
> >>>>>>>>>>>>> >> > (20130531) __________
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> > Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz - poskozeny archiv
> >>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>>> >> GB_out_plus_diff_Files.tar
> >>>>>>>>>>>>> >> > - poskozeny archiv
> >>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>>> GB_out_plus_diff_Files.tar.gz -
> >>>>>>>>>>>>> >> > poskozeny archiv
> >>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>>> GB_out_plus_diff_Files.tar.gz >
> >>>>>>>>>>>>> >> GZIP
> >>>>>>>>>>>>> >> > > GB_out_plus_diff_Files.tar - poskozeny archiv
> >>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>>> GB_out_plus_diff_Files.tar.gz >
> >>>>>>>>>>>>> >> GZIP
> >>>>>>>>>>>>> >> > > GB_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>>> GB_nucleosome-sim3.mdout-full -
> >>>>>>>>>>>>> >> > vyskytl se problem pri cteni archivu
> >>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz - poskozeny archiv
> >>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar - poskozeny archiv
> >>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>>> PME_out_plus_diff_Files.tar.gz -
> >>>>>>>>>>>>> >> > poskozeny archiv
> >>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>>> PME_out_plus_diff_Files.tar.gz >
> >>>>>>>>>>>>> >> > GZIP > PME_out_plus_diff_Files.tar - poskozeny archiv
> >>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
> >>>>>>>>>>>>> >> > PME_out_plus_diff_Filestar > TAR >
> >>>>>>>>>>>>> PME_out_plus_diff_Files.tar.gz
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >> GZIP
> >>>>>>>>>>>>> >> > > PME_out_plus_diff_Files.tar > TAR >
> >>>>>>>>>>>>> >> > PME_JAC_production_NPT-sim3.****mdout-full - vyskytl se
> >>>>>>>>>>>>> problem
> >>>>>>>>>>>>> pri
> >>>>>>>>>>>>> cteni
> >>>>>>>>>>>>> >> > archivu
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >> > http://www.eset.cz
> >>>>>>>>>>>>> >> >
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> --
> >>>>>>>>>>>>> >> Tato zpráva byla vytvořena převratným poštovním klientem
> >>>>>>>>>>>>> Opery:
> >>>>>>>>>>>>> >> http://www.opera.com/mail/
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> >> ______________________________****_________________
> >>>>>>>>>>>>> >> AMBER mailing list
> >>>>>>>>>>>>> >> AMBER.ambermd.org
> >>>>>>>>>>>>> >>
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> > ______________________________****_________________
> >>>>>>>>>>>>> > AMBER mailing list
> >>>>>>>>>>>>> > AMBER.ambermd.org
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> > __________ Informace od ESET NOD32 Antivirus, verze
> >>>>>>>>>>>>> databaze
> >>>>>>>>>>>>> 8398
> >>>>>>>>>>>>> > (20130531) __________
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> > http://www.eset.cz
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> --
> >>>>>>>>>>>>> Tato zpráva byla vytvořena převratným poštovním klientem
> >>>>>>>>>>>>> Opery:
> >>>>>>>>>>>>> http://www.opera.com/mail/
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> ______________________________****_________________
> >>>>>>>>>>>>> AMBER mailing list
> >>>>>>>>>>>>> AMBER.ambermd.org
> >>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<
> http://lists.ambermd.org/**mailman/listinfo/amber>
> >>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>>>>> >
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> __________ Informace od ESET NOD32 Antivirus, verze databaze
> >>>>>>>>>>> 8401
> >>>>>>>>>>> (20130601) __________
> >>>>>>>>>>>
> >>>>>>>>>>> Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>>>>>>
> >>>>>>>>>>> http://www.eset.cz
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>> --
> >>>>>>>>> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> >>>>>>>>> http://www.opera.com/mail/
> >>>>>>>>>
> >>>>>>>>> ______________________________**_________________
> >>>>>>>>> AMBER mailing list
> >>>>>>>>> AMBER.ambermd.org
> >>>>>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>
> >>>>>>> ______________________________**_________________
> >>>>>> AMBER mailing list
> >>>>>> AMBER.ambermd.org
> >>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>
> >>>>>> __________ Informace od ESET NOD32 Antivirus, verze databaze 8403
> >>>>>> (20130602) __________
> >>>>>>
> >>>>>> Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>>
> >>>>>> http://www.eset.cz
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>
> >>>>>
> >>>>
> >>>>
> >>>
> >>>
> >>
> >> --
> >> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> >> http://www.opera.com/mail/
> >>
> >> _______________________________________________
> >> AMBER mailing list
> >> AMBER.ambermd.org
> >> http://lists.ambermd.org/mailman/listinfo/amber
> >>
> >>
> > _______________________________________________
> > AMBER mailing list
> > AMBER.ambermd.org
> > http://lists.ambermd.org/mailman/listinfo/amber
>
>
> --
> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> http://www.opera.com/mail/
>
> _______________________________________________
> AMBER mailing list
> AMBER.ambermd.org
> http://lists.ambermd.org/mailman/listinfo/amber
>
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Wed Jun 19 2013 - 10:30:08 PDT
Custom Search