Re: [AMBER] experiences with EVGA GTX TITAN Superclocked - memtestG80 - UNDERclocking in Linux ?

From: Marek Maly <marek.maly.ujep.cz>
Date: Wed, 19 Jun 2013 18:55:32 +0200

Hi Scott,
thanks for update ! Regarding "GB/Titan" issue it is perhaps still
in debugging phase too (like cuFFT/Titan problem) if I understood well.

     Best,

        Marek




Dne Wed, 19 Jun 2013 17:50:51 +0200 Scott Le Grand <varelse2005.gmail.com>
napsal/-a:

> Hey Marek,
> No updates per se. I had a theory about what was going on that proved to
> be wrong after testing, but I'm still waiting on NVIDIA to report
> something
> beyond having reproed the problem.
>
> Really really really interested in GTX 780 data right now...
>
>
>
> On Wed, Jun 19, 2013 at 8:20 AM, Marek Maly <marek.maly.ujep.cz> wrote:
>
>> Hi all,
>>
>> just a small update from my site.
>>
>> As I have yesterday obtained announcement that the CUDA 5.5 is
>> now available for public (not just for developers).
>>
>> I downloaded it from here:
>>
>> https://developer.nvidia.com/**cuda-pre-production<https://developer.nvidia.com/cuda-pre-production>
>>
>> It is still "just" release candidate ( as all Amber/Titan club members
>> perfectly know :)) ).
>>
>> So I installed this newest release and recompiled Amber cuda code.
>>
>> I was hoping that maybe there was "silently" incorporated some
>> improvement (e.g. in cuFFT) as the result e.g. of Scott's bug report.
>>
>> The results of my 100K tests are attached. It seems that comparing to my
>> latest
>> tests with CUDA 5.5. release candidate from June 3rd (when it was
>> accessible just for CUDA developers in the form of *.run binary
>> installer)
>> there
>> is some slight improvement - e.g. my more stable TITAN was able to
>> finish
>> successfully
>> all the 100K tests including Cellulose twice. But there is still an
>> issue
>> with JAC NVE/NPT irreproducible results. On my "less stable" TITAN the
>> results are slightly better
>> then those older ones as well but still not err free (JAC/CELLULOSE) -
>> see
>> attached file.
>>
>> FACTOR IX NVE/NPT finished again with 100% reproducibility on both GPUs
>> as
>> usually.
>>
>> Scott, do you have any update regarding the "cuFFT"/TITAN issue which
>> you
>> reported/described
>> to NVIDIA guys ? The latest info from you regarding this story was, that
>> they were able to
>> reproduce the "cuFFT"/TITAN error as well. Do you have any more recent
>> information ? How long
>> time it might take to NVIDIA developers to fully solve such problem in
>> your opinion ?
>>
>> Another thing. It seems that you successfully solved the "GB/TITAN"
>> problem in case of bigger molecular systems, here is your relevant
>> message
>> form June 7th.
>>
>> ------------------------------**----------------
>>
>> Really really interesting...
>>
>> I seem to have found a fix for the GB issues on my Titan - not so
>> surprisingly, it's the same fix as on GTX4xx/GTX5xx...
>>
>> But this doesn't yet explain the weirdness with cuFFT so we're not done
>> here yet...
>> ------------------------------**---------------
>>
>> It was already after the latest Amber12 bugfix18 was released and there
>> was no additional
>> bugfix released from that moment. So the "GB/TITAN" patch will be
>> released later maybe as the part of some bigger bugfix ? Or you simply
>> additionally included it into bugfix 18 after it's release ?
>>
>>
>> My last question maybe deserves the new separate thread, but anyway
>> would
>> be interesting
>> to have some information how "Amber-stable" are GTX780 comparing to
>> TITANS
>> (of course based
>> on experience of more users or on testing more than 1 or 2 GTX780 GPUs).
>>
>> Best wishes,
>>
>> Marek
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> Dne Mon, 03 Jun 2013 01:57:36 +0200 Marek Maly <marek.maly.ujep.cz>
>> napsal/-a:
>>
>>
>> Hi here are my results with CUDA 5.5
>>> (Total energy at step 100K(PME)/1000K(GB) (driver 319.23, Amber12
>>> bugfix
>>> 18 applied, cuda 5.5))
>>>
>>>
>>> No significant differences comparing the previous test with CUDA 5.0
>>> (I also added those data to the attached table with CUDA 5.5 test).
>>>
>>> Still the same trend instability in JAC tests, perfect stability and
>>> reproducibility
>>> in FACTOR_IX tests (interesting isn't it ? especially if we consider
>>> 23K
>>> atoms
>>> in JAC case and 90K atoms in case of FACTOR_IX). Again the same
>>> crashes in
>>> CELLULOSE
>>> test now also in case of TITAN_1. Also in stable and reproducible
>>> FACTOR_IX slightly
>>> changed the final energy values comparing to CUDA 5.0 case.
>>>
>>> GB simulations (1M steps) again perfectly stable and reproducible.
>>>
>>> So to conclude, Scott we trust you :)) !
>>>
>>> If you have any idea what to try else (except GPU bios editing, perhaps
>>> too
>>> premature step at this moment) let me know. I got just last idea,
>>> which could be perhaps to try change rand seed and see if it has any
>>> influence in actual trends (e.g. JAC versus FACTOR_IX).
>>>
>>> TO ET : I am curious about your test in single GPU configuration.
>>> Regarding
>>> to your Win tests, in my opinion it is just wasting of time. They
>>> perhaps
>>> tells
>>> you just something about the GPU performance not about the eventual GPU
>>> "soft" errs.
>>>
>>> If intensive memtestG80 and/or cuda_memtest results were negative
>>> there is
>>> in my opinion
>>> very unlikely that Win performace testers will find any errs, but I am
>>> not
>>> an expert
>>> here ...
>>>
>>> Anyway If you learn which tests the ebuyer is using to confirm GPU
>>> errs,
>>> let us know.
>>>
>>> M.
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> Dne Sun, 02 Jun 2013 19:22:54 +0200 Marek Maly <marek.maly.ujep.cz>
>>> napsal/-a:
>>>
>>> Hi so I finally succeeded to compile GPU Amber part under CUDA 5.5
>>>> (after "hacking" of the configure2 file) with common results in
>>>> consequent tests:
>>>>
>>>> ------
>>>> 80 file comparisons passed
>>>> 9 file comparisons failed
>>>> 0 tests experienced errors
>>>> ------
>>>>
>>>> So now I am running the 100K(PME)/1000K(GB) repetitive benchmark tests
>>>> under
>>>> this configuration: drv. 319.23, CUDA 5.5. , bugfix 18 installed
>>>>
>>>> When I finish it I will report results here.
>>>>
>>>> M.
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Dne Sun, 02 Jun 2013 18:44:23 +0200 Marek Maly <marek.maly.ujep.cz>
>>>> napsal/-a:
>>>>
>>>> Hi Scott thanks for the update !
>>>>>
>>>>> Anyway any explanation regarding "cuFFT hypothesis" why there are no
>>>>> problems
>>>>> with GTX 580, GTX 680 or even K20c ???
>>>>>
>>>>>
>>>>> meanwhile I also tried to recompile GPU part of Amber with
>>>>> cuda 5.5 installed before, I have obtained these errs
>>>>> already in configure phase:
>>>>>
>>>>> --------
>>>>> [root.dyn-138-272 amber12]# ./configure -cuda -noX11 gnu
>>>>> Checking for updates...
>>>>> Checking for available patches online. This may take a few seconds...
>>>>>
>>>>> Available AmberTools 13 patches:
>>>>>
>>>>> No patches available
>>>>>
>>>>> Available Amber 12 patches:
>>>>>
>>>>> No patches available
>>>>> Searching for python2... Found python2.6: /usr/bin/python2.6
>>>>> Error: Unsupported CUDA version 5.5 detected.
>>>>> AMBER requires CUDA version == 4.2 .or. 5.0
>>>>> Configure failed due to the errors above!
>>>>> ---------
>>>>>
>>>>> so it seems that Amber is possible to compile only with CUDA 4.2 or
>>>>> 5.0
>>>>> at
>>>>> the moment:
>>>>>
>>>>> and this part of configure2 file has to be edited:
>>>>>
>>>>>
>>>>> -----------
>>>>> nvcc="$CUDA_HOME/bin/nvcc"
>>>>> sm35flags='-gencode arch=compute_35,code=sm_35'
>>>>> sm30flags='-gencode arch=compute_30,code=sm_30'
>>>>> sm20flags='-gencode arch=compute_20,code=sm_20'
>>>>> sm13flags='-gencode arch=compute_13,code=sm_13'
>>>>> nvccflags="$sm13flags $sm20flags"
>>>>> cudaversion=`$nvcc --version | grep 'release' | cut -d' ' -f5 |
>>>>> cut
>>>>> -d',' -f1`
>>>>> if [ "$cudaversion" == "5.0" ]; then
>>>>> echo "CUDA Version $cudaversion detected"
>>>>> nvccflags="$nvccflags $sm30flags $sm35flags"
>>>>> elif [ "$cudaversion" == "4.2" ]; then
>>>>> echo "CUDA Version $cudaversion detected"
>>>>> nvccflags="$nvccflags $sm30flags"
>>>>> else
>>>>> echo "Error: Unsupported CUDA version $cudaversion detected."
>>>>> echo "AMBER requires CUDA version == 4.2 .or. 5.0"
>>>>> exit 1
>>>>> fi
>>>>> nvcc="$nvcc $nvccflags"
>>>>>
>>>>> fi
>>>>>
>>>>> -----------
>>>>>
>>>>> would it be just OK to change
>>>>> "if [ "$cudaversion" == "5.0" ]; then"
>>>>>
>>>>> to
>>>>>
>>>>> "if [ "$cudaversion" == "5.5" ]; then"
>>>>>
>>>>>
>>>>> or some more flags etc. should be defined here to proceed
>>>>> successfully ?
>>>>>
>>>>>
>>>>> BTW it seems Scott, that you are on the way to isolate the problem
>>>>> soon
>>>>> so maybe it's better to wait and not to loose time with cuda 5.5
>>>>> experiments.
>>>>>
>>>>> I just thought that cuda 5.5 might be more "friendly" to Titans :))
>>>>> e.g.
>>>>> in terms of cuFFT function ....
>>>>>
>>>>>
>>>>> I will keep fingers crossed :))
>>>>>
>>>>> M.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> Dne Sun, 02 Jun 2013 18:33:52 +0200 Scott Le Grand
>>>>> <varelse2005.gmail.com>
>>>>> napsal/-a:
>>>>>
>>>>> PS this *might* indicate a software bug in cuFFT, but it needs more
>>>>>> characterization... And things are going to get a little stream of
>>>>>> consciousness from here because you're getting unfiltered raw data,
>>>>>> so
>>>>>> please don't draw any conclusions towards anything yet - I'm just
>>>>>> letting
>>>>>> you guys know what I'm finding out as I find it...
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Sun, Jun 2, 2013 at 9:31 AM, Scott Le Grand
>>>>>> <varelse2005.gmail.com>wrote:
>>>>>>
>>>>>> And bingo...
>>>>>>>
>>>>>>> At the very least, the reciprocal sum is intermittently
>>>>>>> inconsistent...
>>>>>>> This explains the irreproducible behavior...
>>>>>>>
>>>>>>> And here's the level of inconsistency:
>>>>>>> 31989.38940628897399 vs
>>>>>>> 31989.39168370794505
>>>>>>>
>>>>>>> That's error at the level of 1e-7 or a somehow missed
>>>>>>> single-precision
>>>>>>> transaction somewhere...
>>>>>>>
>>>>>>> The next question is figuring out why... This may or may not
>>>>>>> ultimately
>>>>>>> explain the crashes you guys are also seeing...
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Sun, Jun 2, 2013 at 9:07 AM, Scott Le Grand
>>>>>>> <varelse2005.gmail.com>wrote:
>>>>>>>
>>>>>>>
>>>>>>>> Observations:
>>>>>>>> 1. The degree to which the reproducibility is broken *does*
>>>>>>>> appear to
>>>>>>>> vary between individual Titan GPUs. One of my Titans breaks
>>>>>>>> within
>>>>>>>> 10K
>>>>>>>> steps on cellulose, the other one made it to 100K steps twice
>>>>>>>> without
>>>>>>>> doing
>>>>>>>> so leading me to believe it could be trusted (until yesterday
>>>>>>>> where I
>>>>>>>> now
>>>>>>>> see it dies between 50K and 100K steps most of the time).
>>>>>>>>
>>>>>>>> 2. GB hasn't broken (yet). So could you run myoglobin for 500K
>>>>>>>> and
>>>>>>>> TRPcage for 1,000,000 steps and let's see if that's universal.
>>>>>>>>
>>>>>>>> 3. Turning on double-precision mode makes my Titan crash rather
>>>>>>>> than
>>>>>>>> run
>>>>>>>> irreproducibly, sigh...
>>>>>>>>
>>>>>>>> So whatever is going on is triggered by something in PME but not
>>>>>>>> GB.
>>>>>>>> So
>>>>>>>> that's either the radix sort, the FFT, the Ewald grid
>>>>>>>> interpolation,
>>>>>>>> or the
>>>>>>>> neighbor list code. Fixing this involves isolating this and
>>>>>>>> figuring
>>>>>>>> out
>>>>>>>> what exactly goes haywire. It could *still* be software at some
>>>>>>>> very
>>>>>>>> small
>>>>>>>> probability but the combination of both 680 and K20c with ECC off
>>>>>>>> running
>>>>>>>> reliably is really pointing towards the Titans just being clocked
>>>>>>>> too
>>>>>>>> fast.
>>>>>>>>
>>>>>>>> So how long with this take? Asking people how long it takes to
>>>>>>>> fix a
>>>>>>>> bug
>>>>>>>> never really works out well. That said, I found the 480 bug
>>>>>>>> within a
>>>>>>>> week
>>>>>>>> and my usual turnaround for a bug with a solid repro is <24 hours.
>>>>>>>>
>>>>>>>> Scott
>>>>>>>>
>>>>>>>> On Sun, Jun 2, 2013 at 7:58 AM, Marek Maly <marek.maly.ujep.cz>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>> Hi all,
>>>>>>>>>
>>>>>>>>> here are my results after bugfix 18 application (see attachment).
>>>>>>>>>
>>>>>>>>> In principle I don't see any "drastical" changes.
>>>>>>>>>
>>>>>>>>> FACTOR_IX still perfectly stable/reproducible on both cards,
>>>>>>>>>
>>>>>>>>> JAC tests - problems with finishing AND/OR reproducibility the
>>>>>>>>> same CELLULOSE_NVE although here it seems that my TITAN_1
>>>>>>>>> has no problems with this test (but the same same trend I saw
>>>>>>>>> also
>>>>>>>>> before bugfix 18 - see my older 500K steps test).
>>>>>>>>>
>>>>>>>>> But anyway bugfix 18 brought here one change.
>>>>>>>>>
>>>>>>>>> The err
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> #1 ERR writtent in mdout:
>>>>>>>>> ------
>>>>>>>>> | ERROR: max pairlist cutoff must be less than unit cell max
>>>>>>>>> sphere
>>>>>>>>> radius!
>>>>>>>>> ------
>>>>>>>>>
>>>>>>>>> was substituted with err/warning ?
>>>>>>>>>
>>>>>>>>> #0 no ERR writtent in mdout, ERR written in standard output
>>>>>>>>> (nohup.out)
>>>>>>>>> -----
>>>>>>>>> Nonbond cells need to be recalculated, restart simulation from
>>>>>>>>> previous
>>>>>>>>> checkpoint
>>>>>>>>> with a higher value for skinnb.
>>>>>>>>>
>>>>>>>>> -----
>>>>>>>>>
>>>>>>>>> Another thing,
>>>>>>>>>
>>>>>>>>> recently I started on another machine and GTX 580 GPU simulation
>>>>>>>>> of
>>>>>>>>> relatively
>>>>>>>>> big system ( 364275 atoms/PME ). The system is composed also from
>>>>>>>>> the
>>>>>>>>> "exotic" molecules like polymers. ff12SB, gaff, GLYCAM
>>>>>>>>> forcefields
>>>>>>>>> used
>>>>>>>>> here. I had problem even with minimization part here, having big
>>>>>>>>> energy
>>>>>>>>> on the start:
>>>>>>>>>
>>>>>>>>> -----
>>>>>>>>> NSTEP ENERGY RMS GMAX NAME
>>>>>>>>> NUMBER
>>>>>>>>> 1 2.8442E+09 2.1339E+02 1.7311E+04 O
>>>>>>>>> 32998
>>>>>>>>>
>>>>>>>>> BOND = 11051.7467 ANGLE = 17720.4706 DIHED =
>>>>>>>>> 18977.7584
>>>>>>>>> VDWAALS = ************* EEL = -1257709.6203 HBOND =
>>>>>>>>> 0.0000
>>>>>>>>> 1-4 VDW = 7253.7412 1-4 EEL = 149867.0207 RESTRAINT =
>>>>>>>>> 0.0000
>>>>>>>>>
>>>>>>>>> ----
>>>>>>>>>
>>>>>>>>> with no chance to minimize the system even with 50 000 steps in
>>>>>>>>> both
>>>>>>>>> min cycles (with constrained and unconstrained solute) and hence
>>>>>>>>> heating
>>>>>>>>> NVT
>>>>>>>>> crashed immediately even with very small dt. I patched Amber12
>>>>>>>>> here
>>>>>>>>> with
>>>>>>>>> the
>>>>>>>>> bugfix 18 and the minimization was done without any problem with
>>>>>>>>> common
>>>>>>>>> 5000 steps
>>>>>>>>> (obtaining target Energy -1.4505E+06 while that initial was that
>>>>>>>>> written
>>>>>>>>> above).
>>>>>>>>>
>>>>>>>>> So indeed bugfix 18 solved some issues, but unfortunately not
>>>>>>>>> those
>>>>>>>>> related to
>>>>>>>>> Titans.
>>>>>>>>>
>>>>>>>>> Here I will try to install cuda 5.5, recompile GPU Amber part
>>>>>>>>> with
>>>>>>>>> this
>>>>>>>>> new
>>>>>>>>> cuda version and repeat the 100K tests.
>>>>>>>>>
>>>>>>>>> Scott, let us know how finished your experiment with
>>>>>>>>> downclocking of
>>>>>>>>> Titan.
>>>>>>>>> Maybe the best choice would be here to flash Titan directly with
>>>>>>>>> your
>>>>>>>>> K20c bios :))
>>>>>>>>>
>>>>>>>>> M.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Dne Sat, 01 Jun 2013 21:09:46 +0200 Marek Maly
>>>>>>>>> <marek.maly.ujep.cz>
>>>>>>>>> napsal/-a:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> first of all thanks for providing of your test results !
>>>>>>>>>>
>>>>>>>>>> It seems that your results are more or less similar to that of
>>>>>>>>>> mine maybe with the exception of the results on FactorIX tests
>>>>>>>>>> where I had perfect stability and 100% or close to 100%
>>>>>>>>>> reproducibility.
>>>>>>>>>>
>>>>>>>>>> Anyway the type of errs which you reported are the same which I
>>>>>>>>>> obtained.
>>>>>>>>>>
>>>>>>>>>> So let's see if the bugfix 18 will help here (or at least on NPT
>>>>>>>>>> tests)
>>>>>>>>>> or not. As I wrote just before few minutes, it seems that it was
>>>>>>>>>> not
>>>>>>>>>> still
>>>>>>>>>> loaded
>>>>>>>>>> to the given server, although it's description is already
>>>>>>>>>> present
>>>>>>>>>> on
>>>>>>>>>> the
>>>>>>>>>> given
>>>>>>>>>> web page ( see
>>>>>>>>>> http://ambermd.org/bugfixes12.****html<http://ambermd.org/bugfixes12.**html>
>>>>>>>>>> <http://ambermd.org/**bugfixes12.html<http://ambermd.org/bugfixes12.html>
>>>>>>>>>> >).
>>>>>>>>>>
>>>>>>>>>> As you can see, this bugfix contains also changes in CPU code
>>>>>>>>>> although
>>>>>>>>>> the majority is devoted to GPU code, so perhaps the best will
>>>>>>>>>> be to
>>>>>>>>>> recompile
>>>>>>>>>> whole amber with this patch although this patch would be perhaps
>>>>>>>>>> applied
>>>>>>>>>> even after just
>>>>>>>>>> GPU configure command ( i.e. ./configure -cuda -noX11 gnu ) but
>>>>>>>>>> after
>>>>>>>>>> consequent
>>>>>>>>>> building, just the GPU binaries will be updated. Anyway I would
>>>>>>>>>> rather
>>>>>>>>>> recompile
>>>>>>>>>> whole Amber after this patch.
>>>>>>>>>>
>>>>>>>>>> Regarding to GPU test under linux you may try memtestG80
>>>>>>>>>> (please use the updated/patched version from here
>>>>>>>>>> https://github.com/ihaque/****memtestG80<https://github.com/ihaque/**memtestG80>
>>>>>>>>>> <https://github.com/**ihaque/memtestG80<https://github.com/ihaque/memtestG80>
>>>>>>>>>> >
>>>>>>>>>> )
>>>>>>>>>>
>>>>>>>>>> just use git command like:
>>>>>>>>>>
>>>>>>>>>> git clone
>>>>>>>>>> https://github.com/ihaque/****memtestG80.git<https://github.com/ihaque/**memtestG80.git>
>>>>>>>>>> <https://github.**com/ihaque/memtestG80.git<https://github.com/ihaque/memtestG80.git>
>>>>>>>>>> >**PATCHED_MEMTEST-G80
>>>>>>>>>>
>>>>>>>>>> to download all the files and save them into directory named
>>>>>>>>>> PATCHED_MEMTEST-G80.
>>>>>>>>>>
>>>>>>>>>> another possibility is to try perhaps similar (but maybe more
>>>>>>>>>> up to
>>>>>>>>>> date)
>>>>>>>>>> test
>>>>>>>>>> cuda_memtest (
>>>>>>>>>> http://sourceforge.net/****projects/cudagpumemtest/<http://sourceforge.net/**projects/cudagpumemtest/>
>>>>>>>>>> <http:**//sourceforge.net/projects/**cudagpumemtest/<http://sourceforge.net/projects/cudagpumemtest/>
>>>>>>>>>> >).
>>>>>>>>>>
>>>>>>>>>> regarding ig value: If ig is not present in mdin, the default
>>>>>>>>>> value
>>>>>>>>>> is
>>>>>>>>>> used
>>>>>>>>>> (e.g. 71277) if ig=-1 the random seed will be based on the
>>>>>>>>>> current
>>>>>>>>>> date
>>>>>>>>>> and time, and hence will be different for every run (not a good
>>>>>>>>>> variant
>>>>>>>>>> for our testts). I simply deleted eventual ig records from all
>>>>>>>>>> mdins
>>>>>>>>>> so
>>>>>>>>>> I
>>>>>>>>>> assume that in each run the default seed 71277 was automatically
>>>>>>>>>> used.
>>>>>>>>>>
>>>>>>>>>> M.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Dne Sat, 01 Jun 2013 20:26:16 +0200 ET <sketchfoot.gmail.com>
>>>>>>>>>> napsal/-a:
>>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> I've put the graphics card into a machine with the working GTX
>>>>>>>>>>> titan
>>>>>>>>>>> that I
>>>>>>>>>>> mentioned earlier.
>>>>>>>>>>>
>>>>>>>>>>> The Nvidia driver version is: 133.30
>>>>>>>>>>>
>>>>>>>>>>> Amber version is:
>>>>>>>>>>> AmberTools version 13.03
>>>>>>>>>>> Amber version 12.16
>>>>>>>>>>>
>>>>>>>>>>> I ran 50k steps with the amber benchmark using ig=43689 on both
>>>>>>>>>>> cards.
>>>>>>>>>>> For
>>>>>>>>>>> the purpose of discriminating between them, the card I believe
>>>>>>>>>>> (fingers
>>>>>>>>>>> crossed) is working is called GPU-00_TeaNCake, whilst the other
>>>>>>>>>>> one
>>>>>>>>>>> is
>>>>>>>>>>> called GPU-01_008.
>>>>>>>>>>>
>>>>>>>>>>> *When I run the tests on GPU-01_008:*
>>>>>>>>>>>
>>>>>>>>>>> 1) All the tests (across 2x repeats) finish apart from the
>>>>>>>>>>> following
>>>>>>>>>>> which
>>>>>>>>>>> have the errors listed:
>>>>>>>>>>>
>>>>>>>>>>> ------------------------------****--------------
>>>>>>>>>>> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
>>>>>>>>>>> Error: unspecified launch failure launching kernel kNLSkinTest
>>>>>>>>>>> cudaFree GpuBuffer::Deallocate failed unspecified launch
>>>>>>>>>>> failure
>>>>>>>>>>>
>>>>>>>>>>> ------------------------------****--------------
>>>>>>>>>>> CELLULOSE_PRODUCTION_NPT - 408,609 atoms PME
>>>>>>>>>>> cudaMemcpy GpuBuffer::Download failed unspecified launch
>>>>>>>>>>> failure
>>>>>>>>>>>
>>>>>>>>>>> ------------------------------****--------------
>>>>>>>>>>> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
>>>>>>>>>>> Error: unspecified launch failure launching kernel kNLSkinTest
>>>>>>>>>>> cudaFree GpuBuffer::Deallocate failed unspecified launch
>>>>>>>>>>> failure
>>>>>>>>>>>
>>>>>>>>>>> ------------------------------****--------------
>>>>>>>>>>> CELLULOSE_PRODUCTION_NPT - 408,609 atoms PME
>>>>>>>>>>> cudaMemcpy GpuBuffer::Download failed unspecified launch
>>>>>>>>>>> failure
>>>>>>>>>>> grep: mdinfo.1GTX680: No such file or directory
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> 2) The sdiff logs indicate that reproducibility across the two
>>>>>>>>>>> repeats
>>>>>>>>>>> is
>>>>>>>>>>> as follows:
>>>>>>>>>>>
>>>>>>>>>>> *GB_myoglobin: *Reproducible across 50k steps
>>>>>>>>>>> *GB_nucleosome:* Reproducible till step 7400
>>>>>>>>>>> *GB_TRPCage:* Reproducible across 50k steps
>>>>>>>>>>>
>>>>>>>>>>> *PME_JAC_production_NVE: *No reproducibility shown from step
>>>>>>>>>>> 1,000
>>>>>>>>>>> onwards
>>>>>>>>>>> *PME_JAC_production_NPT*: Reproducible till step 1,000. Also
>>>>>>>>>>> outfile
>>>>>>>>>>> is
>>>>>>>>>>> not written properly - blank gaps appear where something should
>>>>>>>>>>> have
>>>>>>>>>>> been
>>>>>>>>>>> written
>>>>>>>>>>>
>>>>>>>>>>> *PME_FactorIX_production_NVE:* Reproducible across 50k steps
>>>>>>>>>>> *PME_FactorIX_production_NPT:* Reproducible across 50k steps
>>>>>>>>>>>
>>>>>>>>>>> *PME_Cellulose_production_NVE:***** Failure means that both
>>>>>>>>>>> runs
>>>>>>>>>>> do
>>>>>>>>>>> not
>>>>>>>>>>> finish
>>>>>>>>>>> (see point1)
>>>>>>>>>>> *PME_Cellulose_production_NPT: *Failure means that both runs do
>>>>>>>>>>> not
>>>>>>>>>>> finish
>>>>>>>>>>> (see point1)
>>>>>>>>>>>
>>>>>>>>>>> ##############################****############################**
>>>>>>>>>>> ##**
>>>>>>>>>>> ###########################
>>>>>>>>>>>
>>>>>>>>>>> *When I run the tests on * *GPU-00_TeaNCake:*
>>>>>>>>>>> *
>>>>>>>>>>> *
>>>>>>>>>>> 1) All the tests (across 2x repeats) finish apart from the
>>>>>>>>>>> following
>>>>>>>>>>> which
>>>>>>>>>>> have the errors listed:
>>>>>>>>>>> ------------------------------****-------
>>>>>>>>>>> JAC_PRODUCTION_NPT - 23,558 atoms PME
>>>>>>>>>>> PMEMD Terminated Abnormally!
>>>>>>>>>>> ------------------------------****-------
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> 2) The sdiff logs indicate that reproducibility across the two
>>>>>>>>>>> repeats
>>>>>>>>>>> is
>>>>>>>>>>> as follows:
>>>>>>>>>>>
>>>>>>>>>>> *GB_myoglobin:* Reproducible across 50k steps
>>>>>>>>>>> *GB_nucleosome:* Reproducible across 50k steps
>>>>>>>>>>> *GB_TRPCage:* Reproducible across 50k steps
>>>>>>>>>>>
>>>>>>>>>>> *PME_JAC_production_NVE:* No reproducibility shown from step
>>>>>>>>>>> 10,000
>>>>>>>>>>> onwards
>>>>>>>>>>> *PME_JAC_production_NPT: * No reproducibility shown from step
>>>>>>>>>>> 10,000
>>>>>>>>>>> onwards. Also outfile is not written properly - blank gaps
>>>>>>>>>>> appear
>>>>>>>>>>> where
>>>>>>>>>>> something should have been written. Repeat 2 Crashes with error
>>>>>>>>>>> noted
>>>>>>>>>>> in
>>>>>>>>>>> 1.
>>>>>>>>>>>
>>>>>>>>>>> *PME_FactorIX_production_NVE:* No reproducibility shown from
>>>>>>>>>>> step
>>>>>>>>>>> 9,000
>>>>>>>>>>> onwards
>>>>>>>>>>> *PME_FactorIX_production_NPT: *Reproducible across 50k steps
>>>>>>>>>>>
>>>>>>>>>>> *PME_Cellulose_production_NVE: *No reproducibility shown from
>>>>>>>>>>> step
>>>>>>>>>>> 5,000
>>>>>>>>>>> onwards
>>>>>>>>>>> *PME_Cellulose_production_NPT: ** *No reproducibility shown
>>>>>>>>>>> from
>>>>>>>>>>> step
>>>>>>>>>>> 29,000 onwards. Also outfile is not written properly - blank
>>>>>>>>>>> gaps
>>>>>>>>>>> appear
>>>>>>>>>>> where something should have been written.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Out files and sdiff files are included as attatchments
>>>>>>>>>>>
>>>>>>>>>>> ##############################****###################
>>>>>>>>>>>
>>>>>>>>>>> So I'm going to update my nvidia driver to the latest version
>>>>>>>>>>> and
>>>>>>>>>>> patch
>>>>>>>>>>> amber to the latest version and rerun the tests to see if
>>>>>>>>>>> there is
>>>>>>>>>>> any
>>>>>>>>>>> improvement. Could someone let me know if it is necessary to
>>>>>>>>>>> recompile
>>>>>>>>>>> any
>>>>>>>>>>> or all of AMBER after applying the bugfixes?
>>>>>>>>>>>
>>>>>>>>>>> Additionally, I'm going to run memory tests and heaven
>>>>>>>>>>> benchmarks
>>>>>>>>>>> on
>>>>>>>>>>> the
>>>>>>>>>>> cards to check whether they are faulty or not.
>>>>>>>>>>>
>>>>>>>>>>> I'm thinking that there is a mix of hardware
>>>>>>>>>>> error/configuration
>>>>>>>>>>> (esp
>>>>>>>>>>> in
>>>>>>>>>>> the case of GPU-01_008) and amber software error in this
>>>>>>>>>>> situation.
>>>>>>>>>>> What
>>>>>>>>>>> do
>>>>>>>>>>> you guys think?
>>>>>>>>>>>
>>>>>>>>>>> Also am I right in thinking (from what Scott was saying) that
>>>>>>>>>>> all
>>>>>>>>>>> the
>>>>>>>>>>> benchmarks should be reproducible across 50k steps but begin to
>>>>>>>>>>> diverge
>>>>>>>>>>> at
>>>>>>>>>>> around 100K steps? Is there any difference from in setting *ig
>>>>>>>>>>> *to
>>>>>>>>>>> an
>>>>>>>>>>> explicit number to removing it from the mdin file?
>>>>>>>>>>>
>>>>>>>>>>> br,
>>>>>>>>>>> g
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On 31 May 2013 23:45, ET <sketchfoot.gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>> I don't need sysadmins, but sysadmins need me as it gives
>>>>>>>>>>> purpose
>>>>>>>>>>> to
>>>>>>>>>>>
>>>>>>>>>>>> their
>>>>>>>>>>>> bureaucratic existence. A encountered evil if working in an
>>>>>>>>>>>> institution
>>>>>>>>>>>> or
>>>>>>>>>>>> comapny IMO. Good science and indiviguality being sacrificed
>>>>>>>>>>>> for
>>>>>>>>>>>> standardisation and mediocrity in the intrerests of maintaing
>>>>>>>>>>>> a
>>>>>>>>>>>> system
>>>>>>>>>>>> that
>>>>>>>>>>>> focusses on maintaining the system and not the objective.
>>>>>>>>>>>>
>>>>>>>>>>>> You need root to move fwd on these things, unfortunately. and
>>>>>>>>>>>> ppl
>>>>>>>>>>>> with
>>>>>>>>>>>> root are kinda like your parents when you try to borrow money
>>>>>>>>>>>> from
>>>>>>>>>>>> them
>>>>>>>>>>>> .
>>>>>>>>>>>> age 12 :D
>>>>>>>>>>>> On May 31, 2013 9:34 PM, "Marek Maly" <marek.maly.ujep.cz>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Sorry why do you need sysadmins :)) ?
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> BTW here is the most recent driver:
>>>>>>>>>>>>>
>>>>>>>>>>>>> http://www.nvidia.com/object/****linux-display-amd64-319.23-**<http://www.nvidia.com/object/**linux-display-amd64-319.23-**>
>>>>>>>>>>>>> driver.html<http://www.nvidia.**com/object/linux-display-**
>>>>>>>>>>>>> amd64-319.23-driver.html<http://www.nvidia.com/object/linux-display-amd64-319.23-driver.html>
>>>>>>>>>>>>> >
>>>>>>>>>>>>>
>>>>>>>>>>>>> I do not remember anything easier than is to install driver
>>>>>>>>>>>>> (especially
>>>>>>>>>>>>> in case of binary (*.run) installer) :))
>>>>>>>>>>>>>
>>>>>>>>>>>>> M.
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Dne Fri, 31 May 2013 22:02:34 +0200 ET <sketchfoot.gmail.com>
>>>>>>>>>>>>> napsal/-a:
>>>>>>>>>>>>>
>>>>>>>>>>>>> > Yup. I know. I replaced a 680 and the everknowing sysadmins
>>>>>>>>>>>>> are
>>>>>>>>>>>>> reluctant
>>>>>>>>>>>>> > to install drivers not in the repositoery as they are
>>>>>>>>>>>>> lame. :(
>>>>>>>>>>>>> > On May 31, 2013 7:14 PM, "Marek Maly" <marek.maly.ujep.cz>
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> As I already wrote you,
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> the first driver which properly/officially supports
>>>>>>>>>>>>> Titans,
>>>>>>>>>>>>> should
>>>>>>>>>>>>> be
>>>>>>>>>>>>> >> 313.26 .
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> Anyway I am curious mainly about your 100K repetitive
>>>>>>>>>>>>> tests
>>>>>>>>>>>>> with
>>>>>>>>>>>>> >> your Titan SC card. Especially in case of these tests (
>>>>>>>>>>>>> JAC_NVE,
>>>>>>>>>>>>> JAC_NPT
>>>>>>>>>>>>> >> and CELLULOSE_NVE ) where
>>>>>>>>>>>>> >> my Titans SC randomly failed or succeeded. In
>>>>>>>>>>>>> FACTOR_IX_NVE,
>>>>>>>>>>>>> >> FACTOR_IX_NPT
>>>>>>>>>>>>> >> tests both
>>>>>>>>>>>>> >> my cards are perfectly stable (independently from drv.
>>>>>>>>>>>>> version)
>>>>>>>>>>>>> and
>>>>>>>>>>>>> also
>>>>>>>>>>>>> >> the runs
>>>>>>>>>>>>> >> are perfectly or almost perfectly reproducible.
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> Also if your test will crash please report the eventual
>>>>>>>>>>>>> errs.
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> To this moment I have this actual library of errs on my
>>>>>>>>>>>>> Titans
>>>>>>>>>>>>> SC
>>>>>>>>>>>>> GPUs.
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> #1 ERR writtent in mdout:
>>>>>>>>>>>>> >> ------
>>>>>>>>>>>>> >> | ERROR: max pairlist cutoff must be less than unit cell
>>>>>>>>>>>>> max
>>>>>>>>>>>>> sphere
>>>>>>>>>>>>> >> radius!
>>>>>>>>>>>>> >> ------
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> #2 no ERR writtent in mdout, ERR written in standard
>>>>>>>>>>>>> output
>>>>>>>>>>>>> (nohup.out)
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> ----
>>>>>>>>>>>>> >> Error: unspecified launch failure launching kernel
>>>>>>>>>>>>> kNLSkinTest
>>>>>>>>>>>>> >> cudaFree GpuBuffer::Deallocate failed unspecified launch
>>>>>>>>>>>>> failure
>>>>>>>>>>>>> >> ----
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> #3 no ERR writtent in mdout, ERR written in standard
>>>>>>>>>>>>> output
>>>>>>>>>>>>> (nohup.out)
>>>>>>>>>>>>> >> ----
>>>>>>>>>>>>> >> cudaMemcpy GpuBuffer::Download failed unspecified launch
>>>>>>>>>>>>> failure
>>>>>>>>>>>>> >> ----
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> Another question, regarding your Titan SC, it is also
>>>>>>>>>>>>> EVGA as
>>>>>>>>>>>>> in
>>>>>>>>>>>>> my
>>>>>>>>>>>>> case
>>>>>>>>>>>>> >> or it is another producer ?
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> Thanks,
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> M.
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> Dne Fri, 31 May 2013 19:17:03 +0200 ET
>>>>>>>>>>>>> <sketchfoot.gmail.com
>>>>>>>>>>>>> >
>>>>>>>>>>>>> napsal/-a:
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> > Well, this is interesting...
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> > I ran 50k steps on the Titan on the other machine with
>>>>>>>>>>>>> driver
>>>>>>>>>>>>> 310.44
>>>>>>>>>>>>> >> and
>>>>>>>>>>>>> >> > it
>>>>>>>>>>>>> >> > passed all the GB steps. i.e totally identical results
>>>>>>>>>>>>> over
>>>>>>>>>>>>> two
>>>>>>>>>>>>> >> repeats.
>>>>>>>>>>>>> >> > However, it failed all the PME tests after step 1000.
>>>>>>>>>>>>> I'm
>>>>>>>>>>>>> going
>>>>>>>>>>>>> to
>>>>>>>>>>>>> > update
>>>>>>>>>>>>> >> > the driver and test it again.
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> > Files included as attachments.
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> > br,
>>>>>>>>>>>>> >> > g
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> > On 31 May 2013 16:40, Marek Maly <marek.maly.ujep.cz>
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> >> One more thing,
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> can you please check under which frequency is running
>>>>>>>>>>>>> that
>>>>>>>>>>>>> your
>>>>>>>>>>>>> >> titan ?
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> As the base frequency of normal Titans is 837MHz and
>>>>>>>>>>>>> the
>>>>>>>>>>>>> Boost
>>>>>>>>>>>>> one
>>>>>>>>>>>>> is
>>>>>>>>>>>>> >> >> 876MHz I
>>>>>>>>>>>>> >> >> assume that yor GPU is running automatically also under
>>>>>>>>>>>>> it's
>>>>>>>>>>>>> boot
>>>>>>>>>>>>> >> >> frequency (876MHz).
>>>>>>>>>>>>> >> >> You can find this information e.g. in Amber mdout file.
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> You also mentioned some crashes in your previous email.
>>>>>>>>>>>>> Your
>>>>>>>>>>>>> ERRs
>>>>>>>>>>>>> >> were
>>>>>>>>>>>>> >> >> something like those here:
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> #1 ERR writtent in mdout:
>>>>>>>>>>>>> >> >> ------
>>>>>>>>>>>>> >> >> | ERROR: max pairlist cutoff must be less than unit
>>>>>>>>>>>>> cell
>>>>>>>>>>>>> max
>>>>>>>>>>>>> sphere
>>>>>>>>>>>>> >> >> radius!
>>>>>>>>>>>>> >> >> ------
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> #2 no ERR writtent in mdout, ERR written in standard
>>>>>>>>>>>>> output
>>>>>>>>>>>>> >> (nohup.out)
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> ----
>>>>>>>>>>>>> >> >> Error: unspecified launch failure launching kernel
>>>>>>>>>>>>> kNLSkinTest
>>>>>>>>>>>>> >> >> cudaFree GpuBuffer::Deallocate failed unspecified
>>>>>>>>>>>>> launch
>>>>>>>>>>>>> failure
>>>>>>>>>>>>> >> >> ----
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> #3 no ERR writtent in mdout, ERR written in standard
>>>>>>>>>>>>> output
>>>>>>>>>>>>> >> (nohup.out)
>>>>>>>>>>>>> >> >> ----
>>>>>>>>>>>>> >> >> cudaMemcpy GpuBuffer::Download failed unspecified
>>>>>>>>>>>>> launch
>>>>>>>>>>>>> failure
>>>>>>>>>>>>> >> >> ----
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> or you obtained some new/additional errs ?
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> M.
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> Dne Fri, 31 May 2013 17:30:57 +0200 filip fratev
>>>>>>>>>>>>> >> <filipfratev.yahoo.com
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> >> napsal/-a:
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> > Hi,
>>>>>>>>>>>>> >> >> > This is what I obtained for 50K tests and "normal"
>>>>>>>>>>>>> GTXTitan:
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > run1:
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> ------------------------------****----------------------------*
>>>>>>>>>>>>> *--**
>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > A V E R A G E S O V E R 50 S T E P S
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
>>>>>>>>>>>>> 299.87
>>>>>>>>>>>>> >> PRESS
>>>>>>>>>>>>> >> >> > = 0.0
>>>>>>>>>>>>> >> >> > Etot = -443237.1079 EKtot = 257679.9750
>>>>>>>>>>>>> EPtot
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > -700917.0829
>>>>>>>>>>>>> >> >> > BOND = 20193.1856 ANGLE = 53517.5432
>>>>>>>>>>>>> DIHED
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 23575.4648
>>>>>>>>>>>>> >> >> > 1-4 NB = 21759.5524 1-4 EEL = 742552.5939
>>>>>>>>>>>>> VDWAALS
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 96286.7714
>>>>>>>>>>>>> >> >> > EELEC = -1658802.1941 EHBOND = 0.0000
>>>>>>>>>>>>> RESTRAINT
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 0.0000
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> ------------------------------****----------------------------
>>>>>>>>>>>>> **--**
>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > R M S F L U C T U A T I O N S
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
>>>>>>>>>>>>> 0.33
>>>>>>>>>>>>> >> PRESS
>>>>>>>>>>>>> >> >> > = 0.0
>>>>>>>>>>>>> >> >> > Etot = 11.2784 EKtot = 284.8999
>>>>>>>>>>>>> EPtot
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 289.0773
>>>>>>>>>>>>> >> >> > BOND = 136.3417 ANGLE = 214.0054
>>>>>>>>>>>>> DIHED
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 59.4893
>>>>>>>>>>>>> >> >> > 1-4 NB = 58.5891 1-4 EEL = 330.5400
>>>>>>>>>>>>> VDWAALS
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 559.2079
>>>>>>>>>>>>> >> >> > EELEC = 743.8771 EHBOND = 0.0000
>>>>>>>>>>>>> RESTRAINT
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 0.0000
>>>>>>>>>>>>> >> >> > |E(PBS) = 21.8119
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> ------------------------------****----------------------------
>>>>>>>>>>>>> **--**
>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > run2:
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> ------------------------------****----------------------------*
>>>>>>>>>>>>> *--**
>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > A V E R A G E S O V E R 50 S T E P S
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
>>>>>>>>>>>>> 299.89
>>>>>>>>>>>>> >> PRESS
>>>>>>>>>>>>> >> >> > = 0.0
>>>>>>>>>>>>> >> >> > Etot = -443240.0999 EKtot = 257700.0950
>>>>>>>>>>>>> EPtot
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > -700940.1949
>>>>>>>>>>>>> >> >> > BOND = 20241.9174 ANGLE = 53644.6694
>>>>>>>>>>>>> DIHED
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 23541.3737
>>>>>>>>>>>>> >> >> > 1-4 NB = 21803.1898 1-4 EEL = 742754.2254
>>>>>>>>>>>>> VDWAALS
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 96298.8308
>>>>>>>>>>>>> >> >> > EELEC = -1659224.4013 EHBOND = 0.0000
>>>>>>>>>>>>> RESTRAINT
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 0.0000
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> ------------------------------****----------------------------
>>>>>>>>>>>>> **--**
>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > R M S F L U C T U A T I O N S
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
>>>>>>>>>>>>> 0.41
>>>>>>>>>>>>> >> PRESS
>>>>>>>>>>>>> >> >> > = 0.0
>>>>>>>>>>>>> >> >> > Etot = 10.7633 EKtot = 348.2819
>>>>>>>>>>>>> EPtot
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 353.9918
>>>>>>>>>>>>> >> >> > BOND = 106.5314 ANGLE = 196.7052
>>>>>>>>>>>>> DIHED
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 69.7476
>>>>>>>>>>>>> >> >> > 1-4 NB = 60.3435 1-4 EEL = 400.7466
>>>>>>>>>>>>> VDWAALS
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 462.7763
>>>>>>>>>>>>> >> >> > EELEC = 651.9857 EHBOND = 0.0000
>>>>>>>>>>>>> RESTRAINT
>>>>>>>>>>>>> =
>>>>>>>>>>>>> >> >> > 0.0000
>>>>>>>>>>>>> >> >> > |E(PBS) = 17.0642
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> ------------------------------****----------------------------
>>>>>>>>>>>>> **--**
>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> ------------------------------****----------------------------*
>>>>>>>>>>>>> *--**
>>>>>>>>>>>>> --------------------
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > ______________________________****__
>>>>>>>>>>>>> >> >> > From: Marek Maly <marek.maly.ujep.cz>
>>>>>>>>>>>>> >> >> > To: AMBER Mailing List <amber.ambermd.org>
>>>>>>>>>>>>> >> >> > Sent: Friday, May 31, 2013 3:34 PM
>>>>>>>>>>>>> >> >> > Subject: Re: [AMBER] experiences with EVGA GTX TITAN
>>>>>>>>>>>>> Superclocked
>>>>>>>>>>>>> -
>>>>>>>>>>>>> >> >> > memtestG80 - UNDERclocking in Linux ?
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > Hi here are my 100K results for driver 313.30 (and
>>>>>>>>>>>>> still
>>>>>>>>>>>>> Cuda
>>>>>>>>>>>>> 5.0).
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > The results are rather similar to those obtained
>>>>>>>>>>>>> >> >> > under my original driver 319.17 (see the first table
>>>>>>>>>>>>> >> >> > which I sent in this thread).
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > M.
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> > Dne Fri, 31 May 2013 12:29:59 +0200 Marek Maly <
>>>>>>>>>>>>> marek.maly.ujep.cz>
>>>>>>>>>>>>> >> >> > napsal/-a:
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >> Hi,
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >> please try to run at lest 100K tests twice to verify
>>>>>>>>>>>>> exact
>>>>>>>>>>>>> >> >> >> reproducibility
>>>>>>>>>>>>> >> >> >> of the results on the given card. If you find in any
>>>>>>>>>>>>> mdin
>>>>>>>>>>>>> file
>>>>>>>>>>>>> >> ig=-1
>>>>>>>>>>>>> >> >> >> just
>>>>>>>>>>>>> >> >> >> delete it to ensure that you are using the identical
>>>>>>>>>>>>> random
>>>>>>>>>>>>> seed
>>>>>>>>>>>>> >> for
>>>>>>>>>>>>> >> >> >> both
>>>>>>>>>>>>> >> >> >> runs. You can eventually omit NUCLEOSOME test
>>>>>>>>>>>>> >> >> >> as it is too much time consuming.
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >> Driver 310.44 ?????
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >> As far as I know the proper support for titans is
>>>>>>>>>>>>> from
>>>>>>>>>>>>> version
>>>>>>>>>>>>> > 313.26
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >> see e.g. here :
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> http://www.geeks3d.com/****20130306/nvidia-releases-r313-****<http://www.geeks3d.com/**20130306/nvidia-releases-r313-**>
>>>>>>>>>>>>> 26-for-linux-with-gtx-titan-****support/<http://www.geeks3d.**
>>>>>>>>>>>>> com/20130306/nvidia-releases-**r313-26-for-linux-with-gtx-**
>>>>>>>>>>>>> titan-support/<http://www.geeks3d.com/20130306/nvidia-releases-r313-26-for-linux-with-gtx-titan-support/>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >> BTW: On my site downgrade to drv. 313.30 did not
>>>>>>>>>>>>> solved
>>>>>>>>>>>>> the
>>>>>>>>>>>>> >> >> situation, I
>>>>>>>>>>>>> >> >> >> will post
>>>>>>>>>>>>> >> >> >> my results soon here.
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >> M.
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >> Dne Fri, 31 May 2013 12:21:21 +0200 ET <
>>>>>>>>>>>>> sketchfoot.gmail.com>
>>>>>>>>>>>>> >> >> napsal/-a:
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >>> ps. I have another install of amber on another
>>>>>>>>>>>>> computer
>>>>>>>>>>>>> with
>>>>>>>>>>>>> a
>>>>>>>>>>>>> >> >> >>> different
>>>>>>>>>>>>> >> >> >>> Titan and different Driver Version: 310.44.
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>> In the interests of thrashing the proverbial horse,
>>>>>>>>>>>>> I'll
>>>>>>>>>>>>> run
>>>>>>>>>>>>> the
>>>>>>>>>>>>> >> >> >>> benchmark
>>>>>>>>>>>>> >> >> >>> for 50k steps. :P
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>> br,
>>>>>>>>>>>>> >> >> >>> g
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>> On 31 May 2013 11:17, ET <sketchfoot.gmail.com>
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>>> Hi, I just ran the Amber benchmark for the default
>>>>>>>>>>>>> (10000
>>>>>>>>>>>>> steps)
>>>>>>>>>>>>> >> >> on my
>>>>>>>>>>>>> >> >> >>>> Titan.
>>>>>>>>>>>>> >> >> >>>>
>>>>>>>>>>>>> >> >> >>>> Using sdiff -sB showed that the two runs were
>>>>>>>>>>>>> completely
>>>>>>>>>>>>> > identical.
>>>>>>>>>>>>> >> >> >>>> I've
>>>>>>>>>>>>> >> >> >>>> attached compressed files of the mdout & diff
>>>>>>>>>>>>> files.
>>>>>>>>>>>>> >> >> >>>>
>>>>>>>>>>>>> >> >> >>>> br,
>>>>>>>>>>>>> >> >> >>>> g
>>>>>>>>>>>>> >> >> >>>>
>>>>>>>>>>>>> >> >> >>>>
>>>>>>>>>>>>> >> >> >>>> On 30 May 2013 23:41, Marek Maly
>>>>>>>>>>>>> <marek.maly.ujep.cz
>>>>>>>>>>>>> >
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>> >> >> >>>>
>>>>>>>>>>>>> >> >> >>>>> OK, let's see. The eventual downclocking I see as
>>>>>>>>>>>>> the
>>>>>>>>>>>>> very
>>>>>>>>>>>>> last
>>>>>>>>>>>>> >> >> >>>>> possibility
>>>>>>>>>>>>> >> >> >>>>> (if I don't decide for RMAing). But now still
>>>>>>>>>>>>> some
>>>>>>>>>>>>> other
>>>>>>>>>>>>> >> >> experiments
>>>>>>>>>>>>> >> >> >>>>> are
>>>>>>>>>>>>> >> >> >>>>> available :))
>>>>>>>>>>>>> >> >> >>>>> I just started 100K tests under 313.30 driver.
>>>>>>>>>>>>> For
>>>>>>>>>>>>> today
>>>>>>>>>>>>> good
>>>>>>>>>>>>> >> >> night
>>>>>>>>>>>>> >> >> >>>>> ...
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> >> >> >>>>> M.
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> >> >> >>>>> Dne Fri, 31 May 2013 00:45:49 +0200 Scott Le
>>>>>>>>>>>>> Grand
>>>>>>>>>>>>> >> >> >>>>> <varelse2005.gmail.com
>>>>>>>>>>>>> >> >> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> napsal/-a:
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> >> >> >>>>> > It will be very interesting if this behavior
>>>>>>>>>>>>> persists
>>>>>>>>>>>>> after
>>>>>>>>>>>>> >> >> >>>>> downclocking.
>>>>>>>>>>>>> >> >> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> > But right now, Titan 0 *looks* hosed and Titan
>>>>>>>>>>>>> 1
>>>>>>>>>>>>> *looks*
>>>>>>>>>>>>> like
>>>>>>>>>>>>> > it
>>>>>>>>>>>>> >> >> >>>>> needs
>>>>>>>>>>>>> >> >> >>>>> > downclocking...
>>>>>>>>>>>>> >> >> >>>>> > On May 30, 2013 3:20 PM, "Marek Maly"
>>>>>>>>>>>>> <marek.maly.ujep.cz>
>>>>>>>>>>>>> >> >> wrote:
>>>>>>>>>>>>> >> >> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >> Hi all,
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> here are my results from the 500K steps 2 x
>>>>>>>>>>>>> repeated
>>>>>>>>>>>>> > benchmarks
>>>>>>>>>>>>> >> >> >>>>> >> under 319.23 driver and still Cuda 5.0 (see
>>>>>>>>>>>>> the
>>>>>>>>>>>>> attached
>>>>>>>>>>>>> >> table
>>>>>>>>>>>>> >> >> ).
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> It is hard to say if the results are better or
>>>>>>>>>>>>> worse
>>>>>>>>>>>>> than
>>>>>>>>>>>>> in
>>>>>>>>>>>>> > my
>>>>>>>>>>>>> >> >> >>>>> >> previous 100K test under driver 319.17.
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> While results from Cellulose test were
>>>>>>>>>>>>> improved
>>>>>>>>>>>>> and
>>>>>>>>>>>>> the
>>>>>>>>>>>>> > TITAN_1
>>>>>>>>>>>>> >> >> >>>>> card
>>>>>>>>>>>>> >> >> >>>>> >> even
>>>>>>>>>>>>> >> >> >>>>> >> successfully finished all 500K steps moreover
>>>>>>>>>>>>> with
>>>>>>>>>>>>> exactly
>>>>>>>>>>>>> >> the
>>>>>>>>>>>>> >> >> >>>>> same
>>>>>>>>>>>>> >> >> >>>>> >> final
>>>>>>>>>>>>> >> >> >>>>> >> energy !
>>>>>>>>>>>>> >> >> >>>>> >> (TITAN_0 at least finished more than 100K
>>>>>>>>>>>>> steps
>>>>>>>>>>>>> and in
>>>>>>>>>>>>> >> RUN_01
>>>>>>>>>>>>> >> >> even
>>>>>>>>>>>>> >> >> >>>>> more
>>>>>>>>>>>>> >> >> >>>>> >> than 400K steps)
>>>>>>>>>>>>> >> >> >>>>> >> In JAC_NPT test no GPU was able to finish at
>>>>>>>>>>>>> least
>>>>>>>>>>>>> 100K
>>>>>>>>>>>>> >> steps
>>>>>>>>>>>>> >> >> and
>>>>>>>>>>>>> >> >> >>>>> the
>>>>>>>>>>>>> >> >> >>>>> >> results from JAC_NVE
>>>>>>>>>>>>> >> >> >>>>> >> test are also not too much convincing.
>>>>>>>>>>>>> FACTOR_IX_NVE
>>>>>>>>>>>>> and
>>>>>>>>>>>>> >> >> >>>>> FACTOR_IX_NPT
>>>>>>>>>>>>> >> >> >>>>> >> were successfully
>>>>>>>>>>>>> >> >> >>>>> >> finished with 100% reproducibility in
>>>>>>>>>>>>> FACTOR_IX_NPT
>>>>>>>>>>>>> case
>>>>>>>>>>>>> >> (on
>>>>>>>>>>>>> >> >> both
>>>>>>>>>>>>> >> >> >>>>> >> cards)
>>>>>>>>>>>>> >> >> >>>>> >> and almost
>>>>>>>>>>>>> >> >> >>>>> >> 100% reproducibility in case of FACTOR_IX_NVE
>>>>>>>>>>>>> (again
>>>>>>>>>>>>> 100%
>>>>>>>>>>>>> in
>>>>>>>>>>>>> >> >> case
>>>>>>>>>>>>> >> >> >>>>> of
>>>>>>>>>>>>> >> >> >>>>> >> TITAN_1). TRPCAGE, MYOGLOBIN
>>>>>>>>>>>>> >> >> >>>>> >> again finished without any problem with 100%
>>>>>>>>>>>>> >> reproducibility.
>>>>>>>>>>>>> >> >> >>>>> NUCLEOSOME
>>>>>>>>>>>>> >> >> >>>>> >> test was not done
>>>>>>>>>>>>> >> >> >>>>> >> this time due to high time requirements. If
>>>>>>>>>>>>> you
>>>>>>>>>>>>> find
>>>>>>>>>>>>> in
>>>>>>>>>>>>> the
>>>>>>>>>>>>> >> >> table
>>>>>>>>>>>>> >> >> >>>>> >> positive
>>>>>>>>>>>>> >> >> >>>>> >> number finishing with
>>>>>>>>>>>>> >> >> >>>>> >> K (which means "thousands") it means the last
>>>>>>>>>>>>> number
>>>>>>>>>>>>> of
>>>>>>>>>>>>> step
>>>>>>>>>>>>> >> >> >>>>> written in
>>>>>>>>>>>>> >> >> >>>>> >> mdout before crash.
>>>>>>>>>>>>> >> >> >>>>> >> Below are all the 3 types of detected errs
>>>>>>>>>>>>> with
>>>>>>>>>>>>> relevant
>>>>>>>>>>>>> >> >> >>>>> systems/rounds
>>>>>>>>>>>>> >> >> >>>>> >> where the given err
>>>>>>>>>>>>> >> >> >>>>> >> appeared.
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> Now I will try just 100K tests under ETs
>>>>>>>>>>>>> favourite
>>>>>>>>>>>>> driver
>>>>>>>>>>>>> >> >> version
>>>>>>>>>>>>> >> >> >>>>> 313.30
>>>>>>>>>>>>> >> >> >>>>> >> :)) and then
>>>>>>>>>>>>> >> >> >>>>> >> I will eventually try to experiment with cuda
>>>>>>>>>>>>> 5.5
>>>>>>>>>>>>> which
>>>>>>>>>>>>> I
>>>>>>>>>>>>> >> >> already
>>>>>>>>>>>>> >> >> >>>>> >> downloaded from the
>>>>>>>>>>>>> >> >> >>>>> >> cuda zone ( I had to become cuda developer for
>>>>>>>>>>>>> this
>>>>>>>>>>>>> :))
>>>>>>>>>>>>> )
>>>>>>>>>>>>> >> BTW
>>>>>>>>>>>>> >> >> ET
>>>>>>>>>>>>> >> >> >>>>> thanks
>>>>>>>>>>>>> >> >> >>>>> >> for the frequency info !
>>>>>>>>>>>>> >> >> >>>>> >> and I am still ( perhaps not alone :)) ) very
>>>>>>>>>>>>> curious
>>>>>>>>>>>>> about
>>>>>>>>>>>>> >> >> your 2
>>>>>>>>>>>>> >> >> >>>>> x
>>>>>>>>>>>>> >> >> >>>>> >> repeated Amber benchmark tests with
>>>>>>>>>>>>> superclocked
>>>>>>>>>>>>> Titan.
>>>>>>>>>>>>> >> Indeed
>>>>>>>>>>>>> >> >> >>>>> that
>>>>>>>>>>>>> >> >> >>>>> I
>>>>>>>>>>>>> >> >> >>>>> am
>>>>>>>>>>>>> >> >> >>>>> >> very curious also about that Ross "hot" patch.
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> M.
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> ERRORS DETECTED DURING THE 500K steps tests
>>>>>>>>>>>>> with
>>>>>>>>>>>>> driver
>>>>>>>>>>>>> >> 319.23
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> #1 ERR writtent in mdout:
>>>>>>>>>>>>> >> >> >>>>> >> ------
>>>>>>>>>>>>> >> >> >>>>> >> | ERROR: max pairlist cutoff must be less
>>>>>>>>>>>>> than
>>>>>>>>>>>>> unit
>>>>>>>>>>>>> cell
>>>>>>>>>>>>> >> max
>>>>>>>>>>>>> >> >> >>>>> sphere
>>>>>>>>>>>>> >> >> >>>>> >> radius!
>>>>>>>>>>>>> >> >> >>>>> >> ------
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_1 JAC_NPT (at least 5000 steps
>>>>>>>>>>>>> successfully
>>>>>>>>>>>>> > done
>>>>>>>>>>>>> >> >> >>>>> before
>>>>>>>>>>>>> >> >> >>>>> >> crash)
>>>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 JAC_NPT (at least 8000 steps
>>>>>>>>>>>>> successfully
>>>>>>>>>>>>> > done
>>>>>>>>>>>>> >> >> >>>>> before
>>>>>>>>>>>>> >> >> >>>>> >> crash)
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> #2 no ERR writtent in mdout, ERR written in
>>>>>>>>>>>>> standard
>>>>>>>>>>>>> output
>>>>>>>>>>>>> >> >> >>>>> (nohup.out)
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> ----
>>>>>>>>>>>>> >> >> >>>>> >> Error: unspecified launch failure launching
>>>>>>>>>>>>> kernel
>>>>>>>>>>>>> >> kNLSkinTest
>>>>>>>>>>>>> >> >> >>>>> >> cudaFree GpuBuffer::Deallocate failed
>>>>>>>>>>>>> unspecified
>>>>>>>>>>>>> launch
>>>>>>>>>>>>> >> >> failure
>>>>>>>>>>>>> >> >> >>>>> >> ----
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_1 CELLULOSE_NVE (at least 437
>>>>>>>>>>>>> 000
>>>>>>>>>>>>> steps
>>>>>>>>>>>>> >> >> successfully
>>>>>>>>>>>>> >> >> >>>>> done
>>>>>>>>>>>>> >> >> >>>>> >> before crash)
>>>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 JAC_NVE (at least 162 000
>>>>>>>>>>>>> steps
>>>>>>>>>>>>> >> successfully
>>>>>>>>>>>>> >> >> done
>>>>>>>>>>>>> >> >> >>>>> >> before
>>>>>>>>>>>>> >> >> >>>>> >> crash)
>>>>>>>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 CELLULOSE_NVE (at least 117
>>>>>>>>>>>>> 000
>>>>>>>>>>>>> steps
>>>>>>>>>>>>> >> >> successfully
>>>>>>>>>>>>> >> >> >>>>> done
>>>>>>>>>>>>> >> >> >>>>> >> before crash)
>>>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_1 JAC_NVE (at least 119 000
>>>>>>>>>>>>> steps
>>>>>>>>>>>>> >> successfully
>>>>>>>>>>>>> >> >> done
>>>>>>>>>>>>> >> >> >>>>> >> before
>>>>>>>>>>>>> >> >> >>>>> >> crash)
>>>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_2 JAC_NVE (at least 43 000
>>>>>>>>>>>>> steps
>>>>>>>>>>>>> successfully
>>>>>>>>>>>>> >> >> done
>>>>>>>>>>>>> >> >> >>>>> before
>>>>>>>>>>>>> >> >> >>>>> >> crash)
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> #3 no ERR writtent in mdout, ERR written in
>>>>>>>>>>>>> standard
>>>>>>>>>>>>> output
>>>>>>>>>>>>> >> >> >>>>> (nohup.out)
>>>>>>>>>>>>> >> >> >>>>> >> ----
>>>>>>>>>>>>> >> >> >>>>> >> cudaMemcpy GpuBuffer::Download failed
>>>>>>>>>>>>> unspecified
>>>>>>>>>>>>> launch
>>>>>>>>>>>>> >> >> failure
>>>>>>>>>>>>> >> >> >>>>> >> ----
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_1 JAC_NPT (at least 77 000
>>>>>>>>>>>>> steps
>>>>>>>>>>>>> successfully
>>>>>>>>>>>>> >> >> done
>>>>>>>>>>>>> >> >> >>>>> before
>>>>>>>>>>>>> >> >> >>>>> >> crash)
>>>>>>>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_2 JAC_NPT (at least 58 000
>>>>>>>>>>>>> steps
>>>>>>>>>>>>> successfully
>>>>>>>>>>>>> >> >> done
>>>>>>>>>>>>> >> >> >>>>> before
>>>>>>>>>>>>> >> >> >>>>> >> crash)
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> Dne Thu, 30 May 2013 21:27:17 +0200 Scott Le
>>>>>>>>>>>>> Grand
>>>>>>>>>>>>> >> >> >>>>> >> <varelse2005.gmail.com>
>>>>>>>>>>>>> >> >> >>>>> >> napsal/-a:
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> Oops meant to send that to Jason...
>>>>>>>>>>>>> >> >> >>>>> >>>
>>>>>>>>>>>>> >> >> >>>>> >>> Anyway, before we all panic, we need to get
>>>>>>>>>>>>> K20's
>>>>>>>>>>>>> behavior
>>>>>>>>>>>>> >> >> >>>>> analyzed
>>>>>>>>>>>>> >> >> >>>>> >>> here.
>>>>>>>>>>>>> >> >> >>>>> >>> If it's deterministic, this truly is a
>>>>>>>>>>>>> hardware
>>>>>>>>>>>>> issue. If
>>>>>>>>>>>>> >> >> not,
>>>>>>>>>>>>> >> >> >>>>> then
>>>>>>>>>>>>> >> >> >>>>> it
>>>>>>>>>>>>> >> >> >>>>> >>> gets interesting because 680 is
>>>>>>>>>>>>> deterministic as
>>>>>>>>>>>>> far
>>>>>>>>>>>>> as I
>>>>>>>>>>>>> >> can
>>>>>>>>>>>>> >> >> >>>>> tell...
>>>>>>>>>>>>> >> >> >>>>> >>> On May 30, 2013 12:24 PM, "Scott Le Grand"
>>>>>>>>>>>>> >> >> >>>>> <varelse2005.gmail.com>
>>>>>>>>>>>>> >> >> >>>>> >>> wrote:
>>>>>>>>>>>>> >> >> >>>>> >>>
>>>>>>>>>>>>> >> >> >>>>> >>> If the errors are not deterministically
>>>>>>>>>>>>> triggered,
>>>>>>>>>>>>> they
>>>>>>>>>>>>> >> >> probably
>>>>>>>>>>>>> >> >> >>>>> >>> won't be
>>>>>>>>>>>>> >> >> >>>>> >>>> fixed by the patch, alas...
>>>>>>>>>>>>> >> >> >>>>> >>>> On May 30, 2013 12:15 PM, "Jason Swails"
>>>>>>>>>>>>> >> >> >>>>> <jason.swails.gmail.com>
>>>>>>>>>>>>> >> >> >>>>> >>>> wrote:
>>>>>>>>>>>>> >> >> >>>>> >>>>
>>>>>>>>>>>>> >> >> >>>>> >>>> Just a reminder to everyone based on what
>>>>>>>>>>>>> Ross
>>>>>>>>>>>>> said:
>>>>>>>>>>>>> >> there
>>>>>>>>>>>>> >> >> is a
>>>>>>>>>>>>> >> >> >>>>> >>>> pending
>>>>>>>>>>>>> >> >> >>>>> >>>>> patch to pmemd.cuda that will be coming out
>>>>>>>>>>>>> shortly
>>>>>>>>>>>>> >> (maybe
>>>>>>>>>>>>> >> >> even
>>>>>>>>>>>>> >> >> >>>>> >>>>> within
>>>>>>>>>>>>> >> >> >>>>> >>>>> hours). It's entirely possible that
>>>>>>>>>>>>> several
>>>>>>>>>>>>> of
>>>>>>>>>>>>> these
>>>>>>>>>>>>> > errors
>>>>>>>>>>>>> >> >> >>>>> are
>>>>>>>>>>>>> >> >> >>>>> >>>>> fixed
>>>>>>>>>>>>> >> >> >>>>> >>>>> by
>>>>>>>>>>>>> >> >> >>>>> >>>>> this patch.
>>>>>>>>>>>>> >> >> >>>>> >>>>>
>>>>>>>>>>>>> >> >> >>>>> >>>>> All the best,
>>>>>>>>>>>>> >> >> >>>>> >>>>> Jason
>>>>>>>>>>>>> >> >> >>>>> >>>>>
>>>>>>>>>>>>> >> >> >>>>> >>>>>
>>>>>>>>>>>>> >> >> >>>>> >>>>> On Thu, May 30, 2013 at 2:46 PM, filip
>>>>>>>>>>>>> fratev
>>>>>>>>>>>>> <
>>>>>>>>>>>>> >> >> >>>>> filipfratev.yahoo.com>
>>>>>>>>>>>>> >> >> >>>>> >>>>> wrote:
>>>>>>>>>>>>> >> >> >>>>> >>>>>
>>>>>>>>>>>>> >> >> >>>>> >>>>> > I have observed the same crashes from
>>>>>>>>>>>>> time
>>>>>>>>>>>>> to
>>>>>>>>>>>>> time. I
>>>>>>>>>>>>> > will
>>>>>>>>>>>>> >> >> >>>>> run
>>>>>>>>>>>>> >> >> >>>>> >>>>> cellulose
>>>>>>>>>>>>> >> >> >>>>> >>>>> > nve for 100k and will past results here.
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > All the best,
>>>>>>>>>>>>> >> >> >>>>> >>>>> > Filip
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > ______________________________******__
>>>>>>>>>>>>> >> >> >>>>> >>>>> > From: Scott Le Grand
>>>>>>>>>>>>> <varelse2005.gmail.com>
>>>>>>>>>>>>> >> >> >>>>> >>>>> > To: AMBER Mailing List
>>>>>>>>>>>>> <amber.ambermd.org>
>>>>>>>>>>>>> >> >> >>>>> >>>>> > Sent: Thursday, May 30, 2013 9:01 PM
>>>>>>>>>>>>> >> >> >>>>> >>>>> > Subject: Re: [AMBER] experiences with
>>>>>>>>>>>>> EVGA
>>>>>>>>>>>>> GTX
>>>>>>>>>>>>> TITAN
>>>>>>>>>>>>> >> >> >>>>> Superclocked
>>>>>>>>>>>>> >> >> >>>>> -
>>>>>>>>>>>>> >> >> >>>>> >>>>> > memtestG80 - UNDERclocking in Linux ?
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > Run cellulose nve for 100k iterations
>>>>>>>>>>>>> twice
>>>>>>>>>>>>> . If
>>>>>>>>>>>>> the
>>>>>>>>>>>>> >> >> final
>>>>>>>>>>>>> >> >> >>>>> >>>>> energies
>>>>>>>>>>>>> >> >> >>>>> >>>>> don't
>>>>>>>>>>>>> >> >> >>>>> >>>>> > match, you have a hardware issue. No
>>>>>>>>>>>>> need
>>>>>>>>>>>>> to
>>>>>>>>>>>>> play
>>>>>>>>>>>>> with
>>>>>>>>>>>>> >> >> ntpr
>>>>>>>>>>>>> >> >> >>>>> or
>>>>>>>>>>>>> >> >> >>>>> any
>>>>>>>>>>>>> >> >> >>>>> >>>>> other
>>>>>>>>>>>>> >> >> >>>>> >>>>> > variable.
>>>>>>>>>>>>> >> >> >>>>> >>>>> > On May 30, 2013 10:58 AM,
>>>>>>>>>>>>> <pavel.banas.upol.cz>
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Dear all,
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > I would also like to share one of my
>>>>>>>>>>>>> experience
>>>>>>>>>>>>> with
>>>>>>>>>>>>> >> >> titan
>>>>>>>>>>>>> >> >> >>>>> >>>>> cards. We
>>>>>>>>>>>>> >> >> >>>>> >>>>> have
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > one gtx titan card and with one system
>>>>>>>>>>>>> (~55k
>>>>>>>>>>>>> atoms,
>>>>>>>>>>>>> > NVT,
>>>>>>>>>>>>> >> >> >>>>> >>>>> RNA+waters)
>>>>>>>>>>>>> >> >> >>>>> >>>>> we
>>>>>>>>>>>>> >> >> >>>>> >>>>> > run
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > into same troubles you are describing.
>>>>>>>>>>>>> I
>>>>>>>>>>>>> was
>>>>>>>>>>>>> also
>>>>>>>>>>>>> >> >> playing
>>>>>>>>>>>>> >> >> >>>>> with
>>>>>>>>>>>>> >> >> >>>>> >>>>> ntpr
>>>>>>>>>>>>> >> >> >>>>> >>>>> to
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > figure out what is going on, step by
>>>>>>>>>>>>> step.
>>>>>>>>>>>>> I
>>>>>>>>>>>>> >> understand
>>>>>>>>>>>>> >> >> >>>>> that
>>>>>>>>>>>>> >> >> >>>>> the
>>>>>>>>>>>>> >> >> >>>>> >>>>> code
>>>>>>>>>>>>> >> >> >>>>> >>>>> is
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > using different routines for
>>>>>>>>>>>>> calculation
>>>>>>>>>>>>> >> >> energies+forces or
>>>>>>>>>>>>> >> >> >>>>> only
>>>>>>>>>>>>> >> >> >>>>> >>>>> forces.
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > The
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > simulations of other systems are
>>>>>>>>>>>>> perfectly
>>>>>>>>>>>>> stable,
>>>>>>>>>>>>> >> >> running
>>>>>>>>>>>>> >> >> >>>>> for
>>>>>>>>>>>>> >> >> >>>>> >>>>> days
>>>>>>>>>>>>> >> >> >>>>> >>>>> and
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > weeks. Only that particular system
>>>>>>>>>>>>> systematically
>>>>>>>>>>>>> >> ends
>>>>>>>>>>>>> >> >> up
>>>>>>>>>>>>> >> >> >>>>> with
>>>>>>>>>>>>> >> >> >>>>> >>>>> this
>>>>>>>>>>>>> >> >> >>>>> >>>>> > error.
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > However, there was one interesting
>>>>>>>>>>>>> issue.
>>>>>>>>>>>>> When
>>>>>>>>>>>>> I
>>>>>>>>>>>>> set
>>>>>>>>>>>>> >> >> >>>>> ntpr=1,
>>>>>>>>>>>>> >> >> >>>>> the
>>>>>>>>>>>>> >> >> >>>>> >>>>> error
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > vanished (systematically in multiple
>>>>>>>>>>>>> runs)
>>>>>>>>>>>>> and
>>>>>>>>>>>>> the
>>>>>>>>>>>>> >> >> >>>>> simulation
>>>>>>>>>>>>> >> >> >>>>> was
>>>>>>>>>>>>> >> >> >>>>> >>>>> able to
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > run for more than millions of steps (I
>>>>>>>>>>>>> was
>>>>>>>>>>>>> not
>>>>>>>>>>>>> let
>>>>>>>>>>>>> it
>>>>>>>>>>>>> >> >> >>>>> running
>>>>>>>>>>>>> >> >> >>>>> for
>>>>>>>>>>>>> >> >> >>>>> >>>>> weeks
>>>>>>>>>>>>> >> >> >>>>> >>>>> > as
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > in the meantime I shifted that
>>>>>>>>>>>>> simulation
>>>>>>>>>>>>> to
>>>>>>>>>>>>> other
>>>>>>>>>>>>> >> card
>>>>>>>>>>>>> >> >> -
>>>>>>>>>>>>> >> >> >>>>> need
>>>>>>>>>>>>> >> >> >>>>> >>>>> data,
>>>>>>>>>>>>> >> >> >>>>> >>>>> not
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > testing). All other setting of ntpr
>>>>>>>>>>>>> failed.
>>>>>>>>>>>>> As
>>>>>>>>>>>>> I
>>>>>>>>>>>>> read
>>>>>>>>>>>>> >> >> this
>>>>>>>>>>>>> >> >> >>>>> >>>>> discussion, I
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > tried to set ene_avg_sampling=1 with
>>>>>>>>>>>>> some
>>>>>>>>>>>>> high
>>>>>>>>>>>>> value
>>>>>>>>>>>>> >> of
>>>>>>>>>>>>> >> >> >>>>> ntpr
>>>>>>>>>>>>> >> >> >>>>> (I
>>>>>>>>>>>>> >> >> >>>>> >>>>> expected
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > that this will shift the code to
>>>>>>>>>>>>> permanently
>>>>>>>>>>>>> use
>>>>>>>>>>>>> the
>>>>>>>>>>>>> >> >> >>>>> >>>>> force+energies
>>>>>>>>>>>>> >> >> >>>>> >>>>> part
>>>>>>>>>>>>> >> >> >>>>> >>>>> > of
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > the code, similarly to ntpr=1), but the
>>>>>>>>>>>>> error
>>>>>>>>>>>>> >> occurred
>>>>>>>>>>>>> >> >> >>>>> again.
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > I know it is not very conclusive for
>>>>>>>>>>>>> finding
>>>>>>>>>>>>> out
>>>>>>>>>>>>> what
>>>>>>>>>>>>> > is
>>>>>>>>>>>>> >> >> >>>>> >>>>> happening,
>>>>>>>>>>>>> >> >> >>>>> >>>>> at
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > least
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > not for me. Do you have any idea, why
>>>>>>>>>>>>> ntpr=1
>>>>>>>>>>>>> might
>>>>>>>>>>>>> > help?
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > best regards,
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Pavel
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > --
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Pavel Banáš
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > pavel.banas.upol.cz
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Department of Physical Chemistry,
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Palacky University Olomouc
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Czech Republic
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > ---------- Původní zpráva ----------
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Od: Jason Swails
>>>>>>>>>>>>> <jason.swails.gmail.com>
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Datum: 29. 5. 2013
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Předmět: Re: [AMBER] experiences with
>>>>>>>>>>>>> EVGA
>>>>>>>>>>>>> GTX
>>>>>>>>>>>>> TITAN
>>>>>>>>>>>>> >> >> >>>>> >>>>> Superclocked -
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > memtestG
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > 80 - UNDERclocking in Linux ?
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > "I'll answer a little bit:
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > NTPR=10 Etot after 2000 steps
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443256.6711
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443256.6711
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > > NTPR=200 Etot after 2000 steps
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443261.0705
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > > -443261.0705
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > > Any idea why energies should depend
>>>>>>>>>>>>> on
>>>>>>>>>>>>> frequency
>>>>>>>>>>>>> of
>>>>>>>>>>>>> >> >> >>>>> energy
>>>>>>>>>>>>> >> >> >>>>> >>>>> records
>>>>>>>>>>>>> >> >> >>>>> >>>>> > (NTPR)
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > ?
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > It is a subtle point, but the answer is
>>>>>>>>>>>>> 'different
>>>>>>>>>>>>> >> code
>>>>>>>>>>>>> >> >> >>>>> paths.'
>>>>>>>>>>>>> >> >> >>>>> >>>>> In
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > general, it is NEVER necessary to
>>>>>>>>>>>>> compute
>>>>>>>>>>>>> the
>>>>>>>>>>>>> actual
>>>>>>>>>>>>> >> >> energy
>>>>>>>>>>>>> >> >> >>>>> of a
>>>>>>>>>>>>> >> >> >>>>> >>>>> molecule
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > during the course of standard molecular
>>>>>>>>>>>>> dynamics
>>>>>>>>>>>>> (by
>>>>>>>>>>>>> >> >> >>>>> analogy, it
>>>>>>>>>>>>> >> >> >>>>> >>>>> is
>>>>>>>>>>>>> >> >> >>>>> >>>>> NEVER
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > necessary to compute atomic forces
>>>>>>>>>>>>> during
>>>>>>>>>>>>> the
>>>>>>>>>>>>> course
>>>>>>>>>>>>> >> of
>>>>>>>>>>>>> >> >> >>>>> random
>>>>>>>>>>>>> >> >> >>>>> >>>>> Monte
>>>>>>>>>>>>> >> >> >>>>> >>>>> > Carlo
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > sampling).
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > For performance's sake, then,
>>>>>>>>>>>>> pmemd.cuda
>>>>>>>>>>>>> computes
>>>>>>>>>>>>> >> only
>>>>>>>>>>>>> >> >> the
>>>>>>>>>>>>> >> >> >>>>> force
>>>>>>>>>>>>> >> >> >>>>> >>>>> when
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > energies are not requested, leading to
>>>>>>>>>>>>> a
>>>>>>>>>>>>> different
>>>>>>>>>>>>> >> >> order of
>>>>>>>>>>>>> >> >> >>>>> >>>>> operations
>>>>>>>>>>>>> >> >> >>>>> >>>>> > for
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > those runs. This difference ultimately
>>>>>>>>>>>>> causes
>>>>>>>>>>>>> >> >> divergence.
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > To test this, try setting the variable
>>>>>>>>>>>>> >> >> ene_avg_sampling=10
>>>>>>>>>>>>> >> >> >>>>> in
>>>>>>>>>>>>> >> >> >>>>> the
>>>>>>>>>>>>> >> >> >>>>> >>>>> &cntrl
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > section. This will force pmemd.cuda to
>>>>>>>>>>>>> compute
>>>>>>>>>>>>> >> energies
>>>>>>>>>>>>> >> >> >>>>> every 10
>>>>>>>>>>>>> >> >> >>>>> >>>>> steps
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > (for energy averaging), which will in
>>>>>>>>>>>>> turn
>>>>>>>>>>>>> make
>>>>>>>>>>>>> the
>>>>>>>>>>>>> >> >> >>>>> followed
>>>>>>>>>>>>> >> >> >>>>> code
>>>>>>>>>>>>> >> >> >>>>> >>>>> path
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > identical for any multiple-of-10 value
>>>>>>>>>>>>> of
>>>>>>>>>>>>> ntpr.
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > --
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Jason M. Swails
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Quantum Theory Project,
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > University of Florida
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > Ph.D. Candidate
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > 352-392-4032
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > ______________________________****
>>>>>>>>>>>>> **_________________
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER mailing list
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER.ambermd.org
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
>>>>>>>>>>>>> mailman/listinfo/amber<http://**lists.ambermd.org/**mailman/**
>>>>>>>>>>>>> listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> <
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> "
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > ______________________________****
>>>>>>>>>>>>> **_________________
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER mailing list
>>>>>>>>>>>>> >> >> >>>>> >>>>> > > AMBER.ambermd.org
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
>>>>>>>>>>>>> mailman/listinfo/amber<http://**lists.ambermd.org/**mailman/**
>>>>>>>>>>>>> listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> <
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > ______________________________****
>>>>>>>>>>>>> **_________________
>>>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER mailing list
>>>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER.ambermd.org
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
>>>>>>>>>>>>> mailman/listinfo/amber<http://**lists.ambermd.org/**mailman/**
>>>>>>>>>>>>> listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> <
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> > ______________________________****
>>>>>>>>>>>>> **_________________
>>>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER mailing list
>>>>>>>>>>>>> >> >> >>>>> >>>>> > AMBER.ambermd.org
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
>>>>>>>>>>>>> mailman/listinfo/amber<http://**lists.ambermd.org/**mailman/**
>>>>>>>>>>>>> listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> <
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>>
>>>>>>>>>>>>> >> >> >>>>> >>>>>
>>>>>>>>>>>>> >> >> >>>>> >>>>>
>>>>>>>>>>>>> >> >> >>>>> >>>>> --
>>>>>>>>>>>>> >> >> >>>>> >>>>> Jason M. Swails
>>>>>>>>>>>>> >> >> >>>>> >>>>> Quantum Theory Project,
>>>>>>>>>>>>> >> >> >>>>> >>>>> University of Florida
>>>>>>>>>>>>> >> >> >>>>> >>>>> Ph.D. Candidate
>>>>>>>>>>>>> >> >> >>>>> >>>>> 352-392-4032
>>>>>>>>>>>>> >> >> >>>>> >>>>> ______________________________****
>>>>>>>>>>>>> **_________________
>>>>>>>>>>>>> >> >> >>>>> >>>>> AMBER mailing list
>>>>>>>>>>>>> >> >> >>>>> >>>>> AMBER.ambermd.org
>>>>>>>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
>>>>>>>>>>>>> mailman/listinfo/amber<http://**lists.ambermd.org/**mailman/**
>>>>>>>>>>>>> listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> <
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>>>
>>>>>>>>>>>>> >> >> >>>>> >>>>>
>>>>>>>>>>>>> >> >> >>>>> >>>> ______________________________****
>>>>>>>>>>>>> **_________________
>>>>>>>>>>>>> >> >> >>>>> >>> AMBER mailing list
>>>>>>>>>>>>> >> >> >>>>> >>> AMBER.ambermd.org
>>>>>>>>>>>>> >> >> >>>>> >>>
>>>>>>>>>>>>> http://lists.ambermd.org/******mailman/listinfo/amber<http://lists.ambermd.org/****mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/**mailman/**listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> <
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>>
>>>>>>>>>>>>> >> >> >>>>> >>> __________ Informace od ESET NOD32 Antivirus,
>>>>>>>>>>>>> verze
>>>>>>>>>>>>> >> databaze
>>>>>>>>>>>>> >> >> 8394
>>>>>>>>>>>>> >> >> >>>>> >>> (20130530) __________
>>>>>>>>>>>>> >> >> >>>>> >>>
>>>>>>>>>>>>> >> >> >>>>> >>> Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>>>>>>>> >> >> >>>>> >>>
>>>>>>>>>>>>> >> >> >>>>> >>> http://www.eset.cz
>>>>>>>>>>>>> >> >> >>>>> >>>
>>>>>>>>>>>>> >> >> >>>>> >>>
>>>>>>>>>>>>> >> >> >>>>> >>>
>>>>>>>>>>>>> >> >> >>>>> >>>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >> --
>>>>>>>>>>>>> >> >> >>>>> >> Tato zpráva byla vytvořena převratným
>>>>>>>>>>>>> poštovním
>>>>>>>>>>>>> klientem
>>>>>>>>>>>>> > Opery:
>>>>>>>>>>>>> >> >> >>>>> >> http://www.opera.com/mail/
>>>>>>>>>>>>> >> >> >>>>> >> ______________________________**
>>>>>>>>>>>>> **_________________
>>>>>>>>>>>>> >> >> >>>>> >> AMBER mailing list
>>>>>>>>>>>>> >> >> >>>>> >> AMBER.ambermd.org
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> >>
>>>>>>>>>>>>> >> >> >>>>> > ______________________________**
>>>>>>>>>>>>> **_________________
>>>>>>>>>>>>> >> >> >>>>> > AMBER mailing list
>>>>>>>>>>>>> >> >> >>>>> > AMBER.ambermd.org
>>>>>>>>>>>>> >> >> >>>>> >
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> > __________ Informace od ESET NOD32 Antivirus,
>>>>>>>>>>>>> verze
>>>>>>>>>>>>> databaze
>>>>>>>>>>>>> >> >> 8394
>>>>>>>>>>>>> >> >> >>>>> > (20130530) __________
>>>>>>>>>>>>> >> >> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>>>>>>>> >> >> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> > http://www.eset.cz
>>>>>>>>>>>>> >> >> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >
>>>>>>>>>>>>> >> >> >>>>> >
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> >> >> >>>>> --
>>>>>>>>>>>>> >> >> >>>>> Tato zpráva byla vytvořena převratným poštovním
>>>>>>>>>>>>> klientem
>>>>>>>>>>>>> Opery:
>>>>>>>>>>>>> >> >> >>>>> http://www.opera.com/mail/
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> ______________________________****_________________
>>>>>>>>>>>>> >> >> >>>>> AMBER mailing list
>>>>>>>>>>>>> >> >> >>>>> AMBER.ambermd.org
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>>>>
>>>>>>>>>>>>> >> >> >>>>
>>>>>>>>>>>>> >> >> >>>>
>>>>>>>>>>>>> >> >> >>> ______________________________****_________________
>>>>>>>>>>>>> >> >> >>> AMBER mailing list
>>>>>>>>>>>>> >> >> >>> AMBER.ambermd.org
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>> __________ Informace od ESET NOD32 Antivirus, verze
>>>>>>>>>>>>> databaze
>>>>>>>>>>>>> 8395
>>>>>>>>>>>>> >> >> >>> (20130531) __________
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>> Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>> http://www.eset.cz
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>>
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >>
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >> >
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> --
>>>>>>>>>>>>> >> >> Tato zpráva byla vytvořena převratným poštovním
>>>>>>>>>>>>> klientem
>>>>>>>>>>>>> Opery:
>>>>>>>>>>>>> >> >> http://www.opera.com/mail/
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >> ______________________________****_________________
>>>>>>>>>>>>> >> >> AMBER mailing list
>>>>>>>>>>>>> >> >> AMBER.ambermd.org
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> >>
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> > __________ Informace od ESET NOD32 Antivirus, verze
>>>>>>>>>>>>> databaze
>>>>>>>>>>>>> 8397
>>>>>>>>>>>>> >> > (20130531) __________
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> > Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz - poskozeny archiv
>>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>>>>>>>> >> GB_out_plus_diff_Files.tar
>>>>>>>>>>>>> >> > - poskozeny archiv
>>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
>>>>>>>>>>>>> GB_out_plus_diff_Files.tar.gz -
>>>>>>>>>>>>> >> > poskozeny archiv
>>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
>>>>>>>>>>>>> GB_out_plus_diff_Files.tar.gz >
>>>>>>>>>>>>> >> GZIP
>>>>>>>>>>>>> >> > > GB_out_plus_diff_Files.tar - poskozeny archiv
>>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
>>>>>>>>>>>>> GB_out_plus_diff_Files.tar.gz >
>>>>>>>>>>>>> >> GZIP
>>>>>>>>>>>>> >> > > GB_out_plus_diff_Files.tar > TAR >
>>>>>>>>>>>>> GB_nucleosome-sim3.mdout-full -
>>>>>>>>>>>>> >> > vyskytl se problem pri cteni archivu
>>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz - poskozeny archiv
>>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar - poskozeny archiv
>>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar > TAR >
>>>>>>>>>>>>> PME_out_plus_diff_Files.tar.gz -
>>>>>>>>>>>>> >> > poskozeny archiv
>>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar > TAR >
>>>>>>>>>>>>> PME_out_plus_diff_Files.tar.gz >
>>>>>>>>>>>>> >> > GZIP > PME_out_plus_diff_Files.tar - poskozeny archiv
>>>>>>>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>>>>>>>> >> > PME_out_plus_diff_Filestar > TAR >
>>>>>>>>>>>>> PME_out_plus_diff_Files.tar.gz
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >> GZIP
>>>>>>>>>>>>> >> > > PME_out_plus_diff_Files.tar > TAR >
>>>>>>>>>>>>> >> > PME_JAC_production_NPT-sim3.****mdout-full - vyskytl se
>>>>>>>>>>>>> problem
>>>>>>>>>>>>> pri
>>>>>>>>>>>>> cteni
>>>>>>>>>>>>> >> > archivu
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >> > http://www.eset.cz
>>>>>>>>>>>>> >> >
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> --
>>>>>>>>>>>>> >> Tato zpráva byla vytvořena převratným poštovním klientem
>>>>>>>>>>>>> Opery:
>>>>>>>>>>>>> >> http://www.opera.com/mail/
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> >> ______________________________****_________________
>>>>>>>>>>>>> >> AMBER mailing list
>>>>>>>>>>>>> >> AMBER.ambermd.org
>>>>>>>>>>>>> >>
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> > ______________________________****_________________
>>>>>>>>>>>>> > AMBER mailing list
>>>>>>>>>>>>> > AMBER.ambermd.org
>>>>>>>>>>>>> >
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> > __________ Informace od ESET NOD32 Antivirus, verze
>>>>>>>>>>>>> databaze
>>>>>>>>>>>>> 8398
>>>>>>>>>>>>> > (20130531) __________
>>>>>>>>>>>>> >
>>>>>>>>>>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>>>>>>>> >
>>>>>>>>>>>>> > http://www.eset.cz
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>> >
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> --
>>>>>>>>>>>>> Tato zpráva byla vytvořena převratným poštovním klientem
>>>>>>>>>>>>> Opery:
>>>>>>>>>>>>> http://www.opera.com/mail/
>>>>>>>>>>>>>
>>>>>>>>>>>>> ______________________________****_________________
>>>>>>>>>>>>> AMBER mailing list
>>>>>>>>>>>>> AMBER.ambermd.org
>>>>>>>>>>>>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>>>>>>>> <http://**lists.ambermd.org/mailman/**listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>>>>> >
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> __________ Informace od ESET NOD32 Antivirus, verze databaze
>>>>>>>>>>> 8401
>>>>>>>>>>> (20130601) __________
>>>>>>>>>>>
>>>>>>>>>>> Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>>>>>>
>>>>>>>>>>> http://www.eset.cz
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
>>>>>>>>> http://www.opera.com/mail/
>>>>>>>>>
>>>>>>>>> ______________________________**_________________
>>>>>>>>> AMBER mailing list
>>>>>>>>> AMBER.ambermd.org
>>>>>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>> ______________________________**_________________
>>>>>> AMBER mailing list
>>>>>> AMBER.ambermd.org
>>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>
>>>>>> __________ Informace od ESET NOD32 Antivirus, verze databaze 8403
>>>>>> (20130602) __________
>>>>>>
>>>>>> Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>
>>>>>> http://www.eset.cz
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>>
>>
>> --
>> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
>> http://www.opera.com/mail/
>>
>> _______________________________________________
>> AMBER mailing list
>> AMBER.ambermd.org
>> http://lists.ambermd.org/mailman/listinfo/amber
>>
>>
> _______________________________________________
> AMBER mailing list
> AMBER.ambermd.org
> http://lists.ambermd.org/mailman/listinfo/amber


-- 
Tato zpráva byla vytvořena převratným poštovním klientem Opery:  
http://www.opera.com/mail/
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Wed Jun 19 2013 - 10:30:05 PDT
Custom Search