Re: [AMBER] experiences with EVGA GTX TITAN Superclocked - memtestG80 - UNDERclocking in Linux ?

From: Scott Le Grand <varelse2005.gmail.com>
Date: Sun, 2 Jun 2013 09:33:52 -0700

PS this *might* indicate a software bug in cuFFT, but it needs more
characterization... And things are going to get a little stream of
consciousness from here because you're getting unfiltered raw data, so
please don't draw any conclusions towards anything yet - I'm just letting
you guys know what I'm finding out as I find it...



On Sun, Jun 2, 2013 at 9:31 AM, Scott Le Grand <varelse2005.gmail.com>wrote:

> And bingo...
>
> At the very least, the reciprocal sum is intermittently inconsistent...
> This explains the irreproducible behavior...
>
> And here's the level of inconsistency:
> 31989.38940628897399 vs
> 31989.39168370794505
>
> That's error at the level of 1e-7 or a somehow missed single-precision
> transaction somewhere...
>
> The next question is figuring out why... This may or may not ultimately
> explain the crashes you guys are also seeing...
>
>
>
> On Sun, Jun 2, 2013 at 9:07 AM, Scott Le Grand <varelse2005.gmail.com>wrote:
>
>>
>> Observations:
>> 1. The degree to which the reproducibility is broken *does* appear to
>> vary between individual Titan GPUs. One of my Titans breaks within 10K
>> steps on cellulose, the other one made it to 100K steps twice without doing
>> so leading me to believe it could be trusted (until yesterday where I now
>> see it dies between 50K and 100K steps most of the time).
>>
>> 2. GB hasn't broken (yet). So could you run myoglobin for 500K and
>> TRPcage for 1,000,000 steps and let's see if that's universal.
>>
>> 3. Turning on double-precision mode makes my Titan crash rather than run
>> irreproducibly, sigh...
>>
>> So whatever is going on is triggered by something in PME but not GB. So
>> that's either the radix sort, the FFT, the Ewald grid interpolation, or the
>> neighbor list code. Fixing this involves isolating this and figuring out
>> what exactly goes haywire. It could *still* be software at some very small
>> probability but the combination of both 680 and K20c with ECC off running
>> reliably is really pointing towards the Titans just being clocked too
>> fast.
>>
>> So how long with this take? Asking people how long it takes to fix a bug
>> never really works out well. That said, I found the 480 bug within a week
>> and my usual turnaround for a bug with a solid repro is <24 hours.
>>
>> Scott
>>
>> On Sun, Jun 2, 2013 at 7:58 AM, Marek Maly <marek.maly.ujep.cz> wrote:
>>
>>> Hi all,
>>>
>>> here are my results after bugfix 18 application (see attachment).
>>>
>>> In principle I don't see any "drastical" changes.
>>>
>>> FACTOR_IX still perfectly stable/reproducible on both cards,
>>>
>>> JAC tests - problems with finishing AND/OR reproducibility the
>>> same CELLULOSE_NVE although here it seems that my TITAN_1
>>> has no problems with this test (but the same same trend I saw also
>>> before bugfix 18 - see my older 500K steps test).
>>>
>>> But anyway bugfix 18 brought here one change.
>>>
>>> The err
>>>
>>>
>>> #1 ERR writtent in mdout:
>>> ------
>>> | ERROR: max pairlist cutoff must be less than unit cell max sphere
>>> radius!
>>> ------
>>>
>>> was substituted with err/warning ?
>>>
>>> #0 no ERR writtent in mdout, ERR written in standard output (nohup.out)
>>> -----
>>> Nonbond cells need to be recalculated, restart simulation from previous
>>> checkpoint
>>> with a higher value for skinnb.
>>>
>>> -----
>>>
>>> Another thing,
>>>
>>> recently I started on another machine and GTX 580 GPU simulation of
>>> relatively
>>> big system ( 364275 atoms/PME ). The system is composed also from the
>>> "exotic" molecules like polymers. ff12SB, gaff, GLYCAM forcefields used
>>> here. I had problem even with minimization part here, having big energy
>>> on the start:
>>>
>>> -----
>>> NSTEP ENERGY RMS GMAX NAME NUMBER
>>> 1 2.8442E+09 2.1339E+02 1.7311E+04 O 32998
>>>
>>> BOND = 11051.7467 ANGLE = 17720.4706 DIHED =
>>> 18977.7584
>>> VDWAALS = ************* EEL = -1257709.6203 HBOND =
>>> 0.0000
>>> 1-4 VDW = 7253.7412 1-4 EEL = 149867.0207 RESTRAINT =
>>> 0.0000
>>>
>>> ----
>>>
>>> with no chance to minimize the system even with 50 000 steps in both
>>> min cycles (with constrained and unconstrained solute) and hence heating
>>> NVT
>>> crashed immediately even with very small dt. I patched Amber12 here with
>>> the
>>> bugfix 18 and the minimization was done without any problem with common
>>> 5000 steps
>>> (obtaining target Energy -1.4505E+06 while that initial was that written
>>> above).
>>>
>>> So indeed bugfix 18 solved some issues, but unfortunately not those
>>> related to
>>> Titans.
>>>
>>> Here I will try to install cuda 5.5, recompile GPU Amber part with this
>>> new
>>> cuda version and repeat the 100K tests.
>>>
>>> Scott, let us know how finished your experiment with downclocking of
>>> Titan.
>>> Maybe the best choice would be here to flash Titan directly with your
>>> K20c bios :))
>>>
>>> M.
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> Dne Sat, 01 Jun 2013 21:09:46 +0200 Marek Maly <marek.maly.ujep.cz>
>>> napsal/-a:
>>>
>>>
>>> Hi,
>>>>
>>>> first of all thanks for providing of your test results !
>>>>
>>>> It seems that your results are more or less similar to that of
>>>> mine maybe with the exception of the results on FactorIX tests
>>>> where I had perfect stability and 100% or close to 100% reproducibility.
>>>>
>>>> Anyway the type of errs which you reported are the same which I
>>>> obtained.
>>>>
>>>> So let's see if the bugfix 18 will help here (or at least on NPT tests)
>>>> or not. As I wrote just before few minutes, it seems that it was not
>>>> still
>>>> loaded
>>>> to the given server, although it's description is already present on the
>>>> given
>>>> web page ( see http://ambermd.org/bugfixes12.**html<http://ambermd.org/bugfixes12.html>).
>>>>
>>>> As you can see, this bugfix contains also changes in CPU code although
>>>> the majority is devoted to GPU code, so perhaps the best will be to
>>>> recompile
>>>> whole amber with this patch although this patch would be perhaps applied
>>>> even after just
>>>> GPU configure command ( i.e. ./configure -cuda -noX11 gnu ) but after
>>>> consequent
>>>> building, just the GPU binaries will be updated. Anyway I would rather
>>>> recompile
>>>> whole Amber after this patch.
>>>>
>>>> Regarding to GPU test under linux you may try memtestG80
>>>> (please use the updated/patched version from here
>>>> https://github.com/ihaque/**memtestG80<https://github.com/ihaque/memtestG80>
>>>> )
>>>>
>>>> just use git command like:
>>>>
>>>> git clone https://github.com/ihaque/**memtestG80.git<https://github.com/ihaque/memtestG80.git>PATCHED_MEMTEST-G80
>>>>
>>>> to download all the files and save them into directory named
>>>> PATCHED_MEMTEST-G80.
>>>>
>>>> another possibility is to try perhaps similar (but maybe more up to
>>>> date)
>>>> test
>>>> cuda_memtest ( http://sourceforge.net/**projects/cudagpumemtest/<http://sourceforge.net/projects/cudagpumemtest/>).
>>>>
>>>> regarding ig value: If ig is not present in mdin, the default value is
>>>> used
>>>> (e.g. 71277) if ig=-1 the random seed will be based on the current date
>>>> and time, and hence will be different for every run (not a good variant
>>>> for our testts). I simply deleted eventual ig records from all mdins so
>>>> I
>>>> assume that in each run the default seed 71277 was automatically used.
>>>>
>>>> M.
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Dne Sat, 01 Jun 2013 20:26:16 +0200 ET <sketchfoot.gmail.com>
>>>> napsal/-a:
>>>>
>>>> Hi,
>>>>>
>>>>> I've put the graphics card into a machine with the working GTX titan
>>>>> that I
>>>>> mentioned earlier.
>>>>>
>>>>> The Nvidia driver version is: 133.30
>>>>>
>>>>> Amber version is:
>>>>> AmberTools version 13.03
>>>>> Amber version 12.16
>>>>>
>>>>> I ran 50k steps with the amber benchmark using ig=43689 on both cards.
>>>>> For
>>>>> the purpose of discriminating between them, the card I believe (fingers
>>>>> crossed) is working is called GPU-00_TeaNCake, whilst the other one is
>>>>> called GPU-01_008.
>>>>>
>>>>> *When I run the tests on GPU-01_008:*
>>>>>
>>>>> 1) All the tests (across 2x repeats) finish apart from the following
>>>>> which
>>>>> have the errors listed:
>>>>>
>>>>> ------------------------------**--------------
>>>>> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
>>>>> Error: unspecified launch failure launching kernel kNLSkinTest
>>>>> cudaFree GpuBuffer::Deallocate failed unspecified launch failure
>>>>>
>>>>> ------------------------------**--------------
>>>>> CELLULOSE_PRODUCTION_NPT - 408,609 atoms PME
>>>>> cudaMemcpy GpuBuffer::Download failed unspecified launch failure
>>>>>
>>>>> ------------------------------**--------------
>>>>> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
>>>>> Error: unspecified launch failure launching kernel kNLSkinTest
>>>>> cudaFree GpuBuffer::Deallocate failed unspecified launch failure
>>>>>
>>>>> ------------------------------**--------------
>>>>> CELLULOSE_PRODUCTION_NPT - 408,609 atoms PME
>>>>> cudaMemcpy GpuBuffer::Download failed unspecified launch failure
>>>>> grep: mdinfo.1GTX680: No such file or directory
>>>>>
>>>>>
>>>>>
>>>>> 2) The sdiff logs indicate that reproducibility across the two repeats
>>>>> is
>>>>> as follows:
>>>>>
>>>>> *GB_myoglobin: *Reproducible across 50k steps
>>>>> *GB_nucleosome:* Reproducible till step 7400
>>>>> *GB_TRPCage:* Reproducible across 50k steps
>>>>>
>>>>> *PME_JAC_production_NVE: *No reproducibility shown from step 1,000
>>>>> onwards
>>>>> *PME_JAC_production_NPT*: Reproducible till step 1,000. Also outfile
>>>>> is
>>>>> not written properly - blank gaps appear where something should have
>>>>> been
>>>>> written
>>>>>
>>>>> *PME_FactorIX_production_NVE:* Reproducible across 50k steps
>>>>> *PME_FactorIX_production_NPT:* Reproducible across 50k steps
>>>>>
>>>>> *PME_Cellulose_production_NVE:*** Failure means that both runs do not
>>>>> finish
>>>>> (see point1)
>>>>> *PME_Cellulose_production_NPT: *Failure means that both runs do not
>>>>> finish
>>>>> (see point1)
>>>>>
>>>>> ##############################**##############################**
>>>>> ###########################
>>>>>
>>>>> *When I run the tests on * *GPU-00_TeaNCake:*
>>>>> *
>>>>> *
>>>>> 1) All the tests (across 2x repeats) finish apart from the following
>>>>> which
>>>>> have the errors listed:
>>>>> ------------------------------**-------
>>>>> JAC_PRODUCTION_NPT - 23,558 atoms PME
>>>>> PMEMD Terminated Abnormally!
>>>>> ------------------------------**-------
>>>>>
>>>>>
>>>>> 2) The sdiff logs indicate that reproducibility across the two repeats
>>>>> is
>>>>> as follows:
>>>>>
>>>>> *GB_myoglobin:* Reproducible across 50k steps
>>>>> *GB_nucleosome:* Reproducible across 50k steps
>>>>> *GB_TRPCage:* Reproducible across 50k steps
>>>>>
>>>>> *PME_JAC_production_NVE:* No reproducibility shown from step 10,000
>>>>> onwards
>>>>> *PME_JAC_production_NPT: * No reproducibility shown from step 10,000
>>>>> onwards. Also outfile is not written properly - blank gaps appear where
>>>>> something should have been written. Repeat 2 Crashes with error noted
>>>>> in
>>>>> 1.
>>>>>
>>>>> *PME_FactorIX_production_NVE:* No reproducibility shown from step 9,000
>>>>> onwards
>>>>> *PME_FactorIX_production_NPT: *Reproducible across 50k steps
>>>>>
>>>>> *PME_Cellulose_production_NVE: *No reproducibility shown from step
>>>>> 5,000
>>>>> onwards
>>>>> *PME_Cellulose_production_NPT: ** *No reproducibility shown from step
>>>>> 29,000 onwards. Also outfile is not written properly - blank gaps
>>>>> appear
>>>>> where something should have been written.
>>>>>
>>>>>
>>>>> Out files and sdiff files are included as attatchments
>>>>>
>>>>> ##############################**###################
>>>>>
>>>>> So I'm going to update my nvidia driver to the latest version and patch
>>>>> amber to the latest version and rerun the tests to see if there is any
>>>>> improvement. Could someone let me know if it is necessary to recompile
>>>>> any
>>>>> or all of AMBER after applying the bugfixes?
>>>>>
>>>>> Additionally, I'm going to run memory tests and heaven benchmarks on
>>>>> the
>>>>> cards to check whether they are faulty or not.
>>>>>
>>>>> I'm thinking that there is a mix of hardware error/configuration (esp
>>>>> in
>>>>> the case of GPU-01_008) and amber software error in this situation.
>>>>> What
>>>>> do
>>>>> you guys think?
>>>>>
>>>>> Also am I right in thinking (from what Scott was saying) that all the
>>>>> benchmarks should be reproducible across 50k steps but begin to diverge
>>>>> at
>>>>> around 100K steps? Is there any difference from in setting *ig *to an
>>>>> explicit number to removing it from the mdin file?
>>>>>
>>>>> br,
>>>>> g
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On 31 May 2013 23:45, ET <sketchfoot.gmail.com> wrote:
>>>>>
>>>>> I don't need sysadmins, but sysadmins need me as it gives purpose to
>>>>>> their
>>>>>> bureaucratic existence. A encountered evil if working in an
>>>>>> institution
>>>>>> or
>>>>>> comapny IMO. Good science and indiviguality being sacrificed for
>>>>>> standardisation and mediocrity in the intrerests of maintaing a system
>>>>>> that
>>>>>> focusses on maintaining the system and not the objective.
>>>>>>
>>>>>> You need root to move fwd on these things, unfortunately. and ppl with
>>>>>> root are kinda like your parents when you try to borrow money from
>>>>>> them
>>>>>> .
>>>>>> age 12 :D
>>>>>> On May 31, 2013 9:34 PM, "Marek Maly" <marek.maly.ujep.cz> wrote:
>>>>>>
>>>>>> Sorry why do you need sysadmins :)) ?
>>>>>>>
>>>>>>> BTW here is the most recent driver:
>>>>>>>
>>>>>>> http://www.nvidia.com/object/**linux-display-amd64-319.23-**
>>>>>>> driver.html<http://www.nvidia.com/object/linux-display-amd64-319.23-driver.html>
>>>>>>>
>>>>>>> I do not remember anything easier than is to install driver
>>>>>>> (especially
>>>>>>> in case of binary (*.run) installer) :))
>>>>>>>
>>>>>>> M.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Dne Fri, 31 May 2013 22:02:34 +0200 ET <sketchfoot.gmail.com>
>>>>>>> napsal/-a:
>>>>>>>
>>>>>>> > Yup. I know. I replaced a 680 and the everknowing sysadmins are
>>>>>>> reluctant
>>>>>>> > to install drivers not in the repositoery as they are lame. :(
>>>>>>> > On May 31, 2013 7:14 PM, "Marek Maly" <marek.maly.ujep.cz> wrote:
>>>>>>> >>
>>>>>>> >> As I already wrote you,
>>>>>>> >>
>>>>>>> >> the first driver which properly/officially supports Titans, should
>>>>>>> be
>>>>>>> >> 313.26 .
>>>>>>> >>
>>>>>>> >> Anyway I am curious mainly about your 100K repetitive tests with
>>>>>>> >> your Titan SC card. Especially in case of these tests ( JAC_NVE,
>>>>>>> JAC_NPT
>>>>>>> >> and CELLULOSE_NVE ) where
>>>>>>> >> my Titans SC randomly failed or succeeded. In FACTOR_IX_NVE,
>>>>>>> >> FACTOR_IX_NPT
>>>>>>> >> tests both
>>>>>>> >> my cards are perfectly stable (independently from drv. version)
>>>>>>> and
>>>>>>> also
>>>>>>> >> the runs
>>>>>>> >> are perfectly or almost perfectly reproducible.
>>>>>>> >>
>>>>>>> >> Also if your test will crash please report the eventual errs.
>>>>>>> >>
>>>>>>> >> To this moment I have this actual library of errs on my Titans SC
>>>>>>> GPUs.
>>>>>>> >>
>>>>>>> >> #1 ERR writtent in mdout:
>>>>>>> >> ------
>>>>>>> >> | ERROR: max pairlist cutoff must be less than unit cell max
>>>>>>> sphere
>>>>>>> >> radius!
>>>>>>> >> ------
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> #2 no ERR writtent in mdout, ERR written in standard output
>>>>>>> (nohup.out)
>>>>>>> >>
>>>>>>> >> ----
>>>>>>> >> Error: unspecified launch failure launching kernel kNLSkinTest
>>>>>>> >> cudaFree GpuBuffer::Deallocate failed unspecified launch failure
>>>>>>> >> ----
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> #3 no ERR writtent in mdout, ERR written in standard output
>>>>>>> (nohup.out)
>>>>>>> >> ----
>>>>>>> >> cudaMemcpy GpuBuffer::Download failed unspecified launch failure
>>>>>>> >> ----
>>>>>>> >>
>>>>>>> >> Another question, regarding your Titan SC, it is also EVGA as in
>>>>>>> my
>>>>>>> case
>>>>>>> >> or it is another producer ?
>>>>>>> >>
>>>>>>> >> Thanks,
>>>>>>> >>
>>>>>>> >> M.
>>>>>>> >>
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> Dne Fri, 31 May 2013 19:17:03 +0200 ET <sketchfoot.gmail.com>
>>>>>>> napsal/-a:
>>>>>>> >>
>>>>>>> >> > Well, this is interesting...
>>>>>>> >> >
>>>>>>> >> > I ran 50k steps on the Titan on the other machine with driver
>>>>>>> 310.44
>>>>>>> >> and
>>>>>>> >> > it
>>>>>>> >> > passed all the GB steps. i.e totally identical results over two
>>>>>>> >> repeats.
>>>>>>> >> > However, it failed all the PME tests after step 1000. I'm going
>>>>>>> to
>>>>>>> > update
>>>>>>> >> > the driver and test it again.
>>>>>>> >> >
>>>>>>> >> > Files included as attachments.
>>>>>>> >> >
>>>>>>> >> > br,
>>>>>>> >> > g
>>>>>>> >> >
>>>>>>> >> >
>>>>>>> >> > On 31 May 2013 16:40, Marek Maly <marek.maly.ujep.cz> wrote:
>>>>>>> >> >
>>>>>>> >> >> One more thing,
>>>>>>> >> >>
>>>>>>> >> >> can you please check under which frequency is running that your
>>>>>>> >> titan ?
>>>>>>> >> >>
>>>>>>> >> >> As the base frequency of normal Titans is 837MHz and the Boost
>>>>>>> one
>>>>>>> is
>>>>>>> >> >> 876MHz I
>>>>>>> >> >> assume that yor GPU is running automatically also under it's
>>>>>>> boot
>>>>>>> >> >> frequency (876MHz).
>>>>>>> >> >> You can find this information e.g. in Amber mdout file.
>>>>>>> >> >>
>>>>>>> >> >> You also mentioned some crashes in your previous email. Your
>>>>>>> ERRs
>>>>>>> >> were
>>>>>>> >> >> something like those here:
>>>>>>> >> >>
>>>>>>> >> >> #1 ERR writtent in mdout:
>>>>>>> >> >> ------
>>>>>>> >> >> | ERROR: max pairlist cutoff must be less than unit cell max
>>>>>>> sphere
>>>>>>> >> >> radius!
>>>>>>> >> >> ------
>>>>>>> >> >>
>>>>>>> >> >>
>>>>>>> >> >> #2 no ERR writtent in mdout, ERR written in standard output
>>>>>>> >> (nohup.out)
>>>>>>> >> >>
>>>>>>> >> >> ----
>>>>>>> >> >> Error: unspecified launch failure launching kernel kNLSkinTest
>>>>>>> >> >> cudaFree GpuBuffer::Deallocate failed unspecified launch
>>>>>>> failure
>>>>>>> >> >> ----
>>>>>>> >> >>
>>>>>>> >> >>
>>>>>>> >> >> #3 no ERR writtent in mdout, ERR written in standard output
>>>>>>> >> (nohup.out)
>>>>>>> >> >> ----
>>>>>>> >> >> cudaMemcpy GpuBuffer::Download failed unspecified launch
>>>>>>> failure
>>>>>>> >> >> ----
>>>>>>> >> >>
>>>>>>> >> >> or you obtained some new/additional errs ?
>>>>>>> >> >>
>>>>>>> >> >>
>>>>>>> >> >>
>>>>>>> >> >> M.
>>>>>>> >> >>
>>>>>>> >> >>
>>>>>>> >> >>
>>>>>>> >> >> Dne Fri, 31 May 2013 17:30:57 +0200 filip fratev
>>>>>>> >> <filipfratev.yahoo.com
>>>>>>> >>
>>>>>>> >> >> napsal/-a:
>>>>>>> >> >>
>>>>>>> >> >> > Hi,
>>>>>>> >> >> > This is what I obtained for 50K tests and "normal" GTXTitan:
>>>>>>> >> >> >
>>>>>>> >> >> > run1:
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >>
>>>>>>> >
>>>>>>> ------------------------------**------------------------------**
>>>>>>> ------------------
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > A V E R A G E S O V E R 50 S T E P S
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
>>>>>>> 299.87
>>>>>>> >> PRESS
>>>>>>> >> >> > = 0.0
>>>>>>> >> >> > Etot = -443237.1079 EKtot = 257679.9750 EPtot
>>>>>>> =
>>>>>>> >> >> > -700917.0829
>>>>>>> >> >> > BOND = 20193.1856 ANGLE = 53517.5432 DIHED
>>>>>>> =
>>>>>>> >> >> > 23575.4648
>>>>>>> >> >> > 1-4 NB = 21759.5524 1-4 EEL = 742552.5939 VDWAALS
>>>>>>> =
>>>>>>> >> >> > 96286.7714
>>>>>>> >> >> > EELEC = -1658802.1941 EHBOND = 0.0000 RESTRAINT
>>>>>>> =
>>>>>>> >> >> > 0.0000
>>>>>>> >> >> >
>>>>>>> >> >>
>>>>>>> >
>>>>>>> ------------------------------**------------------------------**
>>>>>>> ------------------
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > R M S F L U C T U A T I O N S
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
>>>>>>> 0.33
>>>>>>> >> PRESS
>>>>>>> >> >> > = 0.0
>>>>>>> >> >> > Etot = 11.2784 EKtot = 284.8999 EPtot
>>>>>>> =
>>>>>>> >> >> > 289.0773
>>>>>>> >> >> > BOND = 136.3417 ANGLE = 214.0054 DIHED
>>>>>>> =
>>>>>>> >> >> > 59.4893
>>>>>>> >> >> > 1-4 NB = 58.5891 1-4 EEL = 330.5400 VDWAALS
>>>>>>> =
>>>>>>> >> >> > 559.2079
>>>>>>> >> >> > EELEC = 743.8771 EHBOND = 0.0000 RESTRAINT
>>>>>>> =
>>>>>>> >> >> > 0.0000
>>>>>>> >> >> > |E(PBS) = 21.8119
>>>>>>> >> >> >
>>>>>>> >> >>
>>>>>>> >
>>>>>>> ------------------------------**------------------------------**
>>>>>>> ------------------
>>>>>>> >> >> >
>>>>>>> >> >> > run2:
>>>>>>> >> >> >
>>>>>>> >> >>
>>>>>>> >
>>>>>>> ------------------------------**------------------------------**
>>>>>>> ------------------
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > A V E R A G E S O V E R 50 S T E P S
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
>>>>>>> 299.89
>>>>>>> >> PRESS
>>>>>>> >> >> > = 0.0
>>>>>>> >> >> > Etot = -443240.0999 EKtot = 257700.0950 EPtot
>>>>>>> =
>>>>>>> >> >> > -700940.1949
>>>>>>> >> >> > BOND = 20241.9174 ANGLE = 53644.6694 DIHED
>>>>>>> =
>>>>>>> >> >> > 23541.3737
>>>>>>> >> >> > 1-4 NB = 21803.1898 1-4 EEL = 742754.2254 VDWAALS
>>>>>>> =
>>>>>>> >> >> > 96298.8308
>>>>>>> >> >> > EELEC = -1659224.4013 EHBOND = 0.0000 RESTRAINT
>>>>>>> =
>>>>>>> >> >> > 0.0000
>>>>>>> >> >> >
>>>>>>> >> >>
>>>>>>> >
>>>>>>> ------------------------------**------------------------------**
>>>>>>> ------------------
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > R M S F L U C T U A T I O N S
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > NSTEP = 50000 TIME(PS) = 120.020 TEMP(K) =
>>>>>>> 0.41
>>>>>>> >> PRESS
>>>>>>> >> >> > = 0.0
>>>>>>> >> >> > Etot = 10.7633 EKtot = 348.2819 EPtot
>>>>>>> =
>>>>>>> >> >> > 353.9918
>>>>>>> >> >> > BOND = 106.5314 ANGLE = 196.7052 DIHED
>>>>>>> =
>>>>>>> >> >> > 69.7476
>>>>>>> >> >> > 1-4 NB = 60.3435 1-4 EEL = 400.7466 VDWAALS
>>>>>>> =
>>>>>>> >> >> > 462.7763
>>>>>>> >> >> > EELEC = 651.9857 EHBOND = 0.0000 RESTRAINT
>>>>>>> =
>>>>>>> >> >> > 0.0000
>>>>>>> >> >> > |E(PBS) = 17.0642
>>>>>>> >> >> >
>>>>>>> >> >>
>>>>>>> >
>>>>>>> ------------------------------**------------------------------**
>>>>>>> ------------------
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >>
>>>>>>> >
>>>>>>> ------------------------------**------------------------------**
>>>>>>> --------------------
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > ______________________________**__
>>>>>>> >> >> > From: Marek Maly <marek.maly.ujep.cz>
>>>>>>> >> >> > To: AMBER Mailing List <amber.ambermd.org>
>>>>>>> >> >> > Sent: Friday, May 31, 2013 3:34 PM
>>>>>>> >> >> > Subject: Re: [AMBER] experiences with EVGA GTX TITAN
>>>>>>> Superclocked
>>>>>>> -
>>>>>>> >> >> > memtestG80 - UNDERclocking in Linux ?
>>>>>>> >> >> >
>>>>>>> >> >> > Hi here are my 100K results for driver 313.30 (and still Cuda
>>>>>>> 5.0).
>>>>>>> >> >> >
>>>>>>> >> >> > The results are rather similar to those obtained
>>>>>>> >> >> > under my original driver 319.17 (see the first table
>>>>>>> >> >> > which I sent in this thread).
>>>>>>> >> >> >
>>>>>>> >> >> > M.
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > Dne Fri, 31 May 2013 12:29:59 +0200 Marek Maly <
>>>>>>> marek.maly.ujep.cz>
>>>>>>> >> >> > napsal/-a:
>>>>>>> >> >> >
>>>>>>> >> >> >> Hi,
>>>>>>> >> >> >>
>>>>>>> >> >> >> please try to run at lest 100K tests twice to verify exact
>>>>>>> >> >> >> reproducibility
>>>>>>> >> >> >> of the results on the given card. If you find in any mdin
>>>>>>> file
>>>>>>> >> ig=-1
>>>>>>> >> >> >> just
>>>>>>> >> >> >> delete it to ensure that you are using the identical random
>>>>>>> seed
>>>>>>> >> for
>>>>>>> >> >> >> both
>>>>>>> >> >> >> runs. You can eventually omit NUCLEOSOME test
>>>>>>> >> >> >> as it is too much time consuming.
>>>>>>> >> >> >>
>>>>>>> >> >> >> Driver 310.44 ?????
>>>>>>> >> >> >>
>>>>>>> >> >> >> As far as I know the proper support for titans is from
>>>>>>> version
>>>>>>> > 313.26
>>>>>>> >> >> >>
>>>>>>> >> >> >> see e.g. here :
>>>>>>> >> >> >>
>>>>>>> >> >>
>>>>>>> >
>>>>>>> http://www.geeks3d.com/**20130306/nvidia-releases-r313-**
>>>>>>> 26-for-linux-with-gtx-titan-**support/<http://www.geeks3d.com/20130306/nvidia-releases-r313-26-for-linux-with-gtx-titan-support/>
>>>>>>> >> >> >>
>>>>>>> >> >> >> BTW: On my site downgrade to drv. 313.30 did not solved the
>>>>>>> >> >> situation, I
>>>>>>> >> >> >> will post
>>>>>>> >> >> >> my results soon here.
>>>>>>> >> >> >>
>>>>>>> >> >> >> M.
>>>>>>> >> >> >>
>>>>>>> >> >> >>
>>>>>>> >> >> >>
>>>>>>> >> >> >>
>>>>>>> >> >> >>
>>>>>>> >> >> >>
>>>>>>> >> >> >>
>>>>>>> >> >> >>
>>>>>>> >> >> >> Dne Fri, 31 May 2013 12:21:21 +0200 ET <
>>>>>>> sketchfoot.gmail.com>
>>>>>>> >> >> napsal/-a:
>>>>>>> >> >> >>
>>>>>>> >> >> >>> ps. I have another install of amber on another computer
>>>>>>> with
>>>>>>> a
>>>>>>> >> >> >>> different
>>>>>>> >> >> >>> Titan and different Driver Version: 310.44.
>>>>>>> >> >> >>>
>>>>>>> >> >> >>> In the interests of thrashing the proverbial horse, I'll
>>>>>>> run
>>>>>>> the
>>>>>>> >> >> >>> benchmark
>>>>>>> >> >> >>> for 50k steps. :P
>>>>>>> >> >> >>>
>>>>>>> >> >> >>> br,
>>>>>>> >> >> >>> g
>>>>>>> >> >> >>>
>>>>>>> >> >> >>>
>>>>>>> >> >> >>> On 31 May 2013 11:17, ET <sketchfoot.gmail.com> wrote:
>>>>>>> >> >> >>>
>>>>>>> >> >> >>>> Hi, I just ran the Amber benchmark for the default (10000
>>>>>>> steps)
>>>>>>> >> >> on my
>>>>>>> >> >> >>>> Titan.
>>>>>>> >> >> >>>>
>>>>>>> >> >> >>>> Using sdiff -sB showed that the two runs were completely
>>>>>>> > identical.
>>>>>>> >> >> >>>> I've
>>>>>>> >> >> >>>> attached compressed files of the mdout & diff files.
>>>>>>> >> >> >>>>
>>>>>>> >> >> >>>> br,
>>>>>>> >> >> >>>> g
>>>>>>> >> >> >>>>
>>>>>>> >> >> >>>>
>>>>>>> >> >> >>>> On 30 May 2013 23:41, Marek Maly <marek.maly.ujep.cz>
>>>>>>> wrote:
>>>>>>> >> >> >>>>
>>>>>>> >> >> >>>>> OK, let's see. The eventual downclocking I see as the
>>>>>>> very
>>>>>>> last
>>>>>>> >> >> >>>>> possibility
>>>>>>> >> >> >>>>> (if I don't decide for RMAing). But now still some other
>>>>>>> >> >> experiments
>>>>>>> >> >> >>>>> are
>>>>>>> >> >> >>>>> available :))
>>>>>>> >> >> >>>>> I just started 100K tests under 313.30 driver. For today
>>>>>>> good
>>>>>>> >> >> night
>>>>>>> >> >> >>>>> ...
>>>>>>> >> >> >>>>>
>>>>>>> >> >> >>>>> M.
>>>>>>> >> >> >>>>>
>>>>>>> >> >> >>>>> Dne Fri, 31 May 2013 00:45:49 +0200 Scott Le Grand
>>>>>>> >> >> >>>>> <varelse2005.gmail.com
>>>>>>> >> >> >>>>> >
>>>>>>> >> >> >>>>> napsal/-a:
>>>>>>> >> >> >>>>>
>>>>>>> >> >> >>>>> > It will be very interesting if this behavior persists
>>>>>>> after
>>>>>>> >> >> >>>>> downclocking.
>>>>>>> >> >> >>>>> >
>>>>>>> >> >> >>>>> > But right now, Titan 0 *looks* hosed and Titan 1
>>>>>>> *looks*
>>>>>>> like
>>>>>>> > it
>>>>>>> >> >> >>>>> needs
>>>>>>> >> >> >>>>> > downclocking...
>>>>>>> >> >> >>>>> > On May 30, 2013 3:20 PM, "Marek Maly"
>>>>>>> <marek.maly.ujep.cz>
>>>>>>> >> >> wrote:
>>>>>>> >> >> >>>>> >
>>>>>>> >> >> >>>>> >> Hi all,
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> here are my results from the 500K steps 2 x repeated
>>>>>>> > benchmarks
>>>>>>> >> >> >>>>> >> under 319.23 driver and still Cuda 5.0 (see the
>>>>>>> attached
>>>>>>> >> table
>>>>>>> >> >> ).
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> It is hard to say if the results are better or worse
>>>>>>> than
>>>>>>> in
>>>>>>> > my
>>>>>>> >> >> >>>>> >> previous 100K test under driver 319.17.
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> While results from Cellulose test were improved and
>>>>>>> the
>>>>>>> > TITAN_1
>>>>>>> >> >> >>>>> card
>>>>>>> >> >> >>>>> >> even
>>>>>>> >> >> >>>>> >> successfully finished all 500K steps moreover with
>>>>>>> exactly
>>>>>>> >> the
>>>>>>> >> >> >>>>> same
>>>>>>> >> >> >>>>> >> final
>>>>>>> >> >> >>>>> >> energy !
>>>>>>> >> >> >>>>> >> (TITAN_0 at least finished more than 100K steps and in
>>>>>>> >> RUN_01
>>>>>>> >> >> even
>>>>>>> >> >> >>>>> more
>>>>>>> >> >> >>>>> >> than 400K steps)
>>>>>>> >> >> >>>>> >> In JAC_NPT test no GPU was able to finish at least
>>>>>>> 100K
>>>>>>> >> steps
>>>>>>> >> >> and
>>>>>>> >> >> >>>>> the
>>>>>>> >> >> >>>>> >> results from JAC_NVE
>>>>>>> >> >> >>>>> >> test are also not too much convincing. FACTOR_IX_NVE
>>>>>>> and
>>>>>>> >> >> >>>>> FACTOR_IX_NPT
>>>>>>> >> >> >>>>> >> were successfully
>>>>>>> >> >> >>>>> >> finished with 100% reproducibility in FACTOR_IX_NPT
>>>>>>> case
>>>>>>> >> (on
>>>>>>> >> >> both
>>>>>>> >> >> >>>>> >> cards)
>>>>>>> >> >> >>>>> >> and almost
>>>>>>> >> >> >>>>> >> 100% reproducibility in case of FACTOR_IX_NVE (again
>>>>>>> 100%
>>>>>>> in
>>>>>>> >> >> case
>>>>>>> >> >> >>>>> of
>>>>>>> >> >> >>>>> >> TITAN_1). TRPCAGE, MYOGLOBIN
>>>>>>> >> >> >>>>> >> again finished without any problem with 100%
>>>>>>> >> reproducibility.
>>>>>>> >> >> >>>>> NUCLEOSOME
>>>>>>> >> >> >>>>> >> test was not done
>>>>>>> >> >> >>>>> >> this time due to high time requirements. If you find
>>>>>>> in
>>>>>>> the
>>>>>>> >> >> table
>>>>>>> >> >> >>>>> >> positive
>>>>>>> >> >> >>>>> >> number finishing with
>>>>>>> >> >> >>>>> >> K (which means "thousands") it means the last number
>>>>>>> of
>>>>>>> step
>>>>>>> >> >> >>>>> written in
>>>>>>> >> >> >>>>> >> mdout before crash.
>>>>>>> >> >> >>>>> >> Below are all the 3 types of detected errs with
>>>>>>> relevant
>>>>>>> >> >> >>>>> systems/rounds
>>>>>>> >> >> >>>>> >> where the given err
>>>>>>> >> >> >>>>> >> appeared.
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> Now I will try just 100K tests under ETs favourite
>>>>>>> driver
>>>>>>> >> >> version
>>>>>>> >> >> >>>>> 313.30
>>>>>>> >> >> >>>>> >> :)) and then
>>>>>>> >> >> >>>>> >> I will eventually try to experiment with cuda 5.5
>>>>>>> which
>>>>>>> I
>>>>>>> >> >> already
>>>>>>> >> >> >>>>> >> downloaded from the
>>>>>>> >> >> >>>>> >> cuda zone ( I had to become cuda developer for this
>>>>>>> :))
>>>>>>> )
>>>>>>> >> BTW
>>>>>>> >> >> ET
>>>>>>> >> >> >>>>> thanks
>>>>>>> >> >> >>>>> >> for the frequency info !
>>>>>>> >> >> >>>>> >> and I am still ( perhaps not alone :)) ) very curious
>>>>>>> about
>>>>>>> >> >> your 2
>>>>>>> >> >> >>>>> x
>>>>>>> >> >> >>>>> >> repeated Amber benchmark tests with superclocked
>>>>>>> Titan.
>>>>>>> >> Indeed
>>>>>>> >> >> >>>>> that
>>>>>>> >> >> >>>>> I
>>>>>>> >> >> >>>>> am
>>>>>>> >> >> >>>>> >> very curious also about that Ross "hot" patch.
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> M.
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> ERRORS DETECTED DURING THE 500K steps tests with
>>>>>>> driver
>>>>>>> >> 319.23
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> #1 ERR writtent in mdout:
>>>>>>> >> >> >>>>> >> ------
>>>>>>> >> >> >>>>> >> | ERROR: max pairlist cutoff must be less than unit
>>>>>>> cell
>>>>>>> >> max
>>>>>>> >> >> >>>>> sphere
>>>>>>> >> >> >>>>> >> radius!
>>>>>>> >> >> >>>>> >> ------
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_1 JAC_NPT (at least 5000 steps
>>>>>>> successfully
>>>>>>> > done
>>>>>>> >> >> >>>>> before
>>>>>>> >> >> >>>>> >> crash)
>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 JAC_NPT (at least 8000 steps
>>>>>>> successfully
>>>>>>> > done
>>>>>>> >> >> >>>>> before
>>>>>>> >> >> >>>>> >> crash)
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> #2 no ERR writtent in mdout, ERR written in standard
>>>>>>> output
>>>>>>> >> >> >>>>> (nohup.out)
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> ----
>>>>>>> >> >> >>>>> >> Error: unspecified launch failure launching kernel
>>>>>>> >> kNLSkinTest
>>>>>>> >> >> >>>>> >> cudaFree GpuBuffer::Deallocate failed unspecified
>>>>>>> launch
>>>>>>> >> >> failure
>>>>>>> >> >> >>>>> >> ----
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_1 CELLULOSE_NVE (at least 437 000 steps
>>>>>>> >> >> successfully
>>>>>>> >> >> >>>>> done
>>>>>>> >> >> >>>>> >> before crash)
>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 JAC_NVE (at least 162 000 steps
>>>>>>> >> successfully
>>>>>>> >> >> done
>>>>>>> >> >> >>>>> >> before
>>>>>>> >> >> >>>>> >> crash)
>>>>>>> >> >> >>>>> >> TITAN_0 ROUND_2 CELLULOSE_NVE (at least 117 000 steps
>>>>>>> >> >> successfully
>>>>>>> >> >> >>>>> done
>>>>>>> >> >> >>>>> >> before crash)
>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_1 JAC_NVE (at least 119 000 steps
>>>>>>> >> successfully
>>>>>>> >> >> done
>>>>>>> >> >> >>>>> >> before
>>>>>>> >> >> >>>>> >> crash)
>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_2 JAC_NVE (at least 43 000 steps
>>>>>>> successfully
>>>>>>> >> >> done
>>>>>>> >> >> >>>>> before
>>>>>>> >> >> >>>>> >> crash)
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> #3 no ERR writtent in mdout, ERR written in standard
>>>>>>> output
>>>>>>> >> >> >>>>> (nohup.out)
>>>>>>> >> >> >>>>> >> ----
>>>>>>> >> >> >>>>> >> cudaMemcpy GpuBuffer::Download failed unspecified
>>>>>>> launch
>>>>>>> >> >> failure
>>>>>>> >> >> >>>>> >> ----
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_1 JAC_NPT (at least 77 000 steps
>>>>>>> successfully
>>>>>>> >> >> done
>>>>>>> >> >> >>>>> before
>>>>>>> >> >> >>>>> >> crash)
>>>>>>> >> >> >>>>> >> TITAN_1 ROUND_2 JAC_NPT (at least 58 000 steps
>>>>>>> successfully
>>>>>>> >> >> done
>>>>>>> >> >> >>>>> before
>>>>>>> >> >> >>>>> >> crash)
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> Dne Thu, 30 May 2013 21:27:17 +0200 Scott Le Grand
>>>>>>> >> >> >>>>> >> <varelse2005.gmail.com>
>>>>>>> >> >> >>>>> >> napsal/-a:
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> Oops meant to send that to Jason...
>>>>>>> >> >> >>>>> >>>
>>>>>>> >> >> >>>>> >>> Anyway, before we all panic, we need to get K20's
>>>>>>> behavior
>>>>>>> >> >> >>>>> analyzed
>>>>>>> >> >> >>>>> >>> here.
>>>>>>> >> >> >>>>> >>> If it's deterministic, this truly is a hardware
>>>>>>> issue. If
>>>>>>> >> >> not,
>>>>>>> >> >> >>>>> then
>>>>>>> >> >> >>>>> it
>>>>>>> >> >> >>>>> >>> gets interesting because 680 is deterministic as far
>>>>>>> as I
>>>>>>> >> can
>>>>>>> >> >> >>>>> tell...
>>>>>>> >> >> >>>>> >>> On May 30, 2013 12:24 PM, "Scott Le Grand"
>>>>>>> >> >> >>>>> <varelse2005.gmail.com>
>>>>>>> >> >> >>>>> >>> wrote:
>>>>>>> >> >> >>>>> >>>
>>>>>>> >> >> >>>>> >>> If the errors are not deterministically triggered,
>>>>>>> they
>>>>>>> >> >> probably
>>>>>>> >> >> >>>>> >>> won't be
>>>>>>> >> >> >>>>> >>>> fixed by the patch, alas...
>>>>>>> >> >> >>>>> >>>> On May 30, 2013 12:15 PM, "Jason Swails"
>>>>>>> >> >> >>>>> <jason.swails.gmail.com>
>>>>>>> >> >> >>>>> >>>> wrote:
>>>>>>> >> >> >>>>> >>>>
>>>>>>> >> >> >>>>> >>>> Just a reminder to everyone based on what Ross
>>>>>>> said:
>>>>>>> >> there
>>>>>>> >> >> is a
>>>>>>> >> >> >>>>> >>>> pending
>>>>>>> >> >> >>>>> >>>>> patch to pmemd.cuda that will be coming out shortly
>>>>>>> >> (maybe
>>>>>>> >> >> even
>>>>>>> >> >> >>>>> >>>>> within
>>>>>>> >> >> >>>>> >>>>> hours). It's entirely possible that several of
>>>>>>> these
>>>>>>> > errors
>>>>>>> >> >> >>>>> are
>>>>>>> >> >> >>>>> >>>>> fixed
>>>>>>> >> >> >>>>> >>>>> by
>>>>>>> >> >> >>>>> >>>>> this patch.
>>>>>>> >> >> >>>>> >>>>>
>>>>>>> >> >> >>>>> >>>>> All the best,
>>>>>>> >> >> >>>>> >>>>> Jason
>>>>>>> >> >> >>>>> >>>>>
>>>>>>> >> >> >>>>> >>>>>
>>>>>>> >> >> >>>>> >>>>> On Thu, May 30, 2013 at 2:46 PM, filip fratev <
>>>>>>> >> >> >>>>> filipfratev.yahoo.com>
>>>>>>> >> >> >>>>> >>>>> wrote:
>>>>>>> >> >> >>>>> >>>>>
>>>>>>> >> >> >>>>> >>>>> > I have observed the same crashes from time to
>>>>>>> time. I
>>>>>>> > will
>>>>>>> >> >> >>>>> run
>>>>>>> >> >> >>>>> >>>>> cellulose
>>>>>>> >> >> >>>>> >>>>> > nve for 100k and will past results here.
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>> > All the best,
>>>>>>> >> >> >>>>> >>>>> > Filip
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>> > ______________________________****__
>>>>>>> >> >> >>>>> >>>>> > From: Scott Le Grand <varelse2005.gmail.com>
>>>>>>> >> >> >>>>> >>>>> > To: AMBER Mailing List <amber.ambermd.org>
>>>>>>> >> >> >>>>> >>>>> > Sent: Thursday, May 30, 2013 9:01 PM
>>>>>>> >> >> >>>>> >>>>> > Subject: Re: [AMBER] experiences with EVGA GTX
>>>>>>> TITAN
>>>>>>> >> >> >>>>> Superclocked
>>>>>>> >> >> >>>>> -
>>>>>>> >> >> >>>>> >>>>> > memtestG80 - UNDERclocking in Linux ?
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>> > Run cellulose nve for 100k iterations twice . If
>>>>>>> the
>>>>>>> >> >> final
>>>>>>> >> >> >>>>> >>>>> energies
>>>>>>> >> >> >>>>> >>>>> don't
>>>>>>> >> >> >>>>> >>>>> > match, you have a hardware issue. No need to
>>>>>>> play
>>>>>>> with
>>>>>>> >> >> ntpr
>>>>>>> >> >> >>>>> or
>>>>>>> >> >> >>>>> any
>>>>>>> >> >> >>>>> >>>>> other
>>>>>>> >> >> >>>>> >>>>> > variable.
>>>>>>> >> >> >>>>> >>>>> > On May 30, 2013 10:58 AM, <pavel.banas.upol.cz>
>>>>>>> wrote:
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > Dear all,
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > I would also like to share one of my experience
>>>>>>> with
>>>>>>> >> >> titan
>>>>>>> >> >> >>>>> >>>>> cards. We
>>>>>>> >> >> >>>>> >>>>> have
>>>>>>> >> >> >>>>> >>>>> > > one gtx titan card and with one system (~55k
>>>>>>> atoms,
>>>>>>> > NVT,
>>>>>>> >> >> >>>>> >>>>> RNA+waters)
>>>>>>> >> >> >>>>> >>>>> we
>>>>>>> >> >> >>>>> >>>>> > run
>>>>>>> >> >> >>>>> >>>>> > > into same troubles you are describing. I was
>>>>>>> also
>>>>>>> >> >> playing
>>>>>>> >> >> >>>>> with
>>>>>>> >> >> >>>>> >>>>> ntpr
>>>>>>> >> >> >>>>> >>>>> to
>>>>>>> >> >> >>>>> >>>>> > > figure out what is going on, step by step. I
>>>>>>> >> understand
>>>>>>> >> >> >>>>> that
>>>>>>> >> >> >>>>> the
>>>>>>> >> >> >>>>> >>>>> code
>>>>>>> >> >> >>>>> >>>>> is
>>>>>>> >> >> >>>>> >>>>> > > using different routines for calculation
>>>>>>> >> >> energies+forces or
>>>>>>> >> >> >>>>> only
>>>>>>> >> >> >>>>> >>>>> forces.
>>>>>>> >> >> >>>>> >>>>> > > The
>>>>>>> >> >> >>>>> >>>>> > > simulations of other systems are perfectly
>>>>>>> stable,
>>>>>>> >> >> running
>>>>>>> >> >> >>>>> for
>>>>>>> >> >> >>>>> >>>>> days
>>>>>>> >> >> >>>>> >>>>> and
>>>>>>> >> >> >>>>> >>>>> > > weeks. Only that particular system
>>>>>>> systematically
>>>>>>> >> ends
>>>>>>> >> >> up
>>>>>>> >> >> >>>>> with
>>>>>>> >> >> >>>>> >>>>> this
>>>>>>> >> >> >>>>> >>>>> > error.
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > However, there was one interesting issue. When
>>>>>>> I
>>>>>>> set
>>>>>>> >> >> >>>>> ntpr=1,
>>>>>>> >> >> >>>>> the
>>>>>>> >> >> >>>>> >>>>> error
>>>>>>> >> >> >>>>> >>>>> > > vanished (systematically in multiple runs) and
>>>>>>> the
>>>>>>> >> >> >>>>> simulation
>>>>>>> >> >> >>>>> was
>>>>>>> >> >> >>>>> >>>>> able to
>>>>>>> >> >> >>>>> >>>>> > > run for more than millions of steps (I was not
>>>>>>> let
>>>>>>> it
>>>>>>> >> >> >>>>> running
>>>>>>> >> >> >>>>> for
>>>>>>> >> >> >>>>> >>>>> weeks
>>>>>>> >> >> >>>>> >>>>> > as
>>>>>>> >> >> >>>>> >>>>> > > in the meantime I shifted that simulation to
>>>>>>> other
>>>>>>> >> card
>>>>>>> >> >> -
>>>>>>> >> >> >>>>> need
>>>>>>> >> >> >>>>> >>>>> data,
>>>>>>> >> >> >>>>> >>>>> not
>>>>>>> >> >> >>>>> >>>>> > > testing). All other setting of ntpr failed. As
>>>>>>> I
>>>>>>> read
>>>>>>> >> >> this
>>>>>>> >> >> >>>>> >>>>> discussion, I
>>>>>>> >> >> >>>>> >>>>> > > tried to set ene_avg_sampling=1 with some high
>>>>>>> value
>>>>>>> >> of
>>>>>>> >> >> >>>>> ntpr
>>>>>>> >> >> >>>>> (I
>>>>>>> >> >> >>>>> >>>>> expected
>>>>>>> >> >> >>>>> >>>>> > > that this will shift the code to permanently
>>>>>>> use
>>>>>>> the
>>>>>>> >> >> >>>>> >>>>> force+energies
>>>>>>> >> >> >>>>> >>>>> part
>>>>>>> >> >> >>>>> >>>>> > of
>>>>>>> >> >> >>>>> >>>>> > > the code, similarly to ntpr=1), but the error
>>>>>>> >> occurred
>>>>>>> >> >> >>>>> again.
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > I know it is not very conclusive for finding
>>>>>>> out
>>>>>>> what
>>>>>>> > is
>>>>>>> >> >> >>>>> >>>>> happening,
>>>>>>> >> >> >>>>> >>>>> at
>>>>>>> >> >> >>>>> >>>>> > > least
>>>>>>> >> >> >>>>> >>>>> > > not for me. Do you have any idea, why ntpr=1
>>>>>>> might
>>>>>>> > help?
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > best regards,
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > Pavel
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > --
>>>>>>> >> >> >>>>> >>>>> > > Pavel Banáš
>>>>>>> >> >> >>>>> >>>>> > > pavel.banas.upol.cz
>>>>>>> >> >> >>>>> >>>>> > > Department of Physical Chemistry,
>>>>>>> >> >> >>>>> >>>>> > > Palacky University Olomouc
>>>>>>> >> >> >>>>> >>>>> > > Czech Republic
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > ---------- Původní zpráva ----------
>>>>>>> >> >> >>>>> >>>>> > > Od: Jason Swails <jason.swails.gmail.com>
>>>>>>> >> >> >>>>> >>>>> > > Datum: 29. 5. 2013
>>>>>>> >> >> >>>>> >>>>> > > Předmět: Re: [AMBER] experiences with EVGA GTX
>>>>>>> TITAN
>>>>>>> >> >> >>>>> >>>>> Superclocked -
>>>>>>> >> >> >>>>> >>>>> > > memtestG
>>>>>>> >> >> >>>>> >>>>> > > 80 - UNDERclocking in Linux ?
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > "I'll answer a little bit:
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > NTPR=10 Etot after 2000 steps
>>>>>>> >> >> >>>>> >>>>> > > >
>>>>>>> >> >> >>>>> >>>>> > > > -443256.6711
>>>>>>> >> >> >>>>> >>>>> > > > -443256.6711
>>>>>>> >> >> >>>>> >>>>> > > >
>>>>>>> >> >> >>>>> >>>>> > > > NTPR=200 Etot after 2000 steps
>>>>>>> >> >> >>>>> >>>>> > > >
>>>>>>> >> >> >>>>> >>>>> > > > -443261.0705
>>>>>>> >> >> >>>>> >>>>> > > > -443261.0705
>>>>>>> >> >> >>>>> >>>>> > > >
>>>>>>> >> >> >>>>> >>>>> > > > Any idea why energies should depend on
>>>>>>> frequency
>>>>>>> of
>>>>>>> >> >> >>>>> energy
>>>>>>> >> >> >>>>> >>>>> records
>>>>>>> >> >> >>>>> >>>>> > (NTPR)
>>>>>>> >> >> >>>>> >>>>> > > ?
>>>>>>> >> >> >>>>> >>>>> > > >
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > It is a subtle point, but the answer is
>>>>>>> 'different
>>>>>>> >> code
>>>>>>> >> >> >>>>> paths.'
>>>>>>> >> >> >>>>> >>>>> In
>>>>>>> >> >> >>>>> >>>>> > > general, it is NEVER necessary to compute the
>>>>>>> actual
>>>>>>> >> >> energy
>>>>>>> >> >> >>>>> of a
>>>>>>> >> >> >>>>> >>>>> molecule
>>>>>>> >> >> >>>>> >>>>> > > during the course of standard molecular
>>>>>>> dynamics
>>>>>>> (by
>>>>>>> >> >> >>>>> analogy, it
>>>>>>> >> >> >>>>> >>>>> is
>>>>>>> >> >> >>>>> >>>>> NEVER
>>>>>>> >> >> >>>>> >>>>> > > necessary to compute atomic forces during the
>>>>>>> course
>>>>>>> >> of
>>>>>>> >> >> >>>>> random
>>>>>>> >> >> >>>>> >>>>> Monte
>>>>>>> >> >> >>>>> >>>>> > Carlo
>>>>>>> >> >> >>>>> >>>>> > > sampling).
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > For performance's sake, then, pmemd.cuda
>>>>>>> computes
>>>>>>> >> only
>>>>>>> >> >> the
>>>>>>> >> >> >>>>> force
>>>>>>> >> >> >>>>> >>>>> when
>>>>>>> >> >> >>>>> >>>>> > > energies are not requested, leading to a
>>>>>>> different
>>>>>>> >> >> order of
>>>>>>> >> >> >>>>> >>>>> operations
>>>>>>> >> >> >>>>> >>>>> > for
>>>>>>> >> >> >>>>> >>>>> > > those runs. This difference ultimately causes
>>>>>>> >> >> divergence.
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > To test this, try setting the variable
>>>>>>> >> >> ene_avg_sampling=10
>>>>>>> >> >> >>>>> in
>>>>>>> >> >> >>>>> the
>>>>>>> >> >> >>>>> >>>>> &cntrl
>>>>>>> >> >> >>>>> >>>>> > > section. This will force pmemd.cuda to compute
>>>>>>> >> energies
>>>>>>> >> >> >>>>> every 10
>>>>>>> >> >> >>>>> >>>>> steps
>>>>>>> >> >> >>>>> >>>>> > > (for energy averaging), which will in turn make
>>>>>>> the
>>>>>>> >> >> >>>>> followed
>>>>>>> >> >> >>>>> code
>>>>>>> >> >> >>>>> >>>>> path
>>>>>>> >> >> >>>>> >>>>> > > identical for any multiple-of-10 value of ntpr.
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > > --
>>>>>>> >> >> >>>>> >>>>> > > Jason M. Swails
>>>>>>> >> >> >>>>> >>>>> > > Quantum Theory Project,
>>>>>>> >> >> >>>>> >>>>> > > University of Florida
>>>>>>> >> >> >>>>> >>>>> > > Ph.D. Candidate
>>>>>>> >> >> >>>>> >>>>> > > 352-392-4032
>>>>>>> >> >> >>>>> >>>>> > > ______________________________**
>>>>>>> **_________________
>>>>>>> >> >> >>>>> >>>>> > > AMBER mailing list
>>>>>>> >> >> >>>>> >>>>> > > AMBER.ambermd.org
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
>>>>>>> mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>> <
>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >
>>>>>>> >> >> >>>>> >>>>> "
>>>>>>> >> >> >>>>> >>>>> > > ______________________________**
>>>>>>> **_________________
>>>>>>> >> >> >>>>> >>>>> > > AMBER mailing list
>>>>>>> >> >> >>>>> >>>>> > > AMBER.ambermd.org
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
>>>>>>> mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>> <
>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >
>>>>>>> >> >> >>>>> >>>>> > >
>>>>>>> >> >> >>>>> >>>>> > ______________________________**
>>>>>>> **_________________
>>>>>>> >> >> >>>>> >>>>> > AMBER mailing list
>>>>>>> >> >> >>>>> >>>>> > AMBER.ambermd.org
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
>>>>>>> mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>> <
>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >
>>>>>>> >> >> >>>>> >>>>> > ______________________________**
>>>>>>> **_________________
>>>>>>> >> >> >>>>> >>>>> > AMBER mailing list
>>>>>>> >> >> >>>>> >>>>> > AMBER.ambermd.org
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
>>>>>>> mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>> <
>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >
>>>>>>> >> >> >>>>> >>>>> >
>>>>>>> >> >> >>>>> >>>>>
>>>>>>> >> >> >>>>> >>>>>
>>>>>>> >> >> >>>>> >>>>>
>>>>>>> >> >> >>>>> >>>>> --
>>>>>>> >> >> >>>>> >>>>> Jason M. Swails
>>>>>>> >> >> >>>>> >>>>> Quantum Theory Project,
>>>>>>> >> >> >>>>> >>>>> University of Florida
>>>>>>> >> >> >>>>> >>>>> Ph.D. Candidate
>>>>>>> >> >> >>>>> >>>>> 352-392-4032
>>>>>>> >> >> >>>>> >>>>> ______________________________**
>>>>>>> **_________________
>>>>>>> >> >> >>>>> >>>>> AMBER mailing list
>>>>>>> >> >> >>>>> >>>>> AMBER.ambermd.org
>>>>>>> >> >> >>>>> >>>>> http://lists.ambermd.org/****
>>>>>>> mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>> <
>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >
>>>>>>> >> >> >>>>> >>>>>
>>>>>>> >> >> >>>>> >>>>>
>>>>>>> >> >> >>>>> >>>> ______________________________**
>>>>>>> **_________________
>>>>>>> >> >> >>>>> >>> AMBER mailing list
>>>>>>> >> >> >>>>> >>> AMBER.ambermd.org
>>>>>>> >> >> >>>>> >>> http://lists.ambermd.org/****mailman/listinfo/amber<http://lists.ambermd.org/**mailman/listinfo/amber>
>>>>>>> <
>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >
>>>>>>> >> >> >>>>> >>>
>>>>>>> >> >> >>>>> >>> __________ Informace od ESET NOD32 Antivirus, verze
>>>>>>> >> databaze
>>>>>>> >> >> 8394
>>>>>>> >> >> >>>>> >>> (20130530) __________
>>>>>>> >> >> >>>>> >>>
>>>>>>> >> >> >>>>> >>> Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>> >> >> >>>>> >>>
>>>>>>> >> >> >>>>> >>> http://www.eset.cz
>>>>>>> >> >> >>>>> >>>
>>>>>>> >> >> >>>>> >>>
>>>>>>> >> >> >>>>> >>>
>>>>>>> >> >> >>>>> >>>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >> --
>>>>>>> >> >> >>>>> >> Tato zpráva byla vytvořena převratným poštovním
>>>>>>> klientem
>>>>>>> > Opery:
>>>>>>> >> >> >>>>> >> http://www.opera.com/mail/
>>>>>>> >> >> >>>>> >> ______________________________**_________________
>>>>>>> >> >> >>>>> >> AMBER mailing list
>>>>>>> >> >> >>>>> >> AMBER.ambermd.org
>>>>>>> >> >> >>>>> >> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> >>
>>>>>>> >> >> >>>>> > ______________________________**_________________
>>>>>>> >> >> >>>>> > AMBER mailing list
>>>>>>> >> >> >>>>> > AMBER.ambermd.org
>>>>>>> >> >> >>>>> > http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >> >> >>>>> >
>>>>>>> >> >> >>>>> > __________ Informace od ESET NOD32 Antivirus, verze
>>>>>>> databaze
>>>>>>> >> >> 8394
>>>>>>> >> >> >>>>> > (20130530) __________
>>>>>>> >> >> >>>>> >
>>>>>>> >> >> >>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>> >> >> >>>>> >
>>>>>>> >> >> >>>>> > http://www.eset.cz
>>>>>>> >> >> >>>>> >
>>>>>>> >> >> >>>>> >
>>>>>>> >> >> >>>>> >
>>>>>>> >> >> >>>>>
>>>>>>> >> >> >>>>>
>>>>>>> >> >> >>>>> --
>>>>>>> >> >> >>>>> Tato zpráva byla vytvořena převratným poštovním klientem
>>>>>>> Opery:
>>>>>>> >> >> >>>>> http://www.opera.com/mail/
>>>>>>> >> >> >>>>>
>>>>>>> >> >> >>>>> ______________________________**_________________
>>>>>>> >> >> >>>>> AMBER mailing list
>>>>>>> >> >> >>>>> AMBER.ambermd.org
>>>>>>> >> >> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >> >> >>>>>
>>>>>>> >> >> >>>>
>>>>>>> >> >> >>>>
>>>>>>> >> >> >>> ______________________________**_________________
>>>>>>> >> >> >>> AMBER mailing list
>>>>>>> >> >> >>> AMBER.ambermd.org
>>>>>>> >> >> >>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >> >> >>>
>>>>>>> >> >> >>> __________ Informace od ESET NOD32 Antivirus, verze
>>>>>>> databaze
>>>>>>> 8395
>>>>>>> >> >> >>> (20130531) __________
>>>>>>> >> >> >>>
>>>>>>> >> >> >>> Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>> >> >> >>>
>>>>>>> >> >> >>> http://www.eset.cz
>>>>>>> >> >> >>>
>>>>>>> >> >> >>>
>>>>>>> >> >> >>>
>>>>>>> >> >> >>
>>>>>>> >> >> >>
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >>
>>>>>>> >> >>
>>>>>>> >> >> --
>>>>>>> >> >> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
>>>>>>> >> >> http://www.opera.com/mail/
>>>>>>> >> >>
>>>>>>> >> >> ______________________________**_________________
>>>>>>> >> >> AMBER mailing list
>>>>>>> >> >> AMBER.ambermd.org
>>>>>>> >> >> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >> >>
>>>>>>> >> >
>>>>>>> >> >
>>>>>>> >> >
>>>>>>> >> > __________ Informace od ESET NOD32 Antivirus, verze databaze
>>>>>>> 8397
>>>>>>> >> > (20130531) __________
>>>>>>> >> >
>>>>>>> >> > Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>> >> >
>>>>>>> >> > GB_out_plus_diff_Files.tar.gz - poskozeny archiv
>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>> >> GB_out_plus_diff_Files.tar
>>>>>>> >> > - poskozeny archiv
>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
>>>>>>> GB_out_plus_diff_Files.tar.gz -
>>>>>>> >> > poskozeny archiv
>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
>>>>>>> GB_out_plus_diff_Files.tar.gz >
>>>>>>> >> GZIP
>>>>>>> >> > > GB_out_plus_diff_Files.tar - poskozeny archiv
>>>>>>> >> > GB_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>> >> > GB_out_plus_diff_Files.tar > TAR >
>>>>>>> GB_out_plus_diff_Files.tar.gz >
>>>>>>> >> GZIP
>>>>>>> >> > > GB_out_plus_diff_Files.tar > TAR >
>>>>>>> GB_nucleosome-sim3.mdout-full -
>>>>>>> >> > vyskytl se problem pri cteni archivu
>>>>>>> >> > PME_out_plus_diff_Files.tar.gz - poskozeny archiv
>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>> >> > PME_out_plus_diff_Files.tar - poskozeny archiv
>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>> >> > PME_out_plus_diff_Files.tar > TAR >
>>>>>>> PME_out_plus_diff_Files.tar.gz -
>>>>>>> >> > poskozeny archiv
>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>> >> > PME_out_plus_diff_Files.tar > TAR >
>>>>>>> PME_out_plus_diff_Files.tar.gz >
>>>>>>> >> > GZIP > PME_out_plus_diff_Files.tar - poskozeny archiv
>>>>>>> >> > PME_out_plus_diff_Files.tar.gz > GZIP >
>>>>>>> >> > PME_out_plus_diff_Filestar > TAR >
>>>>>>> PME_out_plus_diff_Files.tar.gz
>>>>>>> >
>>>>>>> >> GZIP
>>>>>>> >> > > PME_out_plus_diff_Files.tar > TAR >
>>>>>>> >> > PME_JAC_production_NPT-sim3.**mdout-full - vyskytl se problem
>>>>>>> pri
>>>>>>> cteni
>>>>>>> >> > archivu
>>>>>>> >> >
>>>>>>> >> > http://www.eset.cz
>>>>>>> >> >
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> --
>>>>>>> >> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
>>>>>>> >> http://www.opera.com/mail/
>>>>>>> >>
>>>>>>> >> ______________________________**_________________
>>>>>>> >> AMBER mailing list
>>>>>>> >> AMBER.ambermd.org
>>>>>>> >> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> > ______________________________**_________________
>>>>>>> > AMBER mailing list
>>>>>>> > AMBER.ambermd.org
>>>>>>> > http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>> >
>>>>>>> > __________ Informace od ESET NOD32 Antivirus, verze databaze 8398
>>>>>>> > (20130531) __________
>>>>>>> >
>>>>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>>> >
>>>>>>> > http://www.eset.cz
>>>>>>> >
>>>>>>> >
>>>>>>> >
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
>>>>>>> http://www.opera.com/mail/
>>>>>>>
>>>>>>> ______________________________**_________________
>>>>>>> AMBER mailing list
>>>>>>> AMBER.ambermd.org
>>>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<http://lists.ambermd.org/mailman/listinfo/amber>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> __________ Informace od ESET NOD32 Antivirus, verze databaze 8401
>>>>> (20130601) __________
>>>>>
>>>>> Tuto zpravu proveril ESET NOD32 Antivirus.
>>>>>
>>>>> http://www.eset.cz
>>>>>
>>>>>
>>>>
>>>>
>>>
>>> --
>>> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
>>> http://www.opera.com/mail/
>>>
>>> _______________________________________________
>>> AMBER mailing list
>>> AMBER.ambermd.org
>>> http://lists.ambermd.org/mailman/listinfo/amber
>>>
>>>
>>
>
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Sun Jun 02 2013 - 10:00:03 PDT
Custom Search