Sorry, I missed the attachment because I'm on a cellphone and on a business
trip. Anyway, neither Titan showed fully deterministic behavior and that
is worrisome. Notice that the 680 runs were indeed so. The latter is the
expected behavior and exactly what I see with one of my Titans and my K20.
Which means we need to figure this out. For now, could you take it on
faith that changing ntpr changes the trajectory by changing the code
executed and that it doing so is not a bug? Playing around with it is just
confusing the issue right now.
What would help clarify is if someone could try these tests on K20 or
K20X. I would love for someone to demonstrate this is a coding error on
my part because I can fix that. The evidence just isn't leading me that
way right now.
Scott
On May 29, 2013 2:41 PM, "Marek Maly" <marek.maly.ujep.cz> wrote:
> Hi Scott,
>
> what do you mean by "try running for 100k steps before comparing
> energies". In all
> tests I have actually done I did exactly 100k steps before comparing
> energies
> E_tot(at step 100 000). So you mean to extend tests to 200k steps now ?
>
> M.
>
>
> Dne Wed, 29 May 2013 22:46:58 +0200 Scott Le Grand <varelse2005.gmail.com>
> napsal/-a:
>
> > Ps try running for 100k steps before comparing energies and I suspect no
> > two simulations will match.
> > On May 29, 2013 1:41 PM, "Scott Le Grand" <varelse2005.gmail.com> wrote:
> >
> >> Your Titan setup is hosed. Your results were not 100% deterministic for
> >> the same inputs.
> >>
> >> Energies + Forces use a different subroutine than just Forces hence the
> >> ntpr dependence. Hence changing ntpr effectively is changing the input.
> >>
> >> It's 100% ironclad reproducibility that matters and you demonstrated
> >> it's
> >> not happening.
> >> On May 29, 2013 1:30 PM, "Marek Maly" <marek.maly.ujep.cz> wrote:
> >>
> >>> Hi all,
> >>>
> >>> First of all thanks to Ross for his update ! although it is question
> >>> if it helps to solve all the reported Amber issues with Titan/OC Titan
> >>> GPUs .
> >>> So let's see and hope :))
> >>>
> >>> Here are my results - see the attached TXT file with tables where
> >>> the results from the tests are summarised. I did twice the same
> >>> Amber benchmark tests on each GPU (both titans, GTX 680 and GTX 580)
> >>> to see reproducibility of the results after 100K steps at ig=default
> >>> (so ig not present in mdin file).
> >>>
> >>> The first table contains ns/day estimates obtained for each molecular
> >>> system
> >>> for each TITAN GPU. Interestingly estimates obtained for the same
> >>> system
> >>> in different
> >>> round slightly differ, but maybe that's OK.
> >>>
> >>> In the second table there are values of the total energy after 100k
> >>> steps
> >>> to check
> >>> reproducibility of the results.
> >>>
> >>> Here is summarisation :
> >>>
> >>> #1 - simulation crashes on TITANs
> >>>
> >>> Interestingly there was just one simulation crash in JAC_NPT (TITAN_0,
> >>> ROUND_1) the remaining
> >>> 3 TITAN JAC_NPT simulations were finished. There were also 3 times
> >>> crashes in CELLULOSE_NVE
> >>> test but the last simulation (TITAN_1,ROUND_2) was finished without any
> >>> problem. All the remaining
> >>> simulations were always finished without any problem. So the simulation
> >>> crashes seem to be
> >>> not-reproducible/unpredictible on some moleacular systems/(mdin
> >>> setups).
> >>>
> >>> CRASH ERRORS:
> >>>
> >>> a) JAC_NPT (TITAN_0, ROUND_1)
> >>> Here 11k steps were successfully done before crash, I found this error
> >>> in mdout file:
> >>>
> >>> | ERROR: max pairlist cutoff must be less than unit cell max sphere
> >>> radius!
> >>>
> >>> b) CELLULOSE_NVE (TITAN_0, ROUND_1, ROUND_2; TITAN_1, ROUND_1 )
> >>> Here I did not find any error in mdout file. Just this error was
> >>> written
> >>> on standard output
> >>> (screen/nohup.out file):
> >>>
> >>> ------
> >>> Error: unspecified launch failure launching kernel kNLSkinTest
> >>> cudaFree GpuBuffer::Deallocate failed unspecified launch failure
> >>> grep: mdinfo.1GTX_TITAN: No such file or directory
> >>> -----
> >>>
> >>> in all three cases.
> >>>
> >>> Here on CELLULOSE_NVE case I started to play with NTPR parameter
> >>> (originally just
> >>> on TITAN-0 GPU), to see how many steps were successfully done here
> >>> before
> >>> crash, then this my
> >>> small research started to be more interesting than I ever thought :))
> >>> see
> >>> here
> >>> chronologically my results for E_tot after 2000 steps for different
> >>> GPUs
> >>> (machines) - I repeated calculation several times for the given NTPR
> >>> just
> >>> to be sure.
> >>>
> >>> TITAN-0, Etot after 2000 steps
> >>>
> >>> NTPR=10
> >>>
> >>> -443256.6867
> >>> -443256.6867
> >>> -443256.6867
> >>>
> >>> NTPR=100
> >>>
> >>> -443250.1350
> >>> -443250.1350
> >>> -443250.1350
> >>>
> >>> NTPR=200
> >>>
> >>> -443261.0705
> >>> -443261.0705
> >>> -443072.3097
> >>> -443261.0705
> >>> -443261.0705
> >>> -443261.0705
> >>> -443261.0705
> >>>
> >>> NTPR=10 (again just to verify)
> >>>
> >>> -443256.6867
> >>> -443256.6867
> >>>
> >>>
> >>> Then I tried with TITAN-1
> >>>
> >>> NTPR=10
> >>>
> >>> -443256.6867
> >>> -443256.6867
> >>>
> >>> NTPR=100
> >>>
> >>> -443250.1350
> >>> -443250.1350
> >>>
> >>> NTPR=200
> >>>
> >>> -443261.0705
> >>> -443261.0705
> >>>
> >>>
> >>> Then I tried with GTX-580
> >>>
> >>> NTPR=10
> >>>
> >>> -443256.6867
> >>> -443256.6867
> >>>
> >>> NTPR=200
> >>>
> >>> -443261.0705
> >>> -443261.0705
> >>>
> >>> then I tried with GTX-680
> >>>
> >>> NTPR=10 Etot after 2000 steps
> >>>
> >>> -443256.6711
> >>> -443256.6711
> >>>
> >>> NTPR=200 Etot after 2000 steps
> >>>
> >>> -443261.0705
> >>> -443261.0705
> >>>
> >>> Any idea why energies should depend on frequency of energy records
> >>> (NTPR)
> >>> ?
> >>>
> >>>
> >>>
> >>> #2 - reproducibility on TITANs (see attached table.txt)
> >>>
> >>> Also here are differences depending on concrete systems/setups.
> >>> While in case of FACTOR_IX_NVE, FACTOR_IX_NPT, TRPCAGE, MYOGLOBIN
> >>> systems
> >>> I have obtained
> >>> 100% reproducibility (the results for the given system were identical
> >>> for
> >>> both cards/all ROUNDs)
> >>> on systems JAC_NVE, JAC_NPT, NUCLEOSOME I obtained small differences
> >>> in
> >>> general however in case
> >>> of TITAN_1 GPU also NUCLEOSOME results were 100% reproducible.
> >>> Moreover
> >>> for the TITAN_1 card which succeeded to finish CELLULOSE test at least
> >>> in
> >>> ROUND_2 I did 3rd additional round and I got the identical result as
> >>> from
> >>> the ROUND_2 (i.e. -443246.3206 ) so regarding TITAN_1 GPU I can say
> >>> that
> >>> it is able to 100% reproduce 100k steps CELLULOSE_NVE test result
> >>> perhaps
> >>> on all eventually successfully finished runs :))
> >>>
> >>>
> >>> #3 - GTX-580, GTX-680 controls
> >>>
> >>> Here the simulations were done without any problems and were 100%
> >>> reproducible on each card however
> >>> the results for the given system slightly differ between those two
> >>> cards
> >>> with exception of the
> >>> CELLULOSE system where both cards GTX-580, GTX-680 provided identical
> >>> result which is moreover
> >>> nearly identical with result obtained with TITAN_1 during ROUND_2
> >>> (relative difference 2e-6).
> >>>
> >>>
> >>> TO ET:
> >>> a)
> >>> I had no problems with minimization stages in my own simul. bigger than
> >>> 100k which crashed
> >>> during heat NVT phase.
> >>>
> >>> b)
> >>> 313.30 driver ??? OK so after 319.23 I will try experiment with this a
> >>> bit "outdated" version :))
> >>> Actually I am working under 319.17. (and CUDA 5.0)
> >>>
> >>> c)
> >>> Can you please do at least JAC_NPT, JAC_NVE, NUCLEOSOME and
> >>> CELLULOSE_NVE
> >>> tests using 100 000 steps
> >>> (same random seed e.g. default = ig deleted from mdin if is there)
> >>> twice
> >>> to confirm 100% reproducibility on your TITAN GPU ?
> >>>
> >>> TO Divi:
> >>>
> >>> This is also my usual approach to divide whole simulation into many
> >>> subtrajectories (in my case 0.5 ns = 250k 2fs steps) but it does not
> >>> seem
> >>> to help here it self. Can you please also do the same tests which I
> >>> asked
> >>> ET (point c) )
> >>>
> >>>
> >>> BTW CUDA release candidate 5.5 was just released (
> >>> https://developer.nvidia.com/**cuda-toolkit<
> https://developer.nvidia.com/cuda-toolkit>)
> >>> would it be reasonable idea to try compile/run pmemd.cuda with this
> >>> brand
> >>> new cuda version ?
> >>>
> >>> Thanks !
> >>>
> >>> Best wishes,
> >>>
> >>> Marek
> >>>
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> Dne Wed, 29 May 2013 03:44:33 +0200 Ross Walker <ross.rosswalker.co.uk
> >
> >>> napsal/-a:
> >>>
> >>> Hi All,
> >>>>
> >>>> Just an update that we will have some fixes out soon that address some
> >>>> errors we have been noticing with simulations crashing during NPT
> >>>> runs.
> >>>> It
> >>>> is possible that this is confusing the issue here as to whether the
> >>>> problem is related to the GTX Titan or to a possible bug in the code.
> >>>> I
> >>>> hope to have the patch released within a few days at which point it
> >>>> would
> >>>> be good to repeat these tests and then hopefully we can try to track
> >>>> down
> >>>> what is going on. I find it hard to believe that so many cards are
> >>>> faulty
> >>>> so I suspect that there may be something funky in the code with
> >>>> regards
> >>>> to
> >>>> GTX Titans. We'll try and get it fixed as soon as possible but for now
> >>>> please just wait until we get the update released for AMBER 12 in a
> >>>> few
> >>>> days and see if that helps at all.
> >>>>
> >>>> All the best
> >>>> Ross
> >>>>
> >>>>
> >>>> On 5/28/13 5:12 PM, "Divi/GMAIL" <dvenkatlu.gmail.com> wrote:
> >>>>
> >>>> I have two TITANs in my Gigabyte workstation. I have had similar
> >>>>> issues
> >>>>> of NANs for some of the simulation setups. Never could figure out why
> >>>>> the
> >>>>> simulations failed for no reason. I tried 10, 12 ang. box sizes.
> >>>>> same
> >>>>> random breakdowns. Thought of returning them suspecting memory
> >>>>> errors.
> >>>>> But
> >>>>> some simulations ran perfectly fine. Currently running two
> >>>>> calculations
> >>>>> without any problems. Both are running pretty stable for over 100ns.
> >>>>> I
> >>>>> suspect AMBER CUDA code may have some issues under some simulation
> >>>>> conditions such as NPT. In general, NVT setup is more successful than
> >>>>> NPT,
> >>>>> in my case.
> >>>>>
> >>>>> These are 287426 atoms simulation on one card (9 ns/day)
> >>>>> On other card: 129049 atom setup (20 ns/day)
> >>>>>
> >>>>> Both using same NVT setup. (AMBER12/INTEL-12.x
> >>>>> compilers/CentOS-6.3/Drivers 319.17/CUDA5.0)
> >>>>>
> >>>>> Input is below:
> >>>>> &cntrl
> >>>>> nstlim=500000, dt=0.002,
> >>>>> ntx=5, irest=1, ig=-1,
> >>>>> ntpr=1000, ntwr=10000, ntwx=10000,
> >>>>> ntt=1, tautp=2, ntb=1, ntp=0, ntc=2, ntf=2,
> >>>>> iwrap=1, ioutfm=1, ntxo=2,
> >>>>> &end
> >>>>>
> >>>>> One suggestion If I may add: If you could run short simulations for
> >>>>> no
> >>>>> more
> >>>>> than 500,000 steps (or 1ns with 2 fs), you might find some stability.
> >>>>> Again,
> >>>>> not scientific rationale from my side. But it worked in some cases
> >>>>> for
> >>>>> me.
> >>>>>
> >>>>> This is self-assembled system with GIGABYTE GA-Z77X-UP7 (with core
> >>>>> i5
> >>>>> processor) and 1200W PS/16GB memory.
> >>>>>
> >>>>>
> >>>>> Best regards
> >>>>> Divi
> >>>>>
> >>>>>
> >>>>>
> >>>>> -----Original Message-----
> >>>>> From: Scott Le Grand
> >>>>> Sent: Tuesday, May 28, 2013 4:46 PM
> >>>>> To: AMBER Mailing List
> >>>>> Subject: Re: [AMBER] experiences with EVGA GTX TITAN Superclocked -
> >>>>> memtestG80 - UNDERclocking in Linux ?
> >>>>>
> >>>>> You can play Russian Roulette a whole bunch of rounds without blowing
> >>>>> your
> >>>>> head off.
> >>>>>
> >>>>> Similarly, when you have a GPU that occasionally flips a bit the
> >>>>> wrong
> >>>>> way,
> >>>>> most of the time it will be some low order perturbation to the
> >>>>> coordinates
> >>>>> that does little more than make the trajectory nondeterministic...
> >>>>> Except
> >>>>> when it doesn't...
> >>>>>
> >>>>> You can't even detect this kind of misbehavior in GROMACS, ACEMD, or
> >>>>> NAMD
> >>>>> because *none* of them (to my knowledge) are capable of producing
> >>>>> deterministic output at production-level performance.
> >>>>>
> >>>>> Titans and 680s are consumer cards. I love them to death, but if
> >>>>> you're
> >>>>> going to do production work with them, you need to qual them
> >>>>> thoroughly
> >>>>> before proceeding or you need to pay up and use Teslas instead. I'd
> >>>>> still
> >>>>> build a cluster with Titans myself, but I'd ruthlessly RMA them
> >>>>> until I
> >>>>> got
> >>>>> satisfaction if they couldn't pass a test consisting of running an
> >>>>> AMBER
> >>>>> simulation for 100K iterations without either crashing or producing a
> >>>>> nondeterministic result. The customer is always right.
> >>>>>
> >>>>>
> >>>>> On Tue, May 28, 2013 at 1:20 PM, Marek Maly <marek.maly.ujep.cz>
> >>>>> wrote:
> >>>>>
> >>>>> I would wait for the results of my GOPU0, GPU1 double tests before
> >>>>>> any serious conclusions.
> >>>>>>
> >>>>>> BTW what exactly means "GPU is hosed" ? Something like GPU is
> >>>>>> damaged
> >>>>>> or
> >>>>>> so ?
> >>>>>>
> >>>>>> Also would be strange (not probable) to buy 2 somehow damaged GPUs
> >>>>>> (even
> >>>>>> in the same way).
> >>>>>>
> >>>>>> As I wrote, memtestG80 tests were negative on both cards, if
> >>>>>> moreover
> >>>>>> both cards will perfectly reproduce both repetitions of the Amber
> >>>>>> benchmarks
> >>>>>> and eventually pass some another GPU tests (can you recommend any
> >>>>>> except
> >>>>>> memtestG80 ?)
> >>>>>> I still believe that the GPU cards are OK (also thank to particular
> >>>>>> successes in my Amb. simulations and actual A. benchmarks). So
> >>>>>> maybe I
> >>>>>> will eventually try downclock, but there might be some another
> >>>>>> variables,
> >>>>>> e.g. driver, OS, motherboard (I will perhaps test one card in
> >>>>>> another
> >>>>>> MB
> >>>>>> just to be sure, that problem is not MB based) etc. that's why I
> >>>>>> asked
> >>>>>> before that guy "ET" for the info about driver version, would be
> >>>>>> also
> >>>>>> interesting OS info or MB.
> >>>>>>
> >>>>>> M.
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>> Dne Tue, 28 May 2013 22:13:36 +0200 Scott Le Grand
> >>>>>> <varelse2005.gmail.com>
> >>>>>> napsal/-a:
> >>>>>>
> >>>>>> > Marek,
> >>>>>> > Your GPU is hosed. I don't have anything else to add. I'm not
> >>>>>> going
> >>>>>> to
> >>>>>> > go
> >>>>>> > snark hunting for a bug that doesn't exist.
> >>>>>> >
> >>>>>> >
> >>>>>> >
> >>>>>> > On Tue, May 28, 2013 at 12:24 PM, Marek Maly <marek.maly.ujep.cz>
> >>>>>> wrote:
> >>>>>> >
> >>>>>> >> Hi, just for the curiosity which driver are you using
> >>>>>> >> on that machine with perfectly working with OC TITAN,
> >>>>>> >> 319.17 or some more actual e.g. 319.23 ?
> >>>>>> >>
> >>>>>> >> RMA is a good idea but it could be also long time story and
> >>>>>> >> also to succeed here you need to have strong arguments
> >>>>>> >> especially if you are going to RMA two OC TITANs.
> >>>>>> >>
> >>>>>> >> I am not sure if my arguments "The cards have problems with some
> >>>>>> Amber
> >>>>>> >> calculations"
> >>>>>> >> would be strong enough here. Would be much better to have clear
> >>>>>> results
> >>>>>> >> from
> >>>>>> >> respected GPU tests and as it seems you may do extensive GPU
> >>>>>> tests
> >>>>>> also
> >>>>>> >> with
> >>>>>> >> multiple routines without any errors but still have problems with
> >>>>>> >> particular
> >>>>>> >> Amber simulations...
> >>>>>> >>
> >>>>>> >> BTW I am now doing Amber benchmarks with nstlim=100K and
> >>>>>> ig=default
> >>>>>> for
> >>>>>> >> each card
> >>>>>> >> twice. The tests will be done in cca 3 hours (due to slow
> >>>>>> nucleosome
> >>>>>> GB
> >>>>>> >> test).
> >>>>>> >>
> >>>>>> >> But even now I have interesting results from the first test on
> >>>>>> GPU0
> >>>>>> >> (nucleosome is still running) see below.
> >>>>>> >>
> >>>>>> >> As you can see JAC_NPT crashed around 11000 step, here is the
> >>>>>> last
> >>>>>> >> md.out
> >>>>>> >> record:
> >>>>>> >>
> >>>>>> >> *********
> >>>>>> >>
> >>>>>> >>
> >>>>>>
> >>>>>> ------------------------------**------------------------------**
> >>>>>> -------------
> >>>>>> -----
> >>>>>> >>
> >>>>>> >> check COM velocity, temp: 0.000021 0.00(Removed)
> >>>>>> >>
> >>>>>> >> NSTEP = 11000 TIME(PS) = 28.000 TEMP(K) = 300.39
> >>>>>> PRESS
> >>>>>> >> =
> >>>>>> >> -9.4
> >>>>>> >> Etot = -58092.8958 EKtot = 14440.2520 EPtot =
> >>>>>> >> -72533.1478
> >>>>>> >> BOND = 443.3912 ANGLE = 1253.5177 DIHED =
> >>>>>> >> 970.1275
> >>>>>> >> 1-4 NB = 567.2497 1-4 EEL = 6586.9007 VDWAALS =
> >>>>>> >> 8664.9960
> >>>>>> >> EELEC = -91019.3306 EHBOND = 0.0000 RESTRAINT =
> >>>>>> >> 0.0000
> >>>>>> >> EKCMT = 6274.0354 VIRIAL = 6321.9969 VOLUME =
> >>>>>> >> 236141.9494
> >>>>>> >> Density =
> >>>>>> >> 1.0162
> >>>>>> >>
> >>>>>> >>
> >>>>>>
> >>>>>> ------------------------------**------------------------------**
> >>>>>> -------------
> >>>>>> -----
> >>>>>> >>
> >>>>>> >> | ERROR: max pairlist cutoff must be less than unit cell max
> >>>>>> sphere
> >>>>>> >> radius!
> >>>>>> >>
> >>>>>> >> ********
> >>>>>> >>
> >>>>>> >> Any idea about that ERROR ?
> >>>>>> >>
> >>>>>> >> On the other hand FACTOR_IX_NPT which has much more atoms passed
> >>>>>> >> without
> >>>>>> >> any issue.
> >>>>>> >>
> >>>>>> >> Cellulose crashed on the beginning without any ERROR message in
> >>>>>> md.out
> >>>>>> >> file.
> >>>>>> >>
> >>>>>> >>
> >>>>>> >> I am very curious regarding exact reproducibility of the results
> >>>>>> at
> >>>>>> >> least
> >>>>>> >> in the
> >>>>>> >> framework of both tests on individual cards.
> >>>>>> >>
> >>>>>> >> BTW regarding eventual downclocking, has anyone idea about some
> >>>>>> NVclock
> >>>>>> >> alternative or
> >>>>>> >> I will be really eventually forced to edit frequency value in GPU
> >>>>>> BIOS
> >>>>>> >> ?
> >>>>>> >>
> >>>>>> >> Best,
> >>>>>> >>
> >>>>>> >> Marek
> >>>>>> >>
> >>>>>> >> HERE ARE THE FIRST DATA FROM MY 2x2 Bench tests
> >>>>>> >>
> >>>>>> >> JAC_PRODUCTION_NVE - 23,558 atoms PME
> >>>>>> >> ------------------------------**-------
> >>>>>> >>
> >>>>>> >> 1 x GTX_TITAN: | ns/day = 115.91
> >>>>>> seconds/ns =
> >>>>>> >> 745.39
> >>>>>> >>
> >>>>>> >> JAC_PRODUCTION_NPT - 23,558 atoms PME
> >>>>>> >> ------------------------------**-------
> >>>>>> >>
> >>>>>> >> 1 x GTX_TITAN: STOP PMEMD Terminated Abnormally!
> >>>>>> >> | ns/day = 90.72 seconds/ns = 952.42
> >>>>>> >>
> >>>>>> >> FACTOR_IX_PRODUCTION_NVE - 90,906 atoms PME
> >>>>>> >> ------------------------------**-------------
> >>>>>> >>
> >>>>>> >> 1 x GTX_TITAN: | ns/day = 30.56
> >>>>>> seconds/ns =
> >>>>>> >> 2827.33
> >>>>>> >>
> >>>>>> >> FACTOR_IX_PRODUCTION_NPT - 90,906 atoms PME
> >>>>>> >> ------------------------------**-------------
> >>>>>> >>
> >>>>>> >> 1 x GTX_TITAN: | ns/day = 25.01
> >>>>>> seconds/ns =
> >>>>>> >> 3454.56
> >>>>>> >>
> >>>>>> >> CELLULOSE_PRODUCTION_NVE - 408,609 atoms PME
> >>>>>> >> ------------------------------**--------------
> >>>>>> >>
> >>>>>> >> 1 x GTX_TITAN: Error: unspecified launch failure
> >>>>>> launching
> >>>>>> >> kernel
> >>>>>> >> kNLSkinTest
> >>>>>> >> cudaFree GpuBuffer::Deallocate failed unspecified launch failure
> >>>>>> >> grep: mdinfo.1GTX_TITAN: No such file or directory
> >>>>>> >>
> >>>>>> >> TRPCAGE_PRODUCTION - 304 atoms GB
> >>>>>> >> ------------------------------**---
> >>>>>> >> 1 x GTX_TITAN: | ns/day = 595.09
> >>>>>> seconds/ns =
> >>>>>> >> 145.19
> >>>>>> >>
> >>>>>> >> MYOGLOBIN_PRODUCTION - 2,492 atoms GB
> >>>>>> >> ------------------------------**-------
> >>>>>> >>
> >>>>>> >> 1 x GTX_TITAN: | ns/day = 202.56
> >>>>>> seconds/ns =
> >>>>>> >> 426.53
> >>>>>> >>
> >>>>>> >> NUCLEOSOME_PRODUCTION - 25,095 atoms GB
> >>>>>> >> ------------------------------**---------
> >>>>>> >>
> >>>>>> >> 1 x GTX_TITAN:
> >>>>>> >>
> >>>>>> >>
> >>>>>> >>
> >>>>>> >>
> >>>>>> >>
> >>>>>> >>
> >>>>>> >>
> >>>>>> >> Dne Tue, 28 May 2013 20:42:32 +0200 ET <sketchfoot.gmail.com>
> >>>>>> napsal/-a:
> >>>>>> >>
> >>>>>> >> > Hi,
> >>>>>> >> >
> >>>>>> >> > I just got a superclocked Titan and one at normal freq. The
> >>>>>> first
> >>>>>> one
> >>>>>> >> ran
> >>>>>> >> > like a charm with no issues so far. The other standard clocked
> >>>>>> one
> >>>>>> >> could
> >>>>>> >> > never get past the constant pressure stage in an NPT
> >>>>>> simulation.
> >>>>>> It
> >>>>>> >> kept
> >>>>>> >> > writing NAN or ********* in the outfile. I swapped them about
> >>>>>> in
> >>>>>> the
> >>>>>> >> pcie
> >>>>>> >> > lanes then ran it solo in each one of the lanes. Despite all
> >>>>>> this
> >>>>>> it
> >>>>>> >> was
> >>>>>> >> > still failing the benchmark that the other one had no problems
> >>>>>> with.
> >>>>>> >> >
> >>>>>> >> > I couldn't find any memory errors with GPU-burn either, but as
> >>>>>> they
> >>>>>> >> cost
> >>>>>> >> > near a grand a piece, I RMA'd it today. I recommend you to do
> >>>>>> the
> >>>>>> >> same if
> >>>>>> >> > its not giving you any joy. Life's too short. :)
> >>>>>> >> >
> >>>>>> >> > br,
> >>>>>> >> > g
> >>>>>> >> >
> >>>>>> >> >
> >>>>>> >> > On 28 May 2013 16:57, Scott Le Grand <varelse2005.gmail.com>
> >>>>>> wrote:
> >>>>>> >> >
> >>>>>> >> >> AMBER != NAMD...
> >>>>>> >> >>
> >>>>>> >> >> GTX 680 != GTX Titan...
> >>>>>> >> >>
> >>>>>> >> >> Ian's suggestion is a good one. But even then, you need to
> >>>>>> test
> >>>>>> >> >> your
> >>>>>> >> >> GPUs
> >>>>>> >> >> as the Titans are running right on the edge of stability.
> >>>>>> Like I
> >>>>>> >> told
> >>>>>> >> >> Marek, try running 100K iterations of Cellulose NVE twice with
> >>>>>> the
> >>>>>> >> same
> >>>>>> >> >> random seed. if you don't get identically bit accurate
> >>>>>> output,
> >>>>>> your
> >>>>>> >> >> GPU is
> >>>>>> >> >> not working. Memtest programs do not catch this because (I am
> >>>>>> >> guessing)
> >>>>>> >> >> they are designed for a uniform memory hierarchy and only one
> >>>>>> path
> >>>>>> >> >> to
> >>>>>> >> >> read
> >>>>>> >> >> and write data. I have a stock GTX Titan that cannot pass the
> >>>>>> >> Cellulose
> >>>>>> >> >> NVE test and another one that does. I spent a couple days on
> >>>>>> the
> >>>>>> >> former
> >>>>>> >> >> GPU looking for the imaginary bug that went away like magic
> >>>>>> the
> >>>>>> >> second I
> >>>>>> >> >> switched out the GPU.
> >>>>>> >> >>
> >>>>>> >> >> Scott
> >>>>>> >> >>
> >>>>>> >> >>
> >>>>>> >> >>
> >>>>>> >> >>
> >>>>>> >> >>
> >>>>>> >> >> On Tue, May 28, 2013 at 8:11 AM, Robert Konecny <rok.ucsd.edu
> >
> >>>>>> wrote:
> >>>>>> >> >>
> >>>>>> >> >> > Hi Scott,
> >>>>>> >> >> >
> >>>>>> >> >> > unfortunately we are seeing similar Amber instability on GTX
> >>>>>> >> Titans as
> >>>>>> >> >> > Marek is. We have a box with four GTX Titans (not
> >>>>>> oveclocked)
> >>>>>> >> running
> >>>>>> >> >> > CentOS 6.3 with NVidia 319.17 driver and Amber 12.2. Any
> >>>>>> Amber
> >>>>>> >> >> simulation
> >>>>>> >> >> > longer than 10-15 min eventually crashes on these cards,
> >>>>>> including
> >>>>>> >> >> both
> >>>>>> >> >> JAC
> >>>>>> >> >> > benchmarks (with extended run time). This is reproducible on
> >>>>>> all
> >>>>>> >> four
> >>>>>> >> >> > cards.
> >>>>>> >> >> >
> >>>>>> >> >> > To eliminate the possible hardware error we ran extended GPU
> >>>>>> >> >> > memory
> >>>>>> >> >> tests
> >>>>>> >> >> > on all four Titans with memtestG80, cuda_memtest and also
> >>>>>> gpu_burn
> >>>>>> >> -
> >>>>>> >> >> all
> >>>>>> >> >> > finished without errors. Since I agree that these programs
> >>>>>> may
> >>>>>> not
> >>>>>> >> >> test
> >>>>>> >> >> the
> >>>>>> >> >> > GPU completely we also set up simulations with NAMD. We can
> >>>>>> run
> >>>>>> >> four
> >>>>>> >> >> NAMD
> >>>>>> >> >> > simulations simultaneously for many days without any errors
> >>>>>> on
> >>>>>> >> >> > this
> >>>>>> >> >> > hardware. For reference - we also have exactly the same
> >>>>>> server
> >>>>>> >> >> > with
> >>>>>> >> >> the
> >>>>>> >> >> > same hardware components but with four GTX680s and this
> >>>>>> setup
> >>>>>> >> >> > works
> >>>>>> >> >> just
> >>>>>> >> >> > fine for Amber. So all this leads me to believe that a
> >>>>>> hardware
> >>>>>> >> error
> >>>>>> >> >> is
> >>>>>> >> >> > not very likely.
> >>>>>> >> >> >
> >>>>>> >> >> > I would appreciate your comments on this, perhaps there is
> >>>>>> >> something
> >>>>>> >> >> else
> >>>>>> >> >> > causing these errors which we are not seeing.
> >>>>>> >> >> >
> >>>>>> >> >> > Thanks,
> >>>>>> >> >> >
> >>>>>> >> >> > Robert
> >>>>>> >> >> >
> >>>>>> >> >> >
> >>>>>> >> >> > On Mon, May 27, 2013 at 04:25:24PM -0700, Scott Le Grand
> >>>>>> wrote:
> >>>>>> >> >> > > I have two GTX Titans. One is defective, the other is
> >>>>>> not.
> >>>>>> >> >> > Unfortunately,
> >>>>>> >> >> > > they both pass all standard GPU memory tests.
> >>>>>> >> >> > >
> >>>>>> >> >> > > What the defective one doesn't do is generate reproducibly
> >>>>>> >> >> bit-accurate
> >>>>>> >> >> > > outputs for simulations of Factor IX (90,986 atoms) or
> >>>>>> larger,
> >>>>>> >> >> > > of
> >>>>>> >> >> 100K
> >>>>>> >> >> or
> >>>>>> >> >> > > so iterations.
> >>>>>> >> >> > >
> >>>>>> >> >> > > Which is yet another reason why I insist on MD algorithms
> >>>>>> >> >> (especially
> >>>>>> >> >> on
> >>>>>> >> >> > > GPUS) being deterministic. Besides its ability to find
> >>>>>> software
> >>>>>> >> >> bugs,
> >>>>>> >> >> > and
> >>>>>> >> >> > > fulfilling one of the most important tenets of science,
> >>>>>> it's
> >>>>>> a
> >>>>>> >> great
> >>>>>> >> >> way
> >>>>>> >> >> > to
> >>>>>> >> >> > > diagnose defective hardware with very little effort.
> >>>>>> >> >> > >
> >>>>>> >> >> > > 928 MHz? That's 6% above the boost clock of a stock
> >>>>>> Titan.
> >>>>>> >> Titan
> >>>>>> >> >> is
> >>>>>> >> >> > > pushing the performance envelope as is. If you're going
> >>>>>> to
> >>>>>> pay
> >>>>>> >> the
> >>>>>> >> >> > premium
> >>>>>> >> >> > > for such chips, I'd send them back until you get one that
> >>>>>> runs
> >>>>>> >> >> correctly.
> >>>>>> >> >> > > I'm very curious how fast you can push one of these things
> >>>>>> >> >> > > before
> >>>>>> >> >> they
> >>>>>> >> >> > give
> >>>>>> >> >> > > out.
> >>>>>> >> >> > >
> >>>>>> >> >> > >
> >>>>>> >> >> > >
> >>>>>> >> >> > >
> >>>>>> >> >> > >
> >>>>>> >> >> > >
> >>>>>> >> >> > >
> >>>>>> >> >> > > On Mon, May 27, 2013 at 10:01 AM, Marek Maly
> >>>>>> <marek.maly.ujep.cz
> >>>>>> >
> >>>>>> >> >> wrote:
> >>>>>> >> >> > >
> >>>>>> >> >> > > > Dear all,
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > I have recently bought two "EVGA GTX TITAN Superclocked"
> >>>>>> GPUs.
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > I did the first calculations (pmemd.cuda in Amber12)
> >>>>>> with
> >>>>>> >> systems
> >>>>>> >> >> > around
> >>>>>> >> >> > > > 60K atoms without any problems (NPT, Langevin), but
> >>>>>> when I
> >>>>>> >> later
> >>>>>> >> >> tried
> >>>>>> >> >> > > > with bigger systems (around 100K atoms) I obtained
> >>>>>> "classical"
> >>>>>> >> >> > irritating
> >>>>>> >> >> > > > errors
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > cudaMemcpy GpuBuffer::Download failed unspecified launch
> >>>>>> >> failure
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > just after few thousands of MD steps.
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > So this was obviously the reason for memtestG80 tests.
> >>>>>> >> >> > > > ( https://simtk.org/home/memtest ).
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > So I compiled memtestG80 from sources (
> >>>>>> >> memtestG80-1.1-src.tar.gz
> >>>>>> >> >> )
> >>>>>> >> >> and
> >>>>>> >> >> > > > then tested
> >>>>>> >> >> > > > just small part of memory GPU (200 MB) using 100
> >>>>>> iterations.
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > On both cards I have obtained huge amount of errors but
> >>>>>> "just"
> >>>>>> >> on
> >>>>>> >> >> > > > "Random blocks:". 0 errors in all remaining tests in all
> >>>>>> >> >> iterations.
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > ------THE LAST ITERATION AND FINAL RESULTS-------
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > Test iteration 100 (GPU 0, 200 MiB): 169736847 errors so
> >>>>>> far
> >>>>>> >> >> > > > Moving Inversions (ones and zeros): 0 errors (6
> >>>>>> ms)
> >>>>>> >> >> > > > Memtest86 Walking 8-bit: 0 errors (53 ms)
> >>>>>> >> >> > > > True Walking zeros (8-bit): 0 errors (26 ms)
> >>>>>> >> >> > > > True Walking ones (8-bit): 0 errors (26 ms)
> >>>>>> >> >> > > > Moving Inversions (random): 0 errors (6 ms)
> >>>>>> >> >> > > > Memtest86 Walking zeros (32-bit): 0 errors (105
> >>>>>> ms)
> >>>>>> >> >> > > > Memtest86 Walking ones (32-bit): 0 errors (104
> >>>>>> ms)
> >>>>>> >> >> > > > Random blocks: 1369863 errors (27 ms)
> >>>>>> >> >> > > > Memtest86 Modulo-20: 0 errors (215 ms)
> >>>>>> >> >> > > > Logic (one iteration): 0 errors (4 ms)
> >>>>>> >> >> > > > Logic (4 iterations): 0 errors (8 ms)
> >>>>>> >> >> > > > Logic (shared memory, one iteration): 0 errors
> >>>>>> (8
> >>>>>> ms)
> >>>>>> >> >> > > > Logic (shared-memory, 4 iterations): 0 errors
> >>>>>> (25
> >>>>>> ms)
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > Final error count after 100 iterations over 200 MiB of
> >>>>>> GPU
> >>>>>> >> memory:
> >>>>>> >> >> > > > 171106710 errors
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > ------------------------------**------------
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > I have some questions and would be really grateful for
> >>>>>> any
> >>>>>> >> >> comments.
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > Regarding overclocking, using the deviceQuery I found
> >>>>>> out
> >>>>>> that
> >>>>>> >> >> under
> >>>>>> >> >> > linux
> >>>>>> >> >> > > > both cards run
> >>>>>> >> >> > > > automatically using boost shader/GPU frequency which is
> >>>>>> here
> >>>>>> >> 928
> >>>>>> >> >> MHz
> >>>>>> >> >> > (the
> >>>>>> >> >> > > > basic value for these factory OC cards is 876 MHz).
> >>>>>> >> >> > > > deviceQuery
> >>>>>> >> >> > reported
> >>>>>> >> >> > > > Memory Clock rate is 3004 MHz although "it" should be
> >>>>>> 6008
> >>>>>> MHz
> >>>>>> >> but
> >>>>>> >> >> > maybe
> >>>>>> >> >> > > > the quantity which is reported by deviceQuery "Memory
> >>>>>> Clock
> >>>>>> >> rate"
> >>>>>> >> >> is
> >>>>>> >> >> > > > different from the product specification "Memory Clock"
> >>>>>> .
> >>>>>> It
> >>>>>> >> seems
> >>>>>> >> >> that
> >>>>>> >> >> > > > "Memory Clock rate" = "Memory Clock"/2. Am I right ? Or
> >>>>>> just
> >>>>>> >> >> > deviceQuery
> >>>>>> >> >> > > > is not able to read this spec. properly
> >>>>>> >> >> > > > in Titan GPU ?
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > Anyway for the moment I assume that the problem might be
> >>>>>> due
> >>>>>> >> >> > > > to
> >>>>>> >> >> the
> >>>>>> >> >> > high
> >>>>>> >> >> > > > shader/GPU frequency.
> >>>>>> >> >> > > > (see here :
> >>>>>> http://folding.stanford.edu/**English/DownloadUtils<
> http://folding.stanford.edu/English/DownloadUtils>
> >>>>>> )
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > To verify this hypothesis one should perhaps UNDERclock
> >>>>>> to
> >>>>>> >> basic
> >>>>>> >> >> > frequency
> >>>>>> >> >> > > > which is in this
> >>>>>> >> >> > > > model 876 MHz or even to the TITAN REFERENCE frequency
> >>>>>> which
> >>>>>> >> >> > > > is
> >>>>>> >> >> 837
> >>>>>> >> >> > MHz.
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > Obviously I am working with these cards under linux
> >>>>>> (CentOS
> >>>>>> >> >> > > > 2.6.32-358.6.1.el6.x86_64) and as I found, the OC tools
> >>>>>> under
> >>>>>> >> >> linux
> >>>>>> >> >> > are in
> >>>>>> >> >> > > > fact limited just to NVclock utility, which is
> >>>>>> unfortunately
> >>>>>> >> >> > > > out of date (at least speaking about the GTX Titan ). I
> >>>>>> have
> >>>>>> >> >> obtained
> >>>>>> >> >> > this
> >>>>>> >> >> > > > message when I wanted
> >>>>>> >> >> > > > just to let NVclock utility to read and print shader and
> >>>>>> >> >> > > > memory
> >>>>>> >> >> > > > frequencies of my Titan's:
> >>>>>> >> >> > > >
> >>>>>> >> >> > > >
> >>>>>> >> >>
> >>>>>> ------------------------------**------------------------------**
> >>>>>> -------
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > [root.dyn-138-272 NVCLOCK]# nvclock -s --speeds
> >>>>>> >> >> > > > Card: Unknown Nvidia card
> >>>>>> >> >> > > > Card number: 1
> >>>>>> >> >> > > > Memory clock: -2147483.750 MHz
> >>>>>> >> >> > > > GPU clock: -2147483.750 MHz
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > Card: Unknown Nvidia card
> >>>>>> >> >> > > > Card number: 2
> >>>>>> >> >> > > > Memory clock: -2147483.750 MHz
> >>>>>> >> >> > > > GPU clock: -2147483.750 MHz
> >>>>>> >> >> > > >
> >>>>>> >> >> > > >
> >>>>>> >> >> > > >
> >>>>>> >> >>
> >>>>>> ------------------------------**------------------------------**
> >>>>>> -------
> >>>>>> >> >> > > >
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > I would be really grateful for some tips regarding
> >>>>>> "NVclock
> >>>>>> >> >> > alternatives",
> >>>>>> >> >> > > > but after wasting some hours with googling it seems that
> >>>>>> there
> >>>>>> >> is
> >>>>>> >> >> no
> >>>>>> >> >> > other
> >>>>>> >> >> > > > Linux
> >>>>>> >> >> > > > tool with NVclock functionality. So the only
> >>>>>> possibility is
> >>>>>> >> here
> >>>>>> >> >> > perhaps
> >>>>>> >> >> > > > to edit
> >>>>>> >> >> > > > GPU bios with some Lin/DOS/Win tools like (Kepler BIOS
> >>>>>> >> >> > > > Tweaker,
> >>>>>> >> >> > NVflash)
> >>>>>> >> >> > > > but obviously
> >>>>>> >> >> > > > I would like to rather avoid such approach as using it
> >>>>>> means
> >>>>>> >> >> perhaps
> >>>>>> >> >> > also
> >>>>>> >> >> > > > to void the warranty even if I am going to underclock
> >>>>>> the
> >>>>>> GPUs
> >>>>>> >> >> not to
> >>>>>> >> >> > > > overclock them.
> >>>>>> >> >> > > > So before this eventual step (GPU bios editing) I would
> >>>>>> like
> >>>>>> >> >> > > > to
> >>>>>> >> >> have
> >>>>>> >> >> > some
> >>>>>> >> >> > > > approximative estimate
> >>>>>> >> >> > > > of the probability, that the problems are here really
> >>>>>> because
> >>>>>> >> of
> >>>>>> >> >> the
> >>>>>> >> >> > > > overclocking
> >>>>>> >> >> > > > (too high (boost) default shader frequency).
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > This probability I hope to estimate from the eventual
> >>>>>> >> responses of
> >>>>>> >> >> > another
> >>>>>> >> >> > > > Amber/Titan SC users, if I am not the only crazy guy who
> >>>>>> >> >> > > > bought
> >>>>>> >> >> this
> >>>>>> >> >> > model
> >>>>>> >> >> > > > for Amber calculations :)) But of course any eventual
> >>>>>> >> experiences
> >>>>>> >> >> with
> >>>>>> >> >> > > > Titan cards related to their memtestG80 results and
> >>>>>> >> >> UNDER/OVERclocking
> >>>>>> >> >> > > > (if possible in Linux OS) are of course welcomed as
> >>>>>> well !
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > My HW/SW configuration
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > motherboard: ASUS P9X79 PRO
> >>>>>> >> >> > > > CPU: Intel Core i7-3930K
> >>>>>> >> >> > > > RAM: CRUCIAL Ballistix Sport 32GB (4x8GB) DDR3 1600 VLP
> >>>>>> >> >> > > > CASE: CoolerMaster Dominator CM-690 II Advanced,
> >>>>>> >> >> > > > Power:Enermax PLATIMAX EPM1200EWT 1200W, 80+, Platinum
> >>>>>> >> >> > > > GPUs : 2 x EVGA GTX TITAN Superclocked 6GB
> >>>>>> >> >> > > > cooler: Cooler Master Hyper 412 SLIM
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > OS: CentOS (2.6.32-358.6.1.el6.x86_64)
> >>>>>> >> >> > > > driver version: 319.17
> >>>>>> >> >> > > > cudatoolkit_5.0.35_linux_64_**rhel6.x
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > The computer is in air-conditioned room with permanent
> >>>>>> >> >> > > > external
> >>>>>> >> >> > > > temperature around 18°C
> >>>>>> >> >> > > >
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > Thanks a lot in advance for any comment/experience !
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > Best wishes,
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > Marek
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > --
> >>>>>> >> >> > > > Tato zpráva byla vytvořena převratným poštovním klientem
> >>>>>> >> >> > > > Opery:
> >>>>>> >> >> > > > http://www.opera.com/mail/
> >>>>>> >> >> > > >
> >>>>>> >> >> > > > ______________________________**_________________
> >>>>>> >> >> > > > AMBER mailing list
> >>>>>> >> >> > > > AMBER.ambermd.org
> >>>>>> >> >> > > >
> >>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>> >> >> > > >
> >>>>>> >> >> > > ______________________________**_________________
> >>>>>> >> >> > > AMBER mailing list
> >>>>>> >> >> > > AMBER.ambermd.org
> >>>>>> >> >> > >
> >>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>> >> >> >
> >>>>>> >> >> > ______________________________**_________________
> >>>>>> >> >> > AMBER mailing list
> >>>>>> >> >> > AMBER.ambermd.org
> >>>>>> >> >> >
> >>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>> >> >> >
> >>>>>> >> >> ______________________________**_________________
> >>>>>> >> >> AMBER mailing list
> >>>>>> >> >> AMBER.ambermd.org
> >>>>>> >> >>
> >>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>> >> >>
> >>>>>> >> > ______________________________**_________________
> >>>>>> >> > AMBER mailing list
> >>>>>> >> > AMBER.ambermd.org
> >>>>>> >> >
> >>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>> >> >
> >>>>>> >> > __________ Informace od ESET NOD32 Antivirus, verze databaze
> >>>>>> 8385
> >>>>>> >> > (20130528) __________
> >>>>>> >> >
> >>>>>> >> > Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>> >> >
> >>>>>> >> > http://www.eset.cz
> >>>>>> >> >
> >>>>>> >> >
> >>>>>> >> >
> >>>>>> >>
> >>>>>> >>
> >>>>>> >> --
> >>>>>> >> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> >>>>>> >> http://www.opera.com/mail/
> >>>>>> >>
> >>>>>> >> ______________________________**_________________
> >>>>>> >> AMBER mailing list
> >>>>>> >> AMBER.ambermd.org
> >>>>>> >>
> >>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>> >>
> >>>>>> > ______________________________**_________________
> >>>>>> > AMBER mailing list
> >>>>>> > AMBER.ambermd.org
> >>>>>> >
> >>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>> >
> >>>>>> > __________ Informace od ESET NOD32 Antivirus, verze databaze 8386
> >>>>>> > (20130528) __________
> >>>>>> >
> >>>>>> > Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>>> >
> >>>>>> > http://www.eset.cz
> >>>>>> >
> >>>>>> >
> >>>>>> >
> >>>>>>
> >>>>>>
> >>>>>> --
> >>>>>> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> >>>>>> http://www.opera.com/mail/
> >>>>>>
> >>>>>> ______________________________**_________________
> >>>>>> AMBER mailing list
> >>>>>> AMBER.ambermd.org
> >>>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>>
> >>>>>> ______________________________**_________________
> >>>>> AMBER mailing list
> >>>>> AMBER.ambermd.org
> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>
> >>>>>
> >>>>> ______________________________**_________________
> >>>>> AMBER mailing list
> >>>>> AMBER.ambermd.org
> >>>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>>
> >>>>
> >>>>
> >>>>
> >>>> ______________________________**_________________
> >>>> AMBER mailing list
> >>>> AMBER.ambermd.org
> >>>> http://lists.ambermd.org/**mailman/listinfo/amber<
> http://lists.ambermd.org/mailman/listinfo/amber>
> >>>>
> >>>> __________ Informace od ESET NOD32 Antivirus, verze databaze 8386
> >>>> (20130528) __________
> >>>>
> >>>> Tuto zpravu proveril ESET NOD32 Antivirus.
> >>>>
> >>>> http://www.eset.cz
> >>>>
> >>>>
> >>>>
> >>>>
> >>>
> >>> --
> >>> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> >>> http://www.opera.com/mail/
> >>> _______________________________________________
> >>> AMBER mailing list
> >>> AMBER.ambermd.org
> >>> http://lists.ambermd.org/mailman/listinfo/amber
> >>>
> >>>
> > _______________________________________________
> > AMBER mailing list
> > AMBER.ambermd.org
> > http://lists.ambermd.org/mailman/listinfo/amber
> >
> > __________ Informace od ESET NOD32 Antivirus, verze databaze 8390
> > (20130529) __________
> >
> > Tuto zpravu proveril ESET NOD32 Antivirus.
> >
> > http://www.eset.cz
> >
> >
> >
>
>
> --
> Tato zpráva byla vytvořena převratným poštovním klientem Opery:
> http://www.opera.com/mail/
>
> _______________________________________________
> AMBER mailing list
> AMBER.ambermd.org
> http://lists.ambermd.org/mailman/listinfo/amber
>
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Wed May 29 2013 - 16:30:03 PDT