Re: [AMBER] Anyone running machines with Quad GPU setups

From: ET <sketchfoot.gmail.com>
Date: Tue, 25 Jun 2013 07:09:40 +0100

tell al lie:
http://www.youtube.com/watch?v=ZjvxIEZjL3Y

http://www.youtube.com/watch?v=_9zI1nyv9Dk

Also apparently a backplate reduces load temps by 2-4 degrees on air
cooling, but opinion seems divided.


On 25 June 2013 06:56, ET <sketchfoot.gmail.com> wrote:

> lol. That is a much better solution. :)
>
> haf=d a quick look for a review, but only found one in german for the
> itan, which looks pretty amazing temp wise
>
>
> http://www.hardwareluxx.de/index.php/artikel/hardware/grafikkarten/26227-test-evga-geforce-gtx-titan-hydro-copper-signature.html?start=4
>
>
> On 25 June 2013 06:36, Ross Walker <ross.rosswalker.co.uk> wrote:
>
>> The biggest problem with those rigs is dismantling your circuit breaker
>> board to replace the GFI's with nails. ;-)
>>
>> 15 will work great - but you don't want to be using AMD GPUs with AMBER.
>> ;-)
>>
>> Water cooling is the best option - http://tinyurl.com/p8ppt4c
>>
>> All the best
>> Ross
>>
>>
>>
>> On 6/24/13 10:19 PM, "ET" <sketchfoot.gmail.com> wrote:
>>
>> >ps. Anyone been tempted by a bitcoin mining rig type setup? :P
>> >
>> >http://bitcoinexaminer.org/20-insane-bitcoin-mining-rigs/
>> >
>> >The PCIe extender cable looks good in terms of allowing your card to
>> >breathe if you can custom rig a case that's not going to burn down your
>> >house.
>> >
>> >http://www.ebay.co.uk/bhp/pci-e-riser
>> >
>> >br,
>> >g
>> >
>> >
>> >
>> >
>> >On 25 June 2013 04:06, ET <sketchfoot.gmail.com> wrote:
>> >
>> >> Hi Kevin,
>> >>
>> >> Thanks for the info on the case mods & temps. I decided to go with the
>> >> HAF-X case, which despite it's ugliness has the the ducts that you
>> >> mentioned.. Have got my fingers crossed that the switch connectors
>> >>won't be
>> >> covered by the final GPU. :)
>> >>
>> >> I think the price difference between my build & that of Ross is ~ £250.
>> >> However, as I'm short on time & did not really like the look of AM3+
>> for
>> >> various reasons, I don't mind the cash difference. Additionally, I'm
>> >>prob
>> >> going to sell this machine year end, so kitted it so it would be
>> >>desirable
>> >> for an overclocker.
>> >>
>> >>
>> >> br,
>> >> g
>> >>
>> >>
>> >> On 25 June 2013 00:58, Ross Walker <ross.rosswalker.co.uk> wrote:
>> >>
>> >>> Hi Kevin,
>> >>>
>> >>> Thanks for the detailed info. The GTX680s in our experience have been
>> >>>rock
>> >>> solid stable. 1 out of 20 for infant death sounds about right in my
>> >>> experience. You should be able to just RMA that card and you should be
>> >>> good. Note, you can also build such systems with AMD processors.
>> >>>Attached
>> >>> is a pdf with an Amazon shopping list for a 4 GTX680 system which we
>> >>>have
>> >>> built many of for $3200 - might even be cheaper now with the 680s
>> >>>coming
>> >>> down in price. - This motherboard takes all 4 GPUs without hitting
>> >>> anything - as long as you don't try to connect up a bunch of external
>> >>>USB
>> >>> connectors.
>> >>>
>> >>> This same system should work great for GTX780s as well - we just need
>> >>>to
>> >>> make sure they are giving the correct answers - looking more positive
>> >>>by
>> >>> the day.
>> >>>
>> >>> I've not seen any major issues with 4 GPU cooling in these systems -
>> as
>> >>> long as you have plenty of back airflow as you have ones should be
>> >>>good.
>> >>> 90C is a normal temperature for GTX680 and I've run several flat out
>> >>>for
>> >>> months on end at this temperature.
>> >>>
>> >>> I second the choice of slurm. It's certainly far from "Simple" but it
>> >>>does
>> >>> seem to understand GPUs better than any of the other queuing systems
>> >>>out
>> >>> there. Indeed - the 'certified' clusters I designed with exxact use
>> >>>Slurm.
>> >>> Rocks with the slurm roll works great for GPU clusters.
>> >>>
>> >>> Again, thanks for this info, should be useful to lots of people here.
>> >>>
>> >>> All the best
>> >>> Ross
>> >>>
>> >>>
>> >>> On 6/24/13 4:40 PM, "Kevin Hauser" <84hauser.gmail.com> wrote:
>> >>>
>> >>> >A good bit late to the discussion, but we've been having success with
>> >>>a
>> >>> >relatively cheap setup ~ $3,200 per machine and five machines (about
>> 3
>> >>> >weeks of production burning; acceptable temperatures; see below).
>> >>> >
>> >>> >Bad news first:
>> >>> >We lost one GPU (out of 20). Not bad, considering our expectations
>> >>>from
>> >>> >commodity kit.
>> >>> >
>> >>> >Briefly, we got five quad-GPU boxes running EVGA GTX680-FTW, Intel
>> >>> >i7-3820,
>> >>> >on the GigaByte GA-X79-UD3 LGA2011 mobo. # See slide one of attached
>> >>>*pdf
>> >>> >for overview.
>> >>> >
>> >>> >Discussion on our kit:
>> >>> >The Antec P280 cases are great b/c all metal is rubber coated for
>> >>>quiet
>> >>> >running, there's three fans included, a quick mount for a fourth
>> 120mm
>> >>> >fan,
>> >>> >and space for half-dozen HDDs. It's pretty heavy, though (22 lbs,
>> >>>dry).
>> >>> >--$110
>> >>> >
>> >>> >The CPU is what it is; from our vendor, cpu cooler is not included.
>> We
>> >>> got
>> >>> >the Cooler Master Hyper212 --$300 + $33
>> >>> >
>> >>> >The mobos appear to be well manufactured, especially given the price
>> >>> (half
>> >>> >of ET's Asus kit). BUT, I needed to take apart the power and reset
>> >>> >terminals connecting case to mobo so the last GPU fully seats into
>> the
>> >>> >mobo. -- $230 # See slide two of attached *pdf.
>> >>> >
>> >>> >Good news last:
>> >>> >I took the DHFR test case (mdin below) and ran it for 100 ps. I left
>> >>> >ntpr=1
>> >>> >to see when and where things got funky... Every single GPU in all
>> five
>> >>> >"nodes" produced identical mdouts. Afterwards, we had that one GPU
>> >>>die,
>> >>> >though. I'm getting double the speed we were getting on our center's
>> >>>mega
>> >>> >expensive server (Tesla M2070s).
>> >>> >custom short md
>> >>> > &cntrl
>> >>> > nstlim=100000, ig=11,ioutfm=1,ntxo=2,
>> >>> > ntx=5, irest=1,
>> >>> > ntc=2, ntf=2,
>> >>> > ntpr=1, ntwr=10000,
>> >>> > dt=0.002,
>> >>> > ntt=1, tautp=10.0,
>> >>> > temp0=300.0,
>> >>> > ntb=2,ntp=1,taup=10.0,
>> >>> > /
>> >>> >
>> >>> >
>> >>> >Burn test info:
>> >>> >Overall, our GPUs have not exceeded 90 C, yet. Max sustained we've
>> >>>seen
>> >>> is
>> >>> >87 C. They're in our old, cold server room.
>> >>> >
>> >>> >Only down time last three weeks or so was when we were sorting out
>> our
>> >>> NFS
>> >>> >or PXE or Slurm queuing system (only a few hours, really). Slurm is
>> >>> >actually quite nice, simple, and very free.
>> >>> >
>> >>> >Of course, the intake fans for the GPUs (save bottom one) suck air
>> >>>right
>> >>> >off the heat sink of the GPU below it. The GPUs are cleverly tapered
>> >>> right
>> >>> >where the intake fans face the heat sink, leaving a whopping 2 or so
>> >>> >millimeters for air. On slide one, we installed a massive fan to ram
>> >>> fresh
>> >>> >air into the intakes (mod_1). Tests show that 3 GPUs heat up to 85 C
>> >>>+/-
>> >>> 2
>> >>> >(CVD=0,1,3) and 1 GPU to 76 C (CVD=2). Cooler Master R4-MFJR-07FK-R1
>> >>> 200mm
>> >>> >MegaFlow --$19 # cvd=cuda_visible_device
>> >>> >
>> >>> >On slide 3, you can see I ghetto-fabricated a cardboard box that
>> ducts
>> >>> air
>> >>> >from a 120mm fan directly into the GPUs' intake-tapered section
>> >>>(mod_2).
>> >>> >Tests show that CVD=0 hits 84 C, CDV=1 hits 83 C, CVD=2 hits 74 C,
>> and
>> >>> >CVD=3 hits 82 C. Benefit, the case has a very simple clip-on mount
>> for
>> >>> >120mm. Downside, we needed that duct to realize benefit. We yanked a
>> >>>fan
>> >>> >from the top of the case that was needlessly serving the CPU.
>> >>> >
>> >>> >
>> >>> >HTH,
>> >>> >kevin
>> >>> >
>> >>> >
>> >>> >On Mon, Jun 24, 2013 at 2:48 PM, ET <sketchfoot.gmail.com> wrote:
>> >>> >
>> >>> >> Thanks for the further info Ross! :)
>> >>> >>
>> >>> >> Decided in the end to go for a Asus P8Z77 WS board with a Intel i7
>> >>> >>3770K.
>> >>> >> Slightly overkill, but needed to proof it in the event of resale or
>> >>> >>finding
>> >>> >> another use for it.
>> >>> >>
>> >>> >> br,
>> >>> >> g
>> >>> >>
>> >>> >>
>> >>> >> On 25 June 2013 03:44, <deeptinayar.gmail.com> wrote:
>> >>> >>
>> >>> >> >
>> >>> >> > Sent from my BlackBerry® smartphone from !DEA
>> >>> >> >
>> >>> >> > -----Original Message-----
>> >>> >> > From: ET <sketchfoot.gmail.com>
>> >>> >> > Date: Sat, 22 Jun 2013 20:45:20
>> >>> >> > To: AMBER Mailing List<amber.ambermd.org>
>> >>> >> > Reply-To: AMBER Mailing List <amber.ambermd.org>
>> >>> >> > Subject: Re: [AMBER] Anyone running machines with Quad GPU setups
>> >>> >> >
>> >>> >> > Looks like the Asus P9X79-E WS is for you then Scott! :) Haven't
>> >>> seen
>> >>> >> many
>> >>> >> > (if any!) boards with that anount of bandwidth so far.
>> >>> >> >
>> >>> >> > br,
>> >>> >> > g
>> >>> >> >
>> >>> >> >
>> >>> >> > On 22 June 2013 19:31, Scott Le Grand <varelse2005.gmail.com>
>> >>>wrote:
>> >>> >> >
>> >>> >> > > It may be overkill now but I'm planning to revisit the Multi
>> GPU
>> >>> >>code
>> >>> >> in
>> >>> >> > > the near future and that's why I need a motherboard that can
>> >>>really
>> >>> >> take
>> >>> >> > > advantage of it.
>> >>> >> > > On Jun 22, 2013 10:12 AM, "ET" <sketchfoot.gmail.com> wrote:
>> >>> >> > >
>> >>> >> > > > Hi,
>> >>> >> > > >
>> >>> >> > > > Thanks v much for you specs Divi. :) I've been debating with
>> >>> >>myself
>> >>> >> as
>> >>> >> > my
>> >>> >> > > > board as it looks good and has a very nice spec. From what
>> >>>I've
>> >>> >>read,
>> >>> >> > the
>> >>> >> > > > only problems with it is the higher than average power draw.
>> >>> >> > > >
>> >>> >> > > > .Scott. I believe the board runs in x8/x8/x8/x8 for a 4 GPU
>> >>> >>config -
>> >>> >> so
>> >>> >> > > > effectively PCI2 2.0 x16 rate. Would this present any
>> >>>problems,
>> >>> >>if
>> >>> >> you
>> >>> >> > > > were running the serial GPU code, From what I read on the
>> >>>AMBER
>> >>> >>GPU
>> >>> >> > > > hardware page, this is more important for the parallel GPU
>> >>>code?
>> >>> >> > Though,
>> >>> >> > > I
>> >>> >> > > > imagine having 4x serial ruins going simultaneously would
>> also
>> >>> tax
>> >>> >> the
>> >>> >> > > GPU
>> >>> >> > > > to CPU interface, though how much I'm not sure.
>> >>> >> > > >
>> >>> >> > > > Apparently, if you are going intel, you can only acheive PCIe
>> >>>3.0
>> >>> >> using
>> >>> >> > > at
>> >>> >> > > > least a Sandy Bridge-E or ivy bridge CPU in a socket 155.
>> >>>Please
>> >>> >> > correct
>> >>> >> > > me
>> >>> >> > > > if I have understood this incorrectly though.
>> >>> >> > > >
>> >>> >> > > > http://www.enthusiastpc.net/articles/00003/3.aspx
>> >>> >> > > >
>> >>> >> > > >
>> >>> >> > > > A socket 2011 proposition would be the Asus P9X79-E WES which
>> >>>has
>> >>> >>2x
>> >>> >> > PLX
>> >>> >> > > > PEX 8747 chips so can run at x16/x16/x16/x16 with four GPUs
>> >>> >> > > >
>> >>> >> > > > https://www.asus.com/Motherboards/P9X79E_WS/#specifications
>> >>> >> > > >
>> >>> >> > > > However, I'm unsure whether this is overkill for running
>> >>>4xGPUs
>> >>> >>doing
>> >>> >> > > AMBER
>> >>> >> > > > serial code.
>> >>> >> > > >
>> >>> >> > > > What do you guys think?
>> >>> >> > > >
>> >>> >> > > > br,
>> >>> >> > > > g
>> >>> >> > > >
>> >>> >> > > >
>> >>> >> > > >
>> >>> >> > > > On 22 June 2013 16:15, Scott Le Grand <varelse2005.gmail.com
>> >
>> >>> >>wrote:
>> >>> >> > > >
>> >>> >> > > > > Does this MB support full p2p at 16x PCIE Gen 3 speeds
>> >>>between
>> >>> >>all
>> >>> >> 4
>> >>> >> > > > GPUs?
>> >>> >> > > > > On Jun 21, 2013 4:09 PM, "Divi/GMAIL" <dvenkatlu.gmail.com
>> >
>> >>> >>wrote:
>> >>> >> > > > >
>> >>> >> > > > > >
>> >>> >> > > > > > ET:
>> >>> >> > > > > > I am using GA-Z77X-UP7 that has PLX chipset and
>> >>>supports
>> >>> >>3rd
>> >>> >> Gen
>> >>> >> > > > > LGA1155
>> >>> >> > > > > > socket. Bought together with 2 TITANS sometime in March.
>> >>> >> > > > > > It has been running pretty stable 24/7 since then. I
>> >>> >>thought
>> >>> >> of
>> >>> >> > > > buying
>> >>> >> > > > > > two more titans later to fill all four slots. With so
>> >>>much
>> >>> >>mess
>> >>> >> > > going
>> >>> >> > > > on
>> >>> >> > > > > > with TITANS, I put off that plan until the dust settles.
>> >>>You
>> >>> >> might
>> >>> >> > > > want
>> >>> >> > > > > to
>> >>> >> > > > > > check new 4th GEN cpus and supporting motherboards as the
>> >>> >> Hardware
>> >>> >> > > keep
>> >>> >> > > > > > changing pretty rapidly these days.
>> >>> >> > > > > >
>> >>> >> > > > > > I have i5-processor with 16 GB ram and 256 GB SSD.
>> All
>> >>> >>four
>> >>> >> > PCI-E
>> >>> >> > > > > lanes
>> >>> >> > > > > > are X-16. It also has native X-16 link directly
>> >>>"hardwired"
>> >>> to
>> >>> >> > > > CPU-lanes
>> >>> >> > > > > > that will bypass PLX chipset, in case if you run single
>> >>>GPU.
>> >>> >>This
>> >>> >> > > might
>> >>> >> > > > > > reduce a bit of latency but not much. I get 35ns/day on
>> >>> >>FIX/NVE
>> >>> >> > > > benchmark
>> >>> >> > > > > > bypassing PLX chipset, but get about 34ns/day using PLX
>> >>> >>chipset
>> >>> >> (on
>> >>> >> > > > TITAN
>> >>> >> > > > > > of
>> >>> >> > > > > > course!!). Not a deal breaker..
>> >>> >> > > > > >
>> >>> >> > > > > > Link below:
>> >>> >> > > > > >
>> >>> >> > > > > >
>> >>> >>http://www.gigabyte.com/products/product-page.aspx?pid=4334#ov
>> >>> >> > > > > >
>> >>> >> > > > > > HTH
>> >>> >> > > > > > Divi
>> >>> >> > > > > >
>> >>> >> > > > > > -----Original Message-----
>> >>> >> > > > > > From: ET
>> >>> >> > > > > > Sent: Thursday, June 20, 2013 8:18 PM
>> >>> >> > > > > > To: AMBER Mailing List
>> >>> >> > > > > > Subject: [AMBER] Anyone running machines with Quad GPU
>> >>>setups
>> >>> >> > > > > >
>> >>> >> > > > > > Hi all,
>> >>> >> > > > > >
>> >>> >> > > > > > I was looking at getting a new mobo to run a quad GPU
>> >>>system.
>> >>> >>I
>> >>> >> was
>> >>> >> > > > > > wondering if anyone has done this. If you could post the
>> >>> >>model &
>> >>> >> > make
>> >>> >> > > > of:
>> >>> >> > > > > >
>> >>> >> > > > > > 1) motherboard
>> >>> >> > > > > > 2) CPU
>> >>> >> > > > > > 3) RAM
>> >>> >> > > > > > 4) Case
>> >>> >> > > > > > 5) The aggregate estimate of ns in simulation you have
>> >>>run on
>> >>> >> your
>> >>> >> > > > setup
>> >>> >> > > > > > without issue,
>> >>> >> > > > > >
>> >>> >> > > > > > I would be much obliged! :)
>> >>> >> > > > > >
>> >>> >> > > > > > br,
>> >>> >> > > > > > g
>> >>> >> > > > > > _______________________________________________
>> >>> >> > > > > > AMBER mailing list
>> >>> >> > > > > > AMBER.ambermd.org
>> >>> >> > > > > > http://lists.ambermd.org/mailman/listinfo/amber
>> >>> >> > > > > >
>> >>> >> > > > > >
>> >>> >> > > > > > _______________________________________________
>> >>> >> > > > > > AMBER mailing list
>> >>> >> > > > > > AMBER.ambermd.org
>> >>> >> > > > > > http://lists.ambermd.org/mailman/listinfo/amber
>> >>> >> > > > > >
>> >>> >> > > > > _______________________________________________
>> >>> >> > > > > AMBER mailing list
>> >>> >> > > > > AMBER.ambermd.org
>> >>> >> > > > > http://lists.ambermd.org/mailman/listinfo/amber
>> >>> >> > > > >
>> >>> >> > > > _______________________________________________
>> >>> >> > > > AMBER mailing list
>> >>> >> > > > AMBER.ambermd.org
>> >>> >> > > > http://lists.ambermd.org/mailman/listinfo/amber
>> >>> >> > > >
>> >>> >> > > _______________________________________________
>> >>> >> > > AMBER mailing list
>> >>> >> > > AMBER.ambermd.org
>> >>> >> > > http://lists.ambermd.org/mailman/listinfo/amber
>> >>> >> > >
>> >>> >> > _______________________________________________
>> >>> >> > AMBER mailing list
>> >>> >> > AMBER.ambermd.org
>> >>> >> > http://lists.ambermd.org/mailman/listinfo/amber
>> >>> >> > _______________________________________________
>> >>> >> > AMBER mailing list
>> >>> >> > AMBER.ambermd.org
>> >>> >> > http://lists.ambermd.org/mailman/listinfo/amber
>> >>> >> >
>> >>> >> _______________________________________________
>> >>> >> AMBER mailing list
>> >>> >> AMBER.ambermd.org
>> >>> >> http://lists.ambermd.org/mailman/listinfo/amber
>> >>> >>
>> >>> >
>> >>> >
>> >>> >
>> >>> >--
>> >>> >-- - -
>> >>> >HK
>> >>> >
>> >>> >
>> >>> >════════════════════════════════════════════
>> >>> >Kevin E. Hauser, Ph.D. Candidate
>> >>> >NRSA Fellow, National Institutes of Health
>> >>> >Carlos Simmerling Laboratory
>> >>> >Miguel Garcia-Diaz Laboratory
>> >>> >100 Laufer Center for Physical and Quantitative Biology
>> >>> >Stony Brook, New York 11794-5252
>> >>> >Phone: (631) 632.5394 Email: 84hauser.gmail.com
>> >>> >════════════════════════════════════════════
>> >>> >
>> >>>
>> >>>
>>
>> >>>>***********************************************************************
>> >>>>***
>> >>> >****
>> >>> >This e- mail message, including any attachments,
>> >>> >is for the sole use of the intended recipient(s) and may
>> >>> >contain confidential and privileged information.
>> >>> >Any unauthorized review, use, disclosure or distribution is
>> >>>prohibited.
>> >>> >If you are not the intended recipient, please contact the sender
>> >>> >by e-mail and destroy all copies of the original.
>> >>>
>> >>>
>>
>> >>>>***********************************************************************
>> >>>>***
>> >>> >****
>> >>> >_______________________________________________
>> >>> >AMBER mailing list
>> >>> >AMBER.ambermd.org
>> >>> >http://lists.ambermd.org/mailman/listinfo/amber
>> >>>
>> >>>
>> >>> _______________________________________________
>> >>> AMBER mailing list
>> >>> AMBER.ambermd.org
>> >>> http://lists.ambermd.org/mailman/listinfo/amber
>> >>>
>> >>>
>> >>
>> >_______________________________________________
>> >AMBER mailing list
>> >AMBER.ambermd.org
>> >http://lists.ambermd.org/mailman/listinfo/amber
>>
>>
>>
>> _______________________________________________
>> AMBER mailing list
>> AMBER.ambermd.org
>> http://lists.ambermd.org/mailman/listinfo/amber
>>
>
>
_______________________________________________
AMBER mailing list
AMBER.ambermd.org
http://lists.ambermd.org/mailman/listinfo/amber
Received on Mon Jun 24 2013 - 23:30:02 PDT
Custom Search