CFD Online Discussion Forums

CFD Online Discussion Forums (https://www.cfd-online.com/Forums/)
-   Hardware (https://www.cfd-online.com/Forums/hardware/)
-   -   Unforeseen problems in scaling up a cluster built with desktop parts? (https://www.cfd-online.com/Forums/hardware/88187-unforeseen-problems-scaling-up-cluster-built-desktop-parts.html)

kyle May 10, 2011 16:06

Unforeseen problems in scaling up a cluster built with desktop parts?
 
We currently have a four node cluster built with i7 980X chips and cheap motherboards, RAM and power supplies. This has run fine for us for about six months, and now I am planning on scaling it up with more of the 980X's or with the newer Sandy Bridge i7's. I'm also planning on picking up some used Infiniband equipment.

Is this something anyone else has done at a scale larger than a few nodes? I am looking to add about ~60-80 CPU which could be done for as cheap as $6,000-$7,000 USD using desktop parts. Picking up similar Xeon machines from HP or Dell would be north of $25,000.

I expect the Xeons would scale up better if for no other reason than you can put two CPUs in one motherboard, reducing the traffic between nodes. On the other hand, I could buy 3x as many of the desktop machines and still come out saving money.

Thoughts? Are Infiniband cards happy in desktop motherboards? Will not having ECC RAM randomly blow up my simulations (it hasn't so far). Is there something else I am missing?

sail May 10, 2011 19:41

definitaly an interesting topic, desktop parts are much less expansive... even if you lose the brandwith of server mb and cpu, it still might be well worth.

i cannot give any advice, thought i plan to make a systme like your in the near future. the only problem i can foresee (given my narrow experience) is the interconnect part, but you already planned to buy infiniband. (how much are you paying for a IB card, if you don't mind sharing?)


what system are you using for the logistic and housing? i think not normal pc cases.

i've found http://www.server8.it/index.php this company on the web and it's products look interesting (at least for the european market, hardware in the us is much cheaper :-( ), but other options or opinions would be welcome.

kyle May 11, 2011 22:51

sail,

I was planning on just throwing them in some cheap 2U cases. Space isn't really an issue and I will still be able to fit it all in one rack anyway. I am not sure what the Infiniband equipment is going to cost me. I assume something like $2500 for all of it.

The more I research this the less sense the Xeon machines make. Sandy Bridge desktop chips have the highest memory bandwidth per core available, and you can use 2133mhz memory to increase it even more (here is a CFD benchmark for different memory speeds: http://techreport.com/articles.x/20377/2). You cannot use any memory faster than 1333mhz with Xeons. Supposedly you can easily overclock them to ~4ghz as well.

I understand why larger companies might go with the enterprise class stuff but when hardware is such a significant portion of your budget then desktop parts make much more sense. I'll follow up if I have any issues.

sail May 12, 2011 02:11

Quote:

Originally Posted by kyle (Post 307276)
sail,

I was planning on just throwing them in some cheap 2U cases. Space isn't really an issue and I will still be able to fit it all in one rack anyway. I am not sure what the Infiniband equipment is going to cost me. I assume something like $2500 for all of it.

The more I research this the less sense the Xeon machines make. Sandy Bridge desktop chips have the highest memory bandwidth per core available, and you can use 2133mhz memory to increase it even more (here is a CFD benchmark for different memory speeds: http://techreport.com/articles.x/20377/2). You cannot use any memory faster than 1333mhz with Xeons. Supposedly you can easily overclock them to ~4ghz as well.

I understand why larger companies might go with the enterprise class stuff but when hardware is such a significant portion of your budget then desktop parts make much more sense. I'll follow up if I have any issues.

definitely desktops hardware is faster and cheaper, while server grade equipment is more oriented toward reliability, but for a start-up company the price/performance gain is unbeatable. Can i just suggest to wait 3 weeks until the new amd bulldozer octa-core are launched and to check the benchmarks? they support high freq. ram, and the larger (double compared to SB) number of cores, coupled with the new fpu architecture and high frequencies (3-4 ghz) should make them really interesting for our field of application. but don't forget that server cpu usually have lots of cache, and i don't know how much it impact the performances, being several times faster than ram. i don't know if it is it or the ram the "bottleneck" (relatively speaking)

if you really get IB setup for so cheap please let me know, here in Europe they want 200-300 gbp for a card and approx 3000£ for the smallest switch :mad:. it would be cheaper for me to have an holiday in the us and do the shopping there. :p

about the unforeseen problems, i might recommend asking on the bewoulf mailing list. it is a list which deal about bewoulf-type cluster (cluster of cheap commodity hardware) and you might have some useful insight or recommendations about putting IB cards on desktop mb.

sail May 12, 2011 10:31

and while is not strictly cfd, but i suppose massively parallelized routines, chek out what they did in the Viacom Labs, http://www.anandtech.com/show/4332/v...-gets-bigger/2

http://images.anandtech.com/reviews/...P1000959sm.jpg

Bob2011 May 19, 2011 14:10

very interesting project... the Sandy bridge i7 outperforms most server cpus...

However, i have a question regarding the file i/o, especially when you scale the system up to 60~80 cpu. how you handle the massive files? to build a file i/o node will be quite expensive. you don't want to login into each individual node to copy your file. you have any solutions on this?

Bob

Quote:

Originally Posted by kyle (Post 307065)
We currently have a four node cluster built with i7 980X chips and cheap motherboards, RAM and power supplies. This has run fine for us for about six months, and now I am planning on scaling it up with more of the 980X's or with the newer Sandy Bridge i7's. I'm also planning on picking up some used Infiniband equipment.

Is this something anyone else has done at a scale larger than a few nodes? I am looking to add about ~60-80 CPU which could be done for as cheap as $6,000-$7,000 USD using desktop parts. Picking up similar Xeon machines from HP or Dell would be north of $25,000.

I expect the Xeons would scale up better if for no other reason than you can put two CPUs in one motherboard, reducing the traffic between nodes. On the other hand, I could buy 3x as many of the desktop machines and still come out saving money.

Thoughts? Are Infiniband cards happy in desktop motherboards? Will not having ECC RAM randomly blow up my simulations (it hasn't so far). Is there something else I am missing?


kyle May 19, 2011 17:41

Bob,

You definitely do not need to copy files around to each node. Any HPC cluster, whether built from commodity parts or purchased as a system, should have a networked filesystem that all nodes have access to. My file storage node has 4 2TB drives in RAID 5. That whole setup only cost like $1000. Unless you are doing something funky like saving the entire flow field history at a high time resolution, then your simulation speed is not going to be limited by filesystem I/O.

sail May 26, 2011 21:09

hi there. how is the shopping/planning going?

just happened to stumble upon a discussion about ecc vs non-ecc ram, might be worth a look.

http://www.beowulf.org/archive/2011-May/028799.html

best regards.

kyle May 27, 2011 18:00

sail,

Interesting discussion. They seem pretty split on whether or not ECC is necessary. I think CFD might have an advantage in that a flipped is extremely unlikely to give you a wrong answer... it could only crash the simulation or throw a nonsense value into a cell that would be corrected in the next iteration.

I have ordered 12 i7 2600k with MSI P67 motherboards and 8gb of 2133mhz for each node. I am going to try and get the thing running on ethernet before I go hunting for Infiniband equipment. Scaling should be pretty good up to 4 nodes on gigabit. I'll let you know how it goes. I am hoping to have it up and running by the third week in June.

Attesz May 30, 2011 06:53

Hi there,

it's a really interesting topic. I would build a small cluster my starting small CFD company. To keep the prices down, I would prefer AMD CPU's, namely AMD Phenom2 X6 1100T processors. Do you have any experience using this stuff? I would by a middle category motherboard or maybe server ones (with 2 procs), and I would like to use DDR3 rams at least 1333MHz or more and 16GB's each node. At the first time I would have only 2 nodes, so they will communicate using Gigabit ethernet lan. I would use an other "node" which is a weaker PC (dualcore, 8gb ram) which one would be the file server containing some 2TB HDD's in RAID. I would use OpenFOAM on them. What do you think?

Regards,
Attila

sail June 30, 2011 10:55

Hello there. how is the work going? is your cluster already up and running? any impressions you would like to share?

Attesz July 1, 2011 06:06

Quote:

Originally Posted by sail (Post 314259)
Hello there. how is the work going? is your cluster already up and running? any impressions you would like to share?

I've read about clusters and processors a little bit more, and I found that the Sandy Bridge processors are much better than Phenoms. I would buy them, but now this cluster-building is not actual for me:)

kyle July 1, 2011 14:26

I've got 15 cases, power supplies and processors sitting in an empty office. The rest of the parts should be here in a few days.

AlgoTrader July 6, 2011 10:59

kyle

I am playing now with Phenom ii X6, but would like to know more about your cluster configuration and prices

can you give us contact to your suplier please ? My suplier 60USD per AMD phenom ii X6 3.2Ghz

sail August 16, 2011 01:43

Hi Kyle.

just wondering how it is working your new cluster...

kyle August 16, 2011 10:41

sail,

I got it up and running a couple weeks ago. We went with 15 i7 2600k's with 8gb of RAM each. I bought some old Infiniband equipment off of Ebay for very cheap, and it all worked fine. The scaling is almost perfectly linear on our simulations all the way up to 60 CPU.

I am definitely happy we did it this way. All together we spent just under $10,000 USD. If we had bought an HPC system or a bunch of servers from Dell or HP, it would have been $30,000 minimum.

sail August 16, 2011 22:20

Quote:

Originally Posted by kyle (Post 320312)
sail,

I got it up and running a couple weeks ago. We went with 15 i7 2600k's with 8gb of RAM each. I bought some old Infiniband equipment off of Ebay for very cheap, and it all worked fine. The scaling is almost perfectly linear on our simulations all the way up to 60 CPU.

I am definitely happy we did it this way. All together we spent just under $10,000 USD. If we had bought an HPC system or a bunch of servers from Dell or HP, it would have been $30,000 minimum.

Excellent. great work!

just out of curiosity: you decided to go with IB after running some cases/benchmarks over gigalan or you just bite the bullet and grabbed IB from the start?

kyle August 16, 2011 22:38

We already had a four node cluster with gig-e, and were seeing diminshing returns on the speedup. It was never a thought to go all the way up to 15 nodes without a faster interconnect.

sail August 17, 2011 03:48

Quote:

Originally Posted by kyle (Post 320403)
We already had a four node cluster with gig-e, and were seeing diminshing returns on the speedup. It was never a thought to go all the way up to 15 nodes without a faster interconnect.

Fair enough. thank you for your invaluable testimony. there are not much people that dare to leave the HPC "turnkey" / bunch of servers path and even less that are willing to share their experience.

kudos!

donald November 8, 2011 12:08

Have you ran a HPL benchmark on the system. Is there a measure of the Power consumption. I am very interested in the performance and power efficiency of your setup.

kyle November 8, 2011 17:07

The power consumption is around 130W per machine, so it's pretty much right at 2kW total.

I have not run a generic benchmark, but it is about 15% faster per core for our simulations (~50 million unstructured hex cell incompressible isothermal RANS with Star-CCM+ and OpenFOAM) than the supercomputer that we purchased time on. The supercomputer has dual Xeon 5680 per node with QDR Infiniband. I think we are faster because of better memory bandwidth. The Xeon's are limited to 1333mhz RAM and we are running ours at 1600mhz. I think there are also some inefficiencies in the multiple socket NUMA architecture, as not all memory access will be local to a socket.

I overclocked the chips from 3.4ghz to 4.0ghz and only saw about a 3% increase in solver speed. I think we would see some real gains from overclocking the RAM, but the stuff we bought is of too low quality to even boot beyond 1600mhz.

I'll run HPL.

Docfreezzzz January 18, 2012 13:24

Quote:

Originally Posted by kyle (Post 308464)
Unless you are doing something funky like saving the entire flow field history at a high time resolution, then your simulation speed is not going to be limited by filesystem I/O.

I would suggest otherwise. Our machine is bottlenecked at any point in which we write a flow solution. With a machine as large as 60+ nodes you will have to wait for I/O without some very expensive and carefully considered hardware. The good news is that steady state solutions don't write often but the bad news is that if you run unsteady and want to postprocess the time accurate data, you will likely become pretty frustrated without some pretty quick hardware.

kyle January 18, 2012 13:46

Well yeah, that is the exactly the situation I described as when IO throughput is important...

I would say that this situation does not represent the majority of users. It only applies to people that are 1) running transient simulations, 2) care about the flow-field history, and 3) Cannot post-process on the fly. Sure, those three things apply to a lot of users, but probably not most. Certainly not in industry.


All times are GMT -4. The time now is 20:46.