CFD Online Logo CFD Online URL
www.cfd-online.com
[Sponsors]
Home > Forums > General Forums > Hardware

Regarding Network Switch

Register Blogs Community New Posts Updated Threads Search

Like Tree7Likes
  • 1 Post By flotus1
  • 2 Post By flotus1
  • 2 Post By evcelica
  • 2 Post By flotus1

Reply
 
LinkBack Thread Tools Search this Thread Display Modes
Old   October 5, 2015, 07:12
Default Regarding Network Switch
  #1
Member
 
Mohammed Gowhar
Join Date: Feb 2014
Posts: 48
Rep Power: 12
mdgowhar is on a distinguished road
Hi everyone!

I have ordered High Performance Scientific Computing Servers.

Usage:
STAR-CCM+ [HPC license]
OpenFOAM
Salome for FEA work

Configuration (edited) [ Thanks Alex! I have made mistake while writing.]
Master Node:
2 x AMD Opteron 6308 @3.5GHz 4-core [8 cores]
6 x 16GB RAM 1600MHz ECC Registered [In Total 96 GB]

Compute Node:
4 x AMD Opteron 6308 @3.5GHz 4-core [16 cores]
8 x 16GB RAM 1600MHz ECC Registered [In Total 128 GB]

(1 x 8 cores in Master Node) +(3 x 16 cores in Compute Node) = 56 cores in total

Since buying Infiniband card and switch are out of my budget, I would like to interconnect using Manageable Ethernet switch. Please anyone suggest me which product is best to buy. And Suitable Operating System.

Thanks in advance.

Regards,
Mohammed Gowhar

Last edited by mdgowhar; October 5, 2015 at 08:41. Reason: Thanks Alex! I have made mistake while writing.
mdgowhar is offline   Reply With Quote

Old   October 5, 2015, 08:27
Default
  #2
Super Moderator
 
flotus1's Avatar
 
Alex
Join Date: Jun 2012
Location: Germany
Posts: 3,399
Rep Power: 46
flotus1 has a spectacular aura aboutflotus1 has a spectacular aura about
If I am not mistaken the memory on your compute nodes is poorly configured.
The AMD Opteron 6308 has 4 memory channels. 4 of these CPUs on one node need 16 DIMMs to achive maximum memory bandwidth.
flotus1 is offline   Reply With Quote

Old   October 5, 2015, 08:37
Default
  #3
Member
 
Mohammed Gowhar
Join Date: Feb 2014
Posts: 48
Rep Power: 12
mdgowhar is on a distinguished road
Quote:
Originally Posted by flotus1 View Post
If I am not mistaken the memory on your compute nodes is poorly configured.
The AMD Opteron 6308 has 4 memory channels. 4 of these CPUs on one node need 16 DIMMs to achive maximum memory bandwidth.
Im sorry.. My mistake

In compute node:
8 x 16GB RAM [In Total 128GB} Equally populated.
This board is capable of 1TB ECC registered RAM

In master node:
6 x 16GB RAM [In Total 96GB} Equally populated.
This board is capable of 512GB ECC registered RAM
mdgowhar is offline   Reply With Quote

Old   October 5, 2015, 10:04
Default
  #4
Super Moderator
 
flotus1's Avatar
 
Alex
Join Date: Jun 2012
Location: Germany
Posts: 3,399
Rep Power: 46
flotus1 has a spectacular aura aboutflotus1 has a spectacular aura about
I still think that with only 8 DIMMs per compute node, you are wasting half of the theoretical maximum memory bandwidth of your CPUs. Equally populating the 8 DIMMs among the 4 CPUs will result only in dual-channel mode for each CPU.
See for example the recommendations from Supermicro: http://www.supermicro.com/aplus/memo...%28%2B%29-LN4F
I know this is not the topic of this thread, but I thought it might be worth pointing this out.
Eike likes this.
flotus1 is offline   Reply With Quote

Old   October 6, 2015, 02:56
Default
  #5
Member
 
Mohammed Gowhar
Join Date: Feb 2014
Posts: 48
Rep Power: 12
mdgowhar is on a distinguished road
I strongly agree with you point. But I dont have budget to buy more RAMs.

Moreover If I buy the 8GB RAM with same total memory size, The total modulus occupied by 8GB and it will be difficult to upgrade in the future and become costlier.
Thats why I opt for dual Rank for now. May be after one or two years I can upgrade it with adding additional 16GB RAM to fill the remaining slots.

I really thank you for pointing out this. Definitely I look forward for alternative solution to have better performance.

Since I already placed order, Is it worth to buy additional 8GB module to fill remaining slots?
mdgowhar is offline   Reply With Quote

Old   October 6, 2015, 04:04
Default
  #6
Super Moderator
 
flotus1's Avatar
 
Alex
Join Date: Jun 2012
Location: Germany
Posts: 3,399
Rep Power: 46
flotus1 has a spectacular aura aboutflotus1 has a spectacular aura about
I would avoid this for two reasons.
The first one is stability. Using different types of DIMMs is to some extent a game of chance, especially in a HPC system with a rather complicated memory setup.
The second reason is performance. Even In the ideal case that your system supports this asymmetric quad-socket quad-channel setup, only 2/3 of the total RAM will run in quad-channel mode. The rest can only be accessed much slower which will be a bottleneck for cases that need much RAM. But in reality, this exotic setup could even decrease overall performance.

As far as your original question with the network switches is concerned: maybe buying used Infiniband cards, cables and a switch is an option for you to save some money. It might be worth it.
http://www.hpcadvisorycouncil.com/pd...igE_in_HPC.pdf

Edit: Are you sure that the motherboards you ordered only have 16 DIMM slots? All 4-socket Opteron boards I could find have 32 slots.

Last edited by flotus1; October 6, 2015 at 06:33.
flotus1 is offline   Reply With Quote

Old   October 6, 2015, 07:24
Default
  #7
Member
 
Mohammed Gowhar
Join Date: Feb 2014
Posts: 48
Rep Power: 12
mdgowhar is on a distinguished road
Thanks for your valuable suggestion.
Yes, 4-Socket Opteron Mother Board has 32 slots.

If I installed RAM in dual channel mode and run the simulation within 16 cores, how far it will affect the performance (in terms of %).

Because, Already I have other expenses such as Rack Cabinet [42U], KVM Switches and PDU [Schneider Electrics].
mdgowhar is offline   Reply With Quote

Old   October 6, 2015, 08:28
Default
  #8
Super Moderator
 
flotus1's Avatar
 
Alex
Join Date: Jun 2012
Location: Germany
Posts: 3,399
Rep Power: 46
flotus1 has a spectacular aura aboutflotus1 has a spectacular aura about
Unfortunately I cant find the original thread where this was discussed. But here are some results that show how memory frequency (and thus bandwidth) can affect performance in CFD.
scaling.jpg

Based on these charts alone, quad-channel memory could increase performance by up to 70% compared to dual-channel. But since the results do not include latency which was probably slower for the faster memory frequencies, the benefit of quad-channel memory could even be higher.
flotus1 is offline   Reply With Quote

Old   October 6, 2015, 09:20
Default
  #9
Member
 
Mohammed Gowhar
Join Date: Feb 2014
Posts: 48
Rep Power: 12
mdgowhar is on a distinguished road
Quote:
Originally Posted by flotus1 View Post
Unfortunately I cant find the original thread where this was discussed. But here are some results that show how memory frequency (and thus bandwidth) can affect performance in CFD.
Attachment 42513

Based on these charts alone, quad-channel memory could increase performance by up to 70% compared to dual-channel. But since the results do not include latency which was probably slower for the faster memory frequencies, the benefit of quad-channel memory could even be higher.
I would like to clarify regarding this.

This chart is based on CPU Frequency Speed, not on Memory Frequency.
I am thinking of alternative solutions and found a post on Unbalanced Memory Performance for High Performance Computing by Dell.

Please review this link.
http://en.community.dell.com/techcen...ry-performance
mdgowhar is offline   Reply With Quote

Old   October 6, 2015, 09:25
Default
  #10
Super Moderator
 
flotus1's Avatar
 
Alex
Join Date: Jun 2012
Location: Germany
Posts: 3,399
Rep Power: 46
flotus1 has a spectacular aura aboutflotus1 has a spectacular aura about
Please have a closer look at the chart. Yes, on the horizontal axis we have the CPU frequency. But the data series labeled "1600", "1866" and "2133" indicate memory frequencies.
It is nothing new that performance in CFD applications can correlate with memory bandwidth. I would even say that the only justification to buy Opteron Processors for CFD in 2015 is that they offer the cheapest way to get 16 memory channels on one node. If this is enough to compensate the outdated architecture and the lack of FPUs is a different topic.
Eike and mdgowhar like this.
flotus1 is offline   Reply With Quote

Old   October 6, 2015, 11:41
Default
  #11
Senior Member
 
Erik
Join Date: Feb 2011
Location: Earth (Land portion)
Posts: 1,167
Rep Power: 23
evcelica is on a distinguished road
Like Flotus1 says, You really need to balance your memory for good performance. I ran some benchmarks on a 4 CPU XEON with and without unbalanced memory, and it was 2.18x faster with balanced memory. The performance was horrible with unbalanced memory.

The machine had 6 DIMMS per CPU, so it was running in Quad channel configuration, but 2 channels had 2 DIMMS, and 2 channels only had 1 DIMM. I had them purchase 2 more DIMMS per CPU to populate the channels evenly, and performance skyrocketed.

If you can fill the channels evenly, and use all 4 memory channels on each machine, that would be best.
Just use less machines, but properly configure them if you have to, it doesn't help to have more poorly configured machines.
Attached Images
File Type: jpg BenchmarkGraph.jpg (149.5 KB, 26 views)
Eike and mdgowhar like this.
evcelica is offline   Reply With Quote

Old   October 7, 2015, 05:10
Default
  #12
Member
 
Mohammed Gowhar
Join Date: Feb 2014
Posts: 48
Rep Power: 12
mdgowhar is on a distinguished road
Yea. I got that. Thanks Alex.

Thanks Erik! I would consider your point.

I have a sheet of Memory Configuration for this HPC
Please find the same.

Existing designed is there. Moreover, I have added 4 suggestions.
Kindly review it. If you have any suggestions, kindly modify and send me back.

Existing.png

Suggestion_1.png

Suggestion_2.png

Suggestion_3.png

Suggestion_4.png
mdgowhar is offline   Reply With Quote

Old   October 7, 2015, 05:11
Default
  #13
Member
 
Mohammed Gowhar
Join Date: Feb 2014
Posts: 48
Rep Power: 12
mdgowhar is on a distinguished road
Here is the excel sheet to modify easily.

Memory_Config_HPC.xlsx
mdgowhar is offline   Reply With Quote

Old   October 9, 2015, 08:05
Default
  #14
Super Moderator
 
flotus1's Avatar
 
Alex
Join Date: Jun 2012
Location: Germany
Posts: 3,399
Rep Power: 46
flotus1 has a spectacular aura aboutflotus1 has a spectacular aura about
So you are searching for the "best" way to misconfigure the memory on your system. I strongly discourage buying smaller DIMMs and mixing them with the existing ones, especially if the budget is already exhausted. For now your best option seems to be to use dual-channel mode for all CPUs. If you really cant change the configuration you ordered this appears to be the only reasonable option.
evcelica and Eike like this.
flotus1 is offline   Reply With Quote

Reply


Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is On
HTML code is Off
Trackbacks are Off
Pingbacks are On
Refbacks are On


Similar Threads
Thread Thread Starter Forum Replies Last Post
switch between zerogradient and fixedvalue ahliyu OpenFOAM Pre-Processing 0 January 6, 2014 07:36
IcePak network objects in parallel solver Alexey FLUENT 1 December 25, 2013 08:32
Inlet boundary - Flow switch crevoise STAR-CD 0 August 8, 2013 11:55
48 Core Cluster - GigE Network ronaldalau Hardware 2 September 9, 2011 13:50
uptodate water distribution network fredius,magige,tanzanian,(e.a) Main CFD Forum 0 January 27, 2002 07:10


All times are GMT -4. The time now is 13:48.