CFD Online Discussion Forums

CFD Online Discussion Forums (http://www.cfd-online.com/Forums/)
-   OpenFOAM Installation on Windows, Mac and other Unsupported Platforms (http://www.cfd-online.com/Forums/openfoam-installation-windows-mac/)
-   -   parallel performance on BX900 (http://www.cfd-online.com/Forums/openfoam-installation-windows-mac/83205-parallel-performance-bx900.html)

uzawa December 20, 2010 00:38

parallel performance on BX900
 
1 Attachment(s)
Dear All,

OpenFOAM v1.6 has been successfully installed on a supercomputer at Japan Atomic Energy Agency. The supercomputer system is a hybrid system consisting of three computational server systems, i.e., (I) Large-scale Parallel Computation Unit, (II) Application Development Unit for the Next Generation Supercomputer, and (III) SMP Server. The Large-scale Parallel Computation Unit uses PRIMERGY BX900, which is the Fujitsu's latest blade server with 2134 nodes (4268 CPUs, 17072 cores) connected using the latest InfiniBand QDR high-speed interconnect technology. The details of the Large-scale Parallel Computation Unit are as follows.

CPU: Intel Xeon processor X5570 (2.93GHz)2CPU
level one cache(L1):256K
secondary cache(L2):1MB
third-level cache(L3):8MB
number of cores: 4 cores/CPU
node communication performance:8GB/s
OS: Red Hat Enterprise Linux 5

Based on the LINPACK performance benchmark, the supercomputer achieved performance of 186.1 teraflops, which made it the fastest one in Japan based on the latest TOP500 list of supercomputers at the date of this October.

I would like to report parallel performance up to 256 cores on the Large-scale Parallel Computation Unit. I thought it will be a good idea to share it for supercomputer users in any form. I hope this information helps you if only a little.
Here, a simplified three-dimensional dam break problem is chosen as a test example and the two-phase flow is solved an interFoam solver. Numerical conditions are same in experimental settings as used in Martin[1] and Koshizuka[2].
[1] J.C. Martin and W.J. Moyce, ”PartIV. An experimental study of the collapse of liquid columns on a rigid horizontal plane ”, Phil. Trans. R. Soc. Lond. A, 244, 312-324 (1952).
[2] S. Koshizuka, H. Tamako, Y. Oka, "A particle method for incompressible viscous flow with fluid fragmentation", Computational Fluid Mechanics Journal, 113, 134-147 (1995).

It is found that it scales well for up to 128 cores, yet maintains excellent performance levels even on 256 cores. (Please see the attached file for details.)
Parallel performance up to full cores (17072 cores) will be reported later.

niklas December 20, 2010 01:48

If you instead plot the numbers of cell per core, what would the numbers be?
I usually tries to go for approximately 50k cells / core, lower than that is not worth it

uzawa December 22, 2010 04:29

Dear Niklas Nordin

Thank you very much for your interest in my work. I would be happy to try to answer your question.

Quote:

Originally Posted by niklas (Post 287806)
If you instead plot the numbers of cell per core, what would the numbers be?
I usually tries to go for approximately 50k cells / core, lower than that is not worth it

In this case, number of total cells is approximately 8 million. Consequently, up to 128 cores, this choice meets your requirement. As you indicated, I am planning to perform more simulations by increasing the number of cells from 8 millions to tens of millions. Thank you very much for pointing that out.

lakeat September 5, 2011 16:52

Quote:

Parallel performance up to full cores (17072 cores) will be reported later.
Any updates?


All times are GMT -4. The time now is 11:52.