CFD Online Logo CFD Online URL
www.cfd-online.com
[Sponsors]
Home > Forums > OpenFOAM

parallel processing

Register Blogs Members List Search Today's Posts Mark Forums Read

Reply
 
LinkBack Thread Tools Display Modes
Old   December 8, 2010, 01:10
Default parallel processing
  #1
Member
 
Sandeep
Join Date: Jul 2010
Posts: 47
Rep Power: 6
gandesk is on a distinguished road
Hello

I am new to OpenFOAM and want to run a case in parallel at an high performance cluster. I am able to run the same case on a single node without any hassle but, I am getting the following error attached in the logfile when I run it parallelly.

Any help would be appreciated!!


Sandeep
Attached Files
File Type: txt logfile.txt (68.2 KB, 29 views)
gandesk is offline   Reply With Quote

Old   December 8, 2010, 02:37
Default
  #2
Senior Member
 
Martin
Join Date: Oct 2009
Location: Aachen, Germany
Posts: 252
Rep Power: 11
MartinB is on a distinguished road
Hi Sandeep,

the error log mentions:
Code:
...environment variable MPI_BUFFER_SIZE not defined...
This might be a good point to start with...

Martin
MartinB is offline   Reply With Quote

Old   December 8, 2010, 11:17
Default further doubts
  #3
Member
 
Sandeep
Join Date: Jul 2010
Posts: 47
Rep Power: 6
gandesk is on a distinguished road
Quote:
Originally Posted by MartinB View Post
Hi Sandeep,

the error log mentions:
Code:
...environment variable MPI_BUFFER_SIZE not defined...
This might be a good point to start with...

Martin

thanks martin for your reply......
as am new to parallel processing, i did not understant what MPI_BUFFER_SIZE is...could you send me any other information .....

Actually i am doing my parallel job on a high performance computers on a super computer cluster....now I am including my qsub file also as i am expecting the possible error could be with defining the type of parallel messeging.....
Attached Files
File Type: txt qsubfile.txt (727 Bytes, 19 views)
gandesk is offline   Reply With Quote

Old   December 8, 2010, 11:23
Default
  #4
Member
 
Costas K
Join Date: Nov 2009
Location: UK
Posts: 51
Rep Power: 7
thekay is on a distinguished road
Send a message via ICQ to thekay Send a message via MSN to thekay
In the file you qsub, make sure you set up an adequate buffer size for the mpi, before you execute the solver command. Something like this is what I use:

Code:
MPI_BUFFER_SIZE=2000000000
and after that mpirun -np 64 etc.

Cheers

Edit: By the way, I think openFOAM must use openmpi and not the intel version you set.
Edit2: And something else: you seem to use 32 processor cores in your qsub file
Code:
#PBS -l nodes=16:ppn=2
yet you decompose into 64.
thekay is offline   Reply With Quote

Old   December 8, 2010, 23:12
Default
  #5
Member
 
Sandeep
Join Date: Jul 2010
Posts: 47
Rep Power: 6
gandesk is on a distinguished road
Quote:
Originally Posted by thekay View Post
In the file you qsub, make sure you set up an adequate buffer size for the mpi, before you execute the solver command. Something like this is what I use:

Code:
MPI_BUFFER_SIZE=2000000000
and after that mpirun -np 64 etc.

Cheers

Edit: By the way, I think openFOAM must use openmpi and not the intel version you set.
Edit2: And something else: you seem to use 32 processor cores in your qsub file
Code:
#PBS -l nodes=16:ppn=2
yet you decompose into 64.

Hi Costas,

I am getting the same error message even when i included MPI_BUFFER_SIZE before solver command.

And yeah I got the same message after correcting my qsub file

with openmpi the cluster is not recognizing the command.

I am including my log and qsub file for 4 processors once again after your suggestions were implimented
Attached Files
File Type: txt log.txt (10.1 KB, 9 views)
File Type: txt qsub.txt (742 Bytes, 22 views)
gandesk is offline   Reply With Quote

Old   December 9, 2010, 03:26
Default
  #6
Senior Member
 
Martin
Join Date: Oct 2009
Location: Aachen, Germany
Posts: 252
Rep Power: 11
MartinB is on a distinguished road
Hi Sandeep,
I suppose you must use
export MPI_BUFFER_SIZE=2000000000
in the qsub file.

Martin
MartinB is offline   Reply With Quote

Old   December 9, 2010, 08:06
Default
  #7
Member
 
Costas K
Join Date: Nov 2009
Location: UK
Posts: 51
Rep Power: 7
thekay is on a distinguished road
Send a message via ICQ to thekay Send a message via MSN to thekay
Yes, I think export is right, instead of just setting the value.
thekay is offline   Reply With Quote

Old   December 9, 2010, 12:02
Default
  #8
Member
 
Sandeep
Join Date: Jul 2010
Posts: 47
Rep Power: 6
gandesk is on a distinguished road
thanq guys....it seems to be working now!!
gandesk is offline   Reply With Quote

Old   December 9, 2010, 18:32
Default
  #9
Member
 
Sandeep
Join Date: Jul 2010
Posts: 47
Rep Power: 6
gandesk is on a distinguished road
Quote:
Originally Posted by gandesk View Post
thanq guys....it seems to be working now!!

Hi,

i have got a new problem again....thanks that you have helped me running the job....now the problem is that my job is getting terminated suddenly and restarting again and again which is eventually terminating. I have included the log file in which if you see line 541 and line 3021 look similar........


I am sending the file as html file as i cnnot send more than 100 kb file
http://homepages.uc.edu/~gandesk/log

Can u help me guys...
gandesk is offline   Reply With Quote

Reply

Thread Tools
Display Modes

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is On
HTML code is Off
Trackbacks are On
Pingbacks are On
Refbacks are On


Similar Threads
Thread Thread Starter Forum Replies Last Post
Parallel processing of OpenFOAM cases on multicore processor??? g.akbari OpenFOAM Running, Solving & CFD 24 May 31, 2015 17:58
HP MPI warning...Distributed parallel processing Peter CFX 10 May 14, 2011 06:17
bubbly flow and parallel processing mvee FLUENT 0 September 12, 2007 05:08
FEDORA CORE and PARALLEL processing Tuks CFX 2 August 20, 2005 11:05
About parallel processing in Linux tuks CFX 10 August 8, 2005 08:22


All times are GMT -4. The time now is 23:02.