CFD Online Logo CFD Online URL
www.cfd-online.com
[Sponsors]
Home > Forums > Software User Forums > OpenFOAM > OpenFOAM Running, Solving & CFD

OpenFoam parallel on 2 computers : Cannot find file "points"

Register Blogs Members List Search Today's Posts Mark Forums Read

Reply
 
LinkBack Thread Tools Search this Thread Display Modes
Old   June 3, 2015, 12:59
Default OpenFoam parallel on 2 computers : Cannot find file "points"
  #1
New Member
 
Blue8655's Avatar
 
Join Date: Jun 2015
Location: Montreal
Posts: 11
Rep Power: 11
Blue8655 is on a distinguished road
Hi everyone,

it's my first post on this forum but I'll try to describe my problem with sufficient details.

I've been experimenting with OpenFoam (2.4.0) and Linux Terminal for a few days and so far so good. I've done some tutorials from the UserGuide. I can run OpenFoam on several processors on my computer using mpirun. Paraview shows its just fine. As I try to run OpenFoam in parallel on two computers by wifi however, things go bad. I've been stuck for many hours, although I read posts with similar problems.

Following the steps from tutorial ''Dam Break'' and section 3.4.2. Everything works except the run command.

Steps I did:
Copy the DamBreak directory to DamBreakFine directory
Modify the BlockMeshDict (the tutorial was about refining the mesh...longer simulation)
Reset alpha.water to uniform field with backup
Run SetFields (OK)
Run DecomposePar (OK)
Write a file named machines in the case directory:

Code:
blue@blue-HP-Pavilion-g6-Notebook-PC.local cpu=2
green1@green1-Satellite-A200.local cpu=2
I know that data can be split on several disks but it should also work to only use CPU power of the second computer through mpirun.

Run "mpirun --hostfile machines -np 4 icoFoam -parallel "
Terminal asks for green1 password (slave machine) ... OK
Then..... I get the following error.......

Code:
/*---------------------------------------------------------------------------*\
| =========                 |                                                 |
| \\      /  F ield         | OpenFOAM: The Open Source CFD Toolbox           |
|  \\    /   O peration     | Version:  2.4.0                                 |
|   \\  /    A nd           | Web:      www.OpenFOAM.org                      |
|    \\/     M anipulation  |                                                 |
\*---------------------------------------------------------------------------*/
Build  : 2.4.0-f0842aea0e77
Exec   : icoFoam -parallel
Date   : Jun 03 2015
Time   : 10:11:35
Host   : "blue-HP-Pavilion-g6-Notebook-PC"
PID    : 9695
Case   : /home/blue/OpenFOAM/blue-2.4.0/run/damBreakFine
nProcs : 4
Slaves : 
3
(
"blue-HP-Pavilion-g6-Notebook-PC.9696"
"green1-Satellite-A200.5186"
"green1-Satellite-A200.5187"
)

Pstream initialized with:
    floatTransfer      : 0
    nProcsSimpleSum    : 0
    commsType          : nonBlocking
    polling iterations : 0
sigFpe : Enabling floating point exception trapping (FOAM_SIGFPE).
fileModificationChecking : Monitoring run-time modified files using timeStampMaster
allowSystemOperations : Allowing user-supplied system call operations

// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * //
Create time

Create mesh for time = 0

[2] 
[2] 
[2] --> FOAM FATAL ERROR: 
[2] Cannot find file "points" in directory "polyMesh" in times 0 down to constant
[2] 
[2] --------------------------------------------------------------------------
MPI_ABORT was invoked on rank 2 in communicator MPI_COMM_WORLD 
with errorcode 1.

NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
    From function Time::findInstance(const fileName&, const word&, const IOobject::readOption, const word&)
[2]     in file db/Time/findInstance.C at line 203.
[2] 
FOAM parallel run exiting
[2] 
[3] 
[3] 
[3] --> FOAM FATAL ERROR: 
[3] Cannot find file "points" in directory "polyMesh" in times 0 down to constant
[3] 
[3]     From function Time::findInstance(const fileName&, const word&, const IOobject::readOption, const word&)
[3]     in file db/Time/findInstance.C at line 203.
[3] 
FOAM parallel run exiting
[3] 
--------------------------------------------------------------------------
mpirun has exited due to process rank 2 with PID 5186 on
node green1-Satellite-A200.local exiting improperly. There are two reasons this could occur:

1. this process did not call "init" before exiting, but others in
the job did. This can cause a job to hang indefinitely while it waits
for all processes to call "init". By rule, if one process calls "init",
then ALL processes must call "init" prior to termination.

2. this process called "init", but exited without calling "finalize".
By rule, all processes that call "init" MUST call "finalize" prior to
exiting or it will be considered an "abnormal termination"

This may have caused other processes in the application to be
terminated by signals sent by mpirun (as reported here).
--------------------------------------------------------------------------
[blue-HP-Pavilion-g6-Notebook-PC:09692] 1 more process has sent help message help-mpi-api.txt / mpi-abort
[blue-HP-Pavilion-g6-Notebook-PC:09692]  Set MCA parameter "orte_base_help_aggregate" to 0 to see all help /  error messages
It seems CPU [2] (slave computer) cannot access the required files that are on the main computer, but the file "points" is there in case/constant/polymesh and in processor0/constant/polymesh, processor1, etc. Is it that ssh doesnt give access to the main computer from the second?

Also, I must admit I don't understand this part of the tutorial:
"The < machines > file contains the names of the machines listed one machine per line. The names must correspond to a fully resolved hostname in the /etc/hosts file of the machine on which the openMPI is run."

Your help is very much appreciated.

Fred
Blue8655 is offline   Reply With Quote

Old   June 3, 2015, 21:59
Default
  #2
New Member
 
Qiu Xiaoping
Join Date: Apr 2013
Location: IPE CAS China
Posts: 14
Rep Power: 14
xpqiu is on a distinguished road
Hi, Fred

Quote:
It seems CPU [2] (slave computer) cannot access the required files that are on the main computer, but the file "points" is there in case/constant/polymesh and in processor0/constant/polymesh, processor1, etc. Is it that ssh doesnt give access to the main computer from the second?
Yes, slave cannot access the mesh data on the master, that's why mpirun complain " Cannot find file "points" in directory "polyMesh" "。You should put your case data in a space which can be accessed by both master and slaves, one way is to share a directory (say,"$HOME/shares") on your master to all of your slaves machines. (run command like this as root on your slave machines: mount -t nfs master:$HOME/shares ~/shares ).

Quote:
Also, I must admit I don't understand this part of the tutorial:
"The < machines > file contains the names of the machines listed one machine per line. The names must correspond to a fully resolved hostname in the /etc/hosts file of the machine on which the openMPI is run."
It means that you should make a map between your machine name and IP address of your machines in the /etc/hosts file , for example
Code:
192.168.2.11     green1  
 192.168.2.12      blue
xpqiu is offline   Reply With Quote

Reply

Tags
parallel computing

Thread Tools Search this Thread
Search this Thread:

Advanced Search
Display Modes

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is On
HTML code is Off
Trackbacks are Off
Pingbacks are On
Refbacks are On


Similar Threads
Thread Thread Starter Forum Replies Last Post
[swak4Foam] funkyDoCalc with OF2.3 massflow NiFl OpenFOAM Community Contributions 14 November 25, 2020 03:30
[swak4Foam] groovyBC in openFOAM-2.0 for parabolic velocity bc ofslcm OpenFOAM Community Contributions 25 March 6, 2017 10:03
what is swap4foam ?? AB08 OpenFOAM 28 February 2, 2016 01:22
pisoFoam compiling error with OF 1.7.1 on MAC OSX Greg Givogue OpenFOAM Programming & Development 3 March 4, 2011 17:18
DecomposePar links against liblamso0 with OpenMPI jens_klostermann OpenFOAM Bugs 11 June 28, 2007 17:51


All times are GMT -4. The time now is 22:44.