CFD Online Discussion Forums

CFD Online Discussion Forums (http://www.cfd-online.com/Forums/)
-   OpenFOAM (http://www.cfd-online.com/Forums/openfoam/)
-   -   Running AMI case in parallel (http://www.cfd-online.com/Forums/openfoam/103051-running-ami-case-parallel.html)

Kaskade June 9, 2012 08:47

Running AMI case in parallel
 
Hello,
me again, with a new problem.
I've set up a case involving 3 AMIs which runs fine using MRFSimplefoam (1 core and multiple cores). But when I use the converged solution of the SteadyState-case and use it as the starting point for the transient simulation, I can't run it in parallel, while running on one core works fine.
When I decompose the case (simple/scotch) and run it in parallel I get the following error:
Quote:

[1]
[1]
[1] --> FOAM FATAL IO ERROR:
[1] Read 747304 undisplaced points from "/scratch/kaskade/OpenFOAM/kaskade-2.1.0/run/DA013-Netz3KEpsTransientParallel/processor1/constant/polyMesh/points" but the current mesh has 204035
[1]
[1] file: IOstream::solidBodyMotionFvMeshCoeffs from line 0 to line 0.
[1]
[1] From function solidBodyMotionFvMesh::solidBodyMotionFvMesh(const IOobject&)
[1] in file solidBodyMotionFvMesh/solidBodyMotionFvMesh.C at line 86.
[1]
FOAM parallel run exiting
[1]
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD
with errorcode 1.

NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
[2]
[2]
[2] --> FOAM FATAL IO ERROR:
[2] Read 747304 undisplaced points from "/scratch/kaskade/OpenFOAM/kaskade-2.1.0/run/DA013-Netz3KEpsTransientParallel/processor2/constant/polyMesh/points" but the current mesh has 203734
[2]
[2] file: IOstream::solidBodyMotionFvMeshCoeffs from line 0 to line 0.
[2]
[2] From function solidBodyMotionFvMesh::solidBodyMotionFvMesh(const IOobject&)
[2] in file solidBodyMotionFvMesh/solidBodyMotionFvMesh.C at line 86.
[2]
FOAM parallel run exiting
[2]
[0]
[0]
[0] --> FOAM FATAL IO ERROR:
[0] Read 747304 undisplaced points from "/scratch/kaskade/OpenFOAM/kaskade-2.1.0/run/DA013-Netz3KEpsTransientParallel/processor0/constant/polyMesh/points" but the current mesh has 186025
[0]
[0] file: /scratch/kaskade/OpenFOAM/kaskade-2.1.0/run/DA013-Netz3KEpsTransientParallel/processor0/../constant/dynamicMeshDict::solidBodyMotionFvMeshCoeffs from line 24 to line 30.
[0]
[0] From function solidBodyMotionFvMesh::solidBodyMotionFvMesh(const IOobject&)
[0] in file solidBodyMotionFvMesh/solidBodyMotionFvMesh.C at line 86.
[0]
FOAM parallel run exiting
[0]
[3]
[3]
[3] --> FOAM FATAL IO ERROR:
[3] Read 747304 undisplaced points from "/scratch/kaskade/OpenFOAM/kaskade-2.1.0/run/DA013-Netz3KEpsTransientParallel/processor3/constant/polyMesh/points" but the current mesh has 185112
[3]
[3] file: IOstream::solidBodyMotionFvMeshCoeffs from line 0 to line 0.
[3]
[3] From function solidBodyMotionFvMesh::solidBodyMotionFvMesh(const IOobject&)
[3] in file solidBodyMotionFvMesh/solidBodyMotionFvMesh.C at line 86.
[3]
FOAM parallel run exiting
[3]
--------------------------------------------------------------------------
mpirun has exited due to process rank 1 with PID 4799 on
node node119 exiting improperly. There are two reasons this could occur:

1. this process did not call "init" before exiting, but others in
the job did. This can cause a job to hang indefinitely while it waits
for all processes to call "init". By rule, if one process calls "init",
then ALL processes must call "init" prior to termination.

2. this process called "init", but exited without calling "finalize".
By rule, all processes that call "init" MUST call "finalize" prior to
exiting or it will be considered an "abnormal termination"

This may have caused other processes in the application to be
terminated by signals sent by mpirun (as reported here).
--------------------------------------------------------------------------
[node119:04797] 3 more processes have sent help message help-mpi-api.txt / mpi-abort
[node119:04797] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages
I am using OpenFOAM 2.1.1., my dynamicMeshDict is (apart from the name of the rotating area) identical to the propeller-tutorial.
Thanks in advance.

Kaskade June 23, 2012 05:30

Solution: Someone messed up the decomposePar of OF211. If you decompose the case with OF210's decomposePar, it runs fine with OF211.

Kaskade November 19, 2013 03:24

Just for future reference: When you encounter this errors, because you want to start the DyM-run from a time != 0. You need to run "decomposePar -constant -time time" . I think the mesh for a specific time step is transformed from the original mesh and not from the mesh of the previous time step.


All times are GMT -4. The time now is 04:46.