CFD Online Discussion Forums

CFD Online Discussion Forums (https://www.cfd-online.com/Forums/)
-   OpenFOAM (https://www.cfd-online.com/Forums/openfoam/)
-   -   OPENFOAM Parallel Computing, No solution and Sudden Exit with no Solution (https://www.cfd-online.com/Forums/openfoam/236426-openfoam-parallel-computing-no-solution-sudden-exit-no-solution.html)

bigoqwerty May 29, 2021 10:15

OPENFOAM Parallel Computing, No solution and Sudden Exit with no Solution
 
Hello All,


I try to test Parallel computing with openfoam tutorials. I make sure that Openmpi is installed correctly. So there is no problem there. I also check mpi selector and such(omni_run) to see if it makes any differences. (I have openmpi1 and openmpi2, with Openfoam V1912,2006,2012 installed)



When I try parallel computing to test with various cases, I see that simulation exiting either in recompose phase or at pimpleFoam (or any other) phase after meshing without any solution.


There is no errors. Just solution doesn't start and sudden exit.



It is obviously due to my system (Opensuse leap 15.2). Do you have any suggestions?


Thanks

bigoqwerty May 29, 2021 15:41

This is the log file from pimpleFoam. It appears MPI_INIT doesn't kick in.



Code:

--------------------------------------------------------------------------
It looks like MPI_INIT failed for some reason; your parallel process is
likely to abort.  There are many reasons that a parallel process can
fail during MPI_INIT; some of which are due to configuration or environment
problems.  This failure appears to be an internal failure; here's some
additional information (which may only be relevant to an Open MPI
developer):

  ompi_mpi_init: ompi_rte_init failed
  --> Returned "(null)" (-43) instead of "Success" (0)
--------------------------------------------------------------------------
--------------------------------------------------------------------------
It looks like MPI_INIT failed for some reason; your parallel process is
likely to abort.  There are many reasons that a parallel process can
fail during MPI_INIT; some of which are due to configuration or environment
problems.  This failure appears to be an internal failure; here's some
additional information (which may only be relevant to an Open MPI
developer):

  ompi_mpi_init: ompi_rte_init failed
  --> Returned "(null)" (-43) instead of "Success" (0)
--------------------------------------------------------------------------
[localhost.localdomain:04442] *** An error occurred in MPI_Init_thread
[localhost.localdomain:04442] *** on a NULL communicator
[localhost.localdomain:04442] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
[localhost.localdomain:04442] ***    and potentially your MPI job)
[localhost.localdomain:04441] *** An error occurred in MPI_Init_thread
[localhost.localdomain:04441] *** on a NULL communicator
[localhost.localdomain:04441] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
[localhost.localdomain:04441] ***    and potentially your MPI job)
[localhost.localdomain:04441] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
[localhost.localdomain:04442] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
--------------------------------------------------------------------------
It looks like MPI_INIT failed for some reason; your parallel process is
likely to abort.  There are many reasons that a parallel process can
fail during MPI_INIT; some of which are due to configuration or environment
problems.  This failure appears to be an internal failure; here's some
additional information (which may only be relevant to an Open MPI
developer):

  ompi_mpi_init: ompi_rte_init failed
  --> Returned "(null)" (-43) instead of "Success" (0)
--------------------------------------------------------------------------
[localhost.localdomain:04440] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
[localhost.localdomain:04440] *** An error occurred in MPI_Init_thread
[localhost.localdomain:04440] *** on a NULL communicator
[localhost.localdomain:04440] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
[localhost.localdomain:04440] ***    and potentially your MPI job)
[localhost.localdomain:04443] *** An error occurred in MPI_Init_thread
[localhost.localdomain:04443] *** on a NULL communicator
[localhost.localdomain:04443] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
[localhost.localdomain:04443] ***    and potentially your MPI job)
--------------------------------------------------------------------------
It looks like MPI_INIT failed for some reason; your parallel process is
likely to abort.  There are many reasons that a parallel process can
fail during MPI_INIT; some of which are due to configuration or environment
problems.  This failure appears to be an internal failure; here's some
additional information (which may only be relevant to an Open MPI
developer):

  ompi_mpi_init: ompi_rte_init failed
  --> Returned "(null)" (-43) instead of "Success" (0)
--------------------------------------------------------------------------
[localhost.localdomain:04443] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
-------------------------------------------------------
Primary job  terminated normally, but 1 process returned
a non-zero exit code.. Per user-direction, the job has been aborted.
-------------------------------------------------------
--------------------------------------------------------------------------
mpirun detected that one or more processes exited with non-zero status, thus causing
the job to be terminated. The first process to do so was:

  Process name: [[49625,1],1]
  Exit code:    1
--------------------------------------------------------------------------


bigoqwerty May 29, 2021 17:14

I removed all versions V1912,2006 and 2012 and openmpi and openmpi2. Reinstalled 2012. Running a parallel example SimpleFoam turbineSitting. I see no CPU usage and Running simpleFoam (4 processes) goes on.



If you have an any idea. Pls. share Thanks


All times are GMT -4. The time now is 16:48.