|
[Sponsors] |
mpirun detected that one or more processes exited... on remotely accessed |
|
LinkBack | Thread Tools | Search this Thread | Display Modes |
March 18, 2020, 12:08 |
mpirun detected that one or more processes exited... on remotely accessed
|
#1 |
New Member
Join Date: Mar 2020
Posts: 16
Rep Power: 6 |
Hi,
I have been working with OpenFoam for the last few months. Now I wanted to continue working from home ( Home quarantined ) . I am accessing a workstation via Remote desktop connection from my laptop. A parallel running of a dambreak case on 16 processors gives the following error: " Primary job terminated normally, but 1 process returned a non-zero exit code.. Per user-direction, the job has been aborted. ------------------------------------------------------- -------------------------------------------------------------------------- mpirun detected that one or more processes exited with non-zero status, thus causing the job to be terminated. The first process to do so was: Process name: [[27991,1],0] Exit code: 145 " Any help please! Note that: a) similar cases were working fine when I was working in my office. b) The same case is working fine in my laptop c) The same case with a single processor is ok in the workstation d) I am connected via VPN to the company's internet to access the workstation. |
|
April 10, 2020, 16:05 |
I got the same error
|
#2 |
New Member
Ahsan
Join Date: Nov 2019
Location: Bologna, Italy
Posts: 27
Rep Power: 7 |
I ran:
decomposePar then: mpirun -np 3 interMixingFoam -parallel And then I have got the same error: Build : v1906 OPENFOAM=1906 Arch : "LSB;label=32;scalar=64" Exec : interMixingFoam -parallel Date : Apr 10 2020 Time : 20:37:29 Host : DESKTOP-KVKH8JA PID : 2839 fileName::stripInvalid() called for invalid fileName /mnt/c/Users/AhsanKhan/Documents/Docs/theCase/5thCaseRun/mixingTank3DTurbulentDecomposed For debug level (= 2) > 1 this is considered fatal fileName::stripInvalid() called for invalid fileName /mnt/c/Users/AhsanKhan/Documents/Docs/theCase/5thCaseRun/mixingTank3DTurbulentDecomposed For debug level (= 2) > 1 this is considered fatal fileName::stripInvalid() called for invalid fileName /mnt/c/Users/AhsanKhan/Documents/Docs/theCase/5thCaseRun/mixingTank3DTurbulentDecomposed For debug level (= 2) > 1 this is considered fatal ------------------------------------------------------- Primary job terminated normally, but 1 process returned a non-zero exit code.. Per user-direction, the job has been aborted. ------------------------------------------------------- -------------------------------------------------------------------------- mpirun detected that one or more processes exited with non-zero status, thus causing the job to be terminated. The first process to do so was: Process name: [[5421,1],1] Exit code: 1 Kindly someone help me with this issue. |
|
September 7, 2020, 09:45 |
|
#3 |
New Member
Ahsan
Join Date: Nov 2019
Location: Bologna, Italy
Posts: 27
Rep Power: 7 |
I found the error. My computer's username has a space in it so that was the problem.
Unfortunately, the username can not be changed... |
|
September 7, 2020, 13:10 |
|
#4 | |
Senior Member
Mark Olesen
Join Date: Mar 2009
Location: https://olesenm.github.io/
Posts: 1,714
Rep Power: 40 |
Quote:
In the etc/controlDict Code:
InfoSwitches { // The default ASCII write precision writePrecision 6; ... // Allow space character in fileName (use with caution) // Ignored (always 1) for Windows. // Default: allowSpaceInFileName 0; allowSpaceInFileName 1; } This will likely get it working for you. /mark |
||
February 8, 2021, 14:43 |
|
#5 | ||
New Member
Join Date: Jan 2021
Location: Edmonton
Posts: 4
Rep Power: 5 |
Quote:
Were you able to find a solution? My case is very similar. I am running the program in parallel and remotely with Chrome Remote Desktop. The error appears after several iterations. Quote:
|
|||
February 8, 2021, 17:32 |
|
#6 |
New Member
cem
Join Date: Feb 2021
Posts: 6
Rep Power: 5 |
try to add a hostfile in the case folder and run the case as shown below
mpirun -np 4 --hostfile hostfile interMixingFoam -parallel |
|
February 8, 2021, 23:19 |
|
#7 |
New Member
Join Date: Jan 2021
Location: Edmonton
Posts: 4
Rep Power: 5 |
Thanks for your suggestion Sharkilwo5. I am running in parallel but locally in a multiprocessor machine.
The problem was related to the stability. The case is constrained by the timestep. |
|
February 21, 2021, 04:38 |
|
#8 | |
Member
Mahmoud
Join Date: Nov 2020
Location: United Kingdom
Posts: 43
Rep Power: 6 |
Quote:
I have the same problem now. Could you please explain that what do you mean it is constrained by the timestep? Cheers |
||
July 19, 2022, 21:12 |
|
#9 |
New Member
Basma Maged
Join Date: Dec 2021
Posts: 7
Rep Power: 4 |
||
July 21, 2022, 05:19 |
|
#10 |
New Member
cem
Join Date: Feb 2021
Posts: 6
Rep Power: 5 |
Hi Basma,
Can you explain what error exactly you are getting? I will try to help you. |
|
April 29, 2023, 16:14 |
|
#11 |
New Member
Burak
Join Date: Nov 2012
Posts: 14
Rep Power: 14 |
Hi There
I am running a snappyHexMesh job on remote machine nodes and I have some similar error that shows up. The machines are at the university server remote machines so I cannot touch the installation. Right now for the sake of trying out I am using the STL files from “turbine siting” and increasing the refinement level. Since I use the original tutorial file I know there is no problem with that one. After a refinement level ( in my case it is 5) I get a similar error like the posters previously. My decomposition is hierarchical (4 7 1) (Previous Stuff)..... Shell refinement iteration 2 ---------------------------- Marked for refinement due to distance to explicit features : 0 cells. Marked for refinement due to refinement shells : 2372394 cells. Marked for refinement due to refinement shells : 2372394 cells. -------------------------------------------------------------------------- Primary job terminated normally, but 1 process returned a non-zero exit code. Per user-direction, the job has been aborted. -------------------------------------------------------------------------- -------------------------------------------------------------------------- mpirun noticed that process rank 25 with PID 7837 on node a013 exited on signal 9 (Killed). I am using 28 cores but this stuff happens in even 168 cores.I don’t think computational power is the issue (My size is barely 2 million or so).The command I use is. mpirun -np 28 snappyHexMesh -overwrite -case “......./.../TurbineSitingDryRun_19042023/" I am using openfoam-v2212 but have been experiencing the same problem with openfoam-v1812 so I don’t think it’s a library issue or installation issue. I believe somehow the system is unintentionally caping some computational power which means it’s crashing one note and exiting. Can anyone offer any suggestions at this point? I appreciate all the help. Remember I cannot change that much on the installation so I have to do something on my end Regards Burak K. |
|
October 30, 2023, 02:35 |
Parallel run error while running the simulation using MPPICFoam
|
#12 |
New Member
Join Date: Jun 2023
Posts: 16
Rep Power: 3 |
Hello,
I am getting below error when i run ./Allrun command to solve. fileName::stripInvalid() called for invalid fileName /home/ttdesign/OpenFOAM/ttdesign-v2206/run/EAIPS-NEW/p1_noBladesScaled(copy) For debug level (= 2) > 1 this is considered fatal fileName::stripInvalid() called for invalid fileName /home/ttdesign/OpenFOAM/ttdesign-v2206/run/EAIPS-NEW/p1_noBladesScaled(copy) For debug level (= 2) > 1 this is considered fatal ------------------------------------------------------- Primary job terminated normally, but 1 process returned a non-zero exit code.. Per user-direction, the job has been aborted. ------------------------------------------------------- -------------------------------------------------------------------------- mpiexec detected that one or more processes exited with non-zero status, thus causing the job to be terminated. The first process to do so was: Process name: [[34815,1],1] Exit code: 1 Would be happy to provide the Allrun file as follows for your further investigation. You help would be much appreciated. Thanks Allrun script is as follows: #!/bin/sh cd "${0%/*}" || exit # Run from this directory . ${WM_PROJECT_DIR:?}/bin/tools/RunFunctions # Tutorial run functions #------------------------------------------------------------------------------ # CLEANING THE CASE rm log1_clean rm log2_mesh rm log3_checkMesh rm log4_decomposePar rm log5_parallelRun rm log6_reconstructPar rm -r log5_parallelRun.analyzed rm -r VTK foamCleanTutorials | tee log1_clean # CONVERTING FLUENT MESH TO FOAM ideasUnvToFoam Mesh_1_meters.unv | tee log2.mesh # COPING THE ORIGINAL 0 FOLDER TO MAKE IT EXECUTIBLE 0 FOLDER cp -r 0.orig 0 cp constant/boundary.orig constant/polyMesh/boundary # CHECKIMG THE MESH checkMesh | tee log3_checkMesh #foamToVTK #foamToVTK -faceSet nonOrthoFaces #foamToVTK -pointSet nonManifoldPoints #foamToVTK -pointSet unusedPoints #foamToVTK -faceSet upperTriangularFace # SPLITTING THE CASE FOR 16 PROCESSORS decomposePar | tee log4_decomposePar # RUNNING THE SIMULATION mpiexec -np 16 MPPICFoam -parallel | tee log5_parallelRun #pyFoamPlotWatcher.py log5_parallelRun # RECONSTRUCTING THE CASE reconstructPar | tee log6_reconstructPar |
|
Tags |
mpirun, parallel running |
Thread Tools | Search this Thread |
Display Modes | |
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
PEMFC model with FLUENT | brahimchoice | FLUENT | 22 | April 19, 2020 16:44 |
[ANSYS Meshing] Help with element size | sandri_92 | ANSYS Meshing & Geometry | 14 | November 14, 2018 08:54 |
fluent divergence for no reason | sufjanst | FLUENT | 2 | March 23, 2016 17:08 |
user subroutine error | CFDUSER | CFX | 2 | December 9, 2006 07:31 |
user defined function | cfduser | CFX | 0 | April 29, 2006 11:58 |