|
[Sponsors] |
February 27, 2021, 03:00 |
mpirun error for pimpleFoam case
|
#1 |
New Member
Samridh Sharma
Join Date: Jun 2020
Posts: 3
Rep Power: 6 |
Hello everyone,
I have a pimpleFoam case set up to which running on parallel gives me an error that I can't find solution to, it runs fine when I execute it normally on one core. It asks for missing files as processor01/0/s i don't know what this missing 's' file is, I tried creating dummy 's' files, but still didn't work. Please help me fix this error. It gives following error: Code:
$ mpirun -np 4 mypimpleFoam -parallel /*---------------------------------------------------------------------------*\ ========= | \\ / F ield | OpenFOAM: The Open Source CFD Toolbox \\ / O peration | Website: https://openfoam.org \\ / A nd | Version: 8 \\/ M anipulation | \*---------------------------------------------------------------------------*/ Build : 8-340defec456f Exec : mypimpleFoam -parallel Date : Feb 27 2021 Time : 12:25:30 Host : "samridh-GL63-8RC" PID : 48082 I/O : uncollated Case : /home/samridh/OpenFOAM/samridh-8/run/smr_um nProcs : 4 Slaves : 3 ( "samridh-GL63-8RC.48083" "samridh-GL63-8RC.48084" "samridh-GL63-8RC.48085" ) Pstream initialized with: floatTransfer : 0 nProcsSimpleSum : 0 commsType : nonBlocking polling iterations : 0 sigFpe : Enabling floating point exception trapping (FOAM_SIGFPE). fileModificationChecking : Monitoring run-time modified files using timeStampMaster (fileModificationSkew 10) allowSystemOperations : Allowing user-supplied system call operations // * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * // Create time Create mesh for time = 0 PIMPLE: Convergence criteria found p: tolerance 1e-06 U: tolerance 1e-06 PIMPLE: Operating solver in transient mode with 1 outer corrector PIMPLE: Operating solver in PISO mode Reading field p Reading field U Reading/calculating face flux field phi Selecting incompressible transport model Newtonian Selecting turbulence model type laminar Selecting laminar stress model Stokes No MRF models present No finite volume options present Courant Number mean: 0 max: 0 Starting time loop [3] [3] [3] --> FOAM FATAL ERROR: [3] cannot find file "/home/samridh/OpenFOAM/samridh-8/run/smr_um/processor3/0/s" [0] [0] [0] --> FOAM FATAL ERROR: [0] cannot find file "/home/samridh/OpenFOAM/samridh-8/run/smr_um/processor0/0/s" [0] [0] From function virtual Foam::autoPtr<Foam::ISstream> Foam::fileOperations::uncollatedFileOperation::readStream(Foam::regIOobject&, const Foam::fileName&, const Foam::word&, bool) const [0] [1] [1] [1] --> FOAM FATAL ERROR: [1] cannot find file "/home/samridh/OpenFOAM/samridh-8/run/smr_um/processor1/0/s" [1] [1] From function virtual Foam::autoPtr<Foam::ISstream> Foam::fileOperations::uncollatedFileOperation::readStream(Foam::regIOobject&, const Foam::fileName&, const Foam::word&, bool) const [1] in file global/fileOperations/uncollatedFileOperation/uncollatedFileOperation.C at line 538. [1] FOAM parallel run exiting [1] [3] [3] From function virtual Foam::autoPtr<Foam::ISstream> Foam::fileOperations::uncollatedFileOperation::readStream(Foam::regIOobject&, const Foam::fileName&, const Foam::word&, bool) const [3] in file global/fileOperations/uncollatedFileOperation/uncollatedFileOperation.C at line 538. [3] FOAM parallel run exiting [3] in file global/fileOperations/uncollatedFileOperation/uncollatedFileOperation.C at line 538. [0] FOAM parallel run exiting [0] [2] [2] [2] --> FOAM FATAL ERROR: [2] cannot find file "/home/samridh/OpenFOAM/samridh-8/run/smr_um/processor2/0/s" [2] [2] From function virtual Foam::autoPtr<Foam::ISstream> Foam::fileOperations::uncollatedFileOperation::readStream(Foam::regIOobject&, const Foam::fileName&, const Foam::word&, bool) const [2] in file global/fileOperations/uncollatedFileOperation/uncollatedFileOperation.C at line 538. [2] FOAM parallel run exiting [2] -------------------------------------------------------------------------- MPI_ABORT was invoked on rank 1 in communicator MPI COMMUNICATOR 3 SPLIT FROM 0 with errorcode 1. NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. You may or may not see output from other processes, depending on exactly when Open MPI kills them. -------------------------------------------------------------------------- [samridh-GL63-8RC:48078] 3 more processes have sent help message help-mpi-api.txt / mpi-abort [samridh-GL63-8RC:48078] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages |
|
March 3, 2021, 05:39 |
found solution
|
#2 |
New Member
Samridh Sharma
Join Date: Jun 2020
Posts: 3
Rep Power: 6 |
https://discord.com/channels/3977727...91579618869268
HTML Code:
I think the s file you're referring to is initial condition for scalartransport. Check your controlDict if there is scalartransport included then s is definitely initial condition for passive scalar I commented it out and now the parallel run works. |
|
Tags |
mpirun error |
Thread Tools | Search this Thread |
Display Modes | |
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
Pressure instabilities with interDyMFoam for the floatingObject case | nbadano | OpenFOAM Running, Solving & CFD | 15 | October 15, 2021 07:35 |
[DesignModeler] DesignModeler Scripting: How to get Full Command Access | ANT | ANSYS Meshing & Geometry | 53 | February 16, 2020 16:13 |
[OpenFOAM] Problem postprocessing decomposed case with 'paraFoam -builtin' | bentkj | ParaView | 4 | February 21, 2017 05:22 |
Is Playstation 3 cluster suitable for CFD work | hsieh | OpenFOAM | 9 | August 16, 2015 15:53 |
Error reading new case | montag dp | FLUENT | 5 | September 15, 2011 07:00 |