Hopefully a quick question for
Hopefully a quick question for someone out there to answer.
I have a case with a cell lenght scale of ~0.00002 (not unreasonably small). When decomposed, solvers such as InterDyMFoam fail to get past the first pressure iteration (i.e. the solver goes through the maximum number of iterations).
Yet when run on a single processor, everything proceeds fine.
I currently have 10 decimals of precision set in the controlDict file...
Does anyone know if I need to make some other settings changes to make a small-scale case work in parallel?
What is the case? Refinement/u
What is the case? Refinement/unrefinement or moving mesh? There should be no dependency on scale (only once dimensions/timesteps approach 1e-15 or so). Do you write the mesh binary?
GAMG / floatTransfer 1?
if you're using the GAMG solver probably the parameter "floatTransfer" in the global controlDict (etc/controlDict) is set to 1. You should set it to 0, this will help.
Still having problems... Test Case attached
Hi Mattijs, Nikolaos.
Sorry for the late response to your replies, for some reason my email notification from the forum hasent been working.
I tired doing modifications based on both of your comments, with no luck. Using ascii or binary write formats did not appear to have an impact. I also tried moving the pdRefProbe location to ensure it was within the first processor (which didnt have an effect). Modifying the term in the controlDict file (etc/controDict) also didn't affect the results.
The simulation runs in parallel for a number of iterations, but eventually fails.
I have attached a simple test case in the hopes of obtaining some help in solving this problem (testcase.zip). To minimize the size, I only included the blockmesh dictionary file and boundary file (you'll have to run blockMesh)
The test case is for a droplet in simple shear. Try running the case on a single processor, and on multiple processors (currently set up to decompose into 4).
I have also attached the output from my machine for both the single and parallel runs (decomposeTest.zip).
Please take a look and get back to me with any additional suggetions. This is a rather large problem for me, as I have some large grids (multi-million cell) that I need to run cases on, which would take an unreasonable amount of time on a single processor.
I know that the problem is not limited to the interDyMFoam solver, as I have another custom solver that experiences similar problems. It also uses the GAMG solver for the pressure field.
Still having problems... Additional Note
As an additional Note, there is no adaptive meshing in this case. I am using interDyMFoam as an example since I need it for other similar cases.
From your runParFoam.o56029 file:
Pstream initialized with:
floatTransfer : 1
If you look at the error traceback it is due to a sigfpe - usually a division by 0 or an overflow. A division by 0 would show up in the traceback as originating from '/' so it is an overflow. See also http://www.cfd-online.com/Forums/ope...tml#post210878
|All times are GMT -4. The time now is 16:50.|