
[Sponsors] 
October 27, 2011, 02:28 
Harmonic mean of a volScalarField in parallel

#1 
New Member
Austin Kimbrell
Join Date: Feb 2011
Location: Tennessee, USA
Posts: 8
Rep Power: 7 
Hello all,
I need to multiply a quantity by a harmonic average of a volScalarField in my calculation. The quantity being averaged is defined at cell centers. I have come up with the following algorithm which seems to work properly in serial: Code:
volScalarField Mult = test; volScalarField reciprocal = 1/(test+dSMALL); forAll(mesh.C(), celli) { scalar sum = SMALL; const labelList& neighbors = mesh.cellCells()[celli]; forAll(neighbors, n) { label cell = neighbors[n]; sum += reciprocal[cell]; } sum += reciprocal[celli]; Mult[celli] *= (neighbors.size()+1)/sum; } Also might there be a better way to perform this calculation? Or maybe I should ask, if someone has a better way might they share it? And before someone mentions it I believe the 'harmonic' interpolation does not apply to my case, from my understanding it is performing a harmonic interpolation between cell P and each neighbor separately, I need the harmonic mean of the center cell and all neighbors. Thanks! 

October 27, 2011, 03:37 

#2 
Senior Member
Anton Kidess
Join Date: May 2009
Location: Delft, Netherlands
Posts: 1,113
Rep Power: 20 
You only need a reduce if you want a global quantity, e.g. the global sum or the global average. In this case, reduce will reduce e.g. all the local processor maximums to a single global maximum. Since you are only summing up locally on cells, you will always stay on the same processor domain.
Now, what might be a problem is that on cells that are next to a processor boundary, I believe you will have less neighbors, because a processor boundary is a patch, not a layer of ghost cells. Reduce won't help you here though, instead you'll have to use the gather and scatter functions of PStream. Unfortunately I don't know any more than that. 

October 27, 2011, 11:31 

#3 
New Member
Austin Kimbrell
Join Date: Feb 2011
Location: Tennessee, USA
Posts: 8
Rep Power: 7 
Ok thank you, now I understand what the reduce operation is supposed to be used for. This is clearly not what I need to use as I am performing a local operation on each cell.
I'll try looking into PStream and see what information I can find, if anyone else has experience with this sort of thing please share. 

October 28, 2011, 17:54 

#4 
Member
Ivor Clifford
Join Date: Mar 2009
Location: Switzerland
Posts: 91
Rep Power: 9 
I think you might be missing something here. Since you are looping over cell neighbours you will need to specifically include those neighbours that lie on the opposite side of coupled patches, i.e. on the other processors. You can forget about gather, scatter, etc. since the processor and other coupled boundaries handle all the communication for you and all you need to do is access the neighbouring cell values using patchNeighbourField().
Something along these lines might do the trick... Code:
volScalarField Mult = test; volScalarField reciprocal = 1/(test+SMALL); scalarField sum(mesh.nCells(), 0.0); labelList count(mesh.nCells(), 0); forAll(mesh.C(), celli) { sum[celli] = SMALL; count[celli] = 0; const labelList& neighbors = mesh.cellCells()[celli]; forAll(neighbors, n) { label cell = neighbors[n]; sum[celli] += reciprocal[cell]; } sum[celli] += reciprocal[celli]; count[celli] += neighbors.size()+1; } // Now we include the boundaries forAll(mesh.boundaryMesh(), patchI) { const fvPatchField& pf = reciprocal.boundaryField()[patchI]; const unallocLabelList& faceCells = pf.patch().faceCells(); // Coupled boundaries (processor, cyclic, etc) if (pf.coupled()) { scalarField neighborReciprocal = pf.patchNeighbourField(); forAll(faceCells, faceI) { sum[faceCells[faceI]] += neighbourReciprocal[faceI]; count[faceCells[faceI] ++; } } // Standard boundaries  if you want to include them else { forAll(faceCells, faceI) { sum[faceCells[faceI]] += pf[faceI]; count[faceCells[faceI] ++; } } } Mult.internalField() *= count/sum; 

October 30, 2011, 11:48 

#5 
New Member
Austin Kimbrell
Join Date: Feb 2011
Location: Tennessee, USA
Posts: 8
Rep Power: 7 
Ivor,
Thank you for the additional code, it makes perfect sense from an application standpoint. I had difficulties, however, compiling the code  it stops at this line: Code:
const fvPatchField& pf = reciprocal.boundaryField()[patchI]; Edit: I think I figured it out. You have to use fvPatchScalarField, for example, rather than using fvPatchField directly. It depends on the type for reciprocal. I will post the code once it is working and debugged. Last edited by akimbrell; October 30, 2011 at 14:02. 

October 31, 2011, 11:36 

#6 
Member
Ivor Clifford
Join Date: Mar 2009
Location: Switzerland
Posts: 91
Rep Power: 9 
Sorry, yes should be fvPatchScalarField. There will probably be some other errors as you go... I didn't debug the code, just sent it as a suggested approach.


November 1, 2011, 00:50 

#7 
New Member
Austin Kimbrell
Join Date: Feb 2011
Location: Tennessee, USA
Posts: 8
Rep Power: 7 
Well I have finished debugging, testing checks out on a 1D array, it returns the expected answer in parallel as well as serial. The one necessary line that was missing is a correctBoundaryConditions() call for the reciprocal field  otherwise the processors do not return the correct value for the neighbors and it messes up the calculation. The following is written using the U velocity field, hence the dimensions, but can easily be applied for any harmonic mean:
Code:
volScalarField Result ( IOobject ( "Result", runTime.timeName(), mesh, IOobject::NO_READ, IOobject::AUTO_WRITE ), mesh, dimensionedScalar("zero", dimless, 0.0) ); dimensionedScalar dSMALL("dSMALL", dimVelocity, SMALL); volScalarField reciprocal = 1/(mag(U)+dSMALL); reciprocal.correctBoundaryConditions(); scalarField sum(mesh.nCells(), SMALL); scalarField count(mesh.nCells(), 0); //loop over all cells in each processor forAll(mesh.C(), celli) { const labelList& neighbors = mesh.cellCells()[celli]; forAll(neighbors, n) { sum[celli] += reciprocal[neighbors[n]]; } sum[celli] += reciprocal[celli]; count[celli] += neighbors.size()+1; } //loop over all boundaries to get processor patch neighbors forAll(mesh.boundaryMesh(), patchi) { const fvPatchScalarField& pf = reciprocal.boundaryField()[patchi]; const unallocLabelList& faceCells = pf.patch().faceCells(); //look for coupled boundaries if(pf.coupled()) { scalarField neighborReciprocal = pf.patchNeighbourField(); forAll(faceCells, facei) { sum[faceCells[facei]] += neighborReciprocal[facei]; count[faceCells[facei]]++; } } } Result.internalField() = count/sum; Result.dimensions().reset(dimensionSet(0, 1, 1, 0, 0, 0, 0)); Result.write(); 

Tags 
harmonic mean, not implemented, parallel 
Thread Tools  
Display Modes  


Similar Threads  
Thread  Thread Starter  Forum  Replies  Last Post 
Script to Run Parallel Jobs in Rocks Cluster  asaha  OpenFOAM Running, Solving & CFD  12  July 4, 2012 22:51 
Cumulative continuity error large in parallel simulations  xiao  OpenFOAM  3  March 7, 2011 05:13 
Cyclic patches and parallel postprocessing problems  askjak  OpenFOAM Bugs  18  October 27, 2010 03:35 
Looping over a volScalarField in a parallel run  dohnie  OpenFOAM Programming & Development  4  June 16, 2010 03:58 
Parallel Computing Classes at San Diego Supercomputer Center Jan. 2022  Amitava Majumdar  Main CFD Forum  0  January 5, 1999 13:00 