When an fvModel source introduces fluid into a simulation it should also
create a corresponding source term for all properties transported into
the domain by that injection. The source is, effectively, an alternative
form of inlet boundary, on which all transported properties need an
inlet value specified.
These values are now specified in the property field files. The
following is an example of a 0/U file in which the velocity of fluid
introduced by a fvModel source called "injection1" is set to a fixed
value of (-1 0 0):
dimensions [0 1 -1 0 0 0 0];
internalField uniform (0 0 0);
boundaryField
{
#includeEtc "caseDicts/setConstraintTypes"
wall
{
type noSlip;
}
atmosphere
{
type pressureInletOutletVelocity;
value $internalField;
}
}
// *** NEW ***
sources
{
injection1
{
type uniformFixedValue;
uniformValue (-1 0 0);
}
}
And the following entry in the 0/k file specifies the turbulent kinetic
energy introduced as a fraction of the mean flow kinetic energy:
sources
{
injection1
{
type turbulentIntensityKineticEnergy;
intensity 0.05;
}
}
The specification is directly analogous to boundary conditions. The
conditions are run-time selectable and can be concisely implemented.
They can access each other and be inter-dependent (e.g., the above,
where turbulent kinetic energy depends on velocity). The syntax keeps
field data localised and makes the source model (e.g., massSource,
volumeSource, ...) specification independent from what other models and
fields are present in the simulation. The 'fieldValues' entry previously
required by source models is now no longer required.
If source values need specifying and no source condition has been
supplied in the relevant field file then an error will be generated.
This error is similar to that generated for missing boundary conditions.
This replaces the behaviour where sources such as these would introduce
a value of zero, either silently or with a warning. This is now
considered unacceptable. Zero might be a tolerable default for certain
fields (U, k), but is wholly inappropriate for others (T, epsilon, rho).
This change additionally makes it possible to inject fluid into a
multicomponent solver with a specified temperature. Previously, it was
not possible to do this as there was no means of evaluating the energy
of fluid with the injected composition.
717 lines
19 KiB
C++
717 lines
19 KiB
C++
/*---------------------------------------------------------------------------*\
|
|
========= |
|
|
\\ / F ield | OpenFOAM: The Open Source CFD Toolbox
|
|
\\ / O peration | Website: https://openfoam.org
|
|
\\ / A nd | Copyright (C) 2011-2023 OpenFOAM Foundation
|
|
\\/ M anipulation |
|
|
-------------------------------------------------------------------------------
|
|
License
|
|
This file is part of OpenFOAM.
|
|
|
|
OpenFOAM is free software: you can redistribute it and/or modify it
|
|
under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation, either version 3 of the License, or
|
|
(at your option) any later version.
|
|
|
|
OpenFOAM is distributed in the hope that it will be useful, but WITHOUT
|
|
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with OpenFOAM. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
Application
|
|
redistributePar
|
|
|
|
Description
|
|
Redistributes existing decomposed mesh and fields according to the current
|
|
settings in the decomposeParDict file.
|
|
|
|
Must be run on maximum number of source and destination processors.
|
|
Balances mesh and writes new mesh to new time directory.
|
|
|
|
Can also work like decomposePar:
|
|
\verbatim
|
|
# Create empty processor directories (have to exist for argList)
|
|
mkdir processor0
|
|
..
|
|
mkdir processorN
|
|
|
|
# Copy undecomposed polyMesh
|
|
cp -R constant processor0
|
|
|
|
# Distribute
|
|
mpirun -np ddd redistributePar -parallel
|
|
\endverbatim
|
|
\*---------------------------------------------------------------------------*/
|
|
|
|
#include "argList.H"
|
|
#include "timeSelector.H"
|
|
#include "decompositionMethod.H"
|
|
#include "PstreamReduceOps.H"
|
|
#include "volFields.H"
|
|
#include "fvMeshDistribute.H"
|
|
#include "polyDistributionMap.H"
|
|
#include "IOobjectList.H"
|
|
#include "globalIndex.H"
|
|
#include "loadOrCreateMesh.H"
|
|
#include "extrapolatedCalculatedFvPatchFields.H"
|
|
|
|
using namespace Foam;
|
|
|
|
// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * //
|
|
|
|
void printMeshData(const polyMesh& mesh)
|
|
{
|
|
// Collect all data on master
|
|
|
|
globalIndex globalCells(mesh.nCells());
|
|
labelListList patchNeiProcNo(Pstream::nProcs());
|
|
labelListList patchSize(Pstream::nProcs());
|
|
const labelList& pPatches = mesh.globalData().processorPatches();
|
|
patchNeiProcNo[Pstream::myProcNo()].setSize(pPatches.size());
|
|
patchSize[Pstream::myProcNo()].setSize(pPatches.size());
|
|
forAll(pPatches, i)
|
|
{
|
|
const processorPolyPatch& ppp = refCast<const processorPolyPatch>
|
|
(
|
|
mesh.boundaryMesh()[pPatches[i]]
|
|
);
|
|
patchNeiProcNo[Pstream::myProcNo()][i] = ppp.neighbProcNo();
|
|
patchSize[Pstream::myProcNo()][i] = ppp.size();
|
|
}
|
|
Pstream::gatherList(patchNeiProcNo);
|
|
Pstream::gatherList(patchSize);
|
|
|
|
|
|
// Print stats
|
|
|
|
globalIndex globalBoundaryFaces(mesh.nFaces()-mesh.nInternalFaces());
|
|
|
|
label maxProcCells = 0;
|
|
label totProcFaces = 0;
|
|
label maxProcPatches = 0;
|
|
label totProcPatches = 0;
|
|
label maxProcFaces = 0;
|
|
|
|
for (label proci = 0; proci < Pstream::nProcs(); proci++)
|
|
{
|
|
Info<< endl
|
|
<< "Processor " << proci << nl
|
|
<< " Number of cells = " << globalCells.localSize(proci)
|
|
<< endl;
|
|
|
|
label nProcFaces = 0;
|
|
|
|
const labelList& nei = patchNeiProcNo[proci];
|
|
|
|
forAll(patchNeiProcNo[proci], i)
|
|
{
|
|
Info<< " Number of faces shared with processor "
|
|
<< patchNeiProcNo[proci][i] << " = " << patchSize[proci][i]
|
|
<< endl;
|
|
|
|
nProcFaces += patchSize[proci][i];
|
|
}
|
|
|
|
Info<< " Number of processor patches = " << nei.size() << nl
|
|
<< " Number of processor faces = " << nProcFaces << nl
|
|
<< " Number of boundary faces = "
|
|
<< globalBoundaryFaces.localSize(proci) << endl;
|
|
|
|
maxProcCells = max(maxProcCells, globalCells.localSize(proci));
|
|
totProcFaces += nProcFaces;
|
|
totProcPatches += nei.size();
|
|
maxProcPatches = max(maxProcPatches, nei.size());
|
|
maxProcFaces = max(maxProcFaces, nProcFaces);
|
|
}
|
|
|
|
// Stats
|
|
|
|
scalar avgProcCells = scalar(globalCells.size())/Pstream::nProcs();
|
|
scalar avgProcPatches = scalar(totProcPatches)/Pstream::nProcs();
|
|
scalar avgProcFaces = scalar(totProcFaces)/Pstream::nProcs();
|
|
|
|
// In case of all faces on one processor. Just to avoid division by 0.
|
|
if (totProcPatches == 0)
|
|
{
|
|
avgProcPatches = 1;
|
|
}
|
|
if (totProcFaces == 0)
|
|
{
|
|
avgProcFaces = 1;
|
|
}
|
|
|
|
Info<< nl
|
|
<< "Number of processor faces = " << totProcFaces/2 << nl
|
|
<< "Max number of cells = " << maxProcCells
|
|
<< " (" << 100.0*(maxProcCells-avgProcCells)/avgProcCells
|
|
<< "% above average " << avgProcCells << ")" << nl
|
|
<< "Max number of processor patches = " << maxProcPatches
|
|
<< " (" << 100.0*(maxProcPatches-avgProcPatches)/avgProcPatches
|
|
<< "% above average " << avgProcPatches << ")" << nl
|
|
<< "Max number of faces between processors = " << maxProcFaces
|
|
<< " (" << 100.0*(maxProcFaces-avgProcFaces)/avgProcFaces
|
|
<< "% above average " << avgProcFaces << ")" << nl
|
|
<< endl;
|
|
}
|
|
|
|
|
|
// Debugging: write volScalarField with decomposition for post processing.
|
|
void writeDecomposition
|
|
(
|
|
const word& name,
|
|
const fvMesh& mesh,
|
|
const labelList& decomp
|
|
)
|
|
{
|
|
Info<< "Writing wanted cell distribution to volScalarField " << name
|
|
<< " for postprocessing purposes." << nl << endl;
|
|
|
|
volScalarField procCells
|
|
(
|
|
IOobject
|
|
(
|
|
name,
|
|
mesh.time().name(),
|
|
mesh,
|
|
IOobject::NO_READ,
|
|
IOobject::AUTO_WRITE,
|
|
false // do not register
|
|
),
|
|
mesh,
|
|
dimensionedScalar(name, dimless, -1),
|
|
extrapolatedCalculatedFvPatchScalarField::typeName
|
|
);
|
|
|
|
forAll(procCells, cI)
|
|
{
|
|
procCells[cI] = decomp[cI];
|
|
}
|
|
procCells.correctBoundaryConditions();
|
|
|
|
procCells.write();
|
|
}
|
|
|
|
|
|
template<class GeoField>
|
|
void readFields
|
|
(
|
|
const boolList& haveMesh,
|
|
const typename GeoField::Mesh& mesh,
|
|
const autoPtr<fvMeshSubset>& subsetterPtr,
|
|
IOobjectList& allObjects,
|
|
PtrList<GeoField>& fields
|
|
)
|
|
{
|
|
// Get my objects of type
|
|
IOobjectList objects(allObjects.lookupClass(GeoField::typeName));
|
|
|
|
// Check that we all have all objects
|
|
wordList objectNames = objects.toc();
|
|
|
|
// Get master names
|
|
wordList masterNames(objectNames);
|
|
Pstream::scatter(masterNames);
|
|
|
|
if (haveMesh[Pstream::myProcNo()] && objectNames != masterNames)
|
|
{
|
|
FatalErrorInFunction
|
|
<< "differing fields of type " << GeoField::typeName
|
|
<< " on processors." << endl
|
|
<< "Master has:" << masterNames << endl
|
|
<< Pstream::myProcNo() << " has:" << objectNames
|
|
<< abort(FatalError);
|
|
}
|
|
|
|
fields.setSize(masterNames.size());
|
|
|
|
// Have master send all fields to processors that don't have a mesh
|
|
if (Pstream::master())
|
|
{
|
|
forAll(masterNames, i)
|
|
{
|
|
const word& name = masterNames[i];
|
|
IOobject& io = *objects[name];
|
|
io.writeOpt() = IOobject::AUTO_WRITE;
|
|
|
|
// Load field
|
|
fields.set(i, new GeoField(io, mesh));
|
|
|
|
// Create zero sized field and send
|
|
if (subsetterPtr.valid())
|
|
{
|
|
tmp<GeoField> tsubfld = subsetterPtr().interpolate(fields[i]);
|
|
|
|
// Send to all processors that don't have a mesh
|
|
for (label proci = 1; proci < Pstream::nProcs(); proci++)
|
|
{
|
|
if (!haveMesh[proci])
|
|
{
|
|
OPstream toProc(Pstream::commsTypes::blocking, proci);
|
|
toProc<< tsubfld();
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
else if (!haveMesh[Pstream::myProcNo()])
|
|
{
|
|
// Don't have mesh (nor fields). Receive empty field from master.
|
|
|
|
forAll(masterNames, i)
|
|
{
|
|
const word& name = masterNames[i];
|
|
|
|
// Receive field
|
|
IPstream fromMaster
|
|
(
|
|
Pstream::commsTypes::blocking,
|
|
Pstream::masterNo()
|
|
);
|
|
dictionary fieldDict(fromMaster);
|
|
|
|
fields.set
|
|
(
|
|
i,
|
|
new GeoField
|
|
(
|
|
IOobject
|
|
(
|
|
name,
|
|
mesh.thisDb().time().name(),
|
|
mesh.thisDb(),
|
|
IOobject::NO_READ,
|
|
IOobject::AUTO_WRITE
|
|
),
|
|
mesh,
|
|
fieldDict
|
|
)
|
|
);
|
|
|
|
//// Write it for next time round (since mesh gets written as well)
|
|
// fields[i].write();
|
|
}
|
|
}
|
|
else
|
|
{
|
|
// Have mesh so just try to load
|
|
forAll(masterNames, i)
|
|
{
|
|
const word& name = masterNames[i];
|
|
IOobject& io = *objects[name];
|
|
io.writeOpt() = IOobject::AUTO_WRITE;
|
|
|
|
// Load field
|
|
fields.set(i, new GeoField(io, mesh));
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
int main(int argc, char *argv[])
|
|
{
|
|
#include "addRegionOption.H"
|
|
#include "addOverwriteOption.H"
|
|
|
|
// Include explicit constant options, have zero from time range
|
|
timeSelector::addOptions();
|
|
|
|
#include "setRootCase.H"
|
|
|
|
if (env("FOAM_SIGFPE"))
|
|
{
|
|
WarningInFunction
|
|
<< "Detected floating point exception trapping (FOAM_SIGFPE)."
|
|
<< " This might give" << nl
|
|
<< " problems when mapping fields. Switch it off in case"
|
|
<< " of problems." << endl;
|
|
}
|
|
|
|
|
|
// Create processor directory if non-existing
|
|
if (!Pstream::master() && !isDir(args.path()))
|
|
{
|
|
Pout<< "Creating case directory " << args.path() << endl;
|
|
mkDir(args.path());
|
|
}
|
|
|
|
|
|
// Make sure we do not use the master-only reading.
|
|
regIOobject::fileModificationChecking = regIOobject::timeStamp;
|
|
#include "createTimeNoFunctionObjects.H"
|
|
// Allow override of time
|
|
instantList times = timeSelector::selectIfPresent(runTime, args);
|
|
runTime.setTime(times[0], 0);
|
|
|
|
word regionName = polyMesh::defaultRegion;
|
|
fileName meshSubDir;
|
|
|
|
if (args.optionReadIfPresent("region", regionName))
|
|
{
|
|
meshSubDir = regionName/polyMesh::meshSubDir;
|
|
}
|
|
else
|
|
{
|
|
meshSubDir = polyMesh::meshSubDir;
|
|
}
|
|
Info<< "Using mesh subdirectory " << meshSubDir << nl << endl;
|
|
|
|
const bool overwrite = args.optionFound("overwrite");
|
|
|
|
|
|
// Get time instance directory. Since not all processors have meshes
|
|
// just use the master one everywhere.
|
|
|
|
fileName masterInstDir;
|
|
if (Pstream::master())
|
|
{
|
|
masterInstDir = runTime.findInstance(meshSubDir, "points");
|
|
}
|
|
Pstream::scatter(masterInstDir);
|
|
|
|
// Check who has a mesh
|
|
const fileName meshPath = runTime.path()/masterInstDir/meshSubDir;
|
|
|
|
Info<< "Found points in " << meshPath << nl << endl;
|
|
|
|
|
|
boolList haveMesh(Pstream::nProcs(), false);
|
|
haveMesh[Pstream::myProcNo()] = isDir(meshPath);
|
|
Pstream::gatherList(haveMesh);
|
|
Pstream::scatterList(haveMesh);
|
|
Info<< "Per processor mesh availability : " << haveMesh << endl;
|
|
const bool allHaveMesh = (findIndex(haveMesh, false) == -1);
|
|
|
|
autoPtr<fvMesh> meshPtr = loadOrCreateMesh
|
|
(
|
|
IOobject
|
|
(
|
|
regionName,
|
|
masterInstDir,
|
|
runTime,
|
|
Foam::IOobject::MUST_READ
|
|
)
|
|
);
|
|
|
|
fvMesh& mesh = meshPtr();
|
|
|
|
// Print some statistics
|
|
Info<< "Before distribution:" << endl;
|
|
printMeshData(mesh);
|
|
|
|
|
|
labelList finalDecomp;
|
|
|
|
// Create decompositionMethod and new decomposition
|
|
{
|
|
autoPtr<decompositionMethod> distributor
|
|
(
|
|
decompositionMethod::NewDistributor
|
|
(
|
|
decompositionMethod::decomposeParDict(runTime)
|
|
)
|
|
);
|
|
|
|
finalDecomp = distributor().decompose(mesh, mesh.cellCentres());
|
|
}
|
|
|
|
// Dump decomposition to volScalarField
|
|
if (!overwrite)
|
|
{
|
|
writeDecomposition("decomposition", mesh, finalDecomp);
|
|
}
|
|
|
|
|
|
// Create 0 sized mesh to do all the generation of zero sized
|
|
// fields on processors that have zero sized meshes. Note that this is
|
|
// only necessary on master but since polyMesh construction with
|
|
// Pstream::parRun does parallel comms we have to do it on all
|
|
// processors
|
|
autoPtr<fvMeshSubset> subsetterPtr;
|
|
|
|
if (!allHaveMesh)
|
|
{
|
|
// Find last non-processor patch.
|
|
const polyBoundaryMesh& patches = mesh.boundaryMesh();
|
|
|
|
label nonProci = -1;
|
|
|
|
forAll(patches, patchi)
|
|
{
|
|
if (isA<processorPolyPatch>(patches[patchi]))
|
|
{
|
|
break;
|
|
}
|
|
nonProci++;
|
|
}
|
|
|
|
if (nonProci == -1)
|
|
{
|
|
FatalErrorInFunction
|
|
<< "Cannot find non-processor patch on processor "
|
|
<< Pstream::myProcNo() << endl
|
|
<< " Current patches:" << patches.names() << abort(FatalError);
|
|
}
|
|
|
|
// Subset 0 cells, no parallel comms. This is used to create zero-sized
|
|
// fields.
|
|
subsetterPtr.reset(new fvMeshSubset(mesh));
|
|
subsetterPtr().setLargeCellSubset(labelHashSet(0), nonProci, false);
|
|
}
|
|
|
|
|
|
// Get original objects (before incrementing time!)
|
|
IOobjectList objects(mesh, runTime.name());
|
|
// We don't want to map the decomposition (mapping already tested when
|
|
// mapping the cell centre field)
|
|
IOobjectList::iterator iter = objects.find("decomposition");
|
|
if (iter != objects.end())
|
|
{
|
|
objects.erase(iter);
|
|
}
|
|
|
|
|
|
// volFields
|
|
|
|
PtrList<volScalarField> volScalarFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
mesh,
|
|
subsetterPtr,
|
|
objects,
|
|
volScalarFields
|
|
);
|
|
|
|
PtrList<volVectorField> volVectorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
mesh,
|
|
subsetterPtr,
|
|
objects,
|
|
volVectorFields
|
|
);
|
|
|
|
PtrList<volSphericalTensorField> volSphereTensorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
mesh,
|
|
subsetterPtr,
|
|
objects,
|
|
volSphereTensorFields
|
|
);
|
|
|
|
PtrList<volSymmTensorField> volSymmTensorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
mesh,
|
|
subsetterPtr,
|
|
objects,
|
|
volSymmTensorFields
|
|
);
|
|
|
|
PtrList<volTensorField> volTensorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
mesh,
|
|
subsetterPtr,
|
|
objects,
|
|
volTensorFields
|
|
);
|
|
|
|
|
|
// surfaceFields
|
|
|
|
PtrList<surfaceScalarField> surfScalarFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
mesh,
|
|
subsetterPtr,
|
|
objects,
|
|
surfScalarFields
|
|
);
|
|
|
|
PtrList<surfaceVectorField> surfVectorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
mesh,
|
|
subsetterPtr,
|
|
objects,
|
|
surfVectorFields
|
|
);
|
|
|
|
PtrList<surfaceSphericalTensorField> surfSphereTensorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
mesh,
|
|
subsetterPtr,
|
|
objects,
|
|
surfSphereTensorFields
|
|
);
|
|
|
|
PtrList<surfaceSymmTensorField> surfSymmTensorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
mesh,
|
|
subsetterPtr,
|
|
objects,
|
|
surfSymmTensorFields
|
|
);
|
|
|
|
PtrList<surfaceTensorField> surfTensorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
mesh,
|
|
subsetterPtr,
|
|
objects,
|
|
surfTensorFields
|
|
);
|
|
|
|
|
|
// pointFields
|
|
|
|
pointMesh& pMesh =
|
|
const_cast<pointMesh&>(pointMesh::New(mesh));
|
|
PtrList<pointScalarField> pointScalarFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
pMesh,
|
|
subsetterPtr,
|
|
objects,
|
|
pointScalarFields
|
|
);
|
|
|
|
PtrList<pointVectorField> pointVectorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
pMesh,
|
|
subsetterPtr,
|
|
objects,
|
|
pointVectorFields
|
|
);
|
|
|
|
PtrList<pointSphericalTensorField> pointSphereTensorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
pMesh,
|
|
subsetterPtr,
|
|
objects,
|
|
pointSphereTensorFields
|
|
);
|
|
|
|
PtrList<pointSymmTensorField> pointSymmTensorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
pMesh,
|
|
subsetterPtr,
|
|
objects,
|
|
pointSymmTensorFields
|
|
);
|
|
|
|
PtrList<pointTensorField> pointTensorFields;
|
|
readFields
|
|
(
|
|
haveMesh,
|
|
pMesh,
|
|
subsetterPtr,
|
|
objects,
|
|
pointTensorFields
|
|
);
|
|
|
|
// Debugging: Create additional volField that will be mapped.
|
|
// Used to test correctness of mapping
|
|
// volVectorField mapCc("mapCc", 1*mesh.C());
|
|
|
|
// Mesh distribution engine
|
|
fvMeshDistribute distributor(mesh);
|
|
|
|
// Pout<< "Wanted distribution:"
|
|
// << distributor.countCells(finalDecomp) << nl << endl;
|
|
|
|
// Do actual sending/receiving of mesh
|
|
autoPtr<polyDistributionMap> map = distributor.distribute(finalDecomp);
|
|
|
|
//// Distribute any non-registered data accordingly
|
|
// map().distributeFaceData(faceCc);
|
|
|
|
|
|
// Print some statistics
|
|
Info<< "After distribution:" << endl;
|
|
printMeshData(mesh);
|
|
|
|
|
|
if (!overwrite)
|
|
{
|
|
runTime++;
|
|
}
|
|
else
|
|
{
|
|
mesh.setInstance(masterInstDir);
|
|
}
|
|
Info<< "Writing redistributed mesh to " << runTime.name() << nl << endl;
|
|
mesh.write();
|
|
|
|
|
|
// Print nice message
|
|
// ~~~~~~~~~~~~~~~~~~
|
|
|
|
// Determine which processors remain so we can print nice final message.
|
|
labelList nFaces(Pstream::nProcs());
|
|
nFaces[Pstream::myProcNo()] = mesh.nFaces();
|
|
Pstream::gatherList(nFaces);
|
|
Pstream::scatterList(nFaces);
|
|
|
|
Info<< nl
|
|
<< "You can pick up the redecomposed mesh from the polyMesh directory"
|
|
<< " in " << runTime.name() << "." << nl
|
|
<< "If you redecomposed the mesh to less processors you can delete"
|
|
<< nl
|
|
<< "the processor directories with 0 sized meshes in them." << nl
|
|
<< "Below is a sample set of commands to do this."
|
|
<< " Take care when issuing these" << nl
|
|
<< "commands." << nl << endl;
|
|
|
|
forAll(nFaces, proci)
|
|
{
|
|
fileName procDir = "processor" + name(proci);
|
|
|
|
if (nFaces[proci] == 0)
|
|
{
|
|
Info<< " rm -r " << procDir.c_str() << nl;
|
|
}
|
|
else
|
|
{
|
|
fileName timeDir = procDir/runTime.name()/meshSubDir;
|
|
fileName constDir = procDir/runTime.constant()/meshSubDir;
|
|
|
|
Info<< " rm -r " << constDir.c_str() << nl
|
|
<< " mv " << timeDir.c_str()
|
|
<< ' ' << constDir.c_str() << nl;
|
|
}
|
|
}
|
|
Info<< endl;
|
|
|
|
|
|
Info<< "End\n" << endl;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
// ************************************************************************* //
|