mirror of
https://develop.openfoam.com/Development/openfoam.git
synced 2025-11-28 03:28:01 +00:00
Merge branch 'master' of ssh://noisy/home/noisy3/OpenFOAM/OpenFOAM-dev
This commit is contained in:
@ -40,4 +40,3 @@
|
||||
// Momentum corrector
|
||||
U -= fvc::grad(p)/AU;
|
||||
U.correctBoundaryConditions();
|
||||
|
||||
|
||||
@ -40,6 +40,10 @@ geometry
|
||||
{
|
||||
type triSurfaceMesh;
|
||||
|
||||
//tolerance 1E-6; // optional:non-default tolerance on intersections
|
||||
//maxTreeDepth 10; // optional:depth of octree. Decrease only in case
|
||||
// of memory limitations.
|
||||
|
||||
// Per region the patchname. If not provided will be <name>_<region>.
|
||||
regions
|
||||
{
|
||||
|
||||
@ -94,7 +94,7 @@ void backup
|
||||
{
|
||||
if (fromSet.size())
|
||||
{
|
||||
Pout<< " Backing up " << fromName << " into " << toName << endl;
|
||||
Info<< " Backing up " << fromName << " into " << toName << endl;
|
||||
|
||||
topoSet::New(setType, mesh, toName, fromSet)().write();
|
||||
}
|
||||
@ -525,7 +525,7 @@ bool doCommand
|
||||
{
|
||||
topoSet& currentSet = currentSetPtr();
|
||||
|
||||
Pout<< " Set:" << currentSet.name()
|
||||
Info<< " Set:" << currentSet.name()
|
||||
<< " Size:" << currentSet.size()
|
||||
<< " Action:" << actionName
|
||||
<< endl;
|
||||
@ -622,7 +622,7 @@ bool doCommand
|
||||
+ ".vtk"
|
||||
);
|
||||
|
||||
Pout<< " Writing " << currentSet.name()
|
||||
Info<< " Writing " << currentSet.name()
|
||||
<< " (size " << currentSet.size() << ") to "
|
||||
<< currentSet.instance()/currentSet.local()
|
||||
/currentSet.name()
|
||||
@ -634,7 +634,7 @@ bool doCommand
|
||||
}
|
||||
else
|
||||
{
|
||||
Pout<< " Writing " << currentSet.name()
|
||||
Info<< " Writing " << currentSet.name()
|
||||
<< " (size " << currentSet.size() << ") to "
|
||||
<< currentSet.instance()/currentSet.local()
|
||||
/currentSet.name() << endl << endl;
|
||||
@ -683,7 +683,7 @@ enum commandStatus
|
||||
|
||||
void printMesh(const Time& runTime, const polyMesh& mesh)
|
||||
{
|
||||
Pout<< "Time:" << runTime.timeName()
|
||||
Info<< "Time:" << runTime.timeName()
|
||||
<< " cells:" << mesh.nCells()
|
||||
<< " faces:" << mesh.nFaces()
|
||||
<< " points:" << mesh.nPoints()
|
||||
@ -703,19 +703,19 @@ commandStatus parseType
|
||||
{
|
||||
if (setType.empty())
|
||||
{
|
||||
Pout<< "Type 'help' for usage information" << endl;
|
||||
Info<< "Type 'help' for usage information" << endl;
|
||||
|
||||
return INVALID;
|
||||
}
|
||||
else if (setType == "help")
|
||||
{
|
||||
printHelp(Pout);
|
||||
printHelp(Info);
|
||||
|
||||
return INVALID;
|
||||
}
|
||||
else if (setType == "list")
|
||||
{
|
||||
printAllSets(mesh, Pout);
|
||||
printAllSets(mesh, Info);
|
||||
|
||||
return INVALID;
|
||||
}
|
||||
@ -726,7 +726,7 @@ commandStatus parseType
|
||||
|
||||
label nearestIndex = Time::findClosestTimeIndex(Times, requestedTime);
|
||||
|
||||
Pout<< "Changing time from " << runTime.timeName()
|
||||
Info<< "Changing time from " << runTime.timeName()
|
||||
<< " to " << Times[nearestIndex].name()
|
||||
<< endl;
|
||||
|
||||
@ -737,24 +737,24 @@ commandStatus parseType
|
||||
{
|
||||
case polyMesh::UNCHANGED:
|
||||
{
|
||||
Pout<< " mesh not changed." << endl;
|
||||
Info<< " mesh not changed." << endl;
|
||||
break;
|
||||
}
|
||||
case polyMesh::POINTS_MOVED:
|
||||
{
|
||||
Pout<< " points moved; topology unchanged." << endl;
|
||||
Info<< " points moved; topology unchanged." << endl;
|
||||
break;
|
||||
}
|
||||
case polyMesh::TOPO_CHANGE:
|
||||
{
|
||||
Pout<< " topology changed; patches unchanged." << nl
|
||||
Info<< " topology changed; patches unchanged." << nl
|
||||
<< " ";
|
||||
printMesh(runTime, mesh);
|
||||
break;
|
||||
}
|
||||
case polyMesh::TOPO_PATCH_CHANGE:
|
||||
{
|
||||
Pout<< " topology changed and patches changed." << nl
|
||||
Info<< " topology changed and patches changed." << nl
|
||||
<< " ";
|
||||
printMesh(runTime, mesh);
|
||||
|
||||
@ -773,7 +773,7 @@ commandStatus parseType
|
||||
}
|
||||
else if (setType == "quit")
|
||||
{
|
||||
Pout<< "Quitting ..." << endl;
|
||||
Info<< "Quitting ..." << endl;
|
||||
|
||||
return QUIT;
|
||||
}
|
||||
@ -864,7 +864,7 @@ int main(int argc, char *argv[])
|
||||
printMesh(runTime, mesh);
|
||||
|
||||
// Print current sets
|
||||
printAllSets(mesh, Pout);
|
||||
printAllSets(mesh, Info);
|
||||
|
||||
|
||||
|
||||
@ -874,7 +874,7 @@ int main(int argc, char *argv[])
|
||||
{
|
||||
fileName batchFile(args.option("batch"));
|
||||
|
||||
Pout<< "Reading commands from file " << batchFile << endl;
|
||||
Info<< "Reading commands from file " << batchFile << endl;
|
||||
|
||||
// we cannot handle .gz files
|
||||
if (!isFile(batchFile, false))
|
||||
@ -888,11 +888,11 @@ int main(int argc, char *argv[])
|
||||
#if READLINE != 0
|
||||
else if (!read_history(historyFile))
|
||||
{
|
||||
Pout<< "Successfully read history from " << historyFile << endl;
|
||||
Info<< "Successfully read history from " << historyFile << endl;
|
||||
}
|
||||
#endif
|
||||
|
||||
Pout<< "Please type 'help', 'quit' or a set command after prompt." << endl;
|
||||
Info<< "Please type 'help', 'quit' or a set command after prompt." << endl;
|
||||
|
||||
bool ok = true;
|
||||
|
||||
@ -916,7 +916,7 @@ int main(int argc, char *argv[])
|
||||
{
|
||||
if (!fileStreamPtr->good())
|
||||
{
|
||||
Pout<< "End of batch file" << endl;
|
||||
Info<< "End of batch file" << endl;
|
||||
break;
|
||||
}
|
||||
|
||||
@ -924,7 +924,7 @@ int main(int argc, char *argv[])
|
||||
|
||||
if (rawLine.size())
|
||||
{
|
||||
Pout<< "Doing:" << rawLine << endl;
|
||||
Info<< "Doing:" << rawLine << endl;
|
||||
}
|
||||
}
|
||||
else
|
||||
@ -945,7 +945,7 @@ int main(int argc, char *argv[])
|
||||
}
|
||||
# else
|
||||
{
|
||||
Pout<< "Command>" << flush;
|
||||
Info<< "Command>" << flush;
|
||||
std::getline(std::cin, rawLine);
|
||||
}
|
||||
# endif
|
||||
@ -992,7 +992,7 @@ int main(int argc, char *argv[])
|
||||
delete fileStreamPtr;
|
||||
}
|
||||
|
||||
Pout<< "\nEnd\n" << endl;
|
||||
Info<< "\nEnd\n" << endl;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
@ -96,6 +96,7 @@ void checkPatch(const polyBoundaryMesh& bMesh, const word& name)
|
||||
int main(int argc, char *argv[])
|
||||
{
|
||||
argList::noParallel();
|
||||
# include "addRegionOption.H"
|
||||
|
||||
argList::validArgs.append("masterPatch");
|
||||
argList::validArgs.append("slavePatch");
|
||||
@ -109,7 +110,7 @@ int main(int argc, char *argv[])
|
||||
# include "setRootCase.H"
|
||||
# include "createTime.H"
|
||||
runTime.functionObjects().off();
|
||||
# include "createMesh.H"
|
||||
# include "createNamedMesh.H"
|
||||
const word oldInstance = mesh.pointsInstance();
|
||||
|
||||
|
||||
|
||||
@ -31,6 +31,14 @@ Description
|
||||
|
||||
Mainly used to convert binary mesh/field files to ASCII.
|
||||
|
||||
Problem: any zero-size List written binary gets written as '0'. When
|
||||
reading the file as a dictionary this is interpreted as a label. This
|
||||
is (usually) not a problem when doing patch fields since these get the
|
||||
'uniform', 'nonuniform' prefix. However zone contents are labelLists
|
||||
not labelFields and these go wrong. For now hacked a solution where
|
||||
we detect the keywords in zones and redo the dictionary entries
|
||||
to be labelLists.
|
||||
|
||||
\*---------------------------------------------------------------------------*/
|
||||
|
||||
#include "argList.H"
|
||||
@ -56,6 +64,82 @@ namespace Foam
|
||||
}
|
||||
|
||||
|
||||
// Hack to do zones which have Lists in them. See above.
|
||||
bool writeZones(const word& name, Time& runTime)
|
||||
{
|
||||
IOobject io
|
||||
(
|
||||
name,
|
||||
runTime.timeName(),
|
||||
polyMesh::meshSubDir,
|
||||
runTime,
|
||||
IOobject::MUST_READ,
|
||||
IOobject::NO_WRITE,
|
||||
false
|
||||
);
|
||||
|
||||
bool writeOk = false;
|
||||
|
||||
if (io.headerOk())
|
||||
{
|
||||
Info<< " Reading " << io.headerClassName()
|
||||
<< " : " << name << endl;
|
||||
|
||||
// Switch off type checking (for reading e.g. faceZones as
|
||||
// generic list of dictionaries).
|
||||
const word oldTypeName = IOPtrList<entry>::typeName;
|
||||
const_cast<word&>(IOPtrList<entry>::typeName) = word::null;
|
||||
|
||||
IOPtrList<entry> meshObject(io);
|
||||
|
||||
forAll(meshObject, i)
|
||||
{
|
||||
if (meshObject[i].isDict())
|
||||
{
|
||||
dictionary& d = meshObject[i].dict();
|
||||
|
||||
if (d.found("faceLabels"))
|
||||
{
|
||||
d.set("faceLabels", labelList(d.lookup("faceLabels")));
|
||||
}
|
||||
|
||||
if (d.found("flipMap"))
|
||||
{
|
||||
d.set("flipMap", boolList(d.lookup("flipMap")));
|
||||
}
|
||||
|
||||
if (d.found("cellLabels"))
|
||||
{
|
||||
d.set("cellLabels", labelList(d.lookup("cellLabels")));
|
||||
}
|
||||
|
||||
if (d.found("pointLabels"))
|
||||
{
|
||||
d.set("pointLabels", labelList(d.lookup("pointLabels")));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
const_cast<word&>(IOPtrList<entry>::typeName) = oldTypeName;
|
||||
// Fake type back to what was in field
|
||||
const_cast<word&>(meshObject.type()) = io.headerClassName();
|
||||
|
||||
Info<< " Writing " << name << endl;
|
||||
|
||||
// Force writing as ascii
|
||||
writeOk = meshObject.regIOobject::writeObject
|
||||
(
|
||||
IOstream::ASCII,
|
||||
IOstream::currentVersion,
|
||||
runTime.writeCompression()
|
||||
);
|
||||
}
|
||||
|
||||
return writeOk;
|
||||
}
|
||||
|
||||
|
||||
|
||||
// Main program:
|
||||
|
||||
int main(int argc, char *argv[])
|
||||
@ -76,9 +160,19 @@ int main(int argc, char *argv[])
|
||||
writeMeshObject<labelIOList>("neighbour", runTime);
|
||||
writeMeshObject<faceIOList>("faces", runTime);
|
||||
writeMeshObject<pointIOField>("points", runTime);
|
||||
writeMeshObject<IOPtrList<entry> >("cellZones", runTime);
|
||||
writeMeshObject<IOPtrList<entry> >("faceZones", runTime);
|
||||
writeMeshObject<IOPtrList<entry> >("pointZones", runTime);
|
||||
writeMeshObject<labelIOList>("pointProcAddressing", runTime);
|
||||
writeMeshObject<labelIOList>("faceProcAddressing", runTime);
|
||||
writeMeshObject<labelIOList>("cellProcAddressing", runTime);
|
||||
writeMeshObject<labelIOList>("boundaryProcAddressing", runTime);
|
||||
|
||||
if (runTime.writeFormat() == IOstream::ASCII)
|
||||
{
|
||||
// Only do zones when converting from binary to ascii
|
||||
// The other way gives problems since working on dictionary level.
|
||||
writeZones("cellZones", runTime);
|
||||
writeZones("faceZones", runTime);
|
||||
writeZones("pointZones", runTime);
|
||||
}
|
||||
|
||||
// Get list of objects from the database
|
||||
IOobjectList objects(runTime, runTime.timeName());
|
||||
|
||||
@ -1077,83 +1077,92 @@ int main(int argc, char *argv[])
|
||||
{
|
||||
const faceZone& pp = zones[zoneI];
|
||||
|
||||
const indirectPrimitivePatch ipp
|
||||
(
|
||||
IndirectList<face>(mesh.faces(), pp),
|
||||
mesh.points()
|
||||
);
|
||||
if (pp.size() > 0)
|
||||
{
|
||||
const indirectPrimitivePatch ipp
|
||||
(
|
||||
IndirectList<face>(mesh.faces(), pp),
|
||||
mesh.points()
|
||||
);
|
||||
|
||||
writer.writePolygonalZone
|
||||
(
|
||||
pp.name(),
|
||||
strandID++, //1+patchIDs.size()+zoneI, //strandID,
|
||||
ipp,
|
||||
allVarLocation
|
||||
);
|
||||
writer.writePolygonalZone
|
||||
(
|
||||
pp.name(),
|
||||
strandID++, //1+patchIDs.size()+zoneI, //strandID,
|
||||
ipp,
|
||||
allVarLocation
|
||||
);
|
||||
|
||||
// Write coordinates
|
||||
writer.writeField(ipp.localPoints().component(0)());
|
||||
writer.writeField(ipp.localPoints().component(1)());
|
||||
writer.writeField(ipp.localPoints().component(2)());
|
||||
// Write coordinates
|
||||
writer.writeField(ipp.localPoints().component(0)());
|
||||
writer.writeField(ipp.localPoints().component(1)());
|
||||
writer.writeField(ipp.localPoints().component(2)());
|
||||
|
||||
// Write all volfields
|
||||
forAll(vsf, i)
|
||||
{
|
||||
writer.writeField
|
||||
(
|
||||
writer.getFaceField
|
||||
// Write all volfields
|
||||
forAll(vsf, i)
|
||||
{
|
||||
writer.writeField
|
||||
(
|
||||
linearInterpolate(vsf[i])(),
|
||||
pp
|
||||
)()
|
||||
);
|
||||
}
|
||||
forAll(vvf, i)
|
||||
{
|
||||
writer.writeField
|
||||
(
|
||||
writer.getFaceField
|
||||
writer.getFaceField
|
||||
(
|
||||
linearInterpolate(vsf[i])(),
|
||||
pp
|
||||
)()
|
||||
);
|
||||
}
|
||||
forAll(vvf, i)
|
||||
{
|
||||
writer.writeField
|
||||
(
|
||||
linearInterpolate(vvf[i])(),
|
||||
pp
|
||||
)()
|
||||
);
|
||||
}
|
||||
forAll(vSpheretf, i)
|
||||
{
|
||||
writer.writeField
|
||||
(
|
||||
writer.getFaceField
|
||||
writer.getFaceField
|
||||
(
|
||||
linearInterpolate(vvf[i])(),
|
||||
pp
|
||||
)()
|
||||
);
|
||||
}
|
||||
forAll(vSpheretf, i)
|
||||
{
|
||||
writer.writeField
|
||||
(
|
||||
linearInterpolate(vSpheretf[i])(),
|
||||
pp
|
||||
)()
|
||||
);
|
||||
}
|
||||
forAll(vSymmtf, i)
|
||||
{
|
||||
writer.writeField
|
||||
(
|
||||
writer.getFaceField
|
||||
writer.getFaceField
|
||||
(
|
||||
linearInterpolate(vSpheretf[i])(),
|
||||
pp
|
||||
)()
|
||||
);
|
||||
}
|
||||
forAll(vSymmtf, i)
|
||||
{
|
||||
writer.writeField
|
||||
(
|
||||
linearInterpolate(vSymmtf[i])(),
|
||||
pp
|
||||
)()
|
||||
);
|
||||
}
|
||||
forAll(vtf, i)
|
||||
{
|
||||
writer.writeField
|
||||
(
|
||||
writer.getFaceField
|
||||
writer.getFaceField
|
||||
(
|
||||
linearInterpolate(vSymmtf[i])(),
|
||||
pp
|
||||
)()
|
||||
);
|
||||
}
|
||||
forAll(vtf, i)
|
||||
{
|
||||
writer.writeField
|
||||
(
|
||||
linearInterpolate(vtf[i])(),
|
||||
pp
|
||||
)()
|
||||
);
|
||||
}
|
||||
writer.getFaceField
|
||||
(
|
||||
linearInterpolate(vtf[i])(),
|
||||
pp
|
||||
)()
|
||||
);
|
||||
}
|
||||
|
||||
writer.writeConnectivity(ipp);
|
||||
writer.writeConnectivity(ipp);
|
||||
}
|
||||
else
|
||||
{
|
||||
Info<< " Skipping zero sized faceZone " << zoneI
|
||||
<< "\t" << pp.name()
|
||||
<< nl << endl;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
@ -91,7 +91,7 @@ bool Foam::IOobject::readHeader(Istream& is)
|
||||
<< "First token could not be read or is not the keyword 'FoamFile'"
|
||||
<< nl << nl << "Check header is of the form:" << nl << endl;
|
||||
|
||||
writeHeader(SeriousError);
|
||||
writeHeader(Info);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
@ -77,6 +77,9 @@ public:
|
||||
//- Apply and accumulate the effect of the given constraint direction
|
||||
inline void applyConstraint(const vector& cd);
|
||||
|
||||
//- Combine constraints
|
||||
inline void combine(const pointConstraint&);
|
||||
|
||||
//- Return the accumulated constraint transformation tensor
|
||||
inline tensor constraintTransformation() const;
|
||||
};
|
||||
|
||||
@ -69,6 +69,47 @@ void Foam::pointConstraint::applyConstraint(const vector& cd)
|
||||
}
|
||||
|
||||
|
||||
void Foam::pointConstraint::combine(const pointConstraint& pc)
|
||||
{
|
||||
if (first() == 0)
|
||||
{
|
||||
operator=(pc);
|
||||
}
|
||||
else if (first() == 1)
|
||||
{
|
||||
// Save single normal
|
||||
vector n = second();
|
||||
// Apply to supplied point constaint
|
||||
operator=(pc);
|
||||
applyConstraint(n);
|
||||
}
|
||||
else if (first() == 2)
|
||||
{
|
||||
if (pc.first() == 0)
|
||||
{}
|
||||
else if (pc.first() == 1)
|
||||
{
|
||||
applyConstraint(pc.second());
|
||||
}
|
||||
else if (pc.first() == 2)
|
||||
{
|
||||
// Both constrained to line. Same (+-)direction?
|
||||
if (mag(second() & pc.second()) <= (1.0-1e-3))
|
||||
{
|
||||
// Different directions
|
||||
first() = 3;
|
||||
second() = vector::zero;
|
||||
}
|
||||
}
|
||||
else
|
||||
{
|
||||
first() = 3;
|
||||
second() = vector::zero;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
Foam::tensor Foam::pointConstraint::constraintTransformation() const
|
||||
{
|
||||
if (first() == 0)
|
||||
|
||||
@ -132,7 +132,7 @@ public:
|
||||
// associated with any faces
|
||||
virtual const labelList& loneMeshPoints() const;
|
||||
|
||||
//- Return point unit normals. Not impelemented.
|
||||
//- Return point unit normals. Not implemented.
|
||||
virtual const vectorField& pointNormals() const;
|
||||
};
|
||||
|
||||
|
||||
@ -199,13 +199,20 @@ Foam::scalarField Foam::coupledPolyPatch::calcFaceTol
|
||||
|
||||
const face& f = faces[faceI];
|
||||
|
||||
// 1. calculate a typical size of the face. Use maximum distance
|
||||
// to face centre
|
||||
scalar maxLenSqr = -GREAT;
|
||||
// 2. as measure of truncation error when comparing two coordinates
|
||||
// use SMALL * maximum component
|
||||
scalar maxCmpt = -GREAT;
|
||||
|
||||
forAll(f, fp)
|
||||
{
|
||||
maxLenSqr = max(maxLenSqr, magSqr(points[f[fp]] - cc));
|
||||
const point& pt = points[f[fp]];
|
||||
maxLenSqr = max(maxLenSqr, magSqr(pt - cc));
|
||||
maxCmpt = max(maxCmpt, cmptMax(cmptMag(pt)));
|
||||
}
|
||||
tols[faceI] = matchTol * Foam::sqrt(maxLenSqr);
|
||||
tols[faceI] = max(SMALL*maxCmpt, matchTol*Foam::sqrt(maxLenSqr));
|
||||
}
|
||||
return tols;
|
||||
}
|
||||
|
||||
@ -824,7 +824,7 @@ bool Foam::polyMeshGeometry::checkFaceSkewness
|
||||
// (i.e. treat as if mirror cell on other side)
|
||||
|
||||
vector faceNormal = faceAreas[faceI];
|
||||
faceNormal /= mag(faceNormal) + VSMALL;
|
||||
faceNormal /= mag(faceNormal) + ROOTVSMALL;
|
||||
|
||||
vector dOwn = faceCentres[faceI] - cellCentres[own[faceI]];
|
||||
|
||||
@ -834,7 +834,7 @@ bool Foam::polyMeshGeometry::checkFaceSkewness
|
||||
|
||||
scalar skewness =
|
||||
mag(faceCentres[faceI] - faceIntersection)
|
||||
/(2*mag(dWall) + VSMALL);
|
||||
/(2*mag(dWall) + ROOTVSMALL);
|
||||
|
||||
// Check if the skewness vector is greater than the PN vector.
|
||||
// This does not cause trouble but is a good indication of a poor
|
||||
|
||||
@ -142,7 +142,7 @@ public:
|
||||
|
||||
// Access
|
||||
|
||||
//- Retirn local reference cast into the cyclic patch
|
||||
//- Return local reference cast into the cyclic patch
|
||||
const cyclicFvPatch& cyclicPatch() const
|
||||
{
|
||||
return cyclicPatch_;
|
||||
|
||||
@ -95,7 +95,7 @@ protected:
|
||||
class unionEqOp
|
||||
{
|
||||
public:
|
||||
void operator()( labelList& x, const labelList& y ) const;
|
||||
void operator()(labelList& x, const labelList& y) const;
|
||||
};
|
||||
|
||||
//- Collect cell neighbours of faces in global numbering
|
||||
|
||||
@ -2490,13 +2490,18 @@ void Foam::autoLayerDriver::mergePatchFacesUndo
|
||||
<< "---------------------------" << nl
|
||||
<< " - which are on the same patch" << nl
|
||||
<< " - which make an angle < " << layerParams.featureAngle()
|
||||
<< "- which are on the same patch" << nl
|
||||
<< "- which make an angle < " << layerParams.featureAngle()
|
||||
<< " degrees"
|
||||
<< nl
|
||||
<< " (cos:" << minCos << ')' << nl
|
||||
<< " - as long as the resulting face doesn't become concave"
|
||||
<< " (cos:" << minCos << ')' << nl
|
||||
<< "- as long as the resulting face doesn't become concave"
|
||||
<< " by more than "
|
||||
<< layerParams.concaveAngle() << " degrees" << nl
|
||||
<< " (0=straight, 180=fully concave)" << nl
|
||||
<< " (0=straight, 180=fully concave)" << nl
|
||||
<< endl;
|
||||
|
||||
label nChanged = mergePatchFacesUndo(minCos, concaveCos, motionDict);
|
||||
@ -2546,11 +2551,6 @@ void Foam::autoLayerDriver::addLayers
|
||||
);
|
||||
|
||||
|
||||
// Undistorted edge length
|
||||
const scalar edge0Len = meshRefiner_.meshCutter().level0EdgeLength();
|
||||
const labelList& cellLevel = meshRefiner_.meshCutter().cellLevel();
|
||||
|
||||
|
||||
// Point-wise extrusion data
|
||||
// ~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
@ -2612,6 +2612,10 @@ void Foam::autoLayerDriver::addLayers
|
||||
// Disable extrusion on warped faces
|
||||
// ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
// Undistorted edge length
|
||||
const scalar edge0Len = meshRefiner_.meshCutter().level0EdgeLength();
|
||||
const labelList& cellLevel = meshRefiner_.meshCutter().cellLevel();
|
||||
|
||||
handleWarpedFaces
|
||||
(
|
||||
pp,
|
||||
@ -2651,6 +2655,10 @@ void Foam::autoLayerDriver::addLayers
|
||||
}
|
||||
|
||||
|
||||
// Undistorted edge length
|
||||
const scalar edge0Len = meshRefiner_.meshCutter().level0EdgeLength();
|
||||
const labelList& cellLevel = meshRefiner_.meshCutter().cellLevel();
|
||||
|
||||
// Determine (wanted) point-wise layer thickness and expansion ratio
|
||||
scalarField thickness(pp().nPoints());
|
||||
scalarField minThickness(pp().nPoints());
|
||||
|
||||
@ -54,6 +54,7 @@ License
|
||||
#include "Random.H"
|
||||
#include "searchableSurfaces.H"
|
||||
#include "treeBoundBox.H"
|
||||
#include "zeroGradientFvPatchFields.H"
|
||||
|
||||
// * * * * * * * * * * * * * * Static Data Members * * * * * * * * * * * * * //
|
||||
|
||||
@ -2128,7 +2129,8 @@ void Foam::meshRefinement::dumpRefinementLevel() const
|
||||
false
|
||||
),
|
||||
mesh_,
|
||||
dimensionedScalar("zero", dimless, 0)
|
||||
dimensionedScalar("zero", dimless, 0),
|
||||
zeroGradientFvPatchScalarField::typeName
|
||||
);
|
||||
|
||||
const labelList& cellLevel = meshCutter_.cellLevel();
|
||||
|
||||
@ -496,23 +496,63 @@ Foam::labelList Foam::refinementSurfaces::getClosedNamedSurfaces() const
|
||||
}
|
||||
|
||||
|
||||
// Count number of triangles per surface region
|
||||
Foam::labelList Foam::refinementSurfaces::countRegions(const triSurface& s)
|
||||
{
|
||||
const geometricSurfacePatchList& regions = s.patches();
|
||||
// // Count number of triangles per surface region
|
||||
// Foam::labelList Foam::refinementSurfaces::countRegions(const triSurface& s)
|
||||
// {
|
||||
// const geometricSurfacePatchList& regions = s.patches();
|
||||
//
|
||||
// labelList nTris(regions.size(), 0);
|
||||
//
|
||||
// forAll(s, triI)
|
||||
// {
|
||||
// nTris[s[triI].region()]++;
|
||||
// }
|
||||
// return nTris;
|
||||
// }
|
||||
|
||||
labelList nTris(regions.size(), 0);
|
||||
|
||||
forAll(s, triI)
|
||||
{
|
||||
nTris[s[triI].region()]++;
|
||||
}
|
||||
return nTris;
|
||||
}
|
||||
// // Pre-calculate the refinement level for every element
|
||||
// void Foam::refinementSurfaces::wantedRefinementLevel
|
||||
// (
|
||||
// const shellSurfaces& shells,
|
||||
// const label surfI,
|
||||
// const List<pointIndexHit>& info, // Indices
|
||||
// const pointField& ctrs, // Representative coordinate
|
||||
// labelList& minLevelField
|
||||
// ) const
|
||||
// {
|
||||
// const searchableSurface& geom = allGeometry_[surfaces_[surfI]];
|
||||
//
|
||||
// // Get per element the region
|
||||
// labelList region;
|
||||
// geom.getRegion(info, region);
|
||||
//
|
||||
// // Initialise fields to region wise minLevel
|
||||
// minLevelField.setSize(ctrs.size());
|
||||
// minLevelField = -1;
|
||||
//
|
||||
// forAll(minLevelField, i)
|
||||
// {
|
||||
// if (info[i].hit())
|
||||
// {
|
||||
// minLevelField[i] = minLevel(surfI, region[i]);
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// // Find out if triangle inside shell with higher level
|
||||
// // What level does shell want to refine fc to?
|
||||
// labelList shellLevel;
|
||||
// shells.findHigherLevel(ctrs, minLevelField, shellLevel);
|
||||
//
|
||||
// forAll(minLevelField, i)
|
||||
// {
|
||||
// minLevelField[i] = max(minLevelField[i], shellLevel[i]);
|
||||
// }
|
||||
// }
|
||||
|
||||
|
||||
// Precalculate the refinement level for every element of the searchable
|
||||
// surface. This is currently hardcoded for triSurfaceMesh only.
|
||||
// surface.
|
||||
void Foam::refinementSurfaces::setMinLevelFields
|
||||
(
|
||||
const shellSurfaces& shells
|
||||
@ -522,58 +562,55 @@ void Foam::refinementSurfaces::setMinLevelFields
|
||||
{
|
||||
const searchableSurface& geom = allGeometry_[surfaces_[surfI]];
|
||||
|
||||
if (isA<triSurfaceMesh>(geom))
|
||||
// Precalculation only makes sense if there are different regions
|
||||
// (so different refinement levels possible) and there are some
|
||||
// elements. Possibly should have 'enough' elements to have fine
|
||||
// enough resolution but for now just make sure we don't catch e.g.
|
||||
// searchableBox (size=6)
|
||||
if (geom.regions().size() > 1 && geom.globalSize() > 10)
|
||||
{
|
||||
const triSurfaceMesh& triMesh = refCast<const triSurfaceMesh>(geom);
|
||||
// Representative local coordinates
|
||||
const pointField ctrs = geom.coordinates();
|
||||
|
||||
autoPtr<triSurfaceLabelField> minLevelFieldPtr
|
||||
(
|
||||
new triSurfaceLabelField
|
||||
(
|
||||
IOobject
|
||||
(
|
||||
"minLevel",
|
||||
triMesh.objectRegistry::time().timeName(), // instance
|
||||
"triSurface", // local
|
||||
triMesh,
|
||||
IOobject::NO_READ,
|
||||
IOobject::AUTO_WRITE
|
||||
),
|
||||
triMesh,
|
||||
dimless
|
||||
)
|
||||
);
|
||||
triSurfaceLabelField& minLevelField = minLevelFieldPtr();
|
||||
|
||||
const triSurface& s = static_cast<const triSurface&>(triMesh);
|
||||
|
||||
// Initialise fields to region wise minLevel
|
||||
forAll(s, triI)
|
||||
labelList minLevelField(ctrs.size(), -1);
|
||||
{
|
||||
minLevelField[triI] = minLevel(surfI, s[triI].region());
|
||||
// Get the element index in a roundabout way. Problem is e.g.
|
||||
// distributed surface where local indices differ from global
|
||||
// ones (needed for getRegion call)
|
||||
List<pointIndexHit> info;
|
||||
geom.findNearest
|
||||
(
|
||||
ctrs,
|
||||
scalarField(ctrs.size(), sqr(GREAT)),
|
||||
info
|
||||
);
|
||||
|
||||
// Get per element the region
|
||||
labelList region;
|
||||
geom.getRegion(info, region);
|
||||
|
||||
// From the region get the surface-wise refinement level
|
||||
forAll(minLevelField, i)
|
||||
{
|
||||
if (info[i].hit())
|
||||
{
|
||||
minLevelField[i] = minLevel(surfI, region[i]);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Find out if triangle inside shell with higher level
|
||||
pointField fc(s.size());
|
||||
forAll(s, triI)
|
||||
{
|
||||
fc[triI] = s[triI].centre(s.points());
|
||||
}
|
||||
// What level does shell want to refine fc to?
|
||||
labelList shellLevel;
|
||||
shells.findHigherLevel(fc, minLevelField, shellLevel);
|
||||
shells.findHigherLevel(ctrs, minLevelField, shellLevel);
|
||||
|
||||
forAll(minLevelField, triI)
|
||||
forAll(minLevelField, i)
|
||||
{
|
||||
minLevelField[triI] = max
|
||||
(
|
||||
minLevelField[triI],
|
||||
shellLevel[triI]
|
||||
);
|
||||
minLevelField[i] = max(minLevelField[i], shellLevel[i]);
|
||||
}
|
||||
|
||||
// Store field on triMesh
|
||||
minLevelFieldPtr.ptr()->store();
|
||||
// Store minLevelField on surface
|
||||
const_cast<searchableSurface&>(geom).setField(minLevelField);
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -601,44 +638,89 @@ void Foam::refinementSurfaces::findHigherIntersection
|
||||
return;
|
||||
}
|
||||
|
||||
if (surfaces_.size() == 1)
|
||||
{
|
||||
// Optimisation: single segmented surface. No need to duplicate
|
||||
// point storage.
|
||||
|
||||
label surfI = 0;
|
||||
|
||||
const searchableSurface& geom = allGeometry_[surfaces_[surfI]];
|
||||
|
||||
// Do intersection test
|
||||
List<pointIndexHit> intersectionInfo(start.size());
|
||||
geom.findLineAny(start, end, intersectionInfo);
|
||||
|
||||
// See if a cached level field available
|
||||
labelList minLevelField;
|
||||
geom.getField(intersectionInfo, minLevelField);
|
||||
bool haveLevelField =
|
||||
(
|
||||
returnReduce(minLevelField.size(), sumOp<label>())
|
||||
> 0
|
||||
);
|
||||
|
||||
if (!haveLevelField && geom.regions().size() == 1)
|
||||
{
|
||||
minLevelField = labelList
|
||||
(
|
||||
intersectionInfo.size(),
|
||||
minLevel(surfI, 0)
|
||||
);
|
||||
haveLevelField = true;
|
||||
}
|
||||
|
||||
|
||||
if (haveLevelField)
|
||||
{
|
||||
forAll(intersectionInfo, i)
|
||||
{
|
||||
if
|
||||
(
|
||||
intersectionInfo[i].hit()
|
||||
&& minLevelField[i] > currentLevel[i]
|
||||
)
|
||||
{
|
||||
surfaces[i] = surfI; // index of surface
|
||||
surfaceLevel[i] = minLevelField[i];
|
||||
}
|
||||
}
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
// Work arrays
|
||||
labelList hitMap(identity(start.size()));
|
||||
pointField p0(start);
|
||||
pointField p1(end);
|
||||
List<pointIndexHit> hitInfo(start.size());
|
||||
labelList intersectionToPoint(identity(start.size()));
|
||||
List<pointIndexHit> intersectionInfo(start.size());
|
||||
|
||||
forAll(surfaces_, surfI)
|
||||
{
|
||||
const searchableSurface& geom = allGeometry_[surfaces_[surfI]];
|
||||
|
||||
geom.findLineAny(p0, p1, hitInfo);
|
||||
// Do intersection test
|
||||
geom.findLineAny(p0, p1, intersectionInfo);
|
||||
|
||||
// See if a cached level field available
|
||||
labelList minLevelField;
|
||||
if (isA<triSurfaceMesh>(geom))
|
||||
{
|
||||
const triSurfaceMesh& triMesh = refCast<const triSurfaceMesh>(geom);
|
||||
triMesh.getField
|
||||
(
|
||||
"minLevel",
|
||||
hitInfo,
|
||||
minLevelField
|
||||
);
|
||||
}
|
||||
geom.getField(intersectionInfo, minLevelField);
|
||||
|
||||
|
||||
// Copy all hits into arguments, continue with misses
|
||||
label newI = 0;
|
||||
forAll(hitInfo, hitI)
|
||||
// Copy all hits into arguments, In-place compact misses.
|
||||
label missI = 0;
|
||||
forAll(intersectionInfo, i)
|
||||
{
|
||||
// Get the minLevel for the point
|
||||
label minLocalLevel = -1;
|
||||
|
||||
if (hitInfo[hitI].hit())
|
||||
if (intersectionInfo[i].hit())
|
||||
{
|
||||
// Check if minLevelField for this surface.
|
||||
if (minLevelField.size())
|
||||
{
|
||||
minLocalLevel = minLevelField[hitI];
|
||||
minLocalLevel = minLevelField[i];
|
||||
}
|
||||
else
|
||||
{
|
||||
@ -648,36 +730,35 @@ void Foam::refinementSurfaces::findHigherIntersection
|
||||
}
|
||||
}
|
||||
|
||||
label pointI = hitMap[hitI];
|
||||
|
||||
label pointI = intersectionToPoint[i];
|
||||
|
||||
if (minLocalLevel > currentLevel[pointI])
|
||||
{
|
||||
// Mark point for refinement
|
||||
surfaces[pointI] = surfI;
|
||||
surfaceLevel[pointI] = minLocalLevel;
|
||||
}
|
||||
else
|
||||
{
|
||||
if (hitI != newI)
|
||||
{
|
||||
hitMap[newI] = hitMap[hitI];
|
||||
p0[newI] = p0[hitI];
|
||||
p1[newI] = p1[hitI];
|
||||
}
|
||||
newI++;
|
||||
p0[missI] = p0[pointI];
|
||||
p1[missI] = p1[pointI];
|
||||
intersectionToPoint[missI] = pointI;
|
||||
missI++;
|
||||
}
|
||||
}
|
||||
|
||||
// All done? Note that this decision should be synchronised
|
||||
if (returnReduce(newI, sumOp<label>()) == 0)
|
||||
if (returnReduce(missI, sumOp<label>()) == 0)
|
||||
{
|
||||
break;
|
||||
}
|
||||
|
||||
// Trim and continue
|
||||
hitMap.setSize(newI);
|
||||
p0.setSize(newI);
|
||||
p1.setSize(newI);
|
||||
hitInfo.setSize(newI);
|
||||
// Trim misses
|
||||
p0.setSize(missI);
|
||||
p1.setSize(missI);
|
||||
intersectionToPoint.setSize(missI);
|
||||
intersectionInfo.setSize(missI);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@ -218,8 +218,8 @@ public:
|
||||
const shellSurfaces& shells
|
||||
);
|
||||
|
||||
//- Helper: count number of triangles per region
|
||||
static labelList countRegions(const triSurface&);
|
||||
////- Helper: count number of triangles per region
|
||||
//static labelList countRegions(const triSurface&);
|
||||
|
||||
|
||||
// Searching
|
||||
|
||||
@ -957,6 +957,7 @@ Foam::point Foam::indexedOctree<Type>::pushPointIntoFace
|
||||
}
|
||||
else if (nFaces == 1)
|
||||
{
|
||||
// Point is on a single face
|
||||
keepFaceID = faceIndices[0];
|
||||
}
|
||||
else
|
||||
@ -1782,16 +1783,6 @@ Foam::pointIndexHit Foam::indexedOctree<Type>::findLine
|
||||
label i = 0;
|
||||
for (; i < 100000; i++)
|
||||
{
|
||||
if (verbose)
|
||||
{
|
||||
Pout<< "iter:" << i
|
||||
<< " at startPoint:" << hitInfo.rawPoint() << endl
|
||||
<< " node:" << nodeI
|
||||
<< " octant:" << octant
|
||||
<< " bb:" << subBbox(nodeI, octant) << endl;
|
||||
}
|
||||
|
||||
|
||||
// Ray-trace to end of current node. Updates point (either on triangle
|
||||
// in case of hit or on node bounding box in case of miss)
|
||||
|
||||
@ -1808,6 +1799,19 @@ Foam::pointIndexHit Foam::indexedOctree<Type>::findLine
|
||||
)
|
||||
);
|
||||
|
||||
if (verbose)
|
||||
{
|
||||
Pout<< "iter:" << i
|
||||
<< " at current:" << hitInfo.rawPoint()
|
||||
<< " (perturbed:" << startPoint << ")" << endl
|
||||
<< " node:" << nodeI
|
||||
<< " octant:" << octant
|
||||
<< " bb:" << subBbox(nodeI, octant) << endl;
|
||||
}
|
||||
|
||||
|
||||
|
||||
|
||||
// Faces of current bounding box current point is on
|
||||
direction hitFaceID = 0;
|
||||
|
||||
@ -1833,12 +1837,23 @@ Foam::pointIndexHit Foam::indexedOctree<Type>::findLine
|
||||
break;
|
||||
}
|
||||
|
||||
if (hitFaceID == 0)
|
||||
if (hitFaceID == 0 || hitInfo.rawPoint() == treeEnd)
|
||||
{
|
||||
// endpoint inside the tree. Return miss.
|
||||
break;
|
||||
}
|
||||
|
||||
// Create a point on other side of face.
|
||||
point perturbedPoint
|
||||
(
|
||||
pushPoint
|
||||
(
|
||||
octantBb,
|
||||
hitFaceID,
|
||||
hitInfo.rawPoint(),
|
||||
false // push outside of octantBb
|
||||
)
|
||||
);
|
||||
|
||||
if (verbose)
|
||||
{
|
||||
@ -1848,14 +1863,7 @@ Foam::pointIndexHit Foam::indexedOctree<Type>::findLine
|
||||
<< " node:" << nodeI
|
||||
<< " octant:" << octant
|
||||
<< " bb:" << subBbox(nodeI, octant) << nl
|
||||
<< " walking to neighbour containing:"
|
||||
<< pushPoint
|
||||
(
|
||||
octantBb,
|
||||
hitFaceID,
|
||||
hitInfo.rawPoint(),
|
||||
false // push outside of octantBb
|
||||
)
|
||||
<< " walking to neighbour containing:" << perturbedPoint
|
||||
<< endl;
|
||||
}
|
||||
|
||||
@ -1866,13 +1874,7 @@ Foam::pointIndexHit Foam::indexedOctree<Type>::findLine
|
||||
|
||||
bool ok = walkToNeighbour
|
||||
(
|
||||
pushPoint
|
||||
(
|
||||
octantBb,
|
||||
hitFaceID,
|
||||
hitInfo.rawPoint(),
|
||||
false // push outside of octantBb
|
||||
),
|
||||
perturbedPoint,
|
||||
hitFaceID, // face(s) that hitInfo is on
|
||||
|
||||
nodeI,
|
||||
|
||||
@ -463,7 +463,7 @@ const Foam::indexedOctree<Foam::treeDataFace>& Foam::meshSearch::boundaryTree()
|
||||
|
||||
treeBoundBox overallBb(mesh_.points());
|
||||
Random rndGen(123456);
|
||||
overallBb.extend(rndGen, 1E-4);
|
||||
overallBb = overallBb.extend(rndGen, 1E-4);
|
||||
overallBb.min() -= point(ROOTVSMALL, ROOTVSMALL, ROOTVSMALL);
|
||||
overallBb.max() += point(ROOTVSMALL, ROOTVSMALL, ROOTVSMALL);
|
||||
|
||||
@ -497,7 +497,7 @@ const Foam::indexedOctree<Foam::treeDataCell>& Foam::meshSearch::cellTree()
|
||||
|
||||
treeBoundBox overallBb(mesh_.points());
|
||||
Random rndGen(123456);
|
||||
overallBb.extend(rndGen, 1E-4);
|
||||
overallBb = overallBb.extend(rndGen, 1E-4);
|
||||
overallBb.min() -= point(ROOTVSMALL, ROOTVSMALL, ROOTVSMALL);
|
||||
overallBb.max() += point(ROOTVSMALL, ROOTVSMALL, ROOTVSMALL);
|
||||
|
||||
@ -531,7 +531,7 @@ const Foam::indexedOctree<Foam::treeDataPoint>&
|
||||
|
||||
treeBoundBox overallBb(mesh_.cellCentres());
|
||||
Random rndGen(123456);
|
||||
overallBb.extend(rndGen, 1E-4);
|
||||
overallBb = overallBb.extend(rndGen, 1E-4);
|
||||
overallBb.min() -= point(ROOTVSMALL, ROOTVSMALL, ROOTVSMALL);
|
||||
overallBb.max() += point(ROOTVSMALL, ROOTVSMALL, ROOTVSMALL);
|
||||
|
||||
|
||||
@ -912,41 +912,6 @@ Foam::distributedTriSurfaceMesh::independentlyDistributedBbs
|
||||
}
|
||||
|
||||
|
||||
void Foam::distributedTriSurfaceMesh::calcBounds
|
||||
(
|
||||
boundBox& bb,
|
||||
label& nPoints
|
||||
) const
|
||||
{
|
||||
// Unfortunately nPoints constructs meshPoints() so do compact version
|
||||
// ourselves
|
||||
|
||||
PackedBoolList pointIsUsed(points().size());
|
||||
|
||||
nPoints = 0;
|
||||
bb.min() = point(VGREAT, VGREAT, VGREAT);
|
||||
bb.max() = point(-VGREAT, -VGREAT, -VGREAT);
|
||||
|
||||
const triSurface& s = static_cast<const triSurface&>(*this);
|
||||
|
||||
forAll(s, triI)
|
||||
{
|
||||
const labelledTri& f = s[triI];
|
||||
|
||||
forAll(f, fp)
|
||||
{
|
||||
label pointI = f[fp];
|
||||
if (pointIsUsed.set(pointI, 1u))
|
||||
{
|
||||
bb.min() = ::Foam::min(bb.min(), points()[pointI]);
|
||||
bb.max() = ::Foam::max(bb.max(), points()[pointI]);
|
||||
nPoints++;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
// Does any part of triangle overlap bb.
|
||||
bool Foam::distributedTriSurfaceMesh::overlaps
|
||||
(
|
||||
@ -1935,20 +1900,7 @@ void Foam::distributedTriSurfaceMesh::getNormal
|
||||
{
|
||||
if (!Pstream::parRun())
|
||||
{
|
||||
normal.setSize(info.size());
|
||||
|
||||
forAll(info, i)
|
||||
{
|
||||
if (info[i].hit())
|
||||
{
|
||||
normal[i] = faceNormals()[info[i].index()];
|
||||
}
|
||||
else
|
||||
{
|
||||
// Set to what?
|
||||
normal[i] = vector::zero;
|
||||
}
|
||||
}
|
||||
triSurfaceMesh::getNormal(info, normal);
|
||||
return;
|
||||
}
|
||||
|
||||
@ -2006,70 +1958,64 @@ void Foam::distributedTriSurfaceMesh::getNormal
|
||||
|
||||
void Foam::distributedTriSurfaceMesh::getField
|
||||
(
|
||||
const word& fieldName,
|
||||
const List<pointIndexHit>& info,
|
||||
labelList& values
|
||||
) const
|
||||
{
|
||||
const triSurfaceLabelField& fld = lookupObject<triSurfaceLabelField>
|
||||
(
|
||||
fieldName
|
||||
);
|
||||
|
||||
|
||||
if (!Pstream::parRun())
|
||||
{
|
||||
values.setSize(info.size());
|
||||
forAll(info, i)
|
||||
{
|
||||
if (info[i].hit())
|
||||
{
|
||||
values[i] = fld[info[i].index()];
|
||||
}
|
||||
}
|
||||
triSurfaceMesh::getField(info, values);
|
||||
return;
|
||||
}
|
||||
|
||||
|
||||
// Get query data (= local index of triangle)
|
||||
// ~~~~~~~~~~~~~~
|
||||
|
||||
labelList triangleIndex(info.size());
|
||||
autoPtr<mapDistribute> mapPtr
|
||||
(
|
||||
calcLocalQueries
|
||||
(
|
||||
info,
|
||||
triangleIndex
|
||||
)
|
||||
);
|
||||
const mapDistribute& map = mapPtr();
|
||||
|
||||
|
||||
// Do my tests
|
||||
// ~~~~~~~~~~~
|
||||
|
||||
values.setSize(triangleIndex.size());
|
||||
|
||||
forAll(triangleIndex, i)
|
||||
if (foundObject<triSurfaceLabelField>("values"))
|
||||
{
|
||||
label triI = triangleIndex[i];
|
||||
values[i] = fld[triI];
|
||||
const triSurfaceLabelField& fld = lookupObject<triSurfaceLabelField>
|
||||
(
|
||||
"values"
|
||||
);
|
||||
|
||||
|
||||
// Get query data (= local index of triangle)
|
||||
// ~~~~~~~~~~~~~~
|
||||
|
||||
labelList triangleIndex(info.size());
|
||||
autoPtr<mapDistribute> mapPtr
|
||||
(
|
||||
calcLocalQueries
|
||||
(
|
||||
info,
|
||||
triangleIndex
|
||||
)
|
||||
);
|
||||
const mapDistribute& map = mapPtr();
|
||||
|
||||
|
||||
// Do my tests
|
||||
// ~~~~~~~~~~~
|
||||
|
||||
values.setSize(triangleIndex.size());
|
||||
|
||||
forAll(triangleIndex, i)
|
||||
{
|
||||
label triI = triangleIndex[i];
|
||||
values[i] = fld[triI];
|
||||
}
|
||||
|
||||
|
||||
// Send back results
|
||||
// ~~~~~~~~~~~~~~~~~
|
||||
|
||||
map.distribute
|
||||
(
|
||||
Pstream::nonBlocking,
|
||||
List<labelPair>(0),
|
||||
info.size(),
|
||||
map.constructMap(), // what to send
|
||||
map.subMap(), // what to receive
|
||||
values
|
||||
);
|
||||
}
|
||||
|
||||
|
||||
// Send back results
|
||||
// ~~~~~~~~~~~~~~~~~
|
||||
|
||||
map.distribute
|
||||
(
|
||||
Pstream::nonBlocking,
|
||||
List<labelPair>(0),
|
||||
info.size(),
|
||||
map.constructMap(), // what to send
|
||||
map.subMap(), // what to receive
|
||||
values
|
||||
);
|
||||
}
|
||||
|
||||
|
||||
|
||||
@ -217,9 +217,6 @@ private:
|
||||
const triSurface&
|
||||
);
|
||||
|
||||
//- Calculate surface bounding box
|
||||
void calcBounds(boundBox& bb, label& nPoints) const;
|
||||
|
||||
//- Does any part of triangle overlap bb.
|
||||
static bool overlaps
|
||||
(
|
||||
@ -418,7 +415,7 @@ public:
|
||||
// Should really be split into a routine to determine decomposition
|
||||
// and one that does actual distribution but determining
|
||||
// decomposition with duplicate triangle merging requires
|
||||
// same amoun as work as actual distribution.
|
||||
// same amount as work as actual distribution.
|
||||
virtual void distribute
|
||||
(
|
||||
const List<treeBoundBox>&,
|
||||
@ -430,14 +427,9 @@ public:
|
||||
|
||||
// Other
|
||||
|
||||
//- Specific to triSurfaceMesh: from a set of hits (points and
|
||||
//- WIP. From a set of hits (points and
|
||||
// indices) get the specified field. Misses do not get set.
|
||||
virtual void getField
|
||||
(
|
||||
const word& fieldName,
|
||||
const List<pointIndexHit>&,
|
||||
labelList& values
|
||||
) const;
|
||||
virtual void getField(const List<pointIndexHit>&, labelList&) const;
|
||||
|
||||
//- Subset the part of surface that is overlapping bounds.
|
||||
static triSurface overlappingSurface
|
||||
|
||||
@ -229,6 +229,21 @@ const Foam::wordList& Foam::searchableBox::regions() const
|
||||
}
|
||||
|
||||
|
||||
Foam::pointField Foam::searchableBox::coordinates() const
|
||||
{
|
||||
pointField ctrs(6);
|
||||
|
||||
const pointField pts = treeBoundBox::points();
|
||||
const faceList& fcs = treeBoundBox::faces;
|
||||
|
||||
forAll(fcs, i)
|
||||
{
|
||||
ctrs[i] = fcs[i].centre(pts);
|
||||
}
|
||||
return ctrs;
|
||||
}
|
||||
|
||||
|
||||
Foam::pointIndexHit Foam::searchableBox::findNearest
|
||||
(
|
||||
const point& sample,
|
||||
|
||||
@ -127,6 +127,9 @@ public:
|
||||
return 6;
|
||||
}
|
||||
|
||||
//- Get representative set of element coordinates
|
||||
// Usually the element centres (should be of length size()).
|
||||
virtual pointField coordinates() const;
|
||||
|
||||
// Single point queries.
|
||||
|
||||
|
||||
@ -40,6 +40,14 @@ addToRunTimeSelectionTable(searchableSurface, searchableCylinder, dict);
|
||||
|
||||
// * * * * * * * * * * * * * Private Member Functions * * * * * * * * * * * //
|
||||
|
||||
Foam::pointField Foam::searchableCylinder::coordinates() const
|
||||
{
|
||||
pointField ctrs(1, 0.5*(point1_ + point2_));
|
||||
|
||||
return ctrs;
|
||||
}
|
||||
|
||||
|
||||
Foam::pointIndexHit Foam::searchableCylinder::findNearest
|
||||
(
|
||||
const point& sample,
|
||||
|
||||
@ -150,6 +150,10 @@ public:
|
||||
return 1;
|
||||
}
|
||||
|
||||
//- Get representative set of element coordinates
|
||||
// Usually the element centres (should be of length size()).
|
||||
virtual pointField coordinates() const;
|
||||
|
||||
|
||||
// Multiple point queries.
|
||||
|
||||
|
||||
@ -26,7 +26,7 @@ Class
|
||||
Foam::searchablePlane
|
||||
|
||||
Description
|
||||
Searching on plane. See plane.H
|
||||
Searching on (infinite) plane. See plane.H
|
||||
|
||||
SourceFiles
|
||||
searchablePlane.C
|
||||
@ -122,6 +122,14 @@ public:
|
||||
return 1;
|
||||
}
|
||||
|
||||
//- Get representative set of element coordinates
|
||||
// Usually the element centres (should be of length size()).
|
||||
virtual pointField coordinates() const
|
||||
{
|
||||
//notImplemented("searchablePlane::coordinates()")
|
||||
return pointField(1, refPoint());
|
||||
}
|
||||
|
||||
|
||||
// Multiple point queries.
|
||||
|
||||
|
||||
@ -145,6 +145,13 @@ public:
|
||||
return 1;
|
||||
}
|
||||
|
||||
//- Get representative set of element coordinates
|
||||
// Usually the element centres (should be of length size()).
|
||||
virtual pointField coordinates() const
|
||||
{
|
||||
return pointField(1, origin_);
|
||||
}
|
||||
|
||||
|
||||
// Multiple point queries.
|
||||
|
||||
|
||||
@ -133,6 +133,13 @@ public:
|
||||
return 1;
|
||||
}
|
||||
|
||||
//- Get representative set of element coordinates
|
||||
// Usually the element centres (should be of length size()).
|
||||
virtual pointField coordinates() const
|
||||
{
|
||||
return pointField(1, centre_);
|
||||
}
|
||||
|
||||
|
||||
// Multiple point queries.
|
||||
|
||||
|
||||
@ -190,6 +190,10 @@ public:
|
||||
return size();
|
||||
}
|
||||
|
||||
//- Get representative set of element coordinates
|
||||
// Usually the element centres (should be of length size()).
|
||||
virtual pointField coordinates() const = 0;
|
||||
|
||||
|
||||
// Single point queries.
|
||||
|
||||
@ -319,6 +323,18 @@ public:
|
||||
)
|
||||
{}
|
||||
|
||||
//- WIP. Store element-wise field.
|
||||
virtual void setField(const labelList& values)
|
||||
{}
|
||||
|
||||
//- WIP. From a set of hits (points and
|
||||
// indices) get the specified field. Misses do not get set. Return
|
||||
// empty field if not supported.
|
||||
virtual void getField(const List<pointIndexHit>&, labelList& values)
|
||||
const
|
||||
{
|
||||
values.clear();
|
||||
}
|
||||
|
||||
};
|
||||
|
||||
|
||||
@ -101,7 +101,11 @@ void Foam::searchableSurfaceCollection::findNearest
|
||||
minDistSqr[pointI] = distSqr;
|
||||
nearestInfo[pointI].setPoint(globalPt);
|
||||
nearestInfo[pointI].setHit();
|
||||
nearestInfo[pointI].setIndex(hitInfo[pointI].index());
|
||||
nearestInfo[pointI].setIndex
|
||||
(
|
||||
hitInfo[pointI].index()
|
||||
+ indexOffset_[surfI]
|
||||
);
|
||||
nearestSurf[pointI] = surfI;
|
||||
}
|
||||
}
|
||||
@ -110,6 +114,62 @@ void Foam::searchableSurfaceCollection::findNearest
|
||||
}
|
||||
|
||||
|
||||
// Sort hits into per-surface bins. Misses are rejected. Maintains map back
|
||||
// to position
|
||||
void Foam::searchableSurfaceCollection::sortHits
|
||||
(
|
||||
const List<pointIndexHit>& info,
|
||||
List<List<pointIndexHit> >& surfInfo,
|
||||
labelListList& infoMap
|
||||
) const
|
||||
{
|
||||
// Count hits per surface.
|
||||
labelList nHits(subGeom_.size(), 0);
|
||||
|
||||
forAll(info, pointI)
|
||||
{
|
||||
if (info[pointI].hit())
|
||||
{
|
||||
label index = info[pointI].index();
|
||||
label surfI = findLower(indexOffset_, index+1);
|
||||
nHits[surfI]++;
|
||||
}
|
||||
}
|
||||
|
||||
// Per surface the hit
|
||||
surfInfo.setSize(subGeom_.size());
|
||||
// Per surface the original position
|
||||
infoMap.setSize(subGeom_.size());
|
||||
|
||||
forAll(surfInfo, surfI)
|
||||
{
|
||||
surfInfo[surfI].setSize(nHits[surfI]);
|
||||
infoMap[surfI].setSize(nHits[surfI]);
|
||||
}
|
||||
nHits = 0;
|
||||
|
||||
forAll(info, pointI)
|
||||
{
|
||||
if (info[pointI].hit())
|
||||
{
|
||||
label index = info[pointI].index();
|
||||
label surfI = findLower(indexOffset_, index+1);
|
||||
|
||||
// Store for correct surface and adapt indices back to local
|
||||
// ones
|
||||
label localI = nHits[surfI]++;
|
||||
surfInfo[surfI][localI] = pointIndexHit
|
||||
(
|
||||
info[pointI].hit(),
|
||||
info[pointI].rawPoint(),
|
||||
index-indexOffset_[surfI]
|
||||
);
|
||||
infoMap[surfI][localI] = pointI;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
// * * * * * * * * * * * * * * * * Constructors * * * * * * * * * * * * * * //
|
||||
|
||||
Foam::searchableSurfaceCollection::searchableSurfaceCollection
|
||||
@ -123,11 +183,13 @@ Foam::searchableSurfaceCollection::searchableSurfaceCollection
|
||||
scale_(dict.size()),
|
||||
transform_(dict.size()),
|
||||
subGeom_(dict.size()),
|
||||
mergeSubRegions_(dict.lookup("mergeSubRegions"))
|
||||
mergeSubRegions_(dict.lookup("mergeSubRegions")),
|
||||
indexOffset_(dict.size()+1)
|
||||
{
|
||||
Info<< "SearchableCollection : " << name() << endl;
|
||||
|
||||
label surfI = 0;
|
||||
label startIndex = 0;
|
||||
forAllConstIter(dictionary, dict, iter)
|
||||
{
|
||||
if (dict.isDict(iter().keyword()))
|
||||
@ -153,8 +215,24 @@ Foam::searchableSurfaceCollection::searchableSurfaceCollection
|
||||
const searchableSurface& s =
|
||||
io.db().lookupObject<searchableSurface>(subGeomName);
|
||||
|
||||
// I don't know yet how to handle the globalSize combined with
|
||||
// regionOffset. Would cause non-consecutive indices locally
|
||||
// if all indices offset by globalSize() of the local region...
|
||||
if (s.size() != s.globalSize())
|
||||
{
|
||||
FatalErrorIn
|
||||
(
|
||||
"searchableSurfaceCollection::searchableSurfaceCollection"
|
||||
"(const IOobject&, const dictionary&)"
|
||||
) << "Cannot use a distributed surface in a collection."
|
||||
<< exit(FatalError);
|
||||
}
|
||||
|
||||
subGeom_.set(surfI, &const_cast<searchableSurface&>(s));
|
||||
|
||||
indexOffset_[surfI] = startIndex;
|
||||
startIndex += subGeom_[surfI].size();
|
||||
|
||||
Info<< " instance : " << instance_[surfI] << endl;
|
||||
Info<< " surface : " << s.name() << endl;
|
||||
Info<< " scale : " << scale_[surfI] << endl;
|
||||
@ -163,10 +241,13 @@ Foam::searchableSurfaceCollection::searchableSurfaceCollection
|
||||
surfI++;
|
||||
}
|
||||
}
|
||||
indexOffset_[surfI] = startIndex;
|
||||
|
||||
instance_.setSize(surfI);
|
||||
scale_.setSize(surfI);
|
||||
transform_.setSize(surfI);
|
||||
subGeom_.setSize(surfI);
|
||||
indexOffset_.setSize(surfI+1);
|
||||
}
|
||||
|
||||
|
||||
@ -212,12 +293,36 @@ const Foam::wordList& Foam::searchableSurfaceCollection::regions() const
|
||||
|
||||
Foam::label Foam::searchableSurfaceCollection::size() const
|
||||
{
|
||||
label n = 0;
|
||||
return indexOffset_[indexOffset_.size()-1];
|
||||
}
|
||||
|
||||
|
||||
Foam::pointField Foam::searchableSurfaceCollection::coordinates() const
|
||||
{
|
||||
// Get overall size
|
||||
pointField coords(size());
|
||||
|
||||
// Append individual coordinates
|
||||
label coordI = 0;
|
||||
|
||||
forAll(subGeom_, surfI)
|
||||
{
|
||||
n += subGeom_[surfI].size();
|
||||
const pointField subCoords = subGeom_[surfI].coordinates();
|
||||
|
||||
forAll(subCoords, i)
|
||||
{
|
||||
coords[coordI++] = transform_[surfI].globalPosition
|
||||
(
|
||||
cmptMultiply
|
||||
(
|
||||
subCoords[i],
|
||||
scale_[surfI]
|
||||
)
|
||||
);
|
||||
}
|
||||
}
|
||||
return n;
|
||||
|
||||
return coords;
|
||||
}
|
||||
|
||||
|
||||
@ -296,6 +401,11 @@ void Foam::searchableSurfaceCollection::findLine
|
||||
);
|
||||
info[pointI] = hitInfo[pointI];
|
||||
info[pointI].rawPoint() = nearest[pointI];
|
||||
info[pointI].setIndex
|
||||
(
|
||||
hitInfo[pointI].index()
|
||||
+ indexOffset_[surfI]
|
||||
);
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -397,82 +507,42 @@ void Foam::searchableSurfaceCollection::getRegion
|
||||
}
|
||||
else
|
||||
{
|
||||
// Multiple surfaces. Sort by surface.
|
||||
|
||||
// Per surface the hit
|
||||
List<List<pointIndexHit> > surfInfo;
|
||||
// Per surface the original position
|
||||
List<List<label> > infoMap;
|
||||
sortHits(info, surfInfo, infoMap);
|
||||
|
||||
region.setSize(info.size());
|
||||
region = -1;
|
||||
|
||||
// Which region did point come from. Retest for now to see which
|
||||
// surface it originates from - crap solution! Should use global indices
|
||||
// in index inside pointIndexHit to do this better.
|
||||
// Do region tests
|
||||
|
||||
pointField samples(info.size());
|
||||
forAll(info, pointI)
|
||||
if (mergeSubRegions_)
|
||||
{
|
||||
if (info[pointI].hit())
|
||||
// Actually no need for surfInfo. Just take region for surface.
|
||||
forAll(infoMap, surfI)
|
||||
{
|
||||
samples[pointI] = info[pointI].hitPoint();
|
||||
}
|
||||
else
|
||||
{
|
||||
samples[pointI] = vector::zero;
|
||||
}
|
||||
}
|
||||
//scalarField minDistSqr(info.size(), SMALL);
|
||||
scalarField minDistSqr(info.size(), GREAT);
|
||||
|
||||
labelList nearestSurf;
|
||||
List<pointIndexHit> nearestInfo;
|
||||
findNearest
|
||||
(
|
||||
samples,
|
||||
minDistSqr,
|
||||
nearestInfo,
|
||||
nearestSurf
|
||||
);
|
||||
|
||||
// Check
|
||||
{
|
||||
forAll(info, pointI)
|
||||
{
|
||||
if (info[pointI].hit() && nearestSurf[pointI] == -1)
|
||||
const labelList& map = infoMap[surfI];
|
||||
forAll(map, i)
|
||||
{
|
||||
FatalErrorIn
|
||||
(
|
||||
"searchableSurfaceCollection::getRegion(..)"
|
||||
) << "pointI:" << pointI
|
||||
<< " sample:" << samples[pointI]
|
||||
<< " nearest:" << nearestInfo[pointI]
|
||||
<< " nearestsurf:" << nearestSurf[pointI]
|
||||
<< abort(FatalError);
|
||||
region[map[i]] = regionOffset_[surfI];
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
forAll(subGeom_, surfI)
|
||||
else
|
||||
{
|
||||
// Collect points from my surface
|
||||
labelList indices(findIndices(nearestSurf, surfI));
|
||||
|
||||
if (mergeSubRegions_)
|
||||
{
|
||||
forAll(indices, i)
|
||||
{
|
||||
region[indices[i]] = regionOffset_[surfI];
|
||||
}
|
||||
}
|
||||
else
|
||||
forAll(infoMap, surfI)
|
||||
{
|
||||
labelList surfRegion;
|
||||
subGeom_[surfI].getRegion
|
||||
(
|
||||
List<pointIndexHit>
|
||||
(
|
||||
UIndirectList<pointIndexHit>(info, indices)
|
||||
),
|
||||
surfRegion
|
||||
);
|
||||
forAll(indices, i)
|
||||
subGeom_[surfI].getRegion(surfInfo[surfI], surfRegion);
|
||||
|
||||
const labelList& map = infoMap[surfI];
|
||||
forAll(map, i)
|
||||
{
|
||||
region[indices[i]] = regionOffset_[surfI] + surfRegion[i];
|
||||
region[map[i]] = regionOffset_[surfI] + surfRegion[i];
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -494,52 +564,26 @@ void Foam::searchableSurfaceCollection::getNormal
|
||||
}
|
||||
else
|
||||
{
|
||||
// Multiple surfaces. Sort by surface.
|
||||
|
||||
// Per surface the hit
|
||||
List<List<pointIndexHit> > surfInfo;
|
||||
// Per surface the original position
|
||||
List<List<label> > infoMap;
|
||||
sortHits(info, surfInfo, infoMap);
|
||||
|
||||
normal.setSize(info.size());
|
||||
|
||||
// See above - crap retest to find surface point originates from.
|
||||
pointField samples(info.size());
|
||||
forAll(info, pointI)
|
||||
// Do region tests
|
||||
forAll(surfInfo, surfI)
|
||||
{
|
||||
if (info[pointI].hit())
|
||||
{
|
||||
samples[pointI] = info[pointI].hitPoint();
|
||||
}
|
||||
else
|
||||
{
|
||||
samples[pointI] = vector::zero;
|
||||
}
|
||||
}
|
||||
//scalarField minDistSqr(info.size(), SMALL);
|
||||
scalarField minDistSqr(info.size(), GREAT);
|
||||
|
||||
labelList nearestSurf;
|
||||
List<pointIndexHit> nearestInfo;
|
||||
findNearest
|
||||
(
|
||||
samples,
|
||||
minDistSqr,
|
||||
nearestInfo,
|
||||
nearestSurf
|
||||
);
|
||||
|
||||
|
||||
forAll(subGeom_, surfI)
|
||||
{
|
||||
// Collect points from my surface
|
||||
labelList indices(findIndices(nearestSurf, surfI));
|
||||
|
||||
vectorField surfNormal;
|
||||
subGeom_[surfI].getNormal
|
||||
(
|
||||
List<pointIndexHit>
|
||||
(
|
||||
UIndirectList<pointIndexHit>(info, indices)
|
||||
),
|
||||
surfNormal
|
||||
);
|
||||
forAll(indices, i)
|
||||
subGeom_[surfI].getNormal(surfInfo[surfI], surfNormal);
|
||||
|
||||
const labelList& map = infoMap[surfI];
|
||||
forAll(map, i)
|
||||
{
|
||||
normal[indices[i]] = surfNormal[i];
|
||||
normal[map[i]] = surfNormal[i];
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -561,4 +605,99 @@ void Foam::searchableSurfaceCollection::getVolumeType
|
||||
}
|
||||
|
||||
|
||||
void Foam::searchableSurfaceCollection::distribute
|
||||
(
|
||||
const List<treeBoundBox>& bbs,
|
||||
const bool keepNonLocal,
|
||||
autoPtr<mapDistribute>& faceMap,
|
||||
autoPtr<mapDistribute>& pointMap
|
||||
)
|
||||
{
|
||||
forAll(subGeom_, surfI)
|
||||
{
|
||||
// Note:Tranform the bounding boxes? Something like
|
||||
// pointField bbPoints =
|
||||
// cmptDivide
|
||||
// (
|
||||
// transform_[surfI].localPosition
|
||||
// (
|
||||
// bbs[i].points()
|
||||
// ),
|
||||
// scale_[surfI]
|
||||
// );
|
||||
// treeBoundBox newBb(bbPoints);
|
||||
|
||||
// Note: what to do with faceMap, pointMap from multiple surfaces?
|
||||
subGeom_[surfI].distribute
|
||||
(
|
||||
bbs,
|
||||
keepNonLocal,
|
||||
faceMap,
|
||||
pointMap
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
void Foam::searchableSurfaceCollection::setField(const labelList& values)
|
||||
{
|
||||
forAll(subGeom_, surfI)
|
||||
{
|
||||
subGeom_[surfI].setField
|
||||
(
|
||||
static_cast<const labelList&>
|
||||
(
|
||||
SubList<label>
|
||||
(
|
||||
values,
|
||||
subGeom_[surfI].size(),
|
||||
indexOffset_[surfI]
|
||||
)
|
||||
)
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
void Foam::searchableSurfaceCollection::getField
|
||||
(
|
||||
const List<pointIndexHit>& info,
|
||||
labelList& values
|
||||
) const
|
||||
{
|
||||
if (subGeom_.size() == 0)
|
||||
{}
|
||||
else if (subGeom_.size() == 1)
|
||||
{
|
||||
subGeom_[0].getField(info, values);
|
||||
}
|
||||
else
|
||||
{
|
||||
// Multiple surfaces. Sort by surface.
|
||||
|
||||
// Per surface the hit
|
||||
List<List<pointIndexHit> > surfInfo;
|
||||
// Per surface the original position
|
||||
List<List<label> > infoMap;
|
||||
sortHits(info, surfInfo, infoMap);
|
||||
|
||||
values.setSize(info.size());
|
||||
//?Misses do not get set? values = 0;
|
||||
|
||||
// Do surface tests
|
||||
forAll(surfInfo, surfI)
|
||||
{
|
||||
labelList surfValues;
|
||||
subGeom_[surfI].getField(surfInfo[surfI], surfValues);
|
||||
|
||||
const labelList& map = infoMap[surfI];
|
||||
forAll(map, i)
|
||||
{
|
||||
values[map[i]] = surfValues[i];
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
// ************************************************************************* //
|
||||
|
||||
@ -77,6 +77,10 @@ private:
|
||||
|
||||
Switch mergeSubRegions_;
|
||||
|
||||
//- offsets for indices coming from different surfaces
|
||||
// (sized with size() of each surface)
|
||||
labelList indexOffset_;
|
||||
|
||||
//- Region names
|
||||
mutable wordList regions_;
|
||||
//- From individual regions to collection regions
|
||||
@ -95,6 +99,15 @@ private:
|
||||
labelList& nearestSurf
|
||||
) const;
|
||||
|
||||
//- Sort hits into per-surface bins. Misses are rejected.
|
||||
// Maintains map back to position
|
||||
void sortHits
|
||||
(
|
||||
const List<pointIndexHit>& info,
|
||||
List<List<pointIndexHit> >& surfInfo,
|
||||
labelListList& infoMap
|
||||
) const;
|
||||
|
||||
|
||||
//- Disallow default bitwise copy construct
|
||||
searchableSurfaceCollection(const searchableSurfaceCollection&);
|
||||
@ -161,6 +174,10 @@ public:
|
||||
//- Range of local indices that can be returned.
|
||||
virtual label size() const;
|
||||
|
||||
//- Get representative set of element coordinates
|
||||
// Usually the element centres (should be of length size()).
|
||||
virtual pointField coordinates() const;
|
||||
|
||||
|
||||
// Multiple point queries.
|
||||
|
||||
@ -215,6 +232,27 @@ public:
|
||||
List<volumeType>&
|
||||
) const;
|
||||
|
||||
// Other
|
||||
|
||||
//- Set bounds of surface. Bounds currently set as list of
|
||||
// bounding boxes. The bounds are hints to the surface as for
|
||||
// the range of queries it can expect. faceMap/pointMap can be
|
||||
// set if the surface has done any redistribution.
|
||||
virtual void distribute
|
||||
(
|
||||
const List<treeBoundBox>&,
|
||||
const bool keepNonLocal,
|
||||
autoPtr<mapDistribute>& faceMap,
|
||||
autoPtr<mapDistribute>& pointMap
|
||||
);
|
||||
|
||||
//- WIP. Store element-wise field.
|
||||
virtual void setField(const labelList& values);
|
||||
|
||||
//- WIP. From a set of hits (points and
|
||||
// indices) get the specified field. Misses do not get set. Return
|
||||
// empty field if not supported.
|
||||
virtual void getField(const List<pointIndexHit>&, labelList&) const;
|
||||
|
||||
// regIOobject implementation
|
||||
|
||||
|
||||
@ -151,6 +151,13 @@ public:
|
||||
return surface().size();
|
||||
}
|
||||
|
||||
//- Get representative set of element coordinates
|
||||
// Usually the element centres (should be of length size()).
|
||||
virtual pointField coordinates() const
|
||||
{
|
||||
return surface().coordinates();
|
||||
}
|
||||
|
||||
|
||||
// Multiple point queries.
|
||||
|
||||
@ -225,6 +232,41 @@ public:
|
||||
}
|
||||
|
||||
|
||||
// Other
|
||||
|
||||
//- Set bounds of surface. Bounds currently set as list of
|
||||
// bounding boxes. The bounds are hints to the surface as for
|
||||
// the range of queries it can expect. faceMap/pointMap can be
|
||||
// set if the surface has done any redistribution.
|
||||
virtual void distribute
|
||||
(
|
||||
const List<treeBoundBox>& bbs,
|
||||
const bool keepNonLocal,
|
||||
autoPtr<mapDistribute>& faceMap,
|
||||
autoPtr<mapDistribute>& pointMap
|
||||
)
|
||||
{
|
||||
subGeom_[0].distribute(bbs, keepNonLocal, faceMap, pointMap);
|
||||
}
|
||||
|
||||
//- WIP. Store element-wise field.
|
||||
virtual void setField(const labelList& values)
|
||||
{
|
||||
subGeom_[0].setField(values);
|
||||
}
|
||||
|
||||
//- WIP. From a set of hits (points and
|
||||
// indices) get the specified field. Misses do not get set. Return
|
||||
// empty field if not supported.
|
||||
virtual void getField
|
||||
(
|
||||
const List<pointIndexHit>& info,
|
||||
labelList& values
|
||||
) const
|
||||
{
|
||||
surface().getField(info, values);
|
||||
}
|
||||
|
||||
// regIOobject implementation
|
||||
|
||||
bool writeData(Ostream& os) const
|
||||
|
||||
@ -30,6 +30,7 @@ License
|
||||
#include "EdgeMap.H"
|
||||
#include "triSurfaceFields.H"
|
||||
#include "Time.H"
|
||||
#include "PackedBoolList.H"
|
||||
|
||||
// * * * * * * * * * * * * * * Static Data Members * * * * * * * * * * * * * //
|
||||
|
||||
@ -281,6 +282,36 @@ void Foam::triSurfaceMesh::getNextIntersections
|
||||
}
|
||||
|
||||
|
||||
void Foam::triSurfaceMesh::calcBounds(boundBox& bb, label& nPoints) const
|
||||
{
|
||||
// Unfortunately nPoints constructs meshPoints() so do compact version
|
||||
// ourselves
|
||||
|
||||
const triSurface& s = static_cast<const triSurface&>(*this);
|
||||
|
||||
PackedBoolList pointIsUsed(points().size());
|
||||
|
||||
nPoints = 0;
|
||||
bb = boundBox::invertedBox;
|
||||
|
||||
forAll(s, triI)
|
||||
{
|
||||
const labelledTri& f = s[triI];
|
||||
|
||||
forAll(f, fp)
|
||||
{
|
||||
label pointI = f[fp];
|
||||
if (pointIsUsed.set(pointI, 1u))
|
||||
{
|
||||
bb.min() = ::Foam::min(bb.min(), points()[pointI]);
|
||||
bb.max() = ::Foam::max(bb.max(), points()[pointI]);
|
||||
nPoints++;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
// * * * * * * * * * * * * * * * * Constructors * * * * * * * * * * * * * * //
|
||||
|
||||
Foam::triSurfaceMesh::triSurfaceMesh(const IOobject& io, const triSurface& s)
|
||||
@ -301,6 +332,7 @@ Foam::triSurfaceMesh::triSurfaceMesh(const IOobject& io, const triSurface& s)
|
||||
),
|
||||
triSurface(s),
|
||||
tolerance_(indexedOctree<treeDataTriSurface>::perturbTol()),
|
||||
maxTreeDepth_(10),
|
||||
surfaceClosed_(-1)
|
||||
{}
|
||||
|
||||
@ -344,6 +376,7 @@ Foam::triSurfaceMesh::triSurfaceMesh(const IOobject& io)
|
||||
)
|
||||
),
|
||||
tolerance_(indexedOctree<treeDataTriSurface>::perturbTol()),
|
||||
maxTreeDepth_(10),
|
||||
surfaceClosed_(-1)
|
||||
{}
|
||||
|
||||
@ -390,6 +423,7 @@ Foam::triSurfaceMesh::triSurfaceMesh
|
||||
)
|
||||
),
|
||||
tolerance_(indexedOctree<treeDataTriSurface>::perturbTol()),
|
||||
maxTreeDepth_(10),
|
||||
surfaceClosed_(-1)
|
||||
{
|
||||
scalar scaleFactor = 0;
|
||||
@ -410,6 +444,14 @@ Foam::triSurfaceMesh::triSurfaceMesh
|
||||
Info<< searchableSurface::name() << " : using intersection tolerance "
|
||||
<< tolerance_ << endl;
|
||||
}
|
||||
|
||||
|
||||
// Have optional non-standard tree-depth to limit storage.
|
||||
if (dict.readIfPresent("maxTreeDepth", maxTreeDepth_) && maxTreeDepth_ > 0)
|
||||
{
|
||||
Info<< searchableSurface::name() << " : using maximum tree depth "
|
||||
<< maxTreeDepth_ << endl;
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@ -431,6 +473,17 @@ void Foam::triSurfaceMesh::clearOut()
|
||||
|
||||
// * * * * * * * * * * * * * * * Member Functions * * * * * * * * * * * * * //
|
||||
|
||||
Foam::pointField Foam::triSurfaceMesh::coordinates() const
|
||||
{
|
||||
// Use copy to calculate face centres so they don't get stored
|
||||
return PrimitivePatch<labelledTri, SubList, const pointField&>
|
||||
(
|
||||
SubList<labelledTri>(*this, triSurface::size()),
|
||||
triSurface::points()
|
||||
).faceCentres();
|
||||
}
|
||||
|
||||
|
||||
void Foam::triSurfaceMesh::movePoints(const pointField& newPoints)
|
||||
{
|
||||
tree_.clear();
|
||||
@ -444,15 +497,28 @@ const Foam::indexedOctree<Foam::treeDataTriSurface>&
|
||||
{
|
||||
if (tree_.empty())
|
||||
{
|
||||
// Calculate bb without constructing local point numbering.
|
||||
treeBoundBox bb;
|
||||
label nPoints;
|
||||
calcBounds(bb, nPoints);
|
||||
|
||||
if (nPoints != points().size())
|
||||
{
|
||||
WarningIn("triSurfaceMesh::tree() const")
|
||||
<< "Surface " << searchableSurface::name()
|
||||
<< " does not have compact point numbering."
|
||||
<< " Of " << points().size() << " only " << nPoints
|
||||
<< " are used. This might give problems in some routines."
|
||||
<< endl;
|
||||
}
|
||||
|
||||
|
||||
// Random number generator. Bit dodgy since not exactly random ;-)
|
||||
Random rndGen(65431);
|
||||
|
||||
// Slightly extended bb. Slightly off-centred just so on symmetric
|
||||
// geometry there are less face/edge aligned items.
|
||||
treeBoundBox bb
|
||||
(
|
||||
treeBoundBox(points(), meshPoints()).extend(rndGen, 1E-4)
|
||||
);
|
||||
bb = bb.extend(rndGen, 1E-4);
|
||||
bb.min() -= point(ROOTVSMALL, ROOTVSMALL, ROOTVSMALL);
|
||||
bb.max() += point(ROOTVSMALL, ROOTVSMALL, ROOTVSMALL);
|
||||
|
||||
@ -465,9 +531,9 @@ const Foam::indexedOctree<Foam::treeDataTriSurface>&
|
||||
(
|
||||
treeDataTriSurface(*this),
|
||||
bb,
|
||||
10, // maxLevel
|
||||
10, // leafsize
|
||||
3.0 // duplicity
|
||||
maxTreeDepth_, // maxLevel
|
||||
10, // leafsize
|
||||
3.0 // duplicity
|
||||
)
|
||||
);
|
||||
|
||||
@ -494,15 +560,17 @@ const Foam::indexedOctree<Foam::treeDataEdge>&
|
||||
+ nInternalEdges()
|
||||
);
|
||||
|
||||
treeBoundBox bb;
|
||||
label nPoints;
|
||||
calcBounds(bb, nPoints);
|
||||
|
||||
// Random number generator. Bit dodgy since not exactly random ;-)
|
||||
Random rndGen(65431);
|
||||
|
||||
// Slightly extended bb. Slightly off-centred just so on symmetric
|
||||
// geometry there are less face/edge aligned items.
|
||||
treeBoundBox bb
|
||||
(
|
||||
treeBoundBox(points(), meshPoints()).extend(rndGen, 1E-4)
|
||||
);
|
||||
|
||||
bb = bb.extend(rndGen, 1E-4);
|
||||
bb.min() -= point(ROOTVSMALL, ROOTVSMALL, ROOTVSMALL);
|
||||
bb.max() += point(ROOTVSMALL, ROOTVSMALL, ROOTVSMALL);
|
||||
|
||||
@ -517,10 +585,10 @@ const Foam::indexedOctree<Foam::treeDataEdge>&
|
||||
localPoints(), // points
|
||||
bEdges // selected edges
|
||||
),
|
||||
bb, // bb
|
||||
8, // maxLevel
|
||||
10, // leafsize
|
||||
3.0 // duplicity
|
||||
bb, // bb
|
||||
maxTreeDepth_, // maxLevel
|
||||
10, // leafsize
|
||||
3.0 // duplicity
|
||||
)
|
||||
);
|
||||
}
|
||||
@ -754,24 +822,53 @@ void Foam::triSurfaceMesh::getNormal
|
||||
}
|
||||
|
||||
|
||||
void Foam::triSurfaceMesh::setField(const labelList& values)
|
||||
{
|
||||
autoPtr<triSurfaceLabelField> fldPtr
|
||||
(
|
||||
new triSurfaceLabelField
|
||||
(
|
||||
IOobject
|
||||
(
|
||||
"values",
|
||||
objectRegistry::time().timeName(), // instance
|
||||
"triSurface", // local
|
||||
*this,
|
||||
IOobject::NO_READ,
|
||||
IOobject::AUTO_WRITE
|
||||
),
|
||||
*this,
|
||||
dimless,
|
||||
labelField(values)
|
||||
)
|
||||
);
|
||||
|
||||
// Store field on triMesh
|
||||
fldPtr.ptr()->store();
|
||||
}
|
||||
|
||||
|
||||
void Foam::triSurfaceMesh::getField
|
||||
(
|
||||
const word& fieldName,
|
||||
const List<pointIndexHit>& info,
|
||||
labelList& values
|
||||
) const
|
||||
{
|
||||
const triSurfaceLabelField& fld = lookupObject<triSurfaceLabelField>
|
||||
(
|
||||
fieldName
|
||||
);
|
||||
|
||||
values.setSize(info.size());
|
||||
forAll(info, i)
|
||||
if (foundObject<triSurfaceLabelField>("values"))
|
||||
{
|
||||
if (info[i].hit())
|
||||
values.setSize(info.size());
|
||||
|
||||
const triSurfaceLabelField& fld = lookupObject<triSurfaceLabelField>
|
||||
(
|
||||
"values"
|
||||
);
|
||||
|
||||
forAll(info, i)
|
||||
{
|
||||
values[i] = fld[info[i].index()];
|
||||
if (info[i].hit())
|
||||
{
|
||||
values[i] = fld[info[i].index()];
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@ -71,6 +71,9 @@ private:
|
||||
//- Optional tolerance to use in searches
|
||||
scalar tolerance_;
|
||||
|
||||
//- Optional max tree depth of octree
|
||||
label maxTreeDepth_;
|
||||
|
||||
//- Search tree (triangles)
|
||||
mutable autoPtr<indexedOctree<treeDataTriSurface> > tree_;
|
||||
|
||||
@ -129,6 +132,11 @@ private:
|
||||
void operator=(const triSurfaceMesh&);
|
||||
|
||||
|
||||
protected:
|
||||
|
||||
//- Calculate (number of)used points and their bounding box
|
||||
void calcBounds(boundBox& bb, label& nPoints) const;
|
||||
|
||||
public:
|
||||
|
||||
//- Runtime type information
|
||||
@ -185,6 +193,10 @@ public:
|
||||
return triSurface::size();
|
||||
}
|
||||
|
||||
//- Get representative set of element coordinates
|
||||
// Usually the element centres (should be of length size()).
|
||||
virtual pointField coordinates() const;
|
||||
|
||||
virtual void findNearest
|
||||
(
|
||||
const pointField& sample,
|
||||
@ -236,6 +248,8 @@ public:
|
||||
List<volumeType>&
|
||||
) const;
|
||||
|
||||
// Other
|
||||
|
||||
//- Set bounds of surface. Bounds currently set as list of
|
||||
// bounding boxes. The bounds are hints to the surface as for
|
||||
// the range of queries it can expect. faceMap/pointMap can be
|
||||
@ -249,16 +263,12 @@ public:
|
||||
)
|
||||
{}
|
||||
|
||||
// Other
|
||||
//- WIP. Store element-wise field.
|
||||
virtual void setField(const labelList& values);
|
||||
|
||||
//- Specific to triSurfaceMesh: from a set of hits (points and
|
||||
//- WIP. From a set of hits (points and
|
||||
// indices) get the specified field. Misses do not get set.
|
||||
virtual void getField
|
||||
(
|
||||
const word& fieldName,
|
||||
const List<pointIndexHit>&,
|
||||
labelList& values
|
||||
) const;
|
||||
virtual void getField(const List<pointIndexHit>&, labelList&) const;
|
||||
|
||||
|
||||
// regIOobject implementation
|
||||
|
||||
@ -22,40 +22,87 @@ License
|
||||
along with OpenFOAM; if not, write to the Free Software Foundation,
|
||||
Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
|
||||
|
||||
Default strategy:
|
||||
Strat=b
|
||||
{
|
||||
job=t,
|
||||
map=t,
|
||||
poli=S,
|
||||
sep=
|
||||
(
|
||||
m
|
||||
From scotch forum:
|
||||
|
||||
By: Francois PELLEGRINI RE: Graph mapping 'strategy' string [ reply ]
|
||||
2008-08-22 10:09 Strategy handling in Scotch is a bit tricky. In order
|
||||
not to be confused, you must have a clear view of how they are built.
|
||||
Here are some rules:
|
||||
|
||||
1- Strategies are made up of "methods" which are combined by means of
|
||||
"operators".
|
||||
|
||||
2- A method is of the form "m{param=value,param=value,...}", where "m"
|
||||
is a single character (this is your first error: "f" is a method name,
|
||||
not a parameter name).
|
||||
|
||||
3- There exist different sort of strategies : bipartitioning strategies,
|
||||
mapping strategies, ordering strategies, which cannot be mixed. For
|
||||
instance, you cannot build a bipartitioning strategy and feed it to a
|
||||
mapping method (this is your second error).
|
||||
|
||||
To use the "mapCompute" routine, you must create a mapping strategy, not
|
||||
a bipartitioning one, and so use stratGraphMap() and not
|
||||
stratGraphBipart(). Your mapping strategy should however be based on the
|
||||
"recursive bipartitioning" method ("b"). For instance, a simple (and
|
||||
hence not very efficient) mapping strategy can be :
|
||||
|
||||
"b{sep=f}"
|
||||
|
||||
which computes mappings with the recursive bipartitioning method "b",
|
||||
this latter using the Fiduccia-Mattheyses method "f" to compute its
|
||||
separators.
|
||||
|
||||
If you want an exact partition (see your previous post), try
|
||||
"b{sep=fx}".
|
||||
|
||||
However, these strategies are not the most efficient, as they do not
|
||||
make use of the multi-level framework.
|
||||
|
||||
To use the multi-level framework, try for instance:
|
||||
|
||||
"b{sep=m{vert=100,low=h,asc=f}x}"
|
||||
|
||||
The current default mapping strategy in Scotch can be seen by using the
|
||||
"-vs" option of program gmap. It is, to date:
|
||||
|
||||
b
|
||||
{
|
||||
job=t,
|
||||
map=t,
|
||||
poli=S,
|
||||
sep=
|
||||
(
|
||||
m
|
||||
{
|
||||
asc=b
|
||||
{
|
||||
asc=b
|
||||
{
|
||||
bnd=d{pass=40,dif=1,rem=1}f{move=80,pass=-1,bal=0.005},
|
||||
org=f{move=80,pass=-1,bal=0.005},width=3
|
||||
},
|
||||
low=h{pass=10}f{move=80,pass=-1,bal=0.0005},
|
||||
type=h,
|
||||
vert=80,
|
||||
rat=0.8
|
||||
}
|
||||
| m
|
||||
bnd=d{pass=40,dif=1,rem=1}f{move=80,pass=-1,bal=0.005},
|
||||
org=f{move=80,pass=-1,bal=0.005},
|
||||
width=3
|
||||
},
|
||||
low=h{pass=10}f{move=80,pass=-1,bal=0.0005},
|
||||
type=h,
|
||||
vert=80,
|
||||
rat=0.8
|
||||
}
|
||||
| m
|
||||
{
|
||||
asc=b
|
||||
{
|
||||
asc=b
|
||||
{
|
||||
bnd=d{pass=40,dif=1,rem=1}f{move=80,pass=-1,bal=0.005},
|
||||
org=f{move=80,pass=-1,bal=0.005},
|
||||
width=3},
|
||||
low=h{pass=10}f{move=80,pass=-1,bal=0.0005},
|
||||
type=h,
|
||||
vert=80,
|
||||
rat=0.8
|
||||
}
|
||||
)
|
||||
}
|
||||
bnd=d{pass=40,dif=1,rem=1}f{move=80,pass=-1,bal=0.005},
|
||||
org=f{move=80,pass=-1,bal=0.005},
|
||||
width=3
|
||||
},
|
||||
low=h{pass=10}f{move=80,pass=-1,bal=0.0005},
|
||||
type=h,
|
||||
vert=80,
|
||||
rat=0.8
|
||||
}
|
||||
)
|
||||
}
|
||||
|
||||
|
||||
\*---------------------------------------------------------------------------*/
|
||||
|
||||
#include "scotchDecomp.H"
|
||||
@ -126,6 +173,51 @@ Foam::label Foam::scotchDecomp::decompose
|
||||
List<int>& finalDecomp
|
||||
)
|
||||
{
|
||||
// Dump graph
|
||||
if (decompositionDict_.found("scotchCoeffs"))
|
||||
{
|
||||
const dictionary& scotchCoeffs =
|
||||
decompositionDict_.subDict("scotchCoeffs");
|
||||
|
||||
if (scotchCoeffs.found("writeGraph"))
|
||||
{
|
||||
Switch writeGraph(scotchCoeffs.lookup("writeGraph"));
|
||||
|
||||
if (writeGraph)
|
||||
{
|
||||
OFstream str(mesh_.time().path() / mesh_.name() + ".grf");
|
||||
|
||||
Info<< "Dumping Scotch graph file to " << str.name() << endl
|
||||
<< "Use this in combination with gpart." << endl;
|
||||
|
||||
label version = 0;
|
||||
str << version << nl;
|
||||
// Numer of vertices
|
||||
str << xadj.size()-1 << ' ' << adjncy.size() << nl;
|
||||
// Numbering starts from 0
|
||||
label baseval = 0;
|
||||
// Has weights?
|
||||
label hasEdgeWeights = 0;
|
||||
label hasVertexWeights = 0;
|
||||
label numericflag = 10*hasEdgeWeights+hasVertexWeights;
|
||||
str << baseval << ' ' << numericflag << nl;
|
||||
for (label cellI = 0; cellI < xadj.size()-1; cellI++)
|
||||
{
|
||||
label start = xadj[cellI];
|
||||
label end = xadj[cellI+1];
|
||||
str << end-start;
|
||||
|
||||
for (label i = start; i < end; i++)
|
||||
{
|
||||
str << ' ' << adjncy[i];
|
||||
}
|
||||
str << nl;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
// Strategy
|
||||
// ~~~~~~~~
|
||||
|
||||
@ -206,51 +298,6 @@ Foam::label Foam::scotchDecomp::decompose
|
||||
check(SCOTCH_graphCheck(&grafdat), "SCOTCH_graphCheck");
|
||||
|
||||
|
||||
// Dump graph
|
||||
if (decompositionDict_.found("scotchCoeffs"))
|
||||
{
|
||||
const dictionary& scotchCoeffs =
|
||||
decompositionDict_.subDict("scotchCoeffs");
|
||||
|
||||
if (scotchCoeffs.found("writeGraph"))
|
||||
{
|
||||
Switch writeGraph(scotchCoeffs.lookup("writeGraph"));
|
||||
|
||||
if (writeGraph)
|
||||
{
|
||||
OFstream str(mesh_.time().path() / mesh_.name() + ".grf");
|
||||
|
||||
Info<< "Dumping Scotch graph file to " << str.name() << endl
|
||||
<< "Use this in combination with gpart." << endl;
|
||||
|
||||
label version = 0;
|
||||
str << version << nl;
|
||||
// Numer of vertices
|
||||
str << xadj.size()-1 << ' ' << adjncy.size() << nl;
|
||||
// Numbering starts from 0
|
||||
label baseval = 0;
|
||||
// Has weights?
|
||||
label hasEdgeWeights = 0;
|
||||
label hasVertexWeights = 0;
|
||||
label numericflag = 10*hasEdgeWeights+hasVertexWeights;
|
||||
str << baseval << ' ' << numericflag << nl;
|
||||
for (label cellI = 0; cellI < xadj.size()-1; cellI++)
|
||||
{
|
||||
label start = xadj[cellI];
|
||||
label end = xadj[cellI+1];
|
||||
str << end-start;
|
||||
|
||||
for (label i = start; i < end; i++)
|
||||
{
|
||||
str << ' ' << adjncy[i];
|
||||
}
|
||||
str << nl;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
// Architecture
|
||||
// ~~~~~~~~~~~~
|
||||
// (fully connected network topology since using switch)
|
||||
|
||||
@ -26,6 +26,7 @@ License
|
||||
|
||||
#include "triSurface.H"
|
||||
#include "mergePoints.H"
|
||||
#include "PackedBoolList.H"
|
||||
|
||||
// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * //
|
||||
|
||||
@ -46,46 +47,44 @@ bool triSurface::stitchTriangles
|
||||
pointField newPoints;
|
||||
bool hasMerged = mergePoints(rawPoints, tol, verbose, pointMap, newPoints);
|
||||
|
||||
pointField& ps = storedPoints();
|
||||
|
||||
// Set the coordinates to the merged ones
|
||||
ps.transfer(newPoints);
|
||||
|
||||
if (hasMerged)
|
||||
{
|
||||
if (verbose)
|
||||
{
|
||||
Pout<< "stitchTriangles : Merged from " << rawPoints.size()
|
||||
<< " points down to " << newPoints.size() << endl;
|
||||
<< " points down to " << ps.size() << endl;
|
||||
}
|
||||
|
||||
pointField& ps = storedPoints();
|
||||
|
||||
// Set the coordinates to the merged ones
|
||||
ps = newPoints;
|
||||
|
||||
// Reset the triangle point labels to the unique points array
|
||||
label newTriangleI = 0;
|
||||
forAll(*this, i)
|
||||
{
|
||||
label newA = pointMap[operator[](i)[0]];
|
||||
label newB = pointMap[operator[](i)[1]];
|
||||
label newC = pointMap[operator[](i)[2]];
|
||||
const labelledTri& tri = operator[](i);
|
||||
labelledTri newTri
|
||||
(
|
||||
pointMap[tri[0]],
|
||||
pointMap[tri[1]],
|
||||
pointMap[tri[2]],
|
||||
tri.region()
|
||||
);
|
||||
|
||||
if ((newA != newB) && (newA != newC) && (newB != newC))
|
||||
if ((newTri[0] != newTri[1]) && (newTri[0] != newTri[2]) && (newTri[1] != newTri[2]))
|
||||
{
|
||||
operator[](newTriangleI)[0] = newA;
|
||||
operator[](newTriangleI)[1] = newB;
|
||||
operator[](newTriangleI)[2] = newC;
|
||||
operator[](newTriangleI).region() = operator[](i).region();
|
||||
newTriangleI++;
|
||||
operator[](newTriangleI++) = newTri;
|
||||
}
|
||||
else if (verbose)
|
||||
{
|
||||
Pout<< "stitchTriangles : "
|
||||
<< "Removing triangle " << i << " with non-unique vertices."
|
||||
<< endl
|
||||
<< " vertices :" << newA << ' ' << newB << ' ' << newC
|
||||
<< endl
|
||||
<< " coordinates:" << ps[newA] << ' ' << ps[newB]
|
||||
<< ' ' << ps[newC] << endl;
|
||||
<< "Removing triangle " << i
|
||||
<< " with non-unique vertices." << endl
|
||||
<< " vertices :" << newTri << endl
|
||||
<< " coordinates:" << newTri.points(ps)
|
||||
<< endl;
|
||||
}
|
||||
}
|
||||
|
||||
@ -98,6 +97,58 @@ bool triSurface::stitchTriangles
|
||||
<< " triangles" << endl;
|
||||
}
|
||||
setSize(newTriangleI);
|
||||
|
||||
// And possibly compact out any unused points (since used only
|
||||
// by triangles that have just been deleted)
|
||||
// Done in two passes to save memory (pointField)
|
||||
|
||||
// 1. Detect only
|
||||
PackedBoolList pointIsUsed(ps.size());
|
||||
|
||||
label nPoints = 0;
|
||||
|
||||
forAll(*this, i)
|
||||
{
|
||||
const labelledTri& tri = operator[](i);
|
||||
|
||||
forAll(tri, fp)
|
||||
{
|
||||
label pointI = tri[fp];
|
||||
if (pointIsUsed.set(pointI, 1))
|
||||
{
|
||||
nPoints++;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if (nPoints != ps.size())
|
||||
{
|
||||
// 2. Compact.
|
||||
pointMap.setSize(ps.size());
|
||||
label newPointI = 0;
|
||||
forAll(pointIsUsed, pointI)
|
||||
{
|
||||
if (pointIsUsed[pointI])
|
||||
{
|
||||
ps[newPointI] = ps[pointI];
|
||||
pointMap[pointI] = newPointI++;
|
||||
}
|
||||
}
|
||||
ps.setSize(newPointI);
|
||||
|
||||
newTriangleI = 0;
|
||||
forAll(*this, i)
|
||||
{
|
||||
const labelledTri& tri = operator[](i);
|
||||
operator[](newTriangleI++) = labelledTri
|
||||
(
|
||||
pointMap[tri[0]],
|
||||
pointMap[tri[1]],
|
||||
pointMap[tri[2]],
|
||||
tri.region()
|
||||
);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
Reference in New Issue
Block a user