This is the opensource reference implementation of the SIGGRAPH 2020 paper Incremental Potential Contact: Intersection- and Inversion-free Large Deformation Dynamics.
-
src/
: source code -
cmake/
andCMakeLists.txt
: CMake files -
Format/
: automatic code formatting (requiresclang-format
) -
input/
: input data and scripts needed to rerun all examples in our paper, together with some tutorial examples -
output/
: output data (will be created) -
tests/
: unit-tests -
tools/
: Python and Bash scripts for generating and processing results -
wiki/
: images used in Github wiki page -
build.py
: a python script to automatically build IPC with default compiler -
batch.py
: a python script to automatically run a batch of examples (ininput/n/
by default) withn
threads for each linear solver in the program -
batch_tetgen.py
: a python script to automatically tetrahedralize a batch of closed surface meshes using Tetgen
You can either use python build.py
or
mkdir build
cd build
cmake -DCMAKE_BUILD_TYPE=Release ..
make -j4
- SuiteSparse: must be
installed at a system level
- Specifically IPC requires CHOLMOD, AMD, CAMD, CCOLAMD, and COLAMD which are all part of SuiteSparse and METIS which is sometimes a separate library
- macOS:
brew install suite-sparse
- Alternatively, you can build SuiteSparse and then copy all files under
SuiteSparse/include
andSuiteSparse/lib
to/usr/local/include
and/usr/local/lib
respectively.
- Alternatively, you can build SuiteSparse and then copy all files under
- Ubuntu:
sudo apt-get install libsuitesparse-dev libmetis-dev
- For program efficiency, we strongly recommend compiling SuiteSparse using Intel MKL LAPACK and BLAS (on an Intel machine):
make library BLAS='-lmkl_intel_lp64 -lmkl_core -lmkl_intel_thread -lpthread -lm -lmkl_blas95_lp64 -liomp5' LAPACK='-lmkl_lapack95_lp64' -j 8
- For program efficiency, we strongly recommend compiling SuiteSparse using Intel MKL LAPACK and BLAS (on an Intel machine):
- IPC also supports using Eigen or AMGCL as linear solver, which can be set via
IPC_LINSYSSOLVER
inCMakeLists.txt
. To use custom linear solvers, you can implement a new interface (subclass) to ourLinSysSolver
class.
- libigl, OSQP, AMGCL, and TBB: downloaded and built through CMake
If your system does not have the required OpenGL libraries, you can disable the viewer using the CMake argument -DIPC_WITH_OPENGL=OFF
. It is important to then run IPC in offline mode (see below).
IPC is only implemented for 3D simulation.
If your system uses Slurm like our cluster does, then you can take advantage of our scripts to build IPC as a batch job.
From the IPC root directory use sbatch build.sh
to compile IPC via a
batch job. You can check the status of the job using
squeue -u $USER -n IPC_build
.
It is important to not build on the login node because while it supports AVX2,
other computer nodes may not. Building on the login node will enable AVX2
support which may cause an Illegal instruction
error when run on a compute
node. Alternativly, this can be fixed by using Intel's compilers, icc
and icpc
, with the
flag -axCORE-AVX2
. This enables executables to run on compute nodes with or
without AVX2 instructions. If the compute nodes support AVX2 instructions,
executables will run AVX2 instructions, otherwise it will use other available
instructions such as AVX or SSE4.2.
IPC contains two optional sub-projects for unit tests, debug, and file processing. To enable them use the CMake
flag -DBUILD_<sub-project-name>_PROJECT=On
where sub-project-name
is the name
of the sub-project (e.g. DIAGNOSTIC
, MESH_PROCESSING
). You can also set these interactively using ccmake
.
Please see our quick start guide for a hello world
example with various settings. The output of each script will be saved into a separate folder in output/
.
It is possible to run IPC with and without the viewer. By default, the batch.py
script runs IPC with the viewer. If you provide the argument --offline
to batch.py
then it will run IPC in offline mode (i.e. without the viewer).
If you are running the IPC_bin
executable directly then the first argument controls the mode. Mode 10
runs IPC with the viewer and is the default in the batch.py
. Mode 100
runs IPC in offline mode (without the viewer). See IPC_bin --help
for more detail.
From the IPC root directory use sbatch run.sh
to processes all the input
files in input/12/
using 12 CPU cores.
This will process all input scenes in series. I am currently, working on a script to process all scenes in parallel via separate batch jobs.
The executables for the sub-projects are places in their respective directory
in src/Projects/
.
<n>.obj
: surface mesh of the simulated objects and volumetric/surface kinematic collision objects in n-th time stepn.seg
: mesh file of the segment kinematic collision objects if any in n-th time steppt<n>.obj
: vertex coordinates of the point cloud kinematic collision objects if any in n-th time stepstatus<n>
: saved information of n-th time step that can be used to restart simulation fromACO<m>_<n>.obj
: mesh file of m-th analytical plane if any in n-th (and the following if unchanged) time stepsBC_<n>.txt
: node index of Dirichlet nodes if any in n-th (and the following if unchanged) time steps0.png
: snapshot of initial time stepfinalResult.png
: snapshot of last time stepfinalResult_mesh.msh
: rest shape of simulated objectanim.gif
: animation preview of the simulationconfig.txt
: the config file containing all simulation settings that can be used to rerun the simulationiterStats.txt
: simulation/optimization statistics of each iterationinfo.txt
: timing informationlog.txt
: debug informationsysE.txt
: total energy (kinematic, gravity, and elasticity) of each simulated objectsysL.txt
: linear momentum of each simulated objectsysM.txt
: angular momentum of each simulated object
See our quick start guide for detailed examples of minimal settings needed for a simulation.
energy {FCR | NH}
NH
: Neo-Hookean elasticity energy (default)FCR
: Fixed Co-rotational elasticity energy
timeIntegration {BE | NM}
BE
: Backward Euler (default)NM
: Newmark
dampingStiff <dampingStiff>
- add lagged Rayleigh damping with stiffness
dampingStiff
or not ifdampingStiff
is zero or not specified dampingStiff
must be greater than or equal to zero
- add lagged Rayleigh damping with stiffness
dampingRatio <dampingRatio>
- set
dampingStiff
to be0.75 * dampingRatio * h^3
or0
ifdampingRatio
is zero or not specified dampingRatio
must be greater than or equal to zero
- set
time <num-seconds> <time-step>
- total time in seconds, 5 by default
- time-step, h, in seconds, 0.025 by default
density <density>
- density of the mesh in Kg/m³, 1000 by default
stiffness <young-modulus> <poisson-ratio>
- Young's modulus (E) in Pa, 1e5 by default
- Poisson's ratio (ν), 0.4 by default, must be inside
[0, 0.5)
turnOffGravity
- no gravity if specified
shapes input <num-of-shapes>
- number of input shapes followed by that number of lines specifying the shape paths and initial configuration in model coordinates:
<mesh-file-path> <x> <y> <z> <rot-deg-x> <rot-deg-y> <rot-deg-z> <scale-x> <scale-y> <scale-z> <extra-command> ...
<extra-command>
: assigning different materials, initial velocities, boundary conditions, or scripted motions for each shape individually if anymaterial <density> <young-modulus> <poisson-ratio>
initVel <lvx> <lvy> <lvz> <avx> <avy> <avz>
DBC <left-bottom-back-x> <left-bottom-back-y> <left-bottom-back-z> <right-top-front-x> <right-top-front-y> <right-top-front-z> <lvx> <lvy> <lvz> <avx> <avy> <avz>
NBC <left-bottom-back-x> <left-bottom-back-y> <left-bottom-back-z> <right-top-front-x> <right-top-front-y> <right-top-front-z> <ax> <ay> <az>
linearVelocity <lvx> <lvy> <lvz>
angularVelocity <avx> <avy> <avz>
ground <friction-coefficient> <height>
halfSpace <x> <y> <z> <nx> <ny> <nz> <stiffness> <friction-coefficient>
- create a half-space analytical collision object centered at (x, y, z) with normal (nx, ny, nz).
stiffness
is unused
selfCollisionOff
- self-collision is not handled if specified
selfFric <selfFricCoeff>
: specify the friction coefficient of self-contact, by default0
(no friction)selfFricCoeff
must be larger or equal to0
, setting it larger than1
is allowed
dHat <dHat>
: set the distance that IPC sees objects as attaching and exert contact forces, by default it is1e-3
that of the diagonal length of the bounding box of the scenedHat
must be larger than0
epsv <epsv>
: set the relative tangent velocity that IPC sees the attached objects as not sliding and exert static friction forces, by default it is1e-3
that of the diagonal length of the bounding box of the scene per secondepsv
must be larger than0
fricIterAmt <fricIterAmt>
: set the maximal friction iterations that IPC updates friction tangent and normal forces, by default it is1
- if set to
0
or negative, IPC will update friction tangent and normal forces until convergence (not guaranteed so not recommended)
- if set to
tol positiveInteger
- number of time steps to set for Newton tolerance other than default 1e-2 followed by that number of positiveRealNumber
positiveInteger
is also a relative value on the bounding box diagonal length
constraintSolver {interiorPoint | QP | SQP}
- collision constraint solver
interiorPoint
: IPC (default)QP
: solve elasticity as a single QP with nonlinear contact constraintsSQP
: solve elasticity as a sequence of quadratic programs with nonlinear contact constraints
restart <status-file-path-str>
- restart simulation from the time step in the specified status file (other settings must be the same)
view {orthographic | perspective}
zoom <positive-real-number>
cameraTracking
- automatically place the focus of the camera at the center of the bounding box of the scene at each frame if specified
playBackSpeed <playBackSpeed>
- playback speed of
anim.gif
, which also controls how many frame data will be output (smaller value results in more output (at most output every frame))
- playback speed of
appendStr <str>
- append result folder with the string specified
disableCout
- no command line output if specified
section {interiorPoint | QP | penalty}
- section of settings only used when the specified constraint solver is used
- note: the constraint solver must be declared before the section
- a section ends when the line
section end
is encountered or at the end of a file. - example:
constraintSolver QP # ... section QP constraintType graphics section end
Some of our settings are legacy settings or the ones that are not usually used/recommended. We include them here for completeness as they have been used by our paper examples or tests.
The following settings only affect the QP and SQP contact constraint solvers we use for comparisons.
QPSolver {OSQP | Gurobi}
constraintType {volume | graphics | gapFunction | nonsmoothNewmark | CMR | Verschoor}
- which type of collision constraint to use with the SQP solver (default:
volume
) volume
: measure the volume of a tetrahedron comprised of the colliding pointsgraphics
: distance constraint from Harmon et al. [2008].nonsmoothNewmark
: variation of volume constraints proposed by Kane et al. [1999].gapFunction
: signed distance along the normal direction from the point of intersection.CMR
: constraint manifold refinement (same as standard graphics approach).Verschoor
: Variant of standard graphics approach by Verschoor and Jalba [2019].
- which type of collision constraint to use with the SQP solver (default:
constraintOffset <offset>
- surface thickness used for mesh collision objects when using
constraintSolver QP
/SQP
- surface thickness used for mesh collision objects when using
useActiveSetConvergence
- Optimization can only converge if the active set converged to a fixed set of constraints or no constraints are active.
We conducted a lengthy comparison of our method and several other commercial and open-source software, and "vanilla" implementations of other methods. You can read the full details in Supplement B.
You can find the scenes, meshes, and other files used in these comparisons in
input/paperExamples/supplementB
.
SQPBenchmark
: The collection of scene files used to generate the benchmark of QP and SQP methods.- The all benchmark scenes can be run with default settings using
tools/run-comparison-benchmark-sqp.sh
- The full benchmark sweep can be run using
tools/run-comparison-benchmark-hpc.sh
- The all benchmark scenes can be run with default settings using
COMSOL
: The 2D COMSOL files used in our comparisonHoudini
: The Houdini files used to generate the chain comparisonSOFA
: The SOFA scene files used to compare against SOFAUtopia
: The mesh and IPC script used to compare against Utopia [Krause and Zulian 2016]
The tests
directory contains a very limited number of unit-tests. These tests
use Catch2 which is downloaded through
CMake unless the option -DIPC_WITH_TESTS=OFF
is provided. Currently, these
test contain test for the CCD (both exact and inexact) and the collision
constraints used by the SQP solver.
The tools
directory contains several tools for generating meshes, running
scenes, and processing results.
convert_to_ipc_msh.py
: this script uses PyMesh to convert a tetrahedral mesh to the expected.msh
format of IPC- IPC uses a nonstandard
.msh
format that includes additional surface information - If the input mesh is a triangle mesh, then Tetgen will be used to tetrahedralize it
- IPC uses a nonstandard
geo_to_msh.py
: this script uses PyMesh to convert a GEO (format exported by Houdini) format mesh to a standard MSH mesh- The output is the standard MSH format and is not directly usable by IPC
generate_chain_scene.py
: generate a IPC script for a variable length chainrun-comparison-benchmark-sqp.sh
: run the comparison SQP benchmark with default settingsrun-comparison-benchmark-hpc.sh
: run the comparison full sweep of the SQP benchmark- dispatches jobs by calling the
hpc-job.sh
script
- dispatches jobs by calling the
process_*_results.py
: process the benchmark results, generating a CSV of the table