IPOPT
IPOPT (Interior Point OPTimizer) is an open source interior point optimizer, designed for large-scale nonlinear optimization. The source code can be found here. The latest version we support is 3.13.2.
Installation
IPOPT must be installed separately, then linked to pyOptSparse when building. For the full installation instructions, please see their documentation. OpenMDAO also has a very helpful script which can be used to install IPOPT with other linear solvers. Here we explain a basic setup using MUMPS as the linear solver, together with METIS adapted from the OpenMDAO script.
Download the tarball and extract it to
$IPOPT_DIR
which could be set to for example$HOME/packages/Ipopt
.Install METIS, which can be used to improve the performance of the MUMPS linear solver.
# build METIS cd $IPOPT_DIR git clone https://github.com/coin-or-tools/ThirdParty-Metis.git cd ThirdParty-Metis ./get.Metis ./configure --prefix=$IPOPT_DIR make make install
Install MUMPS
# build MUMPS cd $IPOPT_DIR git clone https://github.com/coin-or-tools/ThirdParty-Mumps.git cd ThirdParty-Mumps ./get.Mumps ./configure --with-metis --with-metis-lflags="-L${IPOPT_DIR}/lib -lcoinmetis" \ --with-metis-cflags="-I${IPOPT_DIR}/include -I${IPOPT_DIR}/include/coin-or -I${IPOPT_DIR}/include/coin-or/metis" \ --prefix=$IPOPT_DIR CFLAGS="-I${IPOPT_DIR}/include -I${IPOPT_DIR}/include/coin-or -I${IPOPT_DIR}/include/coin-or/metis" \ FCFLAGS="-I${IPOPT_DIR}/include -I${IPOPT_DIR}/include/coin-or -I${IPOPT_DIR}/include/coin-or/metis" make make install
Build IPOPT
# build IPOPT cd $IPOPT_DIR mkdir build cd build ../configure --prefix=${IPOPT_DIR} --disable-java --with-mumps --with-mumps-lflags="-L${IPOPT_DIR}/lib -lcoinmumps" \ --with-mumps-cflags="-I${IPOPT_DIR}/include/coin-or/mumps" make make install
You must add the IPOPT library path to the
LD_LIBRARY_PATH
variable for things to work right. This could be done for example by adding the following to your.bashrc
:export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$IPOPT_DIR/lib
Furthermore, the environment variable
$IPOPT_DIR
must be set correctly in order to link to pyOptSparse. Alternatively, you can manually define the variables$IPOPT_LIB
and$IPOPT_INC
for the lib and include paths separately.Now clean build pyOptSparse. Verify that IPOPT works by running the relevant tests.
Options
Please refer to the IPOPT website for complete listing of options. The following are the options which are set by default within pyOptSparse. All other options take the default value with IPOPT unless specified by the user.
Name |
Type |
Default value |
Description |
---|---|---|---|
|
int |
0 |
Printing level |
|
int |
5 |
Printing level for the output file |
|
str |
|
This is an undocumented option which suppresses the IPOPT header from being printed to screen every time. |
|
str |
|
Whether to print the user-modified options |
|
str |
|
The name of the output file from IPOPT |
|
str |
|
The linear solver used. |
Informs
Code |
Description |
---|---|
|
Solve Succeeded |
|
Solved To Acceptable Level |
|
Infeasible Problem Detected |
|
Search Direction Becomes Too Small |
|
Diverging Iterates |
|
User Requested Stop |
|
Feasible Point Found |
|
Maximum Iterations Exceeded |
|
Restoration Failed |
|
Error In Step Computation |
|
Maximum CpuTime Exceeded |
|
Not Enough Degrees Of Freedom |
|
Invalid Problem Definition |
|
Invalid Option |
|
Invalid Number Detected |
|
Unrecoverable Exception |
|
NonIpopt Exception Thrown |
|
Insufficient Memory |
|
Internal Error |
API
- class pyoptsparse.pyIPOPT.pyIPOPT.IPOPT(*args, **kwargs)[source]
IPOPT Optimizer Class - Inherited from Optimizer Abstract Class
IPOPT Optimizer Class Initialization
- __call__(optProb, sens=None, sensStep=None, sensMode=None, storeHistory=None, hotStart=None, storeSens=True)[source]
This is the main routine used to solve the optimization problem.
- Parameters:
- optProbOptimization or Solution class instance
This is the complete description of the optimization problem to be solved by the optimizer
- sensstr or python Function.
Specifiy method to compute sensitivities. To explictly use pyOptSparse gradient class to do the derivatives with finite differenes use ‘FD’. ‘sens’ may also be ‘CS’ which will cause pyOptSpare to compute the derivatives using the complex step method. Finally, ‘sens’ may be a python function handle which is expected to compute the sensitivities directly. For expensive function evaluations and/or problems with large numbers of design variables this is the preferred method.
- sensStepfloat
Set the step size to use for design variables. Defaults to 1e-6 when sens is ‘FD’ and 1e-40j when sens is ‘CS’.
- sensModestr
Use ‘pgc’ for parallel gradient computations. Only available with mpi4py and each objective evaluation is otherwise serial
- storeHistorystr
File name of the history file into which the history of this optimization will be stored
- hotStartstr
File name of the history file to “replay” for the optimziation. The optimization problem used to generate the history file specified in ‘hotStart’ must be IDENTICAL to the currently supplied ‘optProb’. By identical we mean, EVERY SINGLE PARAMETER MUST BE IDENTICAL. As soon as he requested evaluation point does not match the history, function and gradient evaluations revert back to normal evaluations.
- storeSensbool
Flag sepcifying if sensitivities are to be stored in hist. This is necessay for hot-starting only.