Surrogate Models

Some industrial applications require modeling complex processes that can result either in highly nonlinear functions or functions defined by a simulation process. In those contexts, optimization solvers often struggle. The reason may be that relaxations of the nonlinear functions are not good enough to make the solver prove an acceptable bound in a reasonable amount of time. Another issue may be that the solver is not able to represent the functions.

An approach that has been proposed in the literature is to approximate the problematic nonlinear functions via neural networks with ReLU activation and use MIP technology to solve the constructed approximation (see e.g. Heneao Maravelias 2011, Schweitdmann et.al. 2022). This use of neural networks can be motivated by their ability to provide a universal approximation (see e.g. Lu et.al. 2017). This use of ML models to replace complex processes is often referred to as surrogate models.

In the following example, we approximate a nonlinear function via Scikit-learn MLPRegressor and then solve an optimization problem that uses the approximation of the nonlinear function with Gurobi.

The purpose of this example is solely illustrative and doesn’t relate to any particular application.

The function we approximate is the 2D peaks function.

The function is given as

\[\begin{split} \begin{aligned} f(x) = & 3 \cdot (1-x_1)^2 \cdot \exp(-x_1^2 - (x_2+1)^2) - \\ & 10 \cdot (\frac{x_1}{5} - x_1^3 - x_2^5) \cdot \exp(-x_1^2 - x_2^2) - \\ & \frac{1}{3} \cdot \exp(-(x_1+1)^2 - x_2^2). \end{aligned}\end{split}\]

In this example, we want to find the minimum of \(f\) over the interval \([-2, 2]^2\):

\[y = \min \{f(x) : x \in [-2,2]^2\}.\]

The global minimum of this problem can be found numerically to have value \(-6.55113\) at the point \((0.2283, -1.6256)\).

Here to find this minimum of \(f\), we approximate \(f(x)\) through a neural network function \(g(x)\) to obtain a MIP and solve

\[\hat y = \min \{g(x) : x \in [-2,2]^2\} \approx y.\]

First import the necessary packages. Before applying the neural network, we do a preprocessing to extract polynomial features of degree 2. Hopefully this will help us to approximate the smooth function. Besides, gurobipy, numpy and the appropriate sklearn objects, we also use matplotlib to plot the function, and its approximation.

import gurobipy as gp
import numpy as np
from gurobipy import GRB
from matplotlib import cm
from matplotlib import pyplot as plt
from sklearn import metrics
from sklearn.neural_network import MLPRegressor
from sklearn.pipeline import make_pipeline
from sklearn.preprocessing import PolynomialFeatures

from gurobi_ml import add_predictor_constr

Define the nonlinear function of interest

We define the 2D peak function as a python function.

def peak2d(x1, x2):
    return (
        3 * (1 - x1) ** 2.0 * np.exp(-(x1**2) - (x2 + 1) ** 2)
        - 10 * (x1 / 5 - x1**3 - x2**5) * np.exp(-(x1**2) - x2**2)
        - 1 / 3 * np.exp(-((x1 + 1) ** 2) - x2**2)
    )

To train the neural network, we make a uniform sample of the domain of the function in the region of interest using numpy’s arrange function.

We then plot the function with matplotlib.

x1, x2 = np.meshgrid(np.arange(-2, 2, 0.01), np.arange(-2, 2, 0.01))
y = peak2d(x1, x2)

fig, ax = plt.subplots(subplot_kw={"projection": "3d"})
# Plot the surface.
surf = ax.plot_surface(x1, x2, y, cmap=cm.coolwarm, linewidth=0.01, antialiased=False)
# Add a color bar which maps values to colors.
fig.colorbar(surf, shrink=0.5, aspect=5)
plt.show()
example1 2DPeakFunction

Approximate the function

To fit a model, we need to reshape our data. We concatenate the values of x1 and x2 in an array X and make y one dimensional.

X = np.concatenate([x1.ravel().reshape(-1, 1), x2.ravel().reshape(-1, 1)], axis=1)
y = y.ravel()

To approximate the function, we use a Pipeline with polynomial features and a neural-network regressor. We do a relatively small neural-network.

# Run our regression
layers = [30] * 2
regression = MLPRegressor(hidden_layer_sizes=layers, activation="relu")
pipe = make_pipeline(PolynomialFeatures(), regression)
pipe.fit(X=X, y=y)
Pipeline(steps=[('polynomialfeatures', PolynomialFeatures()),
                ('mlpregressor', MLPRegressor(hidden_layer_sizes=[30, 30]))])
In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook.
On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.


To test the accuracy of the approximation, we take a random sample of points, and we print the \(R^2\) value and the maximal error.

X_test = np.random.random((100, 2)) * 4 - 2

r2_score = metrics.r2_score(peak2d(X_test[:, 0], X_test[:, 1]), pipe.predict(X_test))
max_error = metrics.max_error(peak2d(X_test[:, 0], X_test[:, 1]), pipe.predict(X_test))
print("R2 error {}, maximal error {}".format(r2_score, max_error))
R2 error 0.9998672237349817, maximal error 0.06802886791722429

While the \(R^2\) value is good, the maximal error is quite high. For the purpose of this example we still deem it acceptable. We plot the function.

fig, ax = plt.subplots(subplot_kw={"projection": "3d"})
# Plot the surface.
surf = ax.plot_surface(
    x1,
    x2,
    pipe.predict(X).reshape(x1.shape),
    cmap=cm.coolwarm,
    linewidth=0.01,
    antialiased=False,
)
# Add a color bar which maps values to colors.
fig.colorbar(surf, shrink=0.5, aspect=5)
plt.show()
example1 2DPeakFunction

Visually, the approximation looks close enough to the original function.

Build and Solve the Optimization Model

We now turn to the optimization model. For this model we want to find the minimal value of y_approx which is the approximation given by our pipeline on the interval.

Note that in this simple example, we don’t use matrix variables but regular Gurobi variables instead.

m = gp.Model()

x = m.addVars(2, lb=-2, ub=2, name="x")
y_approx = m.addVar(lb=-GRB.INFINITY, name="y")

m.setObjective(y_approx, gp.GRB.MINIMIZE)

# add "surrogate constraint"
pred_constr = add_predictor_constr(m, pipe, x, y_approx)

pred_constr.print_stats()
Restricted license - for non-production use only - expires 2026-11-23
Warning for adding constraints: zero or small (< 1e-13) coefficients, ignored
Model for pipe:
126 variables
61 constraints
6 quadratic constraints
60 general constraints
Input has shape (1, 2)
Output has shape (1, 1)

Pipeline has 2 steps:

--------------------------------------------------------------------------------
Step            Output Shape    Variables              Constraints
                                                Linear    Quadratic      General
================================================================================
poly_feat             (1, 6)            6            0            6            0

dense                (1, 30)           60           30            0           30 (relu)

dense0               (1, 30)           60           30            0           30 (relu)

dense1                (1, 1)            0            1            0            0


--------------------------------------------------------------------------------

Now call optimize. Since we use polynomial features the resulting model is a non-convex quadratic problem. In Gurobi, we need to set the parameter NonConvex to 2 to be able to solve it.

m.Params.TimeLimit = 20
m.Params.MIPGap = 0.1
m.Params.NonConvex = 2

m.optimize()
Set parameter TimeLimit to value 20
Set parameter MIPGap to value 0.1
Set parameter NonConvex to value 2
Gurobi Optimizer version 12.0.2 build v12.0.2rc0 (linux64 - "Ubuntu 24.04 LTS")

CPU model: AMD EPYC 9R14, instruction set [SSE2|AVX|AVX2|AVX512]
Thread count: 2 physical cores, 2 logical processors, using up to 2 threads

Non-default parameters:
TimeLimit  20
MIPGap  0.1
NonConvex  2

Optimize a model with 61 rows, 129 columns and 990 nonzeros
Model fingerprint: 0x1ed78f4d
Model has 6 quadratic constraints
Model has 60 simple general constraints
  60 MAX
Variable types: 129 continuous, 0 integer (0 binary)
Coefficient statistics:
  Matrix range     [2e-12, 2e+00]
  QMatrix range    [1e+00, 1e+00]
  QLMatrix range   [1e+00, 1e+00]
  Objective range  [1e+00, 1e+00]
  Bounds range     [2e+00, 2e+00]
  RHS range        [1e-02, 1e+00]
  QRHS range       [1e+00, 1e+00]
Presolve added 77 rows and 14 columns
Presolve time: 0.00s
Presolved: 148 rows, 144 columns, 1149 nonzeros
Presolved model has 3 bilinear constraint(s)

Solving non-convex MIQCP

Variable types: 102 continuous, 42 integer (42 binary)

Root relaxation: objective -5.601034e+01, 161 iterations, 0.00 seconds (0.00 work units)

    Nodes    |    Current Node    |     Objective Bounds      |     Work
 Expl Unexpl |  Obj  Depth IntInf | Incumbent    BestBd   Gap | It/Node Time

     0     0  -56.01034    0   28          -  -56.01034      -     -    0s
     0     0  -49.23118    0   32          -  -49.23118      -     -    0s
     0     0  -49.13745    0   31          -  -49.13745      -     -    0s
     0     0  -48.56884    0   31          -  -48.56884      -     -    0s
     0     0  -42.03167    0   38          -  -42.03167      -     -    0s
     0     0  -41.36273    0   37          -  -41.36273      -     -    0s
     0     0  -40.64429    0   37          -  -40.64429      -     -    0s
     0     0  -40.37376    0   39          -  -40.37376      -     -    0s
     0     0  -38.94292    0   40          -  -38.94292      -     -    0s
     0     0  -38.89711    0   39          -  -38.89711      -     -    0s
     0     0  -38.45086    0   37          -  -38.45086      -     -    0s
     0     0  -38.37159    0   38          -  -38.37159      -     -    0s
     0     0  -38.22795    0   39          -  -38.22795      -     -    0s
     0     0  -38.19175    0   39          -  -38.19175      -     -    0s
     0     0  -38.12088    0   39          -  -38.12088      -     -    0s
     0     0  -38.09255    0   39          -  -38.09255      -     -    0s
H    0     0                       0.7874787  -38.09255  4937%     -    0s
H    0     0                       0.2182125  -38.09255      -     -    0s
     0     2  -38.09255    0   39    0.21821  -38.09255      -     -    0s
H   64    62                       0.1058747  -35.42299      -  27.6    0s
H   99    96                       0.1058743  -33.36613      -  23.8    0s
H  109    96                       0.1058742  -33.36613      -  23.3    0s
*  195   136              34      -2.9224191  -31.37789   974%  18.2    0s
H  540   357                      -2.9224215  -25.28455   765%  15.1    0s
*  847   431              40      -3.9061980  -22.24559   469%  14.5    0s
*  850   431              42      -3.9078901  -22.24559   469%  14.4    0s
*  852   431              43      -3.9083528  -22.24559   469%  14.4    0s
*  853   431              43      -3.9084045  -22.24559   469%  14.4    0s
H  881   448                      -3.9084048  -22.24559   469%  14.1    0s
H  994   422                      -5.9899582  -20.79698   247%  14.3    0s
H 1430   428                      -6.5148077  -17.02139   161%  14.6    0s
H 2447   445                      -6.5585199  -13.70670   109%  14.6    0s
H 3941   226                      -6.5587535  -10.25950  56.4%  14.9    1s
* 3946   226              39      -6.5587688  -10.25950  56.4%  14.9    1s

Cutting planes:
  Gomory: 3
  Cover: 1
  Implied bound: 37
  MIR: 107
  Flow cover: 34
  RLT: 4

Explored 4395 nodes (65034 simplex iterations) in 1.11 seconds (1.77 work units)
Thread count was 2 (of 2 available processors)

Solution count 10: -6.55877 -6.55875 -6.55852 ... -2.92242
No other solutions better than -6.55877

Optimal solution found (tolerance 1.00e-01)
Best objective -6.558768798169e+00, best bound -6.558768798169e+00, gap 0.0000%

After solving the model, we check the error in the estimate of the Gurobi solution.

print(
    "Maximum error in approximating the regression {:.6}".format(
        np.max(pred_constr.get_error())
    )
)
Maximum error in approximating the regression 8.89795e-07

Finally, we look at the solution and the objective value found.

print(
    f"solution point of the approximated problem ({x[0].X:.4}, {x[1].X:.4}), "
    + f"objective value {m.ObjVal}."
)
print(
    f"Function value at the solution point {peak2d(x[0].X, x[1].X)} error {abs(peak2d(x[0].X, x[1].X) - m.ObjVal)}."
)
solution point of the approximated problem (0.2477, -1.665), objective value -6.558768798169111.
Function value at the solution point -6.52256823452682 error 0.03620056364229107.

The difference between the function and the approximation at the computed solution point is noticeable, but the point we found is reasonably close to the actual global minimum. Depending on the use case this might be deemed acceptable. Of course, training a larger network should result in a better approximation.

Copyright © 2023 Gurobi Optimization, LLC

Total running time of the script: (0 minutes 6.820 seconds)

Gallery generated by Sphinx-Gallery