ReLU#

class gurobi_ml.modeling.neuralnet.activations.ReLU#

Bases: object

Class to apply the ReLU activation on a neural network layer.

Parameters:
  • setbounds (Bool) – Optional flag not to set bounds on the output variables.

  • bigm (Float) – Optional maximal value for bounds use in the formulation

setbounds#

Optional flag not to set bounds on the output variables.

Type:

Bool

bigm#

Optional maximal value for bounds use in the formulation

Type:

Float

Methods

mip_model(layer)

MIP model for ReLU activation on a layer.

mip_model(layer)#

MIP model for ReLU activation on a layer.

Parameters:

layer (AbstractNNLayer) – Layer to which activation is applied.