diff --git a/docs/Dev/contributing/index.html b/docs/Dev/contributing/index.html deleted file mode 100644 index bea0f9eb..00000000 --- a/docs/Dev/contributing/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -How to Contribute · EAGO.jl: Easy Advanced Global Optimization

How to Contribute

How to Contribute

We're always happy to welcome work with additional collaborators and contributors. One of the easy ways for newcomers to contribute is by adding additional relaxations.

If you're interested in contributing in larger ways, please contact: Matthew Wilhelm

If you have any requests for additional functionality, bug fixes, or comments please feel free to open a new issue using Github issue tracker.

diff --git a/docs/Dev/future/index.html b/docs/Dev/future/index.html deleted file mode 100644 index c1dafc62..00000000 --- a/docs/Dev/future/index.html +++ /dev/null @@ -1,104 +0,0 @@ - - - - - - - Future Work · EAGO.jl: Easy Advanced Global Optimization - - - - - - - - - - - - - -
-
- -
-
Future Work
-
-

Future Work

-

Current Activity:

- - -

Other things on the wishlist (but not actively being worked on):

- - - - - -
- - - diff --git a/docs/McCormick/Figure_1.png b/docs/McCormick/Figure_1.png deleted file mode 100644 index b27217f2..00000000 Binary files a/docs/McCormick/Figure_1.png and /dev/null differ diff --git a/docs/McCormick/Figure_2.png b/docs/McCormick/Figure_2.png deleted file mode 100644 index 5b56c3be..00000000 Binary files a/docs/McCormick/Figure_2.png and /dev/null differ diff --git a/docs/McCormick/Figure_3.png b/docs/McCormick/Figure_3.png deleted file mode 100644 index ff43796e..00000000 Binary files a/docs/McCormick/Figure_3.png and /dev/null differ diff --git a/docs/McCormick/implicit/index.html b/docs/McCormick/implicit/index.html deleted file mode 100644 index adef1e28..00000000 --- a/docs/McCormick/implicit/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -Relaxation of Implicit Functions · EAGO.jl: Easy Advanced Global Optimization

Relaxation of Implicit Functions

Relaxation of Implicit Functions

High-level functions

implicit_relax_h!

Relaxes the implicit function determined by h(x,p) with x in X and p in P. The reference point for the affine relaxations is pmid. The parameters generated from the relaxation at pmid are param and the basic parameters of the fixed point method are mc_opt.

implicit_relax_fg

Relaxes the functions f(x,p) and g(x,p) by relaxation the state variable x using the implicit function determined by h(x,p) with x in X and p in P. The reference point for the affine relaxations is pmid. The parameters generated from the relaxation at pmid are param and the basic parameters of the fixed point method are mc_opt.

implicit_relax_f

Relaxes the function f(x,p) by relaxation the state variable x using the implicit function determined by h(x,p) with x in X and p in P. The reference point for the affine relaxations is pmid. The parameters generated from the relaxation at pmid are param and the basic parameters of the fixed point method are mc_opt.

Subroutines

affine_exp!

Computates the affine relaxations of the state variable. Inputs are:

  • x: State variable relaxation
  • p::Vector{MC{N,T}}: Decision variable relaxation
  • p_ref::Vector{MC{N,T}}: Reference variable relaxation
  • xa::Vector{MC{N,T}}: Lower affine relaxation of the state variable
  • xA::Vector{MC{N,T}}: Upper affine relaxation of the state variable
  • z::Vector{MC{N,T}}: Affine function in X
  • nx: Number of state variables
  • lambda: Convex coefficient used to define z as a function of xa, xA

Populats the following vectors with results (xa,xA,z):

  • xa::Vector{MC{N,T}}: Lower affine relaxation of the state variable
  • xA::Vector{MC{N,T}}: Upper affine relaxation of the state variable
  • z::Vector{MC{N,T}}: Affine function in X

correct_exp!

Corrects the relaxation of the state variable x_mc if the affine relaxation, 'z_mc', exceeds the interval bounds xL or xU.

  • z_mc::Vector{MC{N}}: Affine relaxation
  • x_mc::Vector{MC{N}}: Relaxation of state variable
  • X::Vector{IntervalType}: Lower bound on state vector
  • nx::Int64: Size of the state vector
  • epsv::Float64: Tolerance for checking that subgradient exceeds bound
final_cut

An operator that cuts the x_mc object using the x_mc_int bounds in a differentiable or nonsmooth fashion to achieve a composite relaxation within x_mc_int.

gen_expansion_params!

Generates the relaxations at pref_mc that can be subsequentially used to compute affine relaxations for use in relaxing the implicit function.

mc_dense_krawczyk_cw!

Performs a single step of the dense-krawczyk componentwise parametric method after the inputs have been preconditioned.

mc_dense_newton_gs!

Performs a single step of the dense-newton gauss-siedel parametric method after the inputs have been preconditioned.

pmc_kernel!

Peforms the following steps in sequence:

  • Evaluates the function h!(H, x, xp, p, t) in place with x = z_mc, p = p_mc,

xp = xp_mc, t = flt_param and preconditions H using an interval midpoint preconditioner if precond = true.

  • Evaluates the function hj!(J, x, xp, p, t) in place with x = aff_mc, p = p_mc,

xp = xp_mc, t = flt_param and preconditions J using an interval midpoint preconditioner if precond = true.

  • Lastly, applies a Newton-type contractor method. The parametric GS Newton

contractor if cntr = :Newton and the componentwise Krawczyk contractor otherwise.

diff --git a/docs/McCormick/operators/index.html b/docs/McCormick/operators/index.html deleted file mode 100644 index 54bb5cd3..00000000 --- a/docs/McCormick/operators/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -Currently supported operators · EAGO.jl: Easy Advanced Global Optimization

Currently supported operators

Currently supported operators

The operators currently supported are listed below. The operators with a check box have been subject to a large degree of scrutiny and have been implemented for both forward and reverse McCormick relaxations.

Univariate McCormick Operators

Arbitrarily differentiable relaxations can be constructed for the following operators:

Both nonsmooth and Whitney-1 (once differentiable) relaxations are supported:

Bivariate Operators: McCormick & McCormick

The following bivariant operators are supported for two MC objects. Both nonsmooth and Whitney-1 (once differentiable) relaxations are supported.

Arbitrarily differentiable relaxations can be constructed for the following operators:

Bivariate Operators: McCormick & (Integer or Float)

Arbitrarily differentiable relaxations can be constructed for the following operators:

diff --git a/docs/McCormick/overview/index.html b/docs/McCormick/overview/index.html deleted file mode 100644 index a6ff2556..00000000 --- a/docs/McCormick/overview/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -Overview · EAGO.jl: Easy Advanced Global Optimization

Overview

Overview

EAGO provides a library of McCormick relaxations in native Julia code. It supports relaxing functions using both nonsmooth McCormick relaxations (Mitsos2009), smooth McCormick relaxations (Khan2017), multi-variant McCormick relaxations (Tsoukalas2014), as well as subgradient-based interval refinement (Najman2017). For functions with arbitrarily differentiable relaxations, the differentiable can be modified by adjusting a constant value. Additionally, and nonvalidated validated interval bounds are supported via ValidatedNumerics.jl.

diff --git a/docs/McCormick/type/index.html b/docs/McCormick/type/index.html deleted file mode 100644 index 932c3dfe..00000000 --- a/docs/McCormick/type/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -Types · EAGO.jl: Easy Advanced Global Optimization

Types

Types

MC

MC{N, T <: RelaxTag} <: Real is the McCormick (w/ (sub)gradient) structure which is used to overload standard calculations. The fields are:

  • cc::Float64: Concave relaxation
  • cv::Float64: Convex relaxation
  • cc_grad::SVector{N,Float64}: (Sub)gradient of concave relaxation
  • cv_grad::SVector{N,Float64}: (Sub)gradient of convex relaxation
  • Intv::Interval{Float64}: Interval bounds
  • cnst::Bool: Flag for whether the bounds are constant
RelaxTag

An abstract type the subtypes of which define the manner of relaxation that will be performed for each operator applied to the MC object. Currently, the struct NS which specifies that standard (Mitsos 2009) are to be used is fully supported. Limited support is provided for differentiable McCormick relaxations specified by struct Diff (Khan 2017) and struct MV struct MV (Tsoukalas 2011.)

Constructors for MC

EAGO.McCormick.MCMethod.
MC{N,T}(y::Interval{Float64})

Constructs McCormick relaxation with convex relaxation equal to y.lo and concave relaxation equal to y.hi.

MC{N,T}(y::Float64)

Constructs McCormick relaxation with convex relaxation equal to y and concave relaxation equal to y.

MC{N,T}(cv::Float64, cc::Float64)

Constructs McCormick relaxation with convex relaxation equal to cv and concave relaxation equal to cc.

MC{N,T}(val::Float64, Intv::Interval{Float64}, i::Int64)

Constructs McCormick relaxation with convex relaxation equal to val, concave relaxation equal to val, interval bounds of Intv, and a unit subgradient with nonzero's ith dimension of length N.

diff --git a/docs/McCormick/usage/index.html b/docs/McCormick/usage/index.html deleted file mode 100644 index b30075dd..00000000 --- a/docs/McCormick/usage/index.html +++ /dev/null @@ -1,36 +0,0 @@ - -Basic Usage · EAGO.jl: Easy Advanced Global Optimization

Basic Usage

Basic Usage

Bounding a function via smooth McCormick objects

In order to bound a function using a McCormick relaxation. You first construct structure that bounds the input variables then you construct pass these variables two a function.

In the example below, convex/concave relaxations of the function f(x)=sin(2x)+exp(x)-x are calculated at x = 1 on the interval [-2,3].

using EAGO, IntervalArithmetic
-
-# create MC object for x = 2.0 on [1.0,3.0] for relaxing
-# a function f(x) on the interval Intv
-
-f(x) = x*(x-5.0)*sin(x)
-
-x = 2.0                          # value of independent variable x
-Intv = Interval(1.0,4.0)         # define interval to relax over
-
-# create McCormick object
-xMC = MC{1,NS}(x,Intv,1)
-
-fMC = f(xMC)             # relax the function
-
-cv = fMC.cv              # convex relaxation
-cc = fMC.cc              # concave relaxation
-cvgrad = fMC.cv_grad     # subgradient/gradient of convex relaxation
-ccgrad = fMC.cc_grad     # subgradient/gradient of concave relaxation
-Iv = fMC.Intv           # retrieve interval bounds of f(x) on Intv

The plotting the results we can easily generate visual the convex and concave relaxations, interval bounds, and affine bounds constructed using the subgradient at the middle of X.

Figure_1

If we instead use the constructor xMC = MC{1,Diff}(x,Intv,1) in the above code, and then re-plot we arrive at the below graph

set_diff_relax!(1)

Figure_2

This can readily be extended to multivariate functions as shown below


-f(x) = max(x[1],x[2])
-
-x = [2.0 1.0]                                    # values of independent variable x
-Intv = [Interval(-4.0,5.0), Interval(-5.0,3.0)]  # define intervals to relax over
-
-# create McCormick object
-xMC = [MC{2,Diff}(x[i], Intv[i], i) for i=1:2)]
-
-fMC = f(xMC)            # relax the function
-
-cv = fMC.cv              # convex relaxation
-cc = fMC.cc              # concave relaxation
-cvgrad = fMC.cv_grad     # subgradient/gradient of convex relaxation
-ccgrad = fMC.cc_grad     # subgradient/gradient of concave relaxation
-Iv = fMC.Intv            # retrieve interval bounds of f(x) on Intv

Figure_3

diff --git a/docs/Optimizer/bnb_back/index.html b/docs/Optimizer/bnb_back/index.html deleted file mode 100644 index 2aac5009..00000000 --- a/docs/Optimizer/bnb_back/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -EAGO's Branch and Bound Routine · EAGO.jl: Easy Advanced Global Optimization

EAGO's Branch and Bound Routine

EAGO's Branch and Bound Routine

This component is meant to provide a flexible framework for implementing spatial branch-and-bound based optimization routines in Julia. All components of the branch-and-bound routine can be customized by the individual user: lower bounding problem, upper bounding problem.

Branch and Bound Node Storage

EAGO.NodeBBType.
NodeBB

Stores information associated with each node in Branch & Bound tree.

  • lower_variable_bounds::Vector{Float64}: Lower bounds of variable box.
  • upper_variable_bounds::Vector{Float64}: Upper bounds of variable box.
  • lower_bound::Float64: Lower bound of problem solution on nodeBB
  • upper_bound::Float64: Upper bound of problem solution on nodeBB
  • depth::Int64: Depth of node in B&B tree.
  • id::Int64: Unique id for each node.

Customizable subroutines

EAGO.add_cut!Method.
add_cut!(t::ExtensionType, x::Optimizer)

Adds a cut for each constraint and the objective function to the subproblem.

EAGO.branch_node!Method.
branch_node!(t::ExtensionType, x::Optimizer)

Creates two nodes from currentnode using information available the x and stores them to the stack. By default, relative width bisection is perfomed at a point `branchpntwhich is a convex combination (parameter: branch_cvx_factor) of the solution to the relaxation and the midpoint of the node. If this solution lies withinbranchoffset/widthof a bound then the branch point is moved to a distance ofbranchoffset/width` from the bound.

convergence_check(t::ExtensionType, x::Optimizer)

Checks for convergence of algorithm with respect to absolute and/or relative tolerances.

EAGO.cut_conditionMethod.
cut_conditioncut_condition(t::ExtensionType, x::Optimizer)

Checks if a cut should be added and computes a new reference point to add the cut at. If no cut should be added the constraints not modified in place are deleted from the relaxed optimizer and the solution is compared with the interval lower bound. The best lower bound is then used.

EAGO.fathom!Method.
fathom!(t::ExtensionType, d::Optimizer)

Selects and deletes nodes from stack with lower bounds greater than global upper bound.

lower_problem!(t::ExtensionType, x::Optimizer)

Constructs and solves the relaxation using the default EAGO relaxation scheme and optimizer on node y.

relax_objective!(t::ExtensionType, x::Optimizer, x0::Vector{Float64})

A rountine that only relaxes the objective.

relax_problem!(t::ExtensionType, x::Optimizer, v::Vector{Float64}, q::Int64)

A rountine that updates the current node for the Evaluator and relaxes all nonlinear constraints and quadratic constraints.

node_selection!(t::ExtensionType, x::Optimizer)

Selects node with the lowest lower bound in stack.

optimize_hook!(t::ExtensionType, x::Optimizer)

Provides a hook for extensions to EAGO as opposed to standard global, local, or linear solvers.

EAGO.postprocess!Method.
postprocess!(t::ExtensionType, x::Optimizer)

Default postprocess perfoms duality-based bound tightening on the y.

EAGO.preprocess!Method.
preprocess!(t::ExtensionType, x::Optimizer)

Runs interval, linear, quadratic contractor methods followed by obbt and a constraint programming walk up to tolerances specified in EAGO.Optimizer object.

EAGO.repeat_checkMethod.
repeat_check(t::ExtensionType, x::Optimizer)

Checks to see if current node should be reprocessed.

single_storage!(t::ExtensionType, x::Optimizer)

Stores the current node to the stack after updating lower/upper bounds.

termination_check(t::ExtensionType, x::Optimizer)

Checks for termination of algorithm due to satisfying absolute or relative tolerance, infeasibility, or a specified limit, returns a boolean valued true if algorithm should continue.

upper_problem!(t::ExtensionType, x::Optimizer)

Default upper bounding problem which simply calls solve_local_nlp! to solve the nlp locally.

Internal Subroutines

EAGO.cut_updateMethod.
cut_update(x::Optimizer)

Updates the internal storage in the optimizer after a valid feasible cut is added.

default_nlp_heurestic(x::Optimizer, y::NodeBB)

Default check to see if the upper bounding problem should be run. By default, The upper bounding problem is run on every node up to depth upper_bounding_depth and is triggered with a probability of 0.5^(depth - upper_bounding_depth) afterwards.

EAGO.interval_boundFunction.
interval_bound

Computes the natural interval extension of a MathOptInterface function s or ScalarQuadaraticFunction on a node y. Returns the lower bound if flag is true and the upper bound if flag is false.

interval_lower_bound!(x::Optimizer, y::NodeBB)

A fallback lower bounding problem that consists of an natural interval extension calculation. This is called when the optimizer used to compute the lower bound does not return a termination and primal status code indicating that it successfully solved the relaxation to a globally optimal point.

is_globally_optimal(t::MOI.TerminationStatusCode, r::MOI.ResultStatusCode)

Takes an MOI.TerminationStatusCode and a MOI.ResultStatusCode and returns the tuple (valid_result::Bool, feasible::Bool). The value valid_result is true if the pair of codes prove that either the subproblem solution was solved to global optimality or the subproblem solution is infeasible. The value of feasible is true if the problem is feasible and false if the problem is infeasible.

is_feasible_solution(t::MOI.TerminationStatusCode, r::MOI.ResultStatusCode)

Takes an MOI.TerminationStatusCode and a MOI.ResultStatusCode and returns true if this corresponds to a solution that is proven to be feasible. Returns false otherwise.

log_iteration!(x::Optimizer)

If 'loggingon' is true, the 'globallowerbound', 'globalupperbound', 'runtime', and 'nodecount' are stored every 'loginterval'. If 'logsubprobleminfo' then the lower bound, feasibility and run times of the subproblems are logged every 'log_interval'.

EAGO.same_boxMethod.
same_box(x::NodeBB,y::NodeBB, atol::Float64)

Checks that node x and y have equal domains withing a tolerance of atol.

solve_local_nlp!(x::Optimizer)

Constructs and solves the problem locally on on node y updated the upper solution informaton in the optimizer.

EAGO.set_dual!Method.
set_dual!(x::Optimizer)

Retrieves the lower and upper duals for variable bounds from the relaxed_optimizer and sets the appropriate values in the _lower_lvd and _lower_uvd storage fields.

update_relaxed_problem_box!(x::Optimizer, y::NodeBB)

Updates the relaxed constraint by setting the constraint set of v == x*,xLi <= xi, andxi <= xUi` for each such constraint added to the relaxed optimizer.

Functions for generating console output

EAGO.print_iteration!Function.
print_iteration!

Prints the iteration information based on verbosity. The header is displayed every header_interval, the iteration info is displayed every iteration_interval.

EAGO.print_node!Function.
print_node!

Prints node information for the B&B problem. Node id, bound, and interval box.

EAGO.print_results!Function.
print_results!

Prints the results of a single bounding problem.

EAGO.print_solution!Function.
print_solution!

Prints solution information for the B&B problem. Displays first node found, solution value, solution, and time spent solving subproblems.

diff --git a/docs/Optimizer/domain_reduction/index.html b/docs/Optimizer/domain_reduction/index.html deleted file mode 100644 index ea12f2fc..00000000 --- a/docs/Optimizer/domain_reduction/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -Domain Reduction · EAGO.jl: Easy Advanced Global Optimization

Domain Reduction

Domain Reduction

Duality-Based Bound Tightening

Variable bound tightening based on the duality multipliers are supported.

EAGO.variable_dbbt!Function.
variable_dbbt!

Tightens the bounds of the _current_node using the current global upper bound and the duality information obtained from the relaxation.

Special Forms

Bound tightening for linear forms, univariate quadratic forms, and bivariate quadratic forms are also supported.

classify_quadratics!

Classifies constraints as univariate or bivariate and adds them to storage vectors.

EAGO.lp_bound_tightenFunction.
lp_bound_tighten

Performs the linear bound tightening.

univariate_kernel

Kernel of the bound tightening operation on univariant qudaratic functions. Called for each univariate function.

univariate_quadratic

Performs bound tightening on all univariate quadratic functions.

Constraint Propagation

EAGO contains a constraint propagation architecture that supported forward and reverse evaluation of set-valued functions on the directed acyclic graph (DAG). The interval contractor and reverse McCormick relaxation-based contractors are currently available.

EAGO.cpwalkFunction.
cpwalk

Performs forward-reverse pass on directed graph as part of constraint propagation.

Optimization-Based Bound Tightening

EAGO makes use of an optimization-based bound tightening scheme using filtering and greedy ordering as detailed in: Gleixner, A.M., Berthold, T., Müller, B. et al. J Glob Optim (2017) 67: 731. https://doi.org/10.1007/s10898-016-0450-4

aggressive_filtering!

Excludes OBBT on variable indices after a search in a filtering direction.

aggressive_obbt_on_heurestic

Routine that determines if aggressive filtering should be used. Currently, a user-specified option.

EAGO.bool_indx_diffFunction.
bool_indx_diff

Utility function used to set vector of booleans z to x & ~y. Avoids the generation of conversion of the BitArray created by broadcasting logical operators.

EAGO.obbtFunction.
obbt

Performs OBBT with filtering and greedy ordering as detailed in: Gleixner, A.M., Berthold, T., Müller, B. et al. J Glob Optim (2017) 67: 731. https://doi.org/10.1007/s10898-016-0450-4

trivial_filtering!

Excludes OBBT on variable indices that are tight for the solution of the relaxation.

diff --git a/docs/Optimizer/high_performance/index.html b/docs/Optimizer/high_performance/index.html deleted file mode 100644 index 3c06ffb0..00000000 --- a/docs/Optimizer/high_performance/index.html +++ /dev/null @@ -1,15 +0,0 @@ - -High-Performance Configuration · EAGO.jl: Easy Advanced Global Optimization

High-Performance Configuration

High-Performance Configuration

LP Solver Selection

By default, EAGO uses GLPK for solving linear subproblems introduced. Using a commercial linear solver is highly recommended such as Gurobi, CPLEX, or XPRESS is highly recommended. Both Gurobi and CPLEX are free for academics and installation information can be found through http://www.gurobi.com/academia/academia-center and https://www.ibm.com/developerworks/community/blogs/jfp/entry/CPLEXIsFreeForStudents?lang=en, respectively.

A non-default LP solver can then be selected by the user via a series of keyword argument inputs as illustrated in the code snippet below. The relaxed_optimizer contains an instance optimizer with valid relaxations that are made at the root node and is updated with affine relaxations in place. Options can be passed to this optimizer using keyword arguments when initializing EAGO using the withoptimizer syntax in JuMP by defining an Iterators.Pairs structure assigning it to the `relaxedoptimizer_kwargs` keyword argument. MOI.


-# Create opt EAGO Optimizer with CPLEX for use with MOI routines
-opt = EAGO.Optimizer(relaxed_optimizer = Gurobi.Optimizer(OutputFlag=0))
-
-# Create the same model m using an options dictionary in JuMP
-relaxed_optimizer_kwargs = Dict{Symbol, Any}()
-opt_dict[:relaxed_optimizer] = Gurobi.Optimizer()
-opt_dict[:relaxed_optimizer_kwargs] = Iterators.Pairs([:OutputFlag], [0])
-
-m = JuMP.Model(with_optimizer(EAGO.Optimizer; opt_dict...))
-
-# Create the same model m is keyword arguments in JuMP
-m = JuMP.Model(with_optimizer(EAGO.Optimizer; relaxed_optimizer = Gurobi.Optimizer(),
-                                              relaxed_optimizer_kwargs = Iterators.Pairs([:OutputFlag], [0])))

Ipopt Build

Ipopt is the recommended solver for upper bounding problems. Ipopt's performance is highly dependent on the linear algebra package used (up to 30x). By default MUMPS is used. It's recommended that you either compile Ipopt with HSL MA57 or the Pardiso linear algebra packages with a machine specific Blas library (for Intel users the JuliaPro MKL version is recommended). For information on this, see the below links:

diff --git a/docs/Optimizer/optimizer/index.html b/docs/Optimizer/optimizer/index.html deleted file mode 100644 index 6e67dc24..00000000 --- a/docs/Optimizer/optimizer/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -EAGO Optimizer · EAGO.jl: Easy Advanced Global Optimization

EAGO Optimizer

EAGO Optimizer

The EAGO.Optimizer object holds all algorithm solution information. A description of all user-facing options has been provided in the docstring.

EAGO.Optimizer

EAGO.OptimizerType.
Optimizer

The main optimizer object used by EAGO to solve problems during the optimization routine. The following commonly used options are described below and can be set via keyword arguments in the JuMP/MOI model:

  • presolve_scrubber_flag::Bool: Replace code in user-defined functions which may prevent method overloading on Real subtypes (default = false).
  • presolve_to_JuMP_flag::Bool: Create and use DAG representations of user-defined function (default = false).
  • presolve_epigraph_flag::Bool: [FUTURE FEATURE, NOT CURRENTLY IMPLEMENTED] Apply the epigraph reformulation to the problem (default = false).
  • presolve_cse_flag::Bool: [FUTURE FEATURE, NOT CURRENTLY IMPLEMENTED] Enable common subexpression elimination for DAG (default = false).
  • presolve_flatten_flag::Bool: Rerranges the DAG using registered transformations (default = false)
  • cp_depth::Int64: Depth in B&B tree above which constraint propagation should be disabled (default = 1000).
  • cp_repetitions::Int64: Number of repetitions of forward-reverse passes to perform in constraint propagation (default = 3).
  • cp_tolerance::Float64: Disable constraint propagation if the ratio of new node volume to beginning node volume exceeds this number (default = 0.99).
  • cp_interval_only::Bool: Use only valid interval bounds during constraint propagation (default = false).
  • relaxed_optimizer::S: An instance of the optimizer used to solve the relaxed subproblems (default = GLPK.Optimizer()).
  • relaxed_optimizer_kwargs::Base.Iterators.Pairs: Keyword arguments for the relaxed optimizer.
  • obbt_depth::Int64: Depth in B&B tree above which OBBT should be disabled (default = 1000).
  • obbt_repetitions::Int64: Number of repetitions of OBBT to perform in preprocessing (default = 3).
  • obbt_aggressive_on::Bool: Turn aggresive OBBT on (default = false).
  • obbt_aggressive_max_iteration::Int64: Maximum iteration to perform aggresive OBBT (default = 2)
  • obbt_aggressive_min_dimension::Int64: Minimum dimension to perform aggresive OBBT (default = 2)
  • obbt_tolerance::Float64: Tolerance to consider bounds equal (default = 1E-9).
  • obbt_variable_values::Vector{Bool}: Variables to perform OBBT on (default: all variables in nonlinear expressions).
  • lp_depth::Int64: Depth in B&B tree above which linear FBBT should be disabled (default = 1000).
  • lp_repetitions::Int64: Number of repetitions of linear FBBT to perform in preprocessing (default = 3).
  • quad_uni_depth::Int64: Depth in B&B tree above which univariate quadratic FBBT should be disabled (default = -1).
  • quad_uni_repetitions::Int64: Number of repetitions of univariate quadratic FBBT to perform in preprocessing (default = 2).
  • quad_bi_depth::Int64: [FUTURE FEATURE, NOT CURRENTLY IMPLEMENTED] Depth in B&B tree above which bivariate quadratic FBBT should be disabled (default = -1).
  • quad_bi_repetitions::Int64: Number of repetitions of bivariate quadratic FBBT to perform in preprocessing (default = 2).
  • subgrad_tighten::Bool: Perform tightening of interval bounds using subgradients at each factor in each nonlinear tape during a forward-reverse pass (default = true).
  • subgrad_tighten_reverse::Bool: [FUTURE FEATURE, NOT CURRENTLY IMPLEMENTED] Used to enable/disable subgradient tightening of interval bounds on the reverse pass (default = true).
  • cut_max_iterations::Int64
  • cut_cvx::Float64: Convex coefficient used to select point for new added cuts. Branch point is given by (1-cut_cvx)*xmid + cut_cvx*xsol (default = 0.9).
  • cut_tolerance::Float64: Add cut if the L1 distance from the prior cutting point to the new cutting point normalized by the box volume is greater than the tolerance (default = 0.05).
  • objective_cut_on::Bool: Adds an objective cut to the relaxed problem (default = true).
  • upper_optimizer::T: Optimizer used to solve upper bounding problem (default = Ipopt.Optimizer())
  • upper_factory::JuMP.OptimizerFactory: OptimizerFactory used to build optimizer that solves the upper bounding problem (default = with_optimizer(Ipopt.Optimizer, kwargs), check Optimizer constructor for kwargs used).
  • upper_bounding_depth::Int64: Solve upper problem for every node with depth less than upper_bounding_depth and with a probability of (1/2)^(depth-upperboundingdepth) otherwise (default = 4).
  • dbbt_depth::Int64: Depth in B&B tree above which duality-based bound tightening should be disabled (default = 1E10).
  • dbbt_tolerance::Float64: New bound is considered equal to the prior bound if within dbbt_tolerance (default = 1E-9).
  • branch_cvx_factor::Float64: Convex coefficient used to select branch point. Branch point is given by branch_cvx_factor*xmid + (1-branch_cvx_factor)*xsol (default = 0.25)
  • branch_offset::Float64: Minimum distance from bound to have branch point normalized by width of dimension to branch on (default = 0.15)
  • branch_variable::Vector{Bool}: Variables to branch on (default is all nonlinear).
  • branch_max_repetitions::Int64: [FUTURE FEATURE, NOT CURRENTLY IMPLEMENTED] Number of times repeat node processing prior to branching (default = 4).
  • branch_repetition_tol::Float64: [FUTURE FEATURE, NOT CURRENTLY IMPLEMENTED] Volume ratio tolerance required to repeat processing the current node (default = 0.9).
  • rounding_mode::Symbol: Interval rounding mode to use (default = :accurate)
  • node_limit::Int64: Node limit (default = 10^7).
  • time_limit::Float64: Time limit in seconds (default = 3600).
  • iteration_limit::Int64: Iteration limit (default = 3000000).
  • absolute_tolerance::Float64: Absolute tolerance for terminatin (default = 1E-3).
  • relative_tolerance::Float64: Relative tolerance for terminatin (default = 1E-3).
  • local_solve_only::Bool: Perform only a local solve of the problem (default = false).
  • log_on::Bool: Turns logging on records global bounds, node count and run time. Additional options are available for recording information specific to subproblems (default = false).
  • log_subproblem_info::Bool: Turns on logging of times and feasibility of subproblems (default = false).
  • log_interval::Int64: Log data every log_interval iterations (default = 1).
  • verbosity::Int64: The amount of information that should be printed to console while solving values range from 0 - 4: 0 is silent, 1 shows iteration summary statistics only, 2-4 show varying degrees of details about calculations within each iteration (default = 1).
  • output_iterations::Int64: Display summary of iteration to console every output_iterations (default = 10).
  • header_iterations::Int64: Display header for summary to console every output_iterations (default = 100).
  • enable_optimize_hook::Bool: Specifies that the optimize_hook! function should be called rather than throw the problem to the standard B&B routine (default = false).
  • ext::Dict{Symbol, Any}: Holds additional storage needed for constructing extensions to EAGO (default = Dict{Symbol,Any}).
  • ext_type::ExtensionType: Holds an instance of a subtype of EAGO.ExtensionType used to define new custom subroutines (default = DefaultExt()).

Internal Storage Structures

VariableInfo

A structure used to store information related to the bounds assigned to each variable.

  • is_integer::Bool: Is the variable integer valued?
  • lower_bound::Float64: May be -Inf even if haslowerbound == true
  • has_lower_bound::Bool: Implies lower_bound == Inf
  • upper_bound::Float64: May be Inf even if hasupperbound == true
  • has_upper_bound::Bool: Implies upper_bound == Inf
  • is_fixed::Bool: Implies lowerbound == upperbound and !haslowerbound and !hasupperbound.
ExtensionType

An abstract type the subtypes of which are associated with functions method overloaded for for new extensions. An instance of the DefaultExt <:ExtensionType structure to the Optimizer in the ext_type field.

EAGO.LogType.
Log

A structure used to store information on the history of the solution procedure for generating convergence plots and other analysis.

  • current_lower_bound::Vector{Float64}: Storage for lower bound calculated for current node.
  • current_upper_bound::Vector{Float64}: Storage for upper bound calculated for current node.
  • preprocessing_time::Vector{Float64}: Storage for preprocessing time of each iteration.
  • lower_problem_time::Vector{Float64}: Storage for lower bounding time of each iteration.
  • upper_problem_time::Vector{Float64}: Storage for upper bounding time of each iteration.
  • postprocessing_time::Vector{Float64}: Storage for postprocessing time of each iteration.
  • preprocessing_feas::Vector{Bool}: Storage for preprocessing feasibility of each iteration.
  • lower_problem_feas::Vector{Bool}: Storage for lower bounding feasibility of each iteration.
  • upper_problem_feas::Vector{Bool}: Storage for upper bounding feasibility of each iteration.
  • postprocessing_feas::Vector{Bool}: Storage for postprocessing feasibility of each iteration.
  • global_lower_bound::Vector{Float64}: Storage for best (global) lower bound at each iteration.
  • global_upper_bound::Vector{Float64}: Storage for best (global) upper bound at each iteration.

Extending EAGO

Functionality has been included that allows for extension's to EAGO's optimizer to readily be defined. This can be done in two ways first defining a new structure which is a subtype of EAGO.ExtensionType and overloading methods associated with this new structure. An instance of this new structure is provided to the EAGO.Optimizer using the ext_type keyword. This results in EAGO now dispatch to the new methods rather than the generally defined methods for the parent type. For a complete example, the reader is directed to the interval bounding example and quasiconvex example. Alternatively, the user can overload the optimize_hook! for this subtype which will entirely circumvent the default global solution routine. Additional information can be stored in the ext field of EAGO. In order to allow for compatibility between packages the user is encouraged to append their extension name to the start of each variable name (e.g. newext_newdata).

diff --git a/docs/Optimizer/relax_back/index.html b/docs/Optimizer/relax_back/index.html deleted file mode 100644 index 1fc420fb..00000000 --- a/docs/Optimizer/relax_back/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -Relaxation Backend · EAGO.jl: Easy Advanced Global Optimization

Relaxation Backend

Relaxation Backend

Quadratic Relaxations

relax_convex_kernel

Stores the kernel of the calculation required to relax convex quadratic constraints using the immutable dictionary to label terms.

relax_nonconvex_kernel

Stores the kernel of the calculation required to relax nonconvex quadratic constraints using the immutable dictionary to label terms.

relax_quadratic_gen_saf

Default routine for relaxing nonconvex quadratic constraint lower < func < upper on node n. Takes affine bounds of convex part at point x0 and secant line bounds on concave parts.

EAGO.relax_quadratic!Function.
relax_quadratic!

Relaxes all quadratic constraints in x optimizer.

Nonlinear Relaxation

EAGO.relax_nlp!Function.
relax_nlp!

A rountine that relaxes all nonlinear constraints excluding constraints specified as quadratic.

objective_cut_linear!

Adds linear objective cut constraint to the x.relaxed_optimizer.

Nonlinear Storage Structures

FunctionSetStorage

A storage object for both set and number valued data required to compute relaxations which contains the tape used to compute a nonlinear function. The object is parameterized by a {N,T<:RelaxTag} where N corresponds the subgradient size used in the MC object.

  • nd::Vector{JuMP.NodeData}
  • adj::SparseMatrixCSC{Bool,Int64}
  • const_values::Vector{Float64}
  • setstorage::Vector{MC{N,T}}
  • numberstorage::Vector{Float64}
  • numvalued::Vector{Bool}
  • tp1storage::Vector{Float64}
  • tp2storage::Vector{Float64}
  • tp3storage::Vector{Float64}
  • tp4storage::Vector{Float64}
  • tpdict::Dict{Int64,Tuple{Int64,Int64,Int64,Int64}}
  • grad_sparsity::Vector{Int64}
  • hess_I::Vector{Int64}
  • hess_J::Vector{Int64}
  • dependent_subexpressions::Vector{Int64}
SubexpressionSetStorage

A storage object for both set and number valued data required to compute relaxations which contains the tape used to compute a nonlinear subexpression. The object is parameterized by a {N,T<:RelaxTag} where N corresponds the the subgradient size used in the MC object.

  • nd::Vector{JuMP.NodeData}
  • adj::SparseMatrixCSC{Bool,Int64}
  • const_values::Vector{Float64}
  • setstorage::Vector{MC{N,T}}
  • numberstorage::Vector{Float64}
  • numvalued::Vector{Bool}
  • tp1storage::Vector{Float64}
  • tp2storage::Vector{Float64}
  • tp3storage::Vector{Float64}
  • tp4storage::Vector{Float64}
  • tpdict::Dict{Int64,Tuple{Int64,Int64,Int64,Int64}}
  • linearity::JuMP._Derivatives.Linearity

Nonlinear Evaluator

EAGO.EvaluatorType.
Evaluator

MOI.AbstractNLPEvaluator for calculating relaxations of nonlinear terms.

Internal Functions Used by Evaluator

set_current_node!(x::Evaluator, n::NodeBB)

Sets the current node in the Evaluator structure.

eval_objective_lo

Retrieves the lower bound of the objective.

eval_constraint_cc(d::Evaluator, g::Vector{Float64}, y::Vector{Float64})

Populates g with the concave relaxations of the constraints of d evaluated at y.

eval_constraint_lo!(d::Evaluator, g::Vector{Float64})

Populates g with the lower bounds of the constraints of d.

eval_constraint_hi!(d::Evaluator, g::Vector{Float64})

Populates g with the upper bounds of the constraints of d.

eval_constraint_cc_grad(d::Evaluator, g, y)

Populates g with the subgradients of the constraints of d evaluated at y.

get_node_lower(d::FunctionSetStorage, i::Int64)

Retreives the lower bound of ith term in the tape of d.

get_node_upper(d::FunctionSetStorage, i::Int64)

Retreives the upper bound of ith term in the tape of d.

forward_reverse_pass(d::Evaluator, x::Vector{Float64})

Performs a d.fw_repeats forward passes of the set-value evaluator each followed by a reverse pass if d.has_reverse as long as the node between passes differs by more that d.fw_atol at each iteration.

diff --git a/docs/Optimizer/starting/index.html b/docs/Optimizer/starting/index.html deleted file mode 100644 index c7685b9e..00000000 --- a/docs/Optimizer/starting/index.html +++ /dev/null @@ -1,126 +0,0 @@ - - - - - - - Solving simple example: an artificial neural network with EAGO · EAGO.jl: Easy Advanced Global Optimization - - - - - - - - - - - - - -
-
- -
-
Solving simple example: an artificial neural network with EAGO
-
-

Solving simple example: an artificial neural network with EAGO

-

Matthew Wilhelm Department of Chemical and Biomolecular Engineering, University of Connecticut

-

In [1,2], a surrogate ANN model of bioreactor productivity was constructed by fitting results from computationally expensive CFD simulations. The author then optimized this surrogate model to obtain ideal processing conditions. This - optimization problem is given by:

- -
using JuMP, EAGO
-
-# Box constraints for input variables
-xLBD = [0.623   0.093   0.259   6.56   1114   0.013   0.127   0.004]
-xUBD = [5.89    0.5     1.0     90     25000  0.149   0.889   0.049]
-
-# Weights associated with the hidden layer
-W = [ 0.54  -1.97  0.09  -2.14  1.01  -0.58  0.45  0.26;
-     -0.81  -0.74  0.63  -1.60 -0.56  -1.05  1.23  0.93;
-     -0.11  -0.38 -1.19   0.43  1.21   2.78 -0.06  0.40]
-
-# Weights associated with the output layer
-D = [-0.91 0.11 0.52]
-
-# Bias associated with the hidden layer
-B1 = [-2.698 0.012 2.926]
-
-# Bias associated with the output layer
-B2 = -0.46
-

Construct the JuMP model and optimize

-

We now formulate the problem using standard JuMP[3] syntax and optimize it. Note that we are forming an NLexpression object to handle the summation term to keep the code visually simple but this could be placed directly in the JuMP - expressions instead.

-
# Model construction
-model = Model(with_optimizer(EAGO.Optimizer, absolute_tolerance = 0.001))
-@variable(model, xLBD[i] <= x[i=1:8] <= xUBD[i])
-@NLexpression(model, prop[i=1:3], B1[i] + sum(W[i,j]*x[i] for j in 1:8))
-@NLobjective(model, Max, B2 + sum(D[i]*(2/(1+exp(-2*prop[i]))) for i=1:3))
-
-# Solves the model
-optimize!(model)
-

Retrieve results

-

We then recover the objective value, the solution value, and termination status codes using standard JuMP syntax.

-
# Access calculated values
-fval = JuMP.objective_value(model)
-xsol = JuMP.value.(x)
-status_term = JuMP.termination_status(model)
-status_prim = JuMP.primal_status(model)
-
-println("EAGO terminated with a status of $status_term and a result code of $status_prim")
-println("The optimal value is: $fval, the solution found is $xsol.")
-

Reference:

-
    -
  1. J. D. Smith, A. A. Neto, S. Cremaschi, and D. W. Crunkleton, CFD-based optimization of a flooded bed algae bioreactor, Industrial & Engineering Chemistry Research, 52 (2012), pp. 7181–7188
  2. -
  3. A. M. Schweidtmann and A. Mitsos. Global Deterministic Optimization with Artificial Neural Networks Embedded https://arxiv.org/pdf/1801.07114.pdf
  4. -
  5. Iain Dunning and Joey Huchette and Miles Lubin. JuMP: A Modeling Language for Mathematical Optimization, SIAM Review, 59 (2017), pp. 295-320.
  6. -
- -
- - - diff --git a/docs/Optimizer/udf_utilities/index.html b/docs/Optimizer/udf_utilities/index.html deleted file mode 100644 index e3fc9fd1..00000000 --- a/docs/Optimizer/udf_utilities/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -User-Define Functions and DAG Utilities · EAGO.jl: Easy Advanced Global Optimization

User-Define Functions and DAG Utilities

User-Define Functions and DAG Utilities

EAGO has included basic functionality to manipulate user-defined functions. These features are largely experimental and we're interested in providing additional for novel use cases. Please contact us by opening an issue in the Github issue tracker with any questions or requests for additional features.

DAG Substitution and Flattening

dag_flattening!

Flattens (usually) the dag by making all registered substitutions for every nonlinear term in the Optimizer.

flatten_expression!

Flattens (usually) the dag by making all registered substitutions for the expression expr::_NonlinearExprData. Performs a depth-first search through the expression adding the terminal node to the stack, then checking to determine if it matches a registered substitution pattern. If it doesn't not then node is added to the new expression graph representation and it's children are added to the queue. If an expression (node) is identified as a pattern then it is substituted and any children expression nodes are then checked for patterns until the depth first search is exhausted.

register_substitution!

Specifies that the src::Template_Graph should be subsituted out for the trg::Template_Graph.

Conventions for substition, the expression to be checked always appears at key 1 in the Template_Graph and operations are ordered from low value to high value left to right so if 1 is a -, and 4 => 1, 3 => 1 then the expression is 4 - 3

Template_Graph

Holds a list of Template_Nodes, set of directed edges, lengths, an adjacency matrix and the number of children.

Template_Node

A structure which holds a symbol indicating whether the node is an operator, a number, or an expression type, a value which identifies the function or symbol value, potentially a numeric value num_value, and a check that can be run to verify the node is correct check.

User-Defined Function (UDF) Scrubber

EAGO.Script.scrubFunction.
scrub(f::Function, n::Int, inplace = false)

Replaces storage objects and circumvents assertions that present a UDF from being overloaded with subtype Real objects by creating a function which overdubs f in ScrubCtx.

EAGO.Script.scrub!Function.
scrub!(d::_NLPData)

Applies scrub to every user-defined function in the a _NLPData structure.

diff --git a/docs/SemiInfinite/SIPProbFormulation.png b/docs/SemiInfinite/SIPProbFormulation.png deleted file mode 100644 index 212e1c5c..00000000 Binary files a/docs/SemiInfinite/SIPProbFormulation.png and /dev/null differ diff --git a/docs/SemiInfinite/SIPformulation.png b/docs/SemiInfinite/SIPformulation.png deleted file mode 100644 index f0164192..00000000 Binary files a/docs/SemiInfinite/SIPformulation.png and /dev/null differ diff --git a/docs/SemiInfinite/semiinfinite/SIPProbFormulation.png b/docs/SemiInfinite/semiinfinite/SIPProbFormulation.png deleted file mode 100644 index 95419e4d..00000000 Binary files a/docs/SemiInfinite/semiinfinite/SIPProbFormulation.png and /dev/null differ diff --git a/docs/SemiInfinite/semiinfinite/SIPformulation.png b/docs/SemiInfinite/semiinfinite/SIPformulation.png deleted file mode 100644 index 8c9256fe..00000000 Binary files a/docs/SemiInfinite/semiinfinite/SIPformulation.png and /dev/null differ diff --git a/docs/SemiInfinite/semiinfinite/index.html b/docs/SemiInfinite/semiinfinite/index.html deleted file mode 100644 index 609b6cde..00000000 --- a/docs/SemiInfinite/semiinfinite/index.html +++ /dev/null @@ -1,130 +0,0 @@ - - - - - - - Semi-Infinite Programming · EAGO.jl: Easy Advanced Global Optimization - - - - - - - - - - - - - -
-
- -
-
Semi-Infinite Programming
-
-

Solving Semi-Infinite Programming

-

Matthew Wilhelm Department of Chemical and Biomolecular Engineering, University of Connecticut

-

Using EAGO to solve a SIP

-

Semi-infinite programming remains an active area of research. In general, the solution of semi-infinite programs with nonconvex semi-infinite constraints of the below form are extremely challenging:

- -

EAGO implements the SIPres of [1] to determine a globally optimal solution to problems of the above form. This accomplished using the explicit_sip_solve function which returns the optimal value, the solution, and a boolean - feasibility value. To illustrate the functions use, a simple example is presented here which solves the below problem:

- -
using EAGO, JuMP
-
-# Define semi-infinite program
-f(x) = (1/3)*x[1]^2 + x[2]^2 + x[1]/2
-gSIP(x,p) = (1.0 - (x[1]^2)*(p[1]^2))^2 - x[1]*p[1]^2 - x[2]^2 + x[2]
-
-x_l = [-1000.0, -1000.0]
-x_u = [1000.0, 1000.0]
-p_l = [0.0]
-p_u = [1.0]
-
-# Solve the SIP  problem
-sip_result = explicit_sip_solve(x_l, x_u, p_l, p_u, f, gSIP)
-
-println("The global minimum of the semi-infinite program is between: $(sip_result.lower_bound) and $(sip_result.upper_bound).")
-println("The global minimum is attained at: x = $(sip_result.xsol).")
-println("Is the problem feasible? (sip_result.feasibility).")
-

Semi-infinite solver

-
- -
-
-

explicitsipsolve

-

Solve an SIP with decision variable bounds x_l to x_u, uncertain variable bounds p_l to p_u, an objective function of f, and gSIP seminfiniite constraint(s).

-
-
-
-
-
EAGO.SIPProblemType.
-
-
-
  SIPProblem
-

Structure storing problem information for the solution routine.

-
-
-
-
-
EAGO.SIPResultType.
-
-
-
  SIPResult
-

Structure storing the results of the SIPres algorithm.

-
-
-
- -
- - - diff --git a/docs/cite/index.html b/docs/cite/index.html deleted file mode 100644 index 64097270..00000000 --- a/docs/cite/index.html +++ /dev/null @@ -1,4 +0,0 @@ - -Citing EAGO · EAGO.jl: Easy Advanced Global Optimization

Citing EAGO

Citing EAGO

A paper about the EAGO software package is currently under preparation. In the meantime, please feel free to cite the conference presentation below:

Wilhelm, Matthew; Stuber, Matthew (October 2017) Easy Advanced Global
-Optimization (EAGO): An Open-Source Platform for Robust and Global Optimization
-in Julia. Presented at the AIChE Annual Meeting in Minneapolis, MN.
diff --git a/docs/ref/index.html b/docs/ref/index.html deleted file mode 100644 index 46b6034a..00000000 --- a/docs/ref/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -References · EAGO.jl: Easy Advanced Global Optimization

References

References

Branch and Bound

Parametric Interval Techniques

Domain Reduction

Generalized McCormick Relaxations

Semi-Infinite Programming

diff --git a/docs/search/index.html b/docs/search/index.html deleted file mode 100644 index 24f5e03a..00000000 --- a/docs/search/index.html +++ /dev/null @@ -1,2 +0,0 @@ - -Search · EAGO.jl: Easy Advanced Global Optimization

Search

Search

Number of results: loading...