A Julia implementation of the Black-Box-Optimization-Benchmarking (BBOB) functions.
The average sucess rate (meaning the optimizer reached the minimum + 1e-6) in function of the number of iterations, in 3 dimension :
BlackBoxOptim's algorithm are performing the best in 3D.
Since some global optimizers have poor final convergence, they were chained into a Nelder-Mead using 10% of the objective function evaluation budget. Note that some algorithm call the objective function several time per iteration, so this plot is not totally fair (it doesn't really impact the results however).
If we look at the sucess rate per function we can see that only a few algorithm are able to solve all the problems :
The script to produce these plots is in scripts/run_benchmark.jl.
Functions with input dimension N can be generated using bbob_suite():
julia> BlackBoxOptimizationBenchmarking.bbob_suite(Val(2))
20-element Vector{BBOBFunction}:
F1 Sphere
F2 Ellipsoidal
F3 Rastrigin
F4 Buche-Rastrigin
F5 Linear Slope
F6 Attractive Sector
F7 Step Ellipsoidal
F8 Rosenbrock
F9 Rosenbrock Rotated
F10 Ellipsoidal 2
F11 Discus
F12 Bent Cigar
F13 Sharp Ridge
F14 Different Powers
F15 Rastrigin 2
F16 Weierstrass
F17 Schaffers F7
F18 Schaffers F7 Ill-Cond
F19 Griewank-Rosenbrock
F20 SchwefelAn indivdual BBOBFunction function f has fields f_opt its minimal value, and x_opt its minimizer, i.e. f(x_opt) = f_opt.
Functions can be plot using :
using Plots
plot(f::BBOBFunction; nlevels = 15, zoom=1)A benchmark for a single optimizer and function can be run with:
b::BenchmarkResults = benchmark(
optimizer, f::BBOBFunction, run_length::AbstractVector{Int};
Ntrials::Int = 20, dimension::Int = 3, Δf::Real = 1e-6, CI_quantile=0.25
)The first argument optimizer must implement Optimization.jl's interface, and
it must be wrapped in a BenchmarkSetup to indicate if the optimizer requires bounds :
BenchmarkSetup(optimizer, isboxed = true)
To test an optimizer on several functions a vector of BBOBFunction's can be passed instead of a single function,
and all the returned statistics will be averaged over functions (with the expection of success_rate_per_function).
The main fields of the returned struct BenchmarkResults are :
run_length: number of iterations the optimizer performedcallcount: number of objective function callssuccess_rate: for each run_length, the fraction of optimization runs that reached the global minimum with a tolerance of Δf
A benchmark can be plot with :
using Plots
plot(b; label = "NelderMead", x = :callcount, showribbon = true)
plot!(another_benchmark)The ribbon indicates the 25% to 95% confidence intervals of the success_rate (the quantile used
can be changed with compute_CI!(b::BenchmarkResults, CI_quantile)).
We can test an algorithm on a function and plot the result using
Δf = 1e-6
f = test_functions[3]
setup = BenchmarkSetup(NLopt.GN_CRS2_LM(), isboxed=true)
sol = [BBOB.solve_problem(setup, f, 3, 5_000) for in in 1:10]
@info [sol.objective < Δf + f.f_opt for sol in sol]
p = plot(f, size = (600,600), zoom = 1.5)
for sol in sol
scatter!(sol.u[1:1], sol.u[2:2], label="", c="blue", marker = :xcross, markersize=5, markerstrokewidth=0)
end
pTo avoid overfiting and test if algorithms are robust with respect to translation and rotations of
the error function, rotation matrices and local minima are randomly generated when calling bbob_suite (controlled by the seed parameters, see docstrings).
https://numbbo.github.io/gforge/downloads/download16.00/bbobdocfunctions.pdf


