[Convex Optimization Study Notes 1] What is optimization, the mathematical expression of optimization, and the classification of optimization problems

Reference materials:
1. Teacher Ling Qing’s Convex Optimization Course (B station)
2. Stephen Boyd’s "Convex Optimization" Chinese translation (Tsinghua University Press)

What is Optimization

Optimization is also called Mathematical Programming.
Optimization: Find the optimal element from a set of feasible solutions.
Three elements are indispensable: the set of feasible solutions, the method of finding, and the optimal elements.

Mathematical Expression Form of Optimization Problem

minimize ⁡ f 0 ( x ) subject to f i ( x ) ⩽ b i , i = 1 , ⋯   , m \begin{array}{ll} \operatorname{minimize} & f_{0}(x) \\ \text {subject to} & f_{i}(x) \leqslant b_{i}, \quad i=1, \cdots, m \end{array} minimizesubject tof0(x)fi(x)bi,i=1,,m


where:
xxx is calledthe optimization variable(Optimization Variable), which is a vectorx = ( x 1 , ⋯ , xn ) x=\left(x_{1}, \cdots, x_{n}\right)x=(x1,,xn)
f 0 ( x ) f_{0}(x) f0( x ) is calledthe objective function(Objective Function), which is a mapping from n-dimensional space to 1-dimensional space, namelyf 0 : R n → R f_{0}: \mathbf{R}^{n} \rightarrow \ mathbf{R}f0:RnR
f i ( x ) f_{i}(x) fi( x ) is a set ofinequality constraints(Inequality Constraint),fi : R n → R f_{i}: \mathbf{R}^{n} \rightarrow \mathbf{R}fi:RnR

Our goal is to find an x ​​⋆ x^{\star}x , is anoptimal solution(Optimal).
Ifx ⋆ x^{\star}x is an Optimal, then it is equivalent to:
there exists any vectorzzz , satisfy all constraintsz ∈ { f 1 ( z ) ⩽ b 1 , ⋯ , fm ( z ) ⩽ bm } z\in\{ f_{1}(z) \leqslant b_{1}, \cdots, f_{m}(z) \leqslant b_{m}\}z{ f1(z)b1,,fm(z)bm} , there are:f 0 ( z ) ⩾ f 0 ( x ⋆ ) f_{0}(z) \geqslant f_{0}\left(x^{\star}\right)f0(z)f0(x)

上面 { f 1 ( z ) ⩽ b 1 , ⋯   , f m ( z ) ⩽ b m } \{ f_{1}(z) \leqslant b_{1}, \cdots, f_{m}(z) \leqslant b_{m}\} { f1(z)b1,,fm(z)bm} set is calleda feasible solution set(feasible set).
The optimal solution may also have an optimal solution set, that is, there is more than one optimal solution.

Example 1

Take the unary function as an example, the figure:
insert image description here

Find the minimum value of the function in the interval [-1,1].

Converted to an optimization problem:
Objective function: f 0 ( x ) = x 2 + 1 f_{0}(x)=x^2+1f0(x)=x2+1
Feasible solution set:x ⩽ 1 , − x ⩽ 1 x\leqslant1,-x\leqslant1x1,x1
Find the optimal solution (the solution isx = 0 x=0x=0

Example 2 (data fitting problem)

There is a set of sample points (two-dimensional coordinate system as an example), and a line needs to be fitted.
insert image description here

We let this line be a binomial y = ax 2 + bx + cy=ax^2+bx+cy=ax2+bx+c , the coefficients a, b, and c of each item are the parameters to be estimated, set the vectorw = ( a , b , c ) w=(a,b,c)w=(a,b,c)

目标函数: f 0 ( w ) = ∑ i = 1 n [ y i − ( a x i 2 + b x i + c ) ] 2 f_{0}(w)=\sum_{\mathrm{i}=1}^{\mathrm{n}} [y_i-(ax_i^2+bx_i+c)]^2 f0(w)=i=1n[yi(axi2+bxi+c)]2 i i i is each sample point.
Finding the optimal solution is a typical least squares problem.

Classification of Optimization Problems

Linear programming, nonlinear programming

If in the optimization problem, the objective function and constraint function f 0 , . . . , fm f_0,...,f_mf0,...,fmAll are linear functions, that is, for any x , y ∈ R nx,y \in\mathbf{R}^{n}x,yRnα , β ∈ R \alpha,\beta\in\mathbf{R}a ,bR, have:
fi ( α x + β y ) = α fi ( x ) + β fi ( y ), i = 0 , 1 , . . . , m f_{i}(\alpha x+\beta y)=\alpha f_{i}(x)+\beta f_{i}(y),i=0,1,...,mfi(αx+βy)=a fi(x)+βfi(y)i=0,1,...,m
Then such an optimization problem is calleda linear programming.
iff 0 , . . . , fm f_0,...,f_mf0,...,fmIf one of them is not a linear function, then the problem is called nonlinear programming .


Give an example of a linear program:

insert image description here
The pentagon in the figure is composed of 5 linear functions f 1 , . . . , f 5 f_1,...,f_5f1,...,f5The feasible solution set constituted, because the constraint functions are all linear functions, so the feasible solution set must grow into such a pentagonal shape.
At the same time, the objective function f 0 f_0f0It is also a linear function. When we put the objective function in the space, it will be presented with different contour lines to represent its value.
If the arrow direction is f 0 f_0f0The direction of descent, then its optimal solution is on the marked black dot.

Convex programming, non-convex programming

People used to define linear programming and nonlinear programming problems as simple and difficult problems, but later found that it is better to define convex programming as simple problems and non-convex programming as difficult problems.
Any linear programming problem must be convex. Convex programming can be seen as an extension of linear programming.

If both the objective function and the constraint function are convex functions, that is, for any x , y ∈ R nx,y \in\mathbf{R}^{n}x,yRnα , β ∈ R \alpha,\beta\in\mathbf{R}a ,bR ,one valueα + β = 1 , α ⩾ 0 , β ⩾ 0 \alpha+\beta=1,\alpha\geqslant0,\beta\geqslant0a+b=1,a0,b0,电影不等式官方:
fi ( α x + β y ) ⩽ α fi ( x ) + β fi ( y ), i = 0 , 1 , . . . , m f_{i}(\alpha x+\beta y)\leqslant\alpha f_{i}(x)+\beta f_{i}(y),i=0,1,...,mfi(αx+βy)a fi(x)+βfi(y)i=0,1,...,
A function such as m is a convex function.
The problem in which both the objective function and the constraint function are convex functions is a convex programming problem, also called a convex optimization problem.

It can be seen that convexity is a more general property than linearity. Because it only needs to satisfy the inequality. Only a value restriction of α and β is added.

Other categories

①Smooth optimization problem, non-smooth optimization problem

smooth and non-smooth for the objective function f 0 f_0f0In terms of.
Smooth function means objective function f 0 f_0f0All points on the domain are differentiable.

② Continuous optimization problem, discontinuous optimization problem

For the feasible domain.
If the feasible region is continuous, it is a continuous optimization problem (such as the pentagon diagram above).
The feasible region is discrete, which means it is a discontinuous optimization problem. Discrete problems are generally hard problems.

③Single-objective problem, multi-objective problem

For multi-objective problems, say we want to optimize two objective functions:
minimize ⁡ f 1 ( x ) , f 2 ( x ) \operatorname{minimize} f_{1}(x), f_{2}(x)minimizef1(x),f2( x )
This is the multi-objective problem.

For f 1 , f 2 f_{1}, f_{2}f1,f2, we can often plot two function values:
insert image description here

This line is also known as the Pareto front. It is difficult for us to find a point to make the two optimal. Generally, we need to find a compromise point.
So we generally need to weight the two to convert multi-objective into a single-objective problem:
minimize ⁡ α 1 f 1 ( x ) + α 2 f 2 ( x ) \operatorname{minimize} \alpha_1 f_{1}(x) +\alpha_2 f_{2}(x)minimizea1f1(x)+a2f2( x )
This is only a workaround and may not necessarily work.

Guess you like

Origin blog.csdn.net/takedachia/article/details/126307734