• Aucun résultat trouvé

A short review of convex analysis and optimization

N/A
N/A
Protected

Academic year: 2022

Partager "A short review of convex analysis and optimization"

Copied!
8
0
0

Texte intégral

(1)

A short review of convex analysis and optimization

Guillaume Obozinski

Ecole des Ponts - ParisTech

Master MVA 2014-2015

(2)

Review: convex analysis

Convex function

∀λ∈[0,1], f(λx+ (1−λ)y)≤λ f(x) + (1−λ)f(y) Strictly convex function

∀λ∈]0,1[, f(λx+ (1−λ)y)< λ f(x) + (1−λ)f(y) Strongly convex function

∃µ >0, s.t. x7→f(x)−µkxk2 is convex Equivalently:

∀λ∈[0,1], f(λx+(1−λ)y)≤λ f(x)+(1−λ)f(y)−µ λ(1−λ)kx−yk2 The largest possibleµ is called the strong convexity constant.

(3)

Minima of convex functions

Proposition (Supporting hyperplane) If f is convex and differentiable at xthen

f(y)≥f(x) +∇f(x)>(y−x)

Convex function

All local minima are global minima.

Strictly convex function

If there is a local minimum, then it is unique and global.

Strongly convex function

There exists a unique local minimum which is also global.

(4)

Minima and stationary points of differentiable functions

Definition (Stationary point)

For f differentiable, we say that xis a stationary point if ∇f(x) = 0.

Theorem (Fermat)

If f is differentiable at x and xis a local minimum, then x is stationary.

Theorem (Stationary point of a convex differentiable function) If f is convex and differentiable at xand x is stationary thenx is a minimum.

Theorem (Stationary points of a twice differentiable functions) For f twice differentiable at x

if x is a local minimum then ∇f(x) = 0 and∇2f(x)0.

conversely if ∇f(x) = 0 and ∇2f(x)0 then xis a strict local

(5)

Brief review of Lagrange duality

Convex optimization problem with linear constraints For

f a convex function,

X ⊂Rp a convex set included in the domain of f, A∈Rn×p, b∈Rn,

minx∈Xf(x) subject to Ax=b (P) Lagrangian

L(x,λ) =f(x) +λT(Ax−b) with λ∈Rn theLagrange multiplier.

(6)

Properties of the Lagrangian

Link between primal and Lagrangian

λ∈maxRn

L(x,λ) =

(f(x) ifAx=b +∞ otherwise.

So that

x∈Xmin:Ax=bf(x) = min

x∈Xmax

λ∈Rn

L(x,λ)

Lagrangian dual objective function g(λ) = min

x∈XL(x,λ)

Dual optimization problem maxng(λ) = max

nminL(x,λ) (D)

(7)

Maxmin-minmax inequality, weak and strong duality

For any f :Rn×Rm and anyw∈Rn and z∈Rm, we have maxz∈Z min

w∈Wf(w, z)≤ min

w∈Wmax

z∈Z f(w, z).

Weak duality d := max

λ∈Rn

g(λ) = max

λ min

x∈XL(x,λ)≤ min

xx∈X max

λ∈Rn

L(x,λ) = min

x∈Xf(x) =:p So that in general, we haved ≤p. This is called weak duality Strong duality

In some cases, we have strong duality:

d=p

Solutions to (P) and (D) are the same

(8)

Slater’s qualification condition

Slater’s qualification condition is a condition on the constraints of a convex optimization problem that guarantees that strong duality holds.

For linear constraints, Slater’s condition is very simple:

Slater’s condition for a cvx opt. pb with lin. constraints If there exists an xin the relative interior of X ∩ {Ax=b} then strong duality holds.

Références

Documents relatifs

An iterative algorithm has been proposed to compute the so- lution of separable convex optimization problems with linear and box constraints.. It is particularly interesting since

The choice of γ t is crucial for the optimization algorithm, indeed a γ t that is too big, makes the algorithm unstable and possibly diverging, since it follows the direction −∇f (x t

Is it possible to define a mapping of Courant-Beltrami type for a con- vex program (non necessarily an ordinary one)? In other words, is it always possible to reduce the problem of

In Section 6, we consider separable optimization problems associated with the Lov´asz extension; these are reinterpreted in Section 7 as separable optimization over the submodular

Figure 1.11 gives an example of a weakly su- pervised segmentation result: Given a set of tags (car, road, building), the algorithm is able to find the regions associated with

However, if on the one hand convex optimization usually refers to minimizing a convex func- tion on a convex set K without precising its representation (g j ) (see e.g. [3, Chapter

Moreover, it is shown that for this class of problems, these versions of Farkas lemma are actually equivalent to several strong duality results of Lagrange, Fenchel-Lagrange or

Rademacher, Uber partielle und totale Differenzierbarkeit von Functionen mehrerer Variabeln und uber die Transformation der Doppelintegrale.. Weaver,