• 검색 결과가 없습니다.

Convex optimization

N/A
N/A
Protected

Academic year: 2022

Share "Convex optimization"

Copied!
24
0
0

로드 중.... (전체 텍스트 보기)

전체 글

(1)

Convex Optimization

A supplementary note to Chapter 4 of Convex Optimization by S. Boyd and L. Vandenberghe

Optimization Lab.

IE department Seoul National University

4th October 2009

(2)

Optimization:

min f

0

(x )

s.t. f

i

(x ) ≤ 0, i = 1, · · · , m, h

j

(x ) = 0, j = 1, · · · , p.

Decision variables: x ∈ Rn Objective function: f0(x )

Constraints (inequality and equality): fi(x ) ≤ 0 and hj(x ) = 0.

A point x ∈ D is feasible if it satisfies all constraints; infeasible, otherwise

Tm Tp

(3)

Optimal value: p= inf{f0(x )|fi(x ) ≤ 0, ∀i , hj(x ) = 0, ∀j }.

Optimal solution: A feasible xis optimal if f0(x) = p. Denote by Xopt the set of optimal points.

A feasible x is -optimal if f0(x ) ≤ p+ .

A feasible x is a local optimum if ∃R > 0 s.t. f0(x ) = inf{f0(z)|

fi(x ) ≤ 0, ∀i , hj(x ) = 0, ∀j , kz − x k2≤ R}.

The feasibility problem: Find an x satisfying all the constraints.

(4)

Expressing problems in standard form

Min-version

min f0(x )

s.t. fi(x ) ≤ 0, i = 1, · · · , m, hj(x ) = 0, j = 1, · · · , p.

Max-version

max f0(x )

s.t. fi(x ) ≥ 0, i = 1, · · · , m, hj(x ) = 0, j = 1, · · · , p.

(5)

Equivalent problems

We say two problems are equivalent if, given a solution of one, we can efficiently find a solution of the other, and vice versa.

Example

Two problems are equivalent if αi > 0, ∀i , βj6= 0, ∀j:

min f0(x ) min α0f0(x )

s.t. fi(x ) ≤ 0, ∀i s.t. αifi(x ) ≤ 0, ∀i hj(x ) = 0, ∀j βjhj(x ) = 0, ∀j

(6)

Change of variables

Suppose φ : Rn→ Rn is one-to-one mapping. Define ˜fi(x ) = fi(φ(x )) and ˜hj(x ) = hj(φ(x )) Then the two problems are equivalent:

min f0(x )

s.t. fi(x ) ≤ 0, i = 1, . . . , m hi(x ) = 0, i = 1, . . . , p,

min ˜f0(x )

s.t. ˜fi(x ) ≤ 0, i = 1, . . . , m

˜hi(x ) = 0, i = 1, . . . , p.

Why?

(7)

Transformation of objective and constraints

Suppose ψ0: R → R is monotone increasing, ψ1, . . . , ψm: R → R satisfy ψi(u) ≤ 0 iff u ≤ 0, and ψm+1, . . . , ψm+p: R → R satisfy ψi(u) = 0 iff u = 0.

Define ˜fi(x ) = ψi(fi(x )), i = 0, . . . , m and ˜hi(x ) = ψm+i(hi(x )), i = 1, . . . , p.

Then following two are equivalent:

min f0(x )

s.t. fi(x ) ≤ 0, i = 1, . . . , m hi(x ) = 0, i = 1, . . . , p,

min ˜f0(x )

s.t. ˜fi(x ) ≤ 0, i = 1, . . . , m

˜hi(x ) = 0, i = 1, . . . , p.

(8)

Least-norm and least-norm-squared

min kAx − bk2versus min kAx − bk22.

(9)

Slack variables

fi(x ) ≤ 0 if and only if ∃ si ≥ 0 such that fi(x ) + si= 0; si is called a slack variable.

The followings are equivalent:

min f0(x )

s.t. fi(x ) ≤ 0, i = 1, . . . , m hi(x ) = 0, i = 1, . . . , p, min f0(x )

s.t. fi(x ) + si = 0, i = 1, . . . , m hi(x ) = 0, i = 1, . . . , p si≥ 0, i = 1, . . . , m.

(10)

Introducing equality constraints

The following two problems are equivalent:

min f0(A0x + b0) min f0(y0)

s.t. fi(Aix + bi) ≤ 0, ∀i , s.t. fi(yi) ≤ 0, ∀i , hj(x ) = 0, ∀j, yi = Axi+ bi ∀i , hj(x ) = 0 ∀j.

(11)

Optimizing over some variables

Since inf

x ,yf (x , y ) = inf

x

˜f (x ), where ˜f (x ) = inf

y f (x , y ), following two are equivalent:

min f0(x1, x2) ⇔ min ˜f0(x1)

s.t. fi(x1) ≤ 0, i = 1, · · · , m1 s.t. fi(x1) ≤ 0, i = 1, · · · , m1

gj(x2) ≤ 0, j = 1, · · · , m2

where ˜f0(x1) = inf{f0(x1, z)|gj(z) ≤ 0, j = 1, · · · , m2}.

Example

Consider a strictly convex quadratic program constrained on some variables: min x1TP11x1+ 2x1TP12x2+ x2TP22x2s.t. fi(x1) ≤ 0, i = 1, . . . , m.

Since, infx2x1TP11x1+ 2x1TP12x2+ x2TP22x2= x1T(P11− P12P22−1P12T)x1, we can obtain equivalent problem:

min x1T(P11− P12P22−1P12T)x1, s.t. fi(x1) ≤ 0, i = 1, . . . , m.

(12)

Epigraph problem form

The following two problems are equivalent:

min f0(x ) min t

s.t. fi(x ) ≤ 0, i = 1, · · · , m s.t. f0(x ) − t ≤ 0, i = 1, · · · , m1

hj(x ) = 0, j = 1, · · · , p fi(x ) ≤ 0, i = 1, · · · , m hj(x ) = 0, j = 1, · · · , p

(13)

Quasiconvex optimization

Min-version

min f0(x ) f0is convex.

s.t. fi(x ) ≤ 0, i = 1, . . . , m, fi are convex.

aTj x = bj, j = 1, . . . , p.

Max-version

max f0(x ) f0is concave.

s.t. fi(x ) ≥ 0, i = 1, . . . , m, fi are concave.

aTj x = bj, j = 1, . . . , p.

Similarly, can define quasiconvex optimization problem: ‘convex’

(‘concave’) is replaced by ‘quasi-convex’ (‘quasi-concave, resp.).

(14)

Quasiconvex optimization

Local and global optima

Theorem

Any local optimum of convex optimization problems is also a global optimum.

Proof Let x be a local optimum: ∃ R > 0 s.t. f0(x ) ≤ f0(z) ∀ feasible z such that kz − x k2≤ R}. Suppose, on the contrary, ∃ z ∈ D such that f (z) < f (x).

Then, ∃y such that ky − x k2< R and y = λx + (1 − λ)z for some 0 < λ < 1.

Since f (y ) ≥ f (x ) > f (z), f (y ) > λf (x ) + (1 − λ)f (z). A contradiction to convexity of f0.

Remark

(15)

Quasiconvex optimization

Theorem

For convex minimization with differentiable f0, feasible x is optimal iff

∇f0(x )T(y − x ) ≥ 0 for any feasible y . Proof

“If.” For any feasible y we have f0(y ) ≥ f0(x ) + ∇f0(x )T(y − x ) ≥ f0(x ).

“Only if.” Suppose not: ∃y feasible such that ∇f0(x )T(y − x ) < 0.

For λ ∈ [0, 1], let g (λ) = f0((1 − λ)x + λy ). Then, d

d λg (λ) λ=0

= ∇f0(x )T(y − x ) < 0,

which implies that for small enough λ > 0, we have g (λ) < g (0). A contradiction to optimality of x .

(16)

Quasiconvex optimization

Corollary

For unconstrained convex minimization, x is optimal iff x is feasible and

∇f0(x ) = 0.

Corollary

For convex minimization with equality constraints Ax = b only, x is optimal iff

∃λ s.t. ATλ = ∇f0(x ), where λ ∈ Rp. Proof

x optimal ⇐⇒ ∀y s.t. Ay = b, ∇f0(x )T(y − x ) ≥ 0

⇐⇒ ∇f0(x )Tz ≥ 0, ∀z ∈ N (A), null space of A

⇐⇒ ∇f0(x )Tz = 0, ∀z ∈ N (A)

(17)

Quasiconvex optimization

Example

Consider the problem

min f0(x ) s.t. x ≥ 0

Optimality condition : x ≥ 0, ∇f0(x )T(y − x ) ≥ 0 for all y ≥ 0.

∇f0(x )Ty is unbounded below on y ≥ 0 unless ∇f0(x ) ≥ 0, which implies −∇f0(x )Tx ≥ 0.

But, as x ≥ 0 and ∇f0(x ) ≥ 0, we get ∇f0(x )Tx = 0.

Thus, optimality condition becomes

x ≥ 0, ∇f0(x ) ≥ 0, xi(∇f0(x ))i = 0, complementary condition.

(18)

Quasiconvex optimization

min f0(x ) f0is quasiconvex,

s.t. fi(x ) ≤ 0, i = 1, . . . , m, fi’s are convex, Ax = b.

1 It may have local optima that are not global optima.

2 A sufficient optimality condition: If x is feasible and

∇f0(x )T(y − x ) > 0, ∀ feasible y 6= x , then x is globally optimal.

(Why?)

(19)

Quasiconvex optimization

Quasiconvex optim. via sequence of convex feasibility

Let φt : Rn→ R, t ∈ R such that

f0(x ) ≤ t ⇐⇒ φt(x ) ≤ 0 and φs(x ) ≤ φt(x ) for s ≥ t.

(For instance, φt(x ) = f0(x ) − t satisfies conditions.) Then, optimal value p can be computed as follows:

find x If the problem is feasible, then p≤ t = UB.

s.t. φt(x ) ≤ 0 Otherwise, p≥ t = LB.

fi(x ) ≤ 0, ∀i t ← (UB + LB)/2

Ax = b Repeat until UB − LB ≤ .

Note that it requires dlog2((UB − LB)/)e iterations for an -suboptimal solution.

(20)

Quadratic program (QP) minimizes convex quadratic over polyhedron.

min 12xTPx + qTx + r s.t. Gx ≤ h

Ax = b, where P ∈ Sn+, G ∈ Rm×n, A ∈ Rp×n.

Quadratically constrained quadratic program (QCQP) minimizes convex quadratic over intersection of ellipsoids.

min 12xTP0x + q0Tx + r0

s.t. 12xTPix + qiTx + r0≤ 0, i = 1, · · · , m Ax = b,

(21)

Minimizing Euclidean norm of affine functions

Least-squares and regression

kAx − bk22= xTATAx − 2bTAx + bTb.

Analytical solution is x = Ab. But, if linear inequality constraints are added, e.g.

min kAx − bk22

s.t. li ≤ xi ≤ ui, i = 1, · · · , n, problem does not have analytical solution and solvable via QP.

Distance between polyhedra For P1= {x |A1x ≤ b1}, P2= {x |A2x ≤ b2}, dist(P1, P2) = inf{kx1− x2k2|x1∈ P1, x2∈ P2},

can be computed via following QP:

min kx1− x2k22 s.t. A1x ≤ b1(x1∈ P1)

A2x ≤ b2(x2∈ P2).

(22)

Linear program with random cost

Suppose that c ∈ Rn is random with mean ¯c and covariance Σ. Then, E(cTx ) = ¯cTx , Var(cTx ) = E(cTx − ¯cTx )2= xTΣx A possible problem is to minimize a weighted sum of expected cost and uncertainty cost. To do so, we can consider γ, risk-sensitive cost so that the problem is formulated as follows:

min c¯Tx + γxTΣx s.t. Gx ≤ h

Ax = b.

(23)

Markowitz portfolio optimization

Let xi be amount of asset or stock i = 1, . . ., n; xi > 0 ↔ long position in asset i , xi < 0 ↔ short position in asset i .

Let pi be increase of price of i during a period; we assume p has mean vector ¯p and covariance matrix Σ.

Classical model minimizes risk guaranteeing a return with no shorting:

min xTΣx s.t. p¯Tx ≥ rmin,

1Tx = 1, x ≥ 0.

Various extensions are possible.

(24)

Homework

4.1, 4.2, 4.3, 4.5, 4.8, 4.13, 4.20, 4.21, 4.22

참조

관련 문서

The proposed approach is based on improved marker tracking by facilitating the use of convex polygons and reducing the sizes of AR markers without

The dimension of a convex set is defined to be the the maximum number of its affinely independent vectors minus

There is another combinatorial invariant for convex polytopes, called the flag vector, which are not relatively well known but obviously generalizes the notion

(c) the same mask pattern can result in a substantial degree of undercutting using substantial degree of undercutting using an etchant with a fast convex undercut rate such

The convex polytopes are the simplest kind of polytopes, and form the basis for several different generalizations of the concept of polytopes.. For

A cone is called strongly convex if it does not contain a nontrivial linear subspace. In this paper, every cone is assumed to be strongly convex. A polyhedral cone is

Also, the quasi-infraversion is larger for the inferior safe insertion limit direction (Fig. In practical design of a screw guiding device, targeted safe angles can

“In fact the great watershed in optimization isn’t between linearity and nonlinearity, but convexity and nonconvexity.” - Rockafellar We can find global optima in polynomial