lect4_removed
lect4_removed
lect4_removed
(EE227A: UC Berkeley)
Lecture 4
(Conjugates, subdifferentials)
31 Jan, 2013
◦
Suvrit Sra
Organizational
♥ HW1 due: 14th Feb 2013 in class.
♥ Please LATEX your solutions (contact TA if this is an issue)
♥ Discussion with classmates is ok
♥ Each person must submit his/her individual solutions
♥ Acknowledge any help you receive
♥ Do not copy!
♥ Make sure you understand the solution you submit
♥ Cite any source that you use
♥ Have fun solving problems!
2 / 30
Recap
I Eigenvalues, singular values, positive definiteness
I Convex sets, θ1 x + θ2 y ∈ C, θ1 + θ2 = 1, θi ≥ 0
I Convex functions, midpoint convex, recognizing convexity
I Norms, mixed-norms, matrix norms, dual norms
I Indicator, distance function, minimum of jointly convex
I Brief mention of other forms of convexity
x+y
f 2 ≤ 12 [f (x) + f (y)] + continuity =⇒ f is cvx
∇2 f (x) 0 implies f is cvx.
3 / 30
Fenchel Conjugate
4 / 30
Dual norms
uT x
Exercise: Verify that we may writekuk∗ = supx6=0 kxk
Exercise: Verify that kuk∗ is a norm.
I ku + vk∗ = sup (u + v)T x | kxk ≤ 1
5 / 30
Fenchel conjugate
Def. The Fenchel conjugate of a function f is
Example Let f (x) = kxk. We have f ∗ (z) = Ik·k∗ ≤1 (z). That is,
conjugate of norm is the indicator function of dual norm ball.
7 / 30
Fenchel conjugate
Example f (x) = ax + b; then,
f ∗ (z) = sup zx − (ax + b)
x
= ∞, if (z − a) 6= 0.
7 / 30
Subdifferentials
8 / 30
First order global underestimator
f (x) i
−y
), x
h∇ f (y
)+
f (y
f (y)
y x
9 / 30
First order global underestimator
f (x)
f (y) x − yi
+ hg 1,
f (y )
g1
g2
g3
y
f (x) ≥ f (y) + hg, x − yi
9 / 30
Subgradients
f (x)
f (y) x − yi
+ hg 1,
f (y )
g1
g2
g3
y
g1 , g2 , g3 are subgradients at y
10 / 30
Subgradients – basic facts
I f is convex, differentiable: ∇f (y) the unique subgradient at y
I A vector g is a subgradient at a point y if and only if
f (y) + hg, x − yi is globally smaller than f (x).
I Usually, one subgradient costs approx. as much as f (x)
I Determining all subgradients at a given point — difficult.
I Subgradient calculus—a great achievement in convex analysis
I Without convexity, things become wild! — advanced course
11 / 30
Subgradients – example
f (x) f1 (x)
f2 (x)
y
? f1 (x) > f2 (x): unique subgradient of f is f10 (x)
? f1 (x) < f2 (x): unique subgradient of f is f20 (x)
? f1 (y) = f2 (y): subgradients, the segment [f10 (y), f20 (y)]
(imagine all supporting lines turning about point y)
12 / 30
Subgradients
13 / 30
Subdifferential
14 / 30
Subdifferential – example
f (x) = |x|
∂f (x)
+1
x
−1
−1
x < 0,
∂|x| = +1 x > 0,
[−1, 1] x = 0.
15 / 30
More examples
Proof.
kzk2 ≥ kxk2 + hg, z − xi
kzk2 ≥ hg, zi
=⇒ kgk2 ≤ 1.
16 / 30
More examples
17 / 30
Calculus
18 / 30
Recall basic calculus
If f and k are differentiable, we know that
Addition: ∇(f + k)(x) = ∇f (x) + ∇k(x)
Scaling: ∇(αf (x)) = α∇f (x)
Chain rule
If f :Rn → Rm , and k : Rm → Rp . Let h : Rn → Rp be the
composition h(x) = (k ◦ f )(x) = k(f (x)). Then,
Dh(x) = Dk(f (x))Df (x).
19 / 30
Subgradient calculus
♠ Finding one subgradient within ∂f (x)
♠ Determining entire subdifferential ∂f (x) at a point x
♠ Do we have the chain rule?
♠ Usually not easy!
20 / 30
Subgradient calculus
H
If f is differentiable, ∂f (x) = {∇f (x)}
H
Scaling α > 0, ∂(αf )(x) = α∂f (x) = {αg | g ∈ ∂f (x)}
Addition∗ : ∂(f + k)(x) = ∂f (x) + ∂k(x) (set addition)
H
21 / 30
Examples
22 / 30
Examples
23 / 30
Example (S. Boyd)
(−1, 1)
+1 (1, 1)
+1
−1 −1
(1, −1)
24 / 30
Rules for subgradients
25 / 30
Subgradient for pointwise sup
f (x) := sup h(x, y)
y∈Y
h(z, y ∗ ) ≥ h(x, y ∗ ) + g T (z − x)
h(z, y ∗ ) ≥ f (x) + g T (z − x)
f (z) ≥ h(z, y) (because of sup)
T
f (z) ≥ f (x) + g (z − x).
26 / 30
Example
Suppose ai ∈ Rn and bi ∈ R. And
27 / 30
Subgradient of expectation
Suppose f = Ef (x, u), where f is convex in x for each u (an r.v.)
Z
f (x) := f (x, u)p(u)du
28 / 30
Subgradient of composition
Suppose h : Rn → R cvx and nondecreasing; each fi cvx
29 / 30
References
1 R. T. Rockafellar. Convex Analysis
2 S. Boyd (Stanford); EE364b Lecture Notes.
30 / 30