Problem 1

Use the Cutting Plane Method to solve the following integer linear programming problem:

{maxz=2x1+x2s.t.x1+x25x1+x206x1+2x221x1,x20,and integers.\begin{cases}\max&z=2x_1+x_2\\\text{s.t.}&x_1+x_2\leq5\\&-x_1+x_2\leq0\\&6x_1+2x_2\leq21\\&x_1,x_2\geq0,\quad\text{and integers}.\end{cases}
  • Consider the relaxation problem, and adding slack variables, the optimal simplex tableau is:
cjc_j\rightarrow c1=2c_1=2 c2=1c_2=1 c3=0c_3=0 c4=0c_4=0 c5=0c_5=0
CBC_B XBX_B bb x1x_1 x2x_2 s1s_1 s2s_2 s3s_3
11 x2x_2 9/49/4 00 11 3/23/2 00 1/4-1/4
00 s2s_2 1/21/2 00 00 2-2 11 1/21/2
22 x1x_1 11/411/4 11 00 1/2-1/2 00 1/41/4
σj\sigma_j 00 00 1/2-1/2 00 1/4-1/4

The optimal solution is x(0)=(11/4,9/4)x^{(0)} = (11/4,9/4), we choose the third constraint to construct a cut:
x112s1+14s3=114x1s12=3412s114s33412s114s302s1s3+s4=3x_1-\frac{1}{2}s_1+\frac14s_3=\frac{11}4\Rightarrow x_1-s_1-2=\frac{3}{4}-\frac{1}{2}s_1-\frac{1}{4}s_3\Rightarrow\frac{3}{4}-\frac{1}{2}s_1-\frac{1}{4}s_3\leq0\Rightarrow-2s_1-s_3+s_4=-3
By applying dual simplex method, we obtain:

(1)(1) cjc_j\rightarrow c1=2c_1=2 c2=1c_2=1 c3=0c_3=0 c4=0c_4=0 c5=0c_5=0 c6=0c_6=0
CBC_B XBX_B bb x1x_1 x2x_2 s1s_1 s2s_2 s3s_3 s4s_4
11 x2x_2 9/49/4 00 11 3/23/2 00 1/4-1/4 00
00 s2s_2 1/21/2 00 00 2-2 11 1/21/2 00
22 x1x_1 11/411/4 11 00 1/2-1/2 00 1/41/4 00
00 s4s_4 3-3 00 00 [2][-2] 00 1-1 11
σj\sigma_j 00 00 1/2-1/2 00 1/4-1/4 00
(2)(2) 11 x2x_2 00 00 11 00 00 1-1 3/43/4
00 s2s_2 7/27/2 00 00 00 11 3/23/2 1-1
22 x1x_1 7/27/2 11 00 00 00 1/21/2 1/4-1/4
00 s1s_1 3/23/2 00 00 11 00 1/21/2 1/2-1/2
σj\sigma_j 00 00 00 00 00 1/4-1/4

The optimal solution is x(1)=(7/2,0)x^{(1)} = (7/2,0), we choose the third constraint to construct a cut:
x1+12s314s4=72x1s43=1212s334s41212s334s402s33s4+s5=2x_1+\frac12s_3-\frac14s_4=\frac72\Rightarrow x_1-s_4-3=\frac12-\frac12s_3-\frac{3}{4}s_4\Rightarrow\frac12-\frac12s_3-\frac{3}{4}s_4\leq0\Rightarrow-2s_3-3s_4+s_5=-2
By applying dual simplex method, we obtain:

(1)(1) cjc_j\rightarrow c1=2c_1=2 c2=1c_2=1 c3=0c_3=0 c4=0c_4=0 c5=0c_5=0 c6=0c_6=0 c7=0c_7=0
CBC_B XBX_B bb x1x_1 x2x_2 s1s_1 s2s_2 s3s_3 s4s_4 s5s_5
11 x2x_2 00 00 11 00 00 1-1 3/43/4 00
00 s2s_2 7/27/2 00 00 00 11 3/23/2 1-1 00
22 x1x_1 7/27/2 11 00 00 00 1/21/2 1/4-1/4 00
00 s1s_1 3/23/2 00 00 11 00 1/21/2 1/2-1/2 00
00 s5s_5 2-2 00 00 8-8 00 [2][-2] 3-3 11
σj\sigma_j 00 00 00 00 00 1/4-1/4 00
(2)(2) 11 x2x_2 11 00 11 00 00 00 9/49/4 1/2-1/2
00 s2s_2 22 00 00 00 11 00 13/4-13/4 3/43/4
22 x1x_1 33 11 00 00 00 00 1-1 1/41/4
00 s1s_1 11 00 00 11 00 00 5/4-5/4 1/41/4
00 s3s_3 11 00 00 00 00 11 3/23/2 1/2-1/2
σj\sigma_j 00 00 00 00 00 1/4-1/4 00

The optimal solution is x(2)=(3,1)x^{(2)}=(3,1), with optimum of 77.


Problem 2

State and prove the first-order necessary condition, second-order necessary condition, and second-order sufficient condition for the existence of extreme points inmultivariate functions.

  • First-order necessary condition: Assume that R\R is an open set of Euclidean space En\mathbb E_n, f(x)f(x) is continuously differentiable and xx^* is a local minimizer of f(x)f(x), then f(x)x1=f(x)x2==f(x)xn=0\frac{\partial f(x^*)}{\partial x_1}=\frac{\partial f(x^*)}{\partial x_2}=\cdots=\frac{\partial f(x^*)}{\partial x_n}=0.
    Proof: Suppose for contradiction that f(x)0∇f (x^∗) \neq 0. Define the vector
    p=f(x)p =−∇f(x^∗) and note that pTf(x)=f(x)2<0p^T∇f (x^∗) = −∥∇f (x^∗)∥^2 < 0. Because f∇f is continuous near xx^∗, there is a scalar T>0T > 0 such that pTf(x+tp)<0p^T∇f (x^∗ +tp) < 0, for all t[0,T]t ∈ [0,T].
    For any t(0,T]\overline t∈ (0,T], we have by Taylor’s theorem that f(x+tp)=f(x)+tpTf(x+tp)f (x^∗ +\overline tp) = f (x^∗)+ \overline tp^T∇f (x^∗ +tp), for some t(0,t)t ∈ (0,\overline t).
    Therefore, f(x+tp)<f(x)f (x^∗ + \overline tp) < f (x^∗) for all t(0,T]\overline t∈ (0,T]. We have found a direction leading away from xx^∗ along which ff decreases, so xx^∗ is not a local minimizer, and we have a contradiction.
    Second-order necessary condition: If xx^∗ is a local minimizer of ff, 2f∇^2f is continuous in an open neighborhood of xx^∗, then f(x)=0∇f(x^∗) = 0, 2f(x)∇^2f(x^∗) is positive semi-definite.
    Proof: We know from First-order necessary condition that f(x)=0∇f (x^∗) = 0. For contradiction, assume that 2f(x)∇^2f (x^∗) is not positive semidefinite. Then we can choose a vector pp such that pT2f(x)p<0p^T∇^2f (x^∗)p < 0, and because 2f∇^2f is continuous near xx^∗, there is a scalar T>0T > 0 such that pT2f(x+tp)p<0p^T∇^2f (x^∗ +tp)p < 0 for all t[0,T]t ∈ [0,T].
    By doing a Taylor series expansion around xx^∗, we have for all t(0,T]\overline t∈ (0,T] and some t(0,t)t ∈ (0,\overline t) that f(x+tp)=f(x)+tpTf(x)+12t2pT2f(x+tp)p<f(x)f (x^∗ +\overline tp) = f (x^∗)+ \overline tp^T∇f (x^∗) + \frac12 \overline t^2p^T∇^2f (x^∗ + tp)p < f (x^∗).
    As in First-order necessary condition, we have found a direction from xx^∗ along which ff is decreasing, and so again, xx^∗ is not a local minimizer.
    Second-order sufficient condition: Assume that R\R is an open set of Euclidean space En\mathbb E_n, f(x)f(x) is twice continuously differentiable in R\R. If xx^∗ is a local minimizer of ff, 2f∇^2f is continuous in an open neighborhood of xx^∗, f(x)=0∇f(x^∗) = 0 and 2f(x)∇^2f(x^∗) is positive definite, then xRx^∗ ∈ \R is a strict local minimizer of f(x)f(x).
    Proof: Because the Hessian is continuous and positive definite at xx^∗, we can
    choose a radius r>0r > 0 so that 2f(x)∇^2f(x) remains positive definite for all xx in the
    open ball D={zzx<r}D = \{z | ∥z −x^∗∥ < r \}. Taking any nonzero vector pp with p<r∥p∥ < r, we have x+pDx^∗ +p ∈ D and so f(x+p)=f(x)+pTf(x)+12pT2f(z)p=f(x)+12pT2f(z)pf (x^∗ +p) = f (x^∗)+p^T∇f (x^∗)+ \frac12p^T∇^2f(z)p= f(x^∗)+ \frac12p^T∇^2f(z)p, where z=x+tpz = x^∗ +tp for some t(0,1)t ∈ (0,1). Since zDz ∈ D, we have pT2f(z)p>0p^T∇^2f(z)p > 0, and therefore f(x+p)>f(x)f (x^∗ + p) > f (x^∗), giving the result.

Problem 3

Prove: Let f(x)f(x) be a differentiable convex function defined on the convex set RcR_c. If there exists a point xRcx^* \in R_c such that, for all xRcx \in R_c, the following condition holds:

f(x)(xx)0\nabla f(x^*)^\top(x-x^*)\geq0

then xx^* is the global minimum point of f(x)f(x) on RcR_c.

  • Using the first-order characterization of convex functions, we have that for all x,yRcx,y∈R_c, f(y)f(x)+f(x)(yx)f(y)≥f(x)+∇f(x)^⊤(y−x). And by the convexity inequality at x=xx=x^∗, we have that for any xRcx∈R_c, f(x)f(x)+f(x)(xx)f(x)≥f(x∗)+∇f(x^∗)^⊤(x−x^∗). Then consider f(x)(xx)0\nabla f(x^*)^\top(x-x^*)\geq0, we have that f(x)f(x)+f(x)(xx)f(x)f(x)≥f(x^∗)+∇f(x^∗)^⊤(x−x^∗)≥f(x^∗), which means xx^∗ is a global minimum of ff on RcR_c.

Problem 4

Consider the optimization problem:

minxRnf(x)subject to gj(x)0,j=1,,l\begin{aligned}&\min_{x\in\mathbb{R}^{n}}f(x)\\&\mathrm{subject~to~}g_{j}(x)\geq0,\quad j=1,\ldots,l\end{aligned}

This problem is called convex programming if f(x)f(x) is a convex function and all gj(x)g_j (x) are concave functions.
Question: Prove that convex programming has the following properties:

1. The feasible solution set is convex.

  • Consider feasible set F={xRngj(x)0, j=1,,l}F = \{ x \in \mathbb{R}^n \mid g_j(x) \ge 0, \ j=1,\dots,l \}. Take any two points x1,x2Fx_1, x_2 \in F, i.e., gj(x1)0g_j(x_1) \ge 0 and gj(x2)0g_j(x_2) \ge 0 for all jj. For any θ[0,1]\theta \in [0,1], consider the convex combination xθ=θx1+(1θ)x2x_\theta = \theta x_1 + (1-\theta) x_2. Since gjg_j is concave, we have that gj(xθ)=gj(θx1+(1θ)x2)θgj(x1)+(1θ)gj(x2)0g_j(x_\theta) = g_j(\theta x_1 + (1-\theta)x_2) \ge \theta g_j(x_1) + (1-\theta) g_j(x_2) \ge 0 for all jj. Therefore, xθFx_\theta \in F, so FF is convex. Also, from Property 1, xθFx_\theta \in F. Therefore, f(xθ)=ff(x_\theta) = f^* and xθXx_\theta \in X^*. Hence, the set of optimal solutions XX^* is convex.

2. The set of optimal solutions is convex.

  • Consider optimal solution set X=xFf(x)=f,f=minxFf(x)X^* = { x \in F \mid f(x) = f^* }, \quad f^* = \min_{x \in F} f(x). Take any two optimal solutions x1,x2Xx_1, x_2 \in X^*. Then f(x1)=f(x2)=ff(x_1) = f(x_2) = f^*. For any θ[0,1]\theta \in [0,1], consider xθ=θx1+(1θ)x2x_\theta = \theta x_1 + (1-\theta)x_2. Since ff is convex, we have that f(xθ)θf(x1)+(1θ)f(x2)=θf+(1θ)f=ff(x_\theta) \le \theta f(x_1) + (1-\theta) f(x_2) = \theta f^* + (1-\theta) f^* = f^*.

3. Any local optimum is also a global optimum.

  • Let xFx^* \in F be a local minimum, i.e., there exists ϵ>0\epsilon > 0 such that f(x)f(x),xF with xx<ϵf(x^*) \le f(x), \quad \forall x \in F \text{ with } |x - x^*| < \epsilon. Since ff is convex and FF is convex, the first-order condition for convex functions applies that f(x)(xx)0,xF\nabla f(x^*)^\top (x - x^*) \ge 0, \quad \forall x \in F. From the theorem we proved earlier (first-order condition implies global minimum for convex functions on a convex set), this implies xx^* is a global minimum. Therefore, every local optimum is globally optimal.

4. If the objective function is strictly convex and a global optimum exists, then it is unique.

  • Suppose ff is strictly convex and there exist two global optima x1x2x_1 \neq x_2 with f(x1)=f(x2)=ff(x_1) = f(x_2) = f^*. Consider xθ=θx1+(1θ)x2x_\theta = \theta x_1 + (1-\theta) x_2 for θ(0,1)\theta \in (0,1). Strict convexity of ff gives that f(xθ)<θf(x1)+(1θ)f(x2)=ff(x_\theta) < \theta f(x_1) + (1-\theta) f(x_2) = f^*. But ff^* is the global minimum, so f(xθ)ff(x_\theta) \ge f^*, which is a contradiction. Hence, the global minimum is unique.