Using Schur Complement Theorem to prove convexity of some SOC

Journal of Nonlinear and Convex Analysis, vol. 13, no. 3, pp. 421-431, 2012
Using Schur Complement Theorem to prove convexity of some
SOC-functions
Jein-Shan Chen 1
Department of Mathematics
National Taiwan Normal University
Taipei 11677, Taiwan
E-mail: [email protected]
Tsun-Ko Liao
Department of Mathematics
National Taiwan Normal University
Taipei 11677, Taiwan
E-mail: [email protected]
Shaohua Pan 2
Department of Mathematics
South China University of Technology
Guangzhou 510640, China
March 10, 2012
Abstract. In this paper, we provide an important application of the Schur Complement
Theorem in establishing convexity of some functions associated with second-order cones
(SOCs), called SOC-trace functions. As illustrated in the paper, these functions play a
key role in the development of penalty and barrier functions methods for second-order
cone programs, and establishment of some important inequalities associated with SOCs.
Key words. Second-order cone, SOC-functions, convexity, positive semidefiniteness
AMS subject classifications. 26A27, 26B05, 26B35, 49J52, 90C33.
1
Member of Mathematics Division, National Center for Theoretical Sciences, Taipei Office. The
author’s work is supported by National Science Council of Taiwan.
2
The author’s work is supported by National Young Natural Science Foundation (No. 10901058) and
Guangdong Natural Science Foundation (No. 9251802902000001). E-mail: [email protected].
1
1
Introduction
The second-order cone (SOC) in IRn , also called Lorentz cone, is a set defined as
n
o
Kn := (x1 , x2 ) ∈ IR × IRn−1 | x1 ≥ kx2 k ,
(1)
where k·k denotes the Euclidean norm. When n = 1, Kn reduces to the set of nonnegative
real numbers IR+ . As shown in [12], Kn is also a set composed of the squared elements
from Jordan algebra (IRn , ◦), where the Jordan product “◦” is a binary operation defined
by
x ◦ y := (hx, yi, x1 y2 + y1 x2 )
(2)
for any x = (x1 , x2 ), y = (y1 , y2 ) ∈ IR × IRn−1 . Unless otherwise stated, in the rest of
this note, we use e = (1, 0, . . . , 0)T ∈ IRn to denote the unit element of Jordan algebra
(IRn , ◦), i.e., e ◦ x = x for any x ∈ IRn , and for any x ∈ IRn , use x1 to denote the first
component of x, and x2 to denote the vector consisting of the rest n − 1 components.
From [11, 12], we recall that each x ∈ IRn admits a spectral factorization associated
with Kn , of the following form
(2)
x = λ1 (x)u(1)
x + λ2 (x)ux ,
(3)
(i)
where λi (x) and ux for i = 1, 2 are the spectral values and the associated spectral vectors
of x, respectively, defined by
1
i
i
(i)
(4)
1, (−1) x̄2 ,
λi (x) = x1 + (−1) kx2 k, ux =
2
with x̄2 = kxx22 k if x2 6= 0, and otherwise x̄2 being any vector in IRn−1 such that kx̄2 k = 1.
When x2 6= 0, the spectral factorization is unique. The determinant and trace of x are
defined as det(x) := λ1 (x)λ2 (x) and tr(x) := λ1 (x) + λ2 (x), respectively.
With the spectral factorization above, for any given scalar function f : J ⊆ IR → IR,
we may define a vector-valued function f soc : S ⊆ IRn → IRn by
(2)
f soc (x) := f (λ1 (x))u(1)
x + f (λ2 (x))ux
(5)
where J is an interval (finite or infinite, open or closed) of IR, and S is the domain of f soc
determined by f . Obviously, f soc is well-defined whether x2 = 0 or not. Assume that f
is differentiable on intJ. Then, by Lemma 2.2 in Section 2, f soc is differentiable on intS,
and moreover, for any x ∈ intS, ∇f soc (x) has a structure which entails the application of
the Schur Complement Theorem in establishing its positive semidefiniteness or positive
definiteness. On the other hand, Lemma 2.2 shows that the SOC-trace function
f tr (x) := f (λ1 (x)) + f (λ2 (x)) = tr(f soc (x)) ∀x ∈ S
2
(6)
is differentiable on intS with ∇f tr (x) = (f 0 )soc (x) for any x ∈ intS, which means that
∇2 f tr (x) = ∇(f 0 )soc (x)
∀x ∈ intS.
(7)
The two sides show that we may establish the convexity of f tr with the help of the Schur
Complement Theorem. In fact, the Schur Complement Theorem is frequently used in the
topics related to semidefinite programming (SDP), but there is no paper to introduce its
application involving SOCs, to the best of our knowledge.
Motivated by this, in this paper we provide such an application of the Schur Complement Theorem in establishing the convexity of SOC-trace functions and the compounds
of SOC-trace functions and real-valued functions. As illustrated in the next section, some
of these functions are the key of penalty and barrier function methods for second-order
cone programs (SOCPs), as well as the establishment of some important inequalities
associated with SOCs, for which the proof of convexity of these functions is a necessity. But, this requires computation of the first and second-order derivatives, which is
technically much more demanding than in the linear and semidefinite cases. As will be
seen, Theorem 2.1 gives a simple way to achieve this objective via the Schur Complement
Theorem, by which one only needs to check the sign of the second-order derivative of a
scalar function.
Throughout this note, for any x, y ∈ IRn , we write x Kn y if x − y ∈ Kn ; and write
x Kn y if x − y ∈ intKn . For a real symmetric matrix A, we write A 0 (respectively,
A 0) if A is positive semidefinite (respectively, positive definite). For any f : J → IR,
f 0 (t) and f 00 (t) denote the first derivative and second-order derivative of f at the differentiable point t ∈ J, respectively; for any F : S ⊆ IRn → IR, ∇F (x) and ∇2 F (x) denote
the gradient and the Hessian matrix of F at the differentiable point x ∈ S, respectively.
2
Main results
The Schur Complement Theorem gives a characterization for the positive semidefiniteness
(definiteness) of a matrix via the positive semidefiniteness (definiteness) of the Schurcomplement with respect to a block partitioning of the matrix, which is stated as below.
Lemma 2.1 [13] (Schur Complement Theorem) Let A ∈ IRm×m be a symmetric positive
definite matrix, C ∈ IRn×n be a symmetric matrix, and B ∈ IRm×n . Then,
A B
0 ⇐⇒ C − B T A−1 B 0
(8)
BT C
and
A B
BT C
0
C − B T A−1 B 0.
⇐⇒
3
(9)
In this section, we focus on on the application of the Schur Complement Theorem in
establishing convexity of SOC-trace functions. To the end, we need the following lemma.
Lemma 2.2 For any given f : J ⊆ IR → IR, let f soc : S → IRn and f tr : S → IR be given
by (5) and (6), respectively. Assume that J is open. Then, the following results hold.
(a) The domain S of f soc and f tr is also open.
(b) If f is (continuously) differentiable on J, then f soc is (continuously) differentiable
on S. Moreover, for any x ∈ S, ∇f soc (x) = f 0 (x1 )I if x2 = 0, and otherwise


xT2
c(x)
 b(x)

kx2 k
soc


∇f (x) = 
(10)
T ,
x2 x2
x2
a(x)I + (b(x) − a(x))
c(x)
kx2 k
kx2 k2
where
b(x) =
f 0 (λ2 (x))− f 0 (λ1 (x))
f (λ2 (x))− f (λ1 (x))
f 0 (λ2 (x))+ f 0 (λ1 (x))
, c(x) =
, a(x) =
.
2
2
λ2 (x) − λ1 (x)
(c) If f is (continuously) differentiable, then f tr is (continuously) differentiable on S
with ∇f tr (x) = (f 0 )soc (x); if f is twice (continuously) differentiable, then f tr is
twice (continuously) differentiable on S with ∇2 f tr (x) = ∇(f 0 )soc (x).
Proof. (a) Fix any x ∈ S. Then λ1 (x), λ2 (x) ∈ J. Since J is an open subset of IR,
there exist δ1 , δ2 > 0 such that {t√∈ IR| |t − λ1 (x)| < δ1 } ⊆ J and {t ∈ IR| |t − λ2 (x)| <
δ2 } ⊆ J. Let δ := min{δ1 , δ2 }/ 2. Then, for any y satisfying ky − xk < δ, we have
|λ1 (y) − λ1 (x)| < δ1 and |λ2 (y) − λ2 (x)| < δ2 by noting that
(λ1 (x) − λ1 (y))2 + (λ2 (x) − λ2 (y))2
= 2(x21 + kx2 k2 ) + 2(y12 + ky2 k2 ) − 4(x1 y1 + kx2 kky2 k)
≤ 2(x21 + kx2 k2 ) + 2(y12 + ky2 k2 ) − 4(x1 y1 + hx2 , y2 i)
= 2 kxk2 + kyk2 − 2hx, yi = 2kx − yk2 ,
and consequently λ1 (y) ∈ J and λ2 (y) ∈ J. Since f is a function from J to IR, this means
that {y ∈ IRn | ky − xk < δ} ⊆ S, and therefore the set S is open.
(b) The proof is direct by using the same arguments as those of [9, Props. 4 and 5].
(c) If f is (continuously) differentiable, then from part (b) and f tr (x) = eT f soc (x) it
follows that f tr is (continuously) differentiable. In addition, a simple computation yields
that ∇f tr (x) = ∇f soc (x)e = (f 0 )soc (x). Similarly, by part (b), the second part follows.
2
4
Theorem 2.1 For any given f : J → IR, let f soc : S → IRn and f tr : S → IR be given by
(5) and (6), respectively. Assume that J is open. If f is twice differentiable on J, then
(a) f 00 (t) ≥ 0 for any t ∈ J ⇐⇒ ∇(f 0 )soc (x) 0 for any x ∈ S ⇐⇒ f tr is convex in S.
(b) f 00 (t) > 0 for any t ∈ J ⇐⇒ ∇(f 0 )soc (x) 0 ∀x ∈ S =⇒ f tr is strictly convex in S.
Proof. (a) By Lemma 2.2(c), ∇2 f tr (x) = ∇(f 0 )soc (x) for any x ∈ S, and the second
equivalence follows by Prop. B.4(a) and (c) of [4]. We next come to the first equivalence.
By Lemma 2.2(b), for any fixed x ∈ S, ∇(f 0 )soc (x) = f 00 (x1 )I if x2 = 0, and otherwise
∇(f 0 )soc (x) has the same expression as in (10) except that
b(x) =
f 00 (λ2 (x))+ f 00 (λ1 (x))
f 00 (λ2 (x))− f 00 (λ1 (x))
f 0 (λ2 (x))− f 0 (λ1 (x))
, c(x) =
, a(x) =
.
2
2
λ2 (x) − λ1 (x)
Assume that ∇(f 0 )soc (x) 0 for any x ∈ S. Then, we readily have b(x) ≥ 0 for any
x ∈ S. Noting that b(x) = f 00 (x1 ) when x2 = 0, we particularly have f 00 (x1 ) ≥ 0 for all
x1 ∈ J, and consequently f 00 (t) ≥ 0 for all t ∈ J. Assume that f 00 (t) ≥ 0 for all t ∈ J. Fix
any x ∈ S. Clearly, b(x) ≥ 0 and a(x) ≥ 0. If b(x) = 0, then f 00 (λ1 (x)) = f 00 (λ2 (x)) = 0,
and consequently c(x) = 0, which in turn implies that
"
#
0
0
0.
∇(f 0 )soc (x) =
(11)
x xT
0 a(x) I − kx22 k22
If b(x) > 0, then by the first equivalence of Lemma 2.1 and the expression of ∇(f 0 )soc (x)
it suffices to argue that the following matrix
a(x)I + (b(x) − a(x))
c2 (x) x2 xT2
x2 xT2
−
kx2 k2
b(x) kx2 k2
(12)
is positive semidefinite. Since the rank-one matrix x2 xT2 has only one nonzero eigenvalue
kx2 k2 , the matrix in (12) has one eigenvalue a(x) of multiplicity n − 1 and one eigenvalue
2 −c(x)2
b(x)2 −c(x)2
of multiplicity 1. Since a(x) ≥ 0 and b(x)b(x)
= f 00 (λ1 (x))f 00 (λ2 (x)) ≥ 0, the
b(x)
matrix in (12) is positive semidefinite. By the arbitrary of x, we have that ∇(f 0 )soc (x) 0
for all x ∈ S.
(b) The first equivalence is direct by using (9) of Lemma (2.1), noting ∇(f 0 )soc (x) 0
implies a(x) > 0 when x2 6= 0, and following the same arguments as part (a). The second
part is due to [4, Prop. B.4(b)].
2
Remark 2.1 Note that the strict convexity of f tr does not necessarily imply the positive
definiteness of ∇2 f tr (x). Consider f (t) = t4 for t ∈ IR. We next show that f tr is strictly
5
convex. Indeed, f tr is convex in IRn by Theorem 2.1(a) since f 00 (t) = 12t2 ≥ 0. Taking
into account that f tr is continuous, it remains to prove that
f tr (x) + f tr (y)
x+y
tr
f
=
=⇒ x = y.
(13)
2
2
Since h(t) = (t0 + t)4 + (t0 − t)4 for some t0 ∈ IR is increasing on [0, +∞), and the
function f (t) = t4 is strictly convex in IR, we have that
f
tr
x+y
2
4
x+y
λ1
+ λ2
2
4 4
x1 + y1 − kx2 + y2 k
x1 + y1 + kx2 + y2 k
+
2
2
4 4
x1 + y1 − kx2 k − ky2 k
x1 + y1 + kx2 k + ky2 k
+
2
2
4 4
λ1 (x) + λ1 (y)
λ2 (x) + λ2 (y)
+
2
2
4
4
4
(λ1 (x)) + (λ1 (y)) + (λ2 (x)) + (λ2 (y))4
2
f tr (x) + f tr (y)
,
2
=
=
≤
=
≤
=
x+y
2
4
and moreover, the above inequalities become the equalities if and only if
kx2 + y2 k = kx2 k + ky2 k, λ1 (x) = λ1 (y), λ2 (x) = λ2 (y).
It is easy to verify that the three equalities hold if and only if x = y. So, the implication
in (13) holds, i.e., f tr is strictly convex. However, by Theorem 2.1(b), ∇(f 0 )soc (x) 0
does not hold for all x ∈ IRn since f 00 (t) > 0 does not hold for all t ∈ IR.
It should be mentioned that the fact that the strict convexity of f implies the strict
convexity of f tr was proved in [2, 8] via the definition of convex function, but here we use
the Shcur Complement Theorem and the relation between ∇(f 0 )soc and ∇2 f tr to establish
the convexity of SOC-trace functions. In addition, we also note that the necessity involved
in the first equivalence of Theorem 2.1(a) was given in [11] via a different way. Next, we
illustrate the application of Theorem 2.1 with some SOC-trace functions.
Proposition 2.1 The following functions associated with Kn are all strictly convex.
(a) F1 (x) = − ln(det(x)) for x ∈ intKn .
(b) F2 (x) = tr(x−1 ) for x ∈ intKn .
6
(c) F3 (x) = tr(φ(x)) for x ∈ intKn , where
( p+1
1−q
x
−e
+ x q−1−e if p ∈ [0, 1], q > 1;
p+1
φ(x) =
xp+1 −e
− ln x if p ∈ [0, 1], q = 1.
p+1
(d) F4 (x) = − ln(det(e − x)) for x ≺Kn e.
(e) F5 (x) = tr((e − x)−1 ◦ x) for x ≺Kn e.
(f ) F6 (x) = tr(exp(x)) for x ∈ IRn .
(g) F7 (x) = ln(det(e + exp(x))) for x ∈ IRn .
x + (x2 + 4e)1/2
for x ∈ IRn .
(h) F8 (x) = tr
2
Proof. Note that F1 (x), F2 (x) and F3 (x) are the SOC-trace functions associated with
f1 (t) = − ln t (t > 0), f2 (t) = t−1 (t > 0) and f3 (t) (t > 0), respectively, where
( p+1
1−q
t
−1
+ t q−1−1 if p ∈ [0, 1], q > 1,
p+1
f3 (t) =
tp+1 −1
− ln t if p ∈ [0, 1], q = 1;
p+1
F4 (x) is the SOC-trace function associated with f4 (t) = − ln(1 − t) (t < 1), F5 (x) is the
t
SOC-trace function associated with f5 (t) = 1−t
(t < 1) by noting that
(e − x)−1 ◦ x =
λ2 (x) (2)
λ1 (x) (1)
ux +
u ;
λ1 (e − x)
λ2 (e − x) x
F6 (x) and F7 (x) are the SOC-trace functions associated with f6 (t) = exp(t) (t ∈ IR)
and f7 (t) = ln(1 + exp(t)) (t ∈ √IR), respectively,
and F8 (x) is the SOC-trace function
associated with f7 (t) = 2−1 t + t2 + 4 (t ∈ IR). It is easy to verify that the functions
f1 -f8 have positive second-order derivatives in their respective domain, and therefore
F1 -F8 are strictly convex functions by Theorem 2.1(b).
2
The functions F1 , F2 and F3 are the popular barrier functions which play a key role in
the development of interior point methods for SOCPs, see, e.g., [6, 7, 14, 15, 17], where
F3 covers a wide range of barrier functions, including the classical logarithmic barrier
function, the self-regular functions and the non-self-regular functions; see [7] for details.
The functions F4 and F5 are the popular shifted barrier functions [1, 2, 3] for SOCPs, and
F6 -F8 can be used as penalty functions for second-order cone programs (SOCPs), and
these functions are added to the objective of SOCPs for forcing the solution to be feasible.
Besides the application in establishing convexity for SOC-trace functions, the Schur
complement theorem can be employed to establish convexity of some compound functions
of SOC-trace functions and scalar-valued functions, which is usually difficult to achieve by
the definition of convex function. The following proposition presents such an application.
7
Proposition 2.2 For any x ∈ Kn , let F9 (x) := −[det(x)]1/p with p > 1. Then,
(a) F9 is twice continuously differentiable in intKn .
(b) F9 is convex when p ≥ 2, and moreover, it is strictly convex when p > 2.
Proof. (a) Note that −F9 (x) = exp (p−1 ln(det(x))) for any x ∈ intKn , and ln(det(x)) =
f tr (x) with f (t) = ln(t) for t ∈ IR++ . By Lemma 2.2(c), ln(det(x)) is twice continuously
differentiable in intKn . Hence −F9 (x) is twice continuously differentiable in intKn . The
result then follows.
(b) In view of the continuity of F9 , we only need to prove its convexity over intKn . By
part (a), we next achieve this goal by proving that the Hessian matrix ∇2 F9 (x) for any
x ∈ intKn is positive semidefinite when p ≥ 2, and positive definite when p > 2. Fix any
x ∈ intKn . From direct computations, we obtain
"
1 −1 #
1
(2x1 ) x21 − kx2 k2 p
∇F9 (x) = −
1
p (−2x2 ) x21 − kx2 k2 p −1
and

∇2 F9 (x) =
1
p−1
(det(x)) p −2 
2
p
4x21
−
2p(x21 −kx2 k2 )
p−1
−4x1 x2
−4x1 xT2
2p(x21 −kx2 k2 )
4x2 xT2 +
I
p−1

.
(14)
Since x ∈ intKn , we have x1 > 0 and det(x) = x21 − kx2 k2 > 0, and therefore
2p (x21 − kx2 k2 )
2p
2p
2
a1 (x) := 4x1 −
= 4−
x21 +
kx2 k2 .
p−1
p−1
p−1
We next proceed the arguments by the following two cases: (1) a1 (x) = 0; (2) a1 (x) > 0.
Case 1: a1 (x) = 0. Since p ≥ 2, under this case we must have x2 = 0, and consequently,
2
0
0
p−1
−4
2
∇ F9 (x) =
(x1 ) p
0.
2p 2
0 p−1
x1 I
p2
Case 2: a1 (x) > 0. Under this case, we calculate that
2p (x21 − kx2 k2 )
2p (x21 − kx2 k2 )
2
T
4x1 −
4x2 x2 +
I − 16x21 x2 xT2
p−1
p−1
2
2
4p (x1 − kx2 k ) p − 2 2
p
2
T
=
x I+
kx2 k I − 2x2 x2 .
p−1
p−1 1
p−1
(15)
Since the rank-one matrix 2x2 xT2 has only one nonzero eigenvalue 2kx2 k2 , the matrix in
the bracket of the right hand side of (15) has one eigenvalue of multiplicity 1 given by
p−2 2
p
p−2 2
x1 +
kx2 k2 − 2kx2 k2 =
x1 − kx2 k2 ≥ 0,
p−1
p−1
p−1
8
p
and one eigenvalue of multiplicity n − 1 given by p−2
x2 + p−1
kx2 k2 ≥ 0. Furthermore, we
p−1 1
see that these eigenvalues must be positive when p > 2 since x21 > 0 and x21 − kx2 k2 >
0. This means that the matrix on the right hand side of (15) is positive semidefinite,
and moreover, it is positive definite when p > 2. Applying Lemma 2.1, we have that
∇2 F9 (x) 0, and furthermore ∇2 F9 (x) 0 when p > 2.
Since a1 (x) > 0 must hold when p > 2, the arguments above show that F9 (x) is
convex over intKn when p ≥ 2, and strictly convex over intKn when p > 2.
2
It is worthwhile to point out that det(x) is neither convex nor concave on Kn , and it
is difficult to argue the convexity of those compound functions involving det(x) by the
definition of convex function. But, as shown in Proposition 2.2, the Schur Complement
Theorem offers a simple way to prove their convexity.
To close section, we take a look at the application of some of convex functions above
in establishing inequalities associated with SOCs. Some of these inequalities have been
used to analyze the properties of SOC-function f soc [10] and the convergence of interior
point methods for SOCPs [2].
Proposition 2.3 For any x Kn 0 and y Kn 0, the following inequalities hold.
(a) det(αx + (1 − α)y) ≥ (det(x))α (det(y))1−α for any 0 < α < 1.
2
(b) det(x + y)1/p ≥ 2 p −1 det(x)1/p + det(y)1/p for any p ≥ 2.
(c) det(x + y) ≥ det(x) + det(y).
(d) det(αx + (1 − α)y) ≥ α2 det(x) + (1 − α)2 det(y) for any 0 < α < 1.
(e) [det(e + x)]1/2 ≥ 1 + det(x)1/2 .
(f ) If x Kn y, then det(x) ≥ det(y).
1
1/2
tr(x ◦ y) : det(y) = 1, y Kn 0 . Furthermore, when x Kn 0,
(g) det(x)
= inf
2
the same relation holds with inf replaced by min.
(h) tr(x ◦ y) ≥ 2det(x)1/2 det(y)1/2 .
Proof. (a) From Prop. 2.1(a), we know that ln(det(x)) is strictly concave in intKn .
Thus,
ln(det(αx + (1 − α)y)) ≥ α ln(det(x)) + (1 − α)det(y)
= ln(det(x)α ) + ln(det(x)1−α )
9
for any 0 < α < 1 and x, y ∈ intKn . This, together with the increasing of ln t (t > 0)
and the continuity of det(x), implies the desired result.
(b) By Prop. 2.2(b), det(x)1/p is concave over Kn . So, for any x, y ∈ Kn , we have that
1/p
x+y
1
det
≥
det(x)1/p + det(y)1/p
2
2
"
#1/p
2 x2 + y2 2
1/p
1/p
x1 + y 1
≥ x21 − kx2 k2
+ y12 − ky2 k2
⇐⇒ 2
−
2
2
1
1/p i
1/p
4p h 2
2
2 1/p
+ y12 − ky2 k2
≥
⇐⇒ (x1 + y1 ) − kx2 + y2 k
x1 − kx2 k2
2
2
−1
1/p
1/p
p
⇐⇒ det(x + y) ≥ 2
det(x) + det(y)1/p ,
which is the desired result.
(c) Using the inequality in part (b) with p = 2, we have
det(x + y)1/2 ≥ det(x)1/2 + det(y)1/2 .
Squaring both sides yields
det(x + y) ≥ det(x) + det(y) + 2det(x)1/2 det(y)1/2 ≥ det(x) + det(y),
where the last inequality is by the nonnegativity of det(x) and det(y) since x, y ∈ Kn .
(d) The inequality is direct by part (c) and the fact det(αx) = α2 det(x).
(e) The inequality follows from part (b) with p = 2 and the fact that det(e) = 1.
(f) Using part (c) and noting that x Kn y, it is easy to verify that
det(x) = det(y + x − y) ≥ det(y) + det(x − y) ≥ det(y).
(g) Using the Cauchy-Schwartz inequality, it is easy to verify that
tr(x ◦ y) ≥ λ1 (x)λ2 (y) + λ1 (y)λ2 (x)
∀x, y ∈ IRn .
For any x, y ∈ Kn , this along with the arithmetic-geometric mean inequality implies that
λ1 (x)λ2 (y) + λ1 (y)λ2 (x)
tr(x ◦ y)
≥
2
2
p
≥
λ1 (x)λ2 (y)λ1 (y)λ2 (x)
= det(x)1/2 det(y)1/2 ,
1
tr(x ◦ y) : det(y) = 1, y Kn 0 = det(x)1/2 for a fixed x ∈
which means that inf
2
−1
Kn . If x Kn 0, then we can verify that the feasible point y ∗ = √x
is such that
det(x)
1
tr(x ◦ y ∗ ) = det(x)1/2 , and the second part follows.
2
10
(h) Using part (g), for any x ∈ Kn and y ∈ int(Kn ), we have that
!
p
tr(x ◦ y)
y
p
= tr x ◦ p
≥ det(x),
2 det(y)
det(y)
which together with the continuity of det(x) and tr(x) implies that
tr(x ◦ y) ≥ 2det(x)1/2 det(y)1/2
Thus, we complete the proof.
∀x, y ∈ Kn .
2
Note that some of the inequalities in Prop. 2.3 were ever established with the help of
the Schwartz-inequality [10], but here we achieve the goal easily by using the convexity
of SOC-functions. These inequalities all have the corresponding counterparts for matrix
inequalities [5, 13, 16]. For example, Prop. 2.3(b) with p = 2, i.e., p equal to the rank of
Jordan algebra (IRn , ◦), corresponds to the Minkowski inequality of matrix case:
det(A + B)1/n ≥ det(A)1/n + det(B)1/n
for any n × n positive semidefinite matrices A and B.
3
Conclusions
We studied an application of the Schur complement theorem in establishing convexity
of SOC-functions, especially for SOC-trace functions, which are the key of penalty and
barrier function methods for SOCPs and some important inequalities associated with
SOCs. One possible future direction is proving self-concordancy for such barrier/penalty
functions associated with SOCs. We also believe that the results in this paper will be
helpful towards establishing further properties of other SOC-functions.
References
[1] A. Auslender, Penalty and barrier methods: a unified framework, SIAM Journal
on Optimization, vol. 10, pp. 211-230, 1999.
[2] A. Auslender, Variational inequalities over the cone of semidefinite positive symmetric matrices and over the Lorentz cone, Optimization Methods and Software, vol.
18, pp. 359-376, 2003.
[3] A. Auslender and H. Ramirez, Penalty and barrier methods for convex semidefinite programming, Mathematical Methods of Operations Research, vol. 63, pp. 195219, 2006.
11
[4] D. P. Bertsekas, Nonlinear Programming, 2nd edition, Athena Scientific, 1999.
[5] R. Bhatia, Matrix Analysis, Springer-Verlag, New York, 1997.
[6] A. Ben-Tal and A. Nemirovski, Lectures on Modern Convex Optimization: Analysis, Algorithms and Engineering Applications, MPS-SIAM Series on Optimization.
SIAM, Philadelphia, USA, 2001.
[7] Y.-Q. Bai and G. Q. Wang, Primal-dual interior-point algorithms for second-order
cone optimization based on a new parametric kernel function, Acta Mathematica
Sinica, vol. 23, pp. 2027-2042, 2007.
[8] H. Bauschke, O. Güler, A. S. Lewis and S. Sendow, Hyperbolic polynomial
and convex analysis, Canadian Journal of Mathematics, vol. 53, pp. 470–488, 2001.
[9] J.-S. Chen, X. Chen and P. Tseng, Analysis of nonsmooth vector-valued functions associated with second-oredr cone, Mathmatical Programming, vol. 101, pp.
95-117, 2004.
[10] J.-S. Chen, The convex and monotone functions associated with second-order cone,
Optimization, vol. 55, pp. 363-385, 2006.
[11] M. Fukushima, Z.-Q Luo and P. Tseng, Smoothing functions for second-order
cone complementarity problems, SIAM Journal on Optimazation, vol. 12, pp. 436-460,
2002.
[12] J. Faraut and A. Korányi, Analysis on symmetric Cones, Oxford Mathematical
Monographs, Oxford University Press, New York, 1994.
[13] R. A. Horn and C. R. Johnson, Matrix Analysis, Cambridge University Press,
Cambridge, 1986.
[14] R. D. C. Monteiro and T. Tsuchiya, Polynomial convergence of primal-dual
algorithms for the second-order cone programs based on the MZ-family of directions,
Mathematical Programming, vol. 88, pp. 61–83, 2000.
[15] J. Peng, C. Roos and T. Terlaky, Self-Regularity, A New Paradigm for PrimalDual Interior-Point Algorithms, Princeton University Press, 2002.
[16] J. Schott, Matrix Analysis for Statistics, John Wiley and Sons, 2nd edition, 2005.
[17] T. Tsuchiya, A convergence analysis of the scaling-invariant primal-dual pathfollowing algorithms for second-order cone programming, Optimization Methods and
Software, vol. 11, pp. 141–182, 1999.
12