Solving Continuous Min-Max Problems by an Iterative

JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS: Vol. 121, No. 3, pp. 597–612, June 2004 (Ó 2004)
Solving Continuous Min-Max Problems by an
Iterative Entropic Regularization Method1,2
R. L. SHEU3
AND
J. Y. LIN4
Communicated by P. Tseng
Abstract. We propose a method of outer approximations, with each
approximate problem smoothed using entropic regularization, to solve
continuous min-max problems. By using a well-known uniform error
estimate for entropic regularization, convergence of the overall method
is shown while allowing each smoothed problem to be solved inexactly.
In the case of convex objective function and linear constraints, an
interior-point algorithm is proposed to solve the smoothed problem
inexactly. Numerical examples are presented to illustrate the behavior of
the proposed method.
Key Words. Min-max problems, entropic regularization, interior-point
algorithms, semi-infinite programming.
1. Introduction
Consider the following continuous min-max problem (P):
min FðuÞ;
u2W
where W is a compact subset of Rn ,
FðuÞ ¼ maxfft ðuÞg;
t2T
T is a compact metric space, and ft ðuÞ is continuous on T W. Notice that
T may be a finite set, which is a compact metric space in the discrete
1
This research work was partially supported by the National Science Council of Taiwan under
Project NSC 89-2115-M-006-009.
2
The authors thank Professor Paul Tseng and two anonymous referees for valuable comments
and suggestions.
3
Professor, Department of Mathematics, National Cheng-Kung University, Tainan, Taiwan.
4
PhD Candidate, Department of Mathematics, National Cheng-Kung University, Tainan,
Taiwan.
597
0022-3239/04/0600-0597/0 Ó 2004 Plenum Publishing Corporation
598
JOTA: VOL. 121, NO. 3, JUNE 2004
topology. In general, finite min-max problems have been studied more
extensively than continuous min-max problems.
A classical approach to solving problem (P) involves approximating T
by a finite subset Tm of m points in T. Correspondingly, F(u) is
approximated by
F m ðuÞ ¼ maxfft ðuÞg:
t2Tm
m
Since F is not differentiable, various smooth approximations Fpm ; p > 0;
with the property that
lim Fpm ðuÞ ¼ F m ðuÞ;
for all u 2 W;
p#0
have been proposed; (see e.g. Refs. 1–3). We will focus on entropic
regularization/smoothing,
(
)
X
m
Fp ðuÞ ¼ ð1=pÞ log
expfpft ðuÞg ;
ð1Þ
t 2 Tm
which admits a well-known uniform error estimate (Refs. 4–5)
0 Fpm ðuÞ F m ðuÞ logðmÞ=p:
ð2Þ
A related approach, proposed by Wu and Fang (Ref. 6), smooths directly
FðuÞ using an integral analog of (1),
Z
Fp ðuÞ ¼ ð1=pÞ log
expðpft ðuÞÞdt :
ð3Þ
T
However, due to the lack of a uniform error estimate such as (2),
convergence has been proven under the assumptions that fft ðuÞg are both
uniformly bounded and superuniformly continuous. These assumptions are
restrictive, since they require effectively the members of fft ðuÞg to behave
alike. Moreover, introducing an integral into the objective function
increases the computational effort.
Another approach, also involving an integration, was due to Polak et al.
(Ref. 7) who proposed a reciprocal barrier function,
Z
pðu; nk Þ ¼
dt=½nk ft ðuÞ;
T¼½0;1
where nk is an estimate of the optimal objective value at the kth iteration. By
minimizing pðu; nk Þ over the domain
Cðnk Þ ¼ fu 2 Rn jFðuÞ < nk g;
their method computes a u kþ1 whose average ft value is minimum. To ensure
that pðu; nk Þ ! 1 as u approaches a boundary point of Cðnk Þ, a Lipschitz
condition on ft in terms of t, uniformly with respect to any compact subset
JOTA: VOL. 121, NO. 3, JUNE 2004
599
in the u-space, is assumed. However, such a uniform Lipschitz condition
could fail to hold p
even
ffiffi on simple examples. For example,
min max u t
u 2 ½0;1 t 2 ½0;1
pffiffi
has an optimal value 0, but ft ðuÞ ¼ u t is not Lipschitz continuous in
t 2 ½0; 1 for any u.
In general, it seems that additional assumptions on ft ; t 2 T, are
unavoidable if one works at each iteration with the full set of functions
fft gt 2 T . In contrast, the method of outer approximations (Ref. 8), which
reduces problem (P) into a sequence of finite min-max problems, needs
typically no additional assumptions for convergence. At each iteration, T is
approximated by a finite subset Tm ¼ ft1 ; t2 ; . . . ; tm g, an approximate
solution
umþ1 2 arg min F m ðuÞ
u2W
is computed, and
tmþ1 2 arg max ft ðumþ1 Þ
t2T
is added to Tm to form Tmþ1 . This method is analogous to the exchange
method in semi-infinite programming (SIP) when applied to the following
problem:
minfnjft ðuÞ n 0; t2T; u2Wg:
ð4Þ
Under certain circumstances, a subset of ft1 ; . . . ; tm g can be dropped so that
each subproblem is kept to a limited size. We refer the readers to Ref. 9 and
the book edited by Reemtsen and Rückmann (Ref. 10) for detailed
discussions of SIP.
In this paper, we study the method of outer approximations, with each
approximate problem smoothed using the entropic regularization (1). We
call this algorithm ‘iterative entropic regularization’ (IER). By using the
uniform error estimate (2), we prove convergence of the IER algorithm,
while allowing minu 2 W Fpm ðuÞ to be solved inexactly. When each ft is smooth
convex and W is a linear constraint set, we propose to use an interior-point
algorithm to compute an inexact solution. Numerical examples are
presented to illustrate the behavior of the IER algorithm.
2. Iterative Entropic Regularization (IER)
Considering any approximation Fpm ðuÞ having a uniform error estimate,
ðm; pÞ Fpm ðuÞ F m ðuÞ ðm; pÞ;
8u 2 W;
ð5Þ
600
JOTA: VOL. 121, NO. 3, JUNE 2004
where ðm; pÞ 0 is a decreasing function of p having the following
properties:
(i) for any fixed m, limp ! 1 ðm; pÞ ¼ 0;
(ii) there exists some pðmÞ such that
limm!1 ðm; pðmÞÞ ¼ 0:
lim m ! 1 pðmÞ ¼ 1 and
In the case of the entropic regularization (1), we may choose
ðm; pÞ ¼ logðmÞ=p; pðmÞ ¼ ðlog mÞ2 :
Algorithm 2.1. IER Algorithm.
Step 1. Select t1 2 T and let T1 ¼ ft1 g; m ¼ k ¼ 1. Choose d 2 ð0; 1Þ;
p > 0:
Step 2. Find um
p 2 W satisfying
Fpm ðupm Þ min Fpm ðuÞ þ dk ;
u2W
ð6Þ
increase the iteration count k by 1.
k
m m
Step 3. If (i) Fðum
p Þ Fp ðup Þ and (ii) d þ ðm; pÞ is below a desired
tolerance, stop. If (i) is violated, then choose any
tmþ1 2 arg maxt2T ft ðum
p Þ; set Tmþ1 ¼ Tm [ ftmþ1 g, increase m
by 1, select p pðmÞ, and go to Step 2. If (ii) is violated,
increase p by a constant factor, and go to Step 2.
Theorem 2.1. The IER algorithm either stops in Step 3 with a
ðdk þ ðm; pÞÞ optimal solution or else it generates an infinite sequence fum
pg
in Wðm ! 1; p pðmÞÞ, any cluster point of which is a global minimum of
(P).
Proof. Suppose that (i) in Step 3 is violated finitely often. Then, m is
and p ! 1, k ! 1, so that ðm;
pÞ ! 0 [by
fixed eventually at some m
property (i) of ðm; pÞ] and dk ! 0. Then, (5) and (6) imply
m
m
Fðum
p Þ Fp ðup Þ
Fpm ðu Þ þ dk
pÞ
F m ðu Þ þ dk þ ðm;
pÞ;
Fðu Þ þ dk þ ðm;
pÞÞ optimal
where u is any optimal solution of (P). Thus, upm is a ðdk þ ðm;
solution.
Suppose that (i) in Step 3 is violated infinitely often. If m ! 1, then
p pðmÞ ensures that ðm; pÞ ðm; pðmÞÞ ! 0, as well as dk ! 0. Also, (5)
and (6) imply
JOTA: VOL. 121, NO. 3, JUNE 2004
601
m m
F m ðum
p Þ Fp ðup Þ þ ðm; pÞ
Fpm ðuÞ þ dk þ ðm; pÞ
F m ðuÞ þ dk þ 2ðm; pÞ
FðuÞ þ dk þ 2ðm; pÞ;
8u 2 W:
ð7Þ
(um
p ; tmþ1 )
t) be any cluster point of
Let (u;
as m ! 1 with p pðmÞ. Then,
there exists an infinite M f1; 2; . . .g such that, as m ! 1; m 2 M, we have
t) in the compact space W T. Then,
ðum
p ; tmþ1 Þ converging to (u;
m
ftmþ1 ðum
p Þ ¼ Fðup Þ
F m ðum
pÞ
ftiðmÞþ1 ðum
p Þ;
ð8Þ
where
iðmÞ ¼ maxfiji 2 M \ f1; 2; . . . ; mgg:
Thus,
tiðmÞþ1 ! t;
as m ! 1; m 2 M:
implying
By the continuity of ft ðuÞ, both sides of (8) tend to ftðuÞ,
m m
m
F ðup Þ ! Fðup Þ:
since F is continuous. Using this in (7) yields in the limit
Also, Fðum
p Þ ! FðuÞ,
FðuÞ; for all u 2 W;
FðuÞ
so u is a global minimum of (P).
(
Suppose that the IER algorithm generates an infinite sequence fum
p g.
For each fixed u 2 W; fF m ðuÞg1
forms
an
increasing
sequence
bounded
m¼1
above by FðuÞ. Define
FðuÞ ¼ lim F m ðuÞ;
m!1
8u 2 W:
Theorem 2.2. Suppose that the IER algorithm generates an infinite
sequence fum
p g. Then, minimizing FðuÞ over W is equivalent to minimizing
FðuÞ
over W. Namely,
min FðuÞ ¼ min FðuÞ:
u2W
u2W
Proof. Replacing the last inequality in (7) by F m ðuÞ FðuÞ, We obtain
as in the proof of Theorem 2.1 that
k
F m ðum
p Þ FðuÞ þ d þ 2ðm; pÞ;
8u 2 W;
602
JOTA: VOL. 121, NO. 3, JUNE 2004
and that, in the limit as m ! 1 with p pðmÞ,
FðuÞ;
FðuÞ
where u is any cluster point of um
p . Since
FðuÞ FðuÞ;
we have
¼ FðuÞ:
FðuÞ
So,
FðuÞ;
FðuÞ
and this shows that u also minimizes F over W.
h
Recall that F is continuous on W. By the Dini theorem (Ref. 11) [the
theorem states that, if an increasing sequence of real-valued continuous
functions fF m g1
m¼1 converges pointwise to a continuous function g on a
compact metric space, then it converges uniformly to g], we have the
uniform convergence property below. This property will be used in
Section 3.
Theorem 2.3. Suppose that the IER algorithm generates an infinite
m 1
sequence fum
p g. Then, fF gm¼1 converges uniformly to F on W.
3. Continuous Convex Min-Max Problems with Linear Constraints
In this section, we consider the special case of (P) in which ft ; t 2 T, are
smooth convex functions and
W ¼ fu 0jAu ¼ bg;
where A is a q n matrix and b 2 Rq . The convexity of ft implies that Fpm
given by (1) is also convex. We propose to use an interior-point algorithm to
implement Step 2 of the IER algorithm. In particular, we use a pathfollowing algorithm (Ref.12) to compute um
p 2 W satisfying (6). Such an
algorithm has been shown to be efficient for solving convex programs with
linear constraints. We call this algorithm ‘IER-interior point’ (IER-IP)
algorithm.
Algorithm 3.1. IER-IP Algorithm.
Step 1. This is the same as in the IER algorithm.
Step 2. Let l ¼ dk =2n. Use any algorithm to finitely generate
m m
um
p > 0; yp ; sp > 0 satisfying
JOTA: VOL. 121, NO. 3, JUNE 2004
603
Au ¼ b; s ¼ rFpm ðuÞ At y; kUs=l ek1 < 1;
ð9Þ
where e ¼ ð1; 1; :::; 1Þt and U ¼ diagðu1 ; u2 ; :::; un Þ. Increase k
by 1.
Step 3. This is the same as in the IER algorithm.
Step 2 can be realized by various interior-point algorithms
(Refs. 3,12–13) or by any feasible-descent method applied to the linearly
constrained convex problem
n
X
min
Fpm ðuÞ l
log ui :
u 0; Au ¼ b
i¼1
In the following theorem, we show that l ¼ dk =2n ensures that um
p generated
by the IER-IP algorithm satisfies (6).
Theorem 3.1. The um
p generated in Step 2 of the IER-IP algorithm
satisfies (6).
Proof. By attaching Lagrange multipliers y 2 Rq ; s 2 Rn to the constraints Au ¼ b; u 0; the dual problem of minu2W Fpm ðuÞ can be written as
sup fFpm ðuÞ rFpm ðuÞt u þ yt bg;
y;s;u
s.t.
s ¼ rFpm ðuÞ At y 0:
m
m
Thus, the pair ym
p ; sp > 0 generated in Step 2 is dual feasible, while up is
primal feasible.
By the weak duality theorem, the objective value of any dual feasible
solution provides a lower bound on that of any primal feasible solution.
Consequently,
m m t m
m t
m
m m
Fpm ðum
p Þ rFp ðup Þ up þ ðyp Þ b min Fp ðuÞ Fp ðup Þ;
u2W
which gives the following error estimate:
m
Fpm ðum
p Þ min Fp ðuÞ
u2W
m m
m m t m
m t
Fpm ðum
p Þ ½Fp ðup Þ rFp ðup Þ up þ ðyp Þ b
m t
m
m t
¼ ðspm Þt um
p þ ðyp Þ Aup ðyp Þ b
¼ ðspm Þt um
p:
on the other hand, since
kUpm spm =l ek1 < 1;
we have also
ð10Þ
604
JOTA: VOL. 121, NO. 3, JUNE 2004
0 ðUpm spm Þi 2l;
8i ¼ 1; 2; . . . ; n:
ð11Þ
Combining (10) and (11) yields
k
m
m t m
Fpm ðum
p Þ min Fp ðuÞ ðsp Þ up 2nl ¼ d :
(
u2W
Theorems 2.1 and 3.1 show that any cluster point of the primal sequence
fum
p g solves (P). The following theorem shows that, in addition, any cluster
m
point of fðym
p ; sp Þg solves the dual problem of (P). In other words, the
primal and dual cluster points satisfy the Karush-Kuhn-Tucker (KKT)
conditions,
Au ¼ b; u 0
0 2 f@FðuÞ At y sg;
ð12aÞ
s 0;
ðuÞt s ¼ 0:
ð12bÞ
ð12cÞ
Theorem 3.2. Suppose that the IER-IP algorithm generates an infinite
m m
m m
sequence of um
p ; yp ; sp . Then, any cluster point of fðyp ; sp Þg, if it exists,
solves the dual of (P).
m
Proof. Let ðy ; s Þ be any cluster point of fðym
p ; sp Þg. By passing to a
m m
subsequence if necessary, we may assume that fðyp ; sp Þg ! ðy ; s Þ: Since
fum
p g W, it has a cluster point u 2 W. By further passing to a subsequence if
necessary, we may assume that fum
p g ! u . We show below that ðu ; y ; s Þ
satisfies (12).
From the Proof of Theorem 3.1, we observe that
k
0 ðspm Þt um
p 2nl ¼ d :
Since dk ! 0; this yields in the limit
ðs Þt u ¼ 0:
m
Also, um
p > 0 and Aup ¼ b yield in the limit that
u 0 and Au ¼ b:
Thus, it remains only to verify the condition
0 2 @Fðu Þ At y s ;
which by Theorem 2.2 is equivalent to
At y þ s 2 @Fðu Þ:
Recall that F is convex. By the convexity of Fpm ðuÞ; we have
m m t
m
Fpm ðuÞ Fpm ðum
p Þ rFp ðup Þ ðu up Þ; 8u 2 W:
ð13Þ
JOTA: VOL. 121, NO. 3, JUNE 2004
605
Based on the following triangular inequality:
jFpm ðum
p Þ Fðu Þj
m m
m m
m
m
jFpm ðum
p Þ F ðup Þj þ jF ðup Þ Fðup Þj þ jFðup Þ Fðu Þj;
ð14Þ
we claim that
Fpm ðum
p Þ ! Fðu Þ:
This is true because the first term of (14) is uniformly bounded by ðm; pÞ
[see (5)], the second term tends to 0 by the uniform convergence of F m ðuÞ to
FðuÞ (Theorem 2.3), while the last term tends to 0 due to the continuity of
FðuÞ. By a similar argument as in (14),
Fpm ðuÞ ! FðuÞ:
Finally,
t m
m
t rFpm ðum
p Þ ¼ A yp þ sp ! A y þ s :
Now, pass to the limit in (13) to obtain
FðuÞ Fðu Þ ðAt y þ s Þt ðu u Þ;
8u 2 W:
This shows that
At y þ s 2 @Fðu Þ
and completes the proof.
h
4. Numerical Examples
In this section, we present three numerical examples to illustrate the
practical behavior of the IER/IER-IP algorithm. The first two are selected
from Polak et al. (Ref. 7), while the last one is a convex linearly constrained
problem.
We implemented the IER Algorithm using MATLAB 6.5 on an Athlon
1.4GHz PC. Key steps of the algorithm include:
k
m m
m
(i) finding um
p that satisfies Fp ðup Þ minu2W Fp ðuÞ þ d ;
(ii) computing tmþ1 ;
(iii) checking the stopping criterion.
For (i), we utilize the MATLAB subroutine fminsearch which allows the
users to set the tolerance dk .
606
JOTA: VOL. 121, NO. 3, JUNE 2004
For (ii), it is a global optimization problem. We subdivide T and select
tmþ1 to be the point on the grid which attains the largest functional value
ft ðum
p Þ. The partition is made finer and finer as the iteration count k increases.
For (iii), p is updated by pnew ¼ pð1 þ eÞ and the algorithm is stopped
when dk g1 and 1=p g2 ; with g1 ; g2 being two preset positive constants.
Although the IER Algorithm does not drop any point from Tm ; various
constraint dropping schemes such as those in Refs. 10, 14–15 can be
employed to reduce the computational effort. The idea is to keep only active
or -active constraints in each subproblem, while ignoring all others. A
second version of the IER algorithm, which drops all ‘nonbinding’
objectives5 from Tm is implemented for comparison. Such constraint
dropping reduces the number of summation terms in (1). A similar issue
arises in SIP and it was suggested (Refs. 10, 16) that the sup-norm should be
used in place of the L1 –norm on large-scale problems.
A greater concern for the IER algorithm, due to the use of exponential
functions in (1), is numerical overflow resulting from large positive values in
pftj ðuÞ; especially when ftj ðuÞ is already large. We resolved this issue by
working with small values of p, thus sacrificing to some extent the precision
of the solutions found. Alternatively, one could replace the exponential
function by a polynomial approximation whenever its argument exceeds a
threshold.
Example 4.1. This example, adopted from Ref. 7, has the following
form:
min maxfhðuÞ; max fhðuÞ þ 100½u1 þ u2 eu3 t þ e2t 2sinð4tÞgg:
u
t 2 ½0;1
ð15Þ
The numerical behavior of the IER algorithm is shown in Figure 1 and
Table 1 below. The initial parameters were set at
t1 ¼ 0; p ¼ 104 ; ¼ 1:3; g1 ¼ 106 ; g2 ¼ 0:005; d ¼ 1=2:
At the first iteration of the algorithm, fminsearch started from
u ¼ ð1; 1; 1Þt ; which is the same initial point used in Ref. 7. As can be
seen from Figure 1, the objective value jumped from 1265.09 to 7229.76
after one iteration and then quickly declined. This is because the IER
algorithm minimizes Fpm ðuÞ; which is a poor approximation of FðuÞ
initially. As m and p increases, this approximation appears to improve
quickly.
5
m
m
0
By nonbinding objectives, we mean ft 0 ðum
p Þ such that ft 0 ðup Þ 6¼ maxt 2 T ft ðup Þ, where t 2 Tm
m
and up is the iterate point to generate tmþ1 .
607
JOTA: VOL. 121, NO. 3, JUNE 2004
Fig. 1. Numerical behavior of the IER algorithm on solving Example 4.1 with X=iterations
and Y=FðuÞ.
Table 1. IER algorithm for Example 4.1.
Dropping
points
Time
u1
u2
u3
Optimal
Value
No
Yes
3.564
2.829
)0.213230
)0.213230
)1.361000
)1.361000
1.854000
1.854000
5.334900
5.334900
In general, the IER algorithm converges rapidly to a neighborhood of
the optimal solution, after which the convergence slows. About half of the
iterations (12 out of 25 iterations) were spent on improving the precision
level by one decimal place. This behavior is typical of cutting-plane
methods, including IER.
Listed in Table 1 are numerical results with and without the dropping rules.
We observe that dropping nonbinding objectives improves the run times.
Example 4.2. This example, also selected from Ref. 7, minimizes the
maximum of two functions f1 ðuÞ and f2 ðuÞ, with
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
2
2
f1 ðuÞ ¼ u1 u1 þ u2 cos
u21 þ u22
þ 0:005 u21 þ u22 ;
608
JOTA: VOL. 121, NO. 3, JUNE 2004
Table 2. IER algorithm for Example 4.2.
e
u1
u2
Optimal Value
Time[sec]
1.3
)4.2352e)020
6.6466e)008
4.4398e)015
0.219
f2 ðuÞ ¼
u2 qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
u21 þ u22 sin
u21 þ u22
þ 0:005 u21 þ u22 :
It was considered to be difficult by Polak et al. (Ref. 7) due to the spiral
contours of the max function. However, using the same starting point
u ¼ ð1:41831; 4:79462Þt as in Ref. 7 to initialize fminsearch, the IER
algorithm moved quickly to a near-optimal solution up to an accuracy of
1015 in less than one second (Table 2). Initial parameters used in this
example were
p ¼ 105 ; g1 ¼ 107 ; g2 ¼ 105 :
Example 4.3. This example has a convex objective function and linear
constraints. We took Au ¼ b; called afiro, from www.netlib.org and added
the constraint 0 u 5 to make W compact. Next, slack variables were
introduced to convert the inequalities u 5 into equalities. Furthermore,
the matrix A was modified to make ðe; 4eÞt an initial feasible solution. The
final form of the A-matrix has a size of 781026 .
To implement Step 2 in the IER-IP algorithm, we use a primal pathfollowing method (e.g., Refs. 3, 13). Starting from an interior feasible
solution um
p ; we compute
m2 t 1
m2
m m
ym
p ¼ ðAUp A Þ AUp ðrFp ðup Þ leÞ;
ð16Þ
t m
spm ¼ rFpm ðum
p Þ A yp ;
ð17Þ
m m
wm
p ¼ ð1=lÞUp sp e:
ð18Þ
If kwm
p k1 1; a line search
m
m m
up :¼ um
p rUp wp
ð19Þ
is invoked, where the step size r is computed by golden section, applied to
!
n
X
m
logui
:
min Fp ðuÞ l
r0
i¼1
m m
u¼um
p rUp wp
In this example, T=[0, a] and the objective function is
6
It can be found at our web site http://math.ncku.edu.tw/jylin/a.mat
609
JOTA: VOL. 121, NO. 3, JUNE 2004
ft ðuÞ ¼
102
X
tðui sin tÞ2 :
i¼1
For any t > 0; ft ðuÞ is convex in u. However, for any u 0; it is a
nonperiodical function of t with different oscillations (Figure 2).
In general, the computational time was insensitive to the choice of a
(Table 3). Moreover, the pattern of convergence is quite similar. Let us take
a=20 and use Figure 3 as an example for explanation. The IER-IP
algorithm started from the point u ¼ ðe; 4eÞt with an objective value about
Fig. 2. Graph of ft ðuÞ with u fixed and t 2 ½0; 50.
Table 3. IER-IP algorithm for Example 4.3 without dropping rule.
T=[0, a]
a=10
a=20
a=30
a=40
a=50
Time for Step 2
Time for Step 3
Total time (sec)
Number of points added
Final objective value
2nl þ logðmÞ=p
Relative error
14.641
16.814
31.487
2
10755
69.682
0.00630
21.608
17.032
38.704
7
23919
192.02
0.00794
19.9
16.857
36.818
5
41227
159.16
0.00381
21.377
17.14
38.58
5
49889
176.97
0.00315
18.082
20.234
38.395
6
67220
701.9
0.01044
610
JOTA: VOL. 121, NO. 3, JUNE 2004
Fig. 3. Behavior of the IER-IP algorithm for the case a=20.
2.565104 . After one interior-point step (16)–(19), a point near the central
path satisfying kwm
p k1 < 1 was found and the objective value was quickly
improved to 2.44104 . Then, the parameters p, l were updated with a new
point t2 adding to T1 , which yielded another drop in the objective value to
about 2.395104 .
Although the newly selected member ftmþ1 shifted the central path, the
time to recompute a solution near the central path is less than before. After
adding more t-points (the fifth row in Table 3), the process quickly
converges.
On this example, numerical overflow is an issue because the objective
function is the sum of 102 terms each of which is in the order of hundreds or
thousands. Thus, we started IER-IP from p ¼ 105 and finished at
g1 ¼ 102 and g2 ¼ 100, namely, p 2 ½105 ; 102 : Other initial parameters
included t1 ¼ 0:1; l ¼ 10; ¼ 3; d ¼ 1=2. According to Theorem 2.1, if the
IER algorithm stops when there is no additional t-point to include, we have
a ðdk þ ðm; pÞÞ optimal solution. Since
dk ¼ 2nl and ðm; pÞ ¼ logðmÞ=p;
the quantity 2nl þ logðmÞ=p provides a bound on the error between the
solution obtained by IER-IP and the true optimal solution; see the final two
611
JOTA: VOL. 121, NO. 3, JUNE 2004
Table 4.
IER-IP algorithm for Example 4.3 with dropping rule.
T=[0, a]
a=10
a=20
a=30
a=40
a=50
Time for Step 2
Time for Step 3
Total time (sec)
Number of points retained
Final objective value
2nl þ 1=p
Relative error
14.467
17.204
31.703
1
10755
99.648
0.00908
17.923
17.282
35.268
1
23919
99.648
0.00408
19.454
17
36.486
1
41227
99.648
0.00237
18.31
17.173
35.515
1
49889
99.648
0.00196
13.018
20.063
33.144
1
67220
392.62
0.00584
rows in Tables 3 and 4, which were obtained using the same initial
parameters. This error is in the hundreds for most cases, indicating that the
solutions obtained by IER-IP are not precise in absolute sense. However,
when 2nl þ logðmÞ=p is divided by the final objective value, the relative
error bound is about 103 , which we feel is acceptable.
Finally, Table 4 suggests that the IER-IP algorithm equipped with the
dropping rule converges faster generally than the original IER-IP, except
when IER-IP adds very few t-points as in the case a=10.
References
1. BERTSEKAS, D. P., Constrained Optimization and Lagrange Multiplier Methods,
Academic Press, New York, NY, 1982.
2. POLYAK, R. A., Smooth Optimization Methods for Minmax Problems, SIAM
Journal on Control and Optimization, Vol. 26, pp. 1274–1286, 1988.
3. SHEU, R. L., and WU, S. Y., Combined Entropic Regularization and PathFollowing Method for Solving Finite Convex Min-Max Problems Subject to
Infinitely Many Linear Constraints, Journal of Optimization Theory and
Applications, Vol. 101, pp. 167–190,1999.
4. CHANG, P. L., A Minimax Approach to Nonlinear Programming, Doctoral
Dissertation, Department of Mathematics, University of Washington, Seattle,
Washington, 1980.
5. BEN-TAL, A., and TEBOULLE, M., A Smoothing Technique for Nondifferentiable
Optimization Problems, Lecture Notes in Mathematics, Springer Verlag, Berlin,
Germany, Vol. 1405, pp. 1–11, 1989.
6. WU, S. Y., and FANG, S. C., Solving Min-Max Problems and Linear Semi- Infinite
Programs, Computers and Mathematics with Applications, Vol. 32, pp. 87–93,
1996.
7. POLAK, E., HIGGINS, J. E., and MAYNE, D. Q., A Barrier Function Method for
Minimax Problems, Mathematical Programming, Vol. 54, pp. 155–176, 1992.
8. POLAK, E., Optimization: Algorithms and Consistent Approximations, Springer
Verlag, New York, NY, 1994.
612
JOTA: VOL. 121, NO. 3, JUNE 2004
9. HETTICH, R., and KORTANEK, K. O., Semi-Infinite Programming: Theory,
Methods, and Applications, SIAM Review, Vol. 35, pp. 380–429, 1993.
10. REEMTSEN, R., and GÖRNER, S., Numerical Methods for Semi-Infinite Programming: A Survey, Semi-Infinite Programming, Edited by R. Reemtsen and
J.J. Rückmann, Kluwer Academic Publishers, Dordrecht, Netherlands, pp. 195–
275, 1998.
11. DIEUDONNE, J., Foundations of Modern Analysis 10–1, Academic Press, New
York, NY, 1969.
12. DEN HERTOG, D., Interior-Point Approach to Linear, Quadratic, and Convex
Programming, Kluwer Academic Publishers, Dordrecht, Netherlands, 1994.
13. SHEU, R. L., A Generalized Interior-Point Barrier Function Approach for Smooth
Convex Programming with Linear Constraints, Journal of Information and
Optimization Sciences, Vol. 20, pp. 187–202, 1999.
14. GONZAGA, C. C., and POLAK, E., On Constraint Dropping Schemes and
Optimality Functions for a Class of Outer Approximations Algorithms, SIAM
Journal on Control and Optimization, Vol. 17, pp. 477–493, 1979.
15. POWELL, M. J. D., A Tolerant Algorithm for Linearly Constrained Optimization
Calculations, Mathematical Programming, Vol. 45, pp. 547–566, 1989.
16. ABBE, L., A Logarithmic Barrier Approach and Its Regularization Applied to
Convex Semi-Infinite Programming Problems, PhD Dissertation, Universität
Trier, Trier, Germany, 2001.