A new class of hybrid extragradient algorithms for solving quasi

J Glob Optim
DOI 10.1007/s10898-011-9814-y
A new class of hybrid extragradient algorithms
for solving quasi-equilibrium problems
Jean Jacques Strodiot · Thi Thu Van Nguyen ·
Van Hien Nguyen
Received: 24 May 2011 / Accepted: 1 November 2011
© Springer Science+Business Media, LLC. 2011
Abstract Generalized Nash equilibrium problems are important examples of quasi-equilibrium problems. The aim of this paper is to study a general class of algorithms for solving
such problems. The method is a hybrid extragradient method whose second step consists in
finding a descent direction for the distance function to the solution set. This is done thanks to
a linesearch. Two descent directions are studied and for each one several steplengths are proposed to obtain the next iterate. A general convergence theorem applicable to each algorithm
of the class is presented. It is obtained under weak assumptions: the pseudomonotonicity of
the equilibrium function and the continuity of the multivalued mapping defining the constraint set of the quasi-equilibrium problem. Finally some preliminary numerical results are
displayed to show the behavior of each algorithm of the class on generalized Nash equilibrium
problems.
Keywords Quasi-equilibrium problems · Quasi-variational inequalities ·
Hybrid extragradient methods · Generalized Nash equilibrium problems
Mathematics Subject Classification (2000)
49J40 · 65K10 · 91A12
This research was supported by the Institute for Computational Science and Technology at Ho Chi Minh
City, Vietnam (ICST HCMC).
J. J. Strodiot · T. T. V. Nguyen · V. H. Nguyen
Institute for Computational Science and Technology, Ho Chi Minh City, Vietnam
T. T. V. Nguyen
e-mail: [email protected]
J. J. Strodiot (B) · V. H. Nguyen
Department of Mathematics, University of Namur, Namur, Belgium
e-mail: [email protected]
V. H. Nguyen
e-mail: [email protected]
123
J Glob Optim
1 Introduction
The aim of this paper is first to present a general double-projection method as well as two
of its concrete realizations for solving a quasi-equilibrium problem in the sense of Blum and
Oettli [2], and then to illustrate numerically these methods on some test problems and in particular, on some generalized Nash equilibrium problems (GNEPs)[4]. Let X be a nonempty
closed convex subset of IRn , let K be a multivalued mapping from X into itself such that for
every x ∈ X, K(x) is a nonempty closed convex subset of X and let f : X × X → IR be a
mapping such that for every x ∈ X, f (x, x) = 0 and f (x, ·) is a convex function on X. The
quasi-equilibrium problem QE(K; f ) consists in finding x ∗ ∈ K(x ∗ ) such that
f (x ∗ , y) ≥ 0 for all y ∈ K(x ∗ ).
We denote the solution set of QE(K; f ) by K ∗ .
When f (x, y) = F (x), y−x where F : X → IRn , problem QE(K; f ) can be expressed
as: Find x ∗ ∈ K(x ∗ ) such that
F (x ∗ ), y − x ∗ ≥ 0 for all y ∈ K(x ∗ ).
This problem is called a quasi-variational inequality problem and is denoted QV I (K; F ).
When K(x) = K for every x ∈ X, problems QE(K; f ) and QV I (K; F ) are denoted
E(K; f ) and V I (K; F ), and called equilibrium problem and variational inequality problem,
respectively.
Problems QE(K; f ) and QV I (K; F ) have a number of important applications, for example, in economics, engineering, and operations research (see, for example, [3,7,11,19]). In
particular the GNEP can be modeled as a problem QE(K; f ) and when the data are differentiable as a problem QV I (K; f ). More precisely, assume there are N players each controlling
the variables xi ∈ IRni . We denote by x the overall vector of all variables: x = (x1 , . . . , xN ),
while we use the notation x−i = (x1 , . . . , xi−1 , xi+1 , . . . , xN ) to denote the vector of all players’ variables except that of player i. The aim of player i, given the other players’ strategies,
is to choose an xi ∈ Ki ⊂ IRni that solves the problem
Minimize fi (xi , x−i )
subject to xi ∈ Ki ,
where fi : IRn → IR and n = N
i=1 ni . We make the blanket assumption that the objective functions fi are, as a function of xi alone, convex, while the sets Ki are all nonempty
closed and convex. A point x is feasible if xi ∈ Ki for all players i. A solution of the Nash
equilibrium problem (NEP) is a feasible point x ∗ such that
∗
∗
fi (xi∗ , x−i
) ≤ fi (xi , x−i
) ∀xi ∈ Ki
holds for each player i = 1, . . . , N . So, a Nash equilibrium is a feasible strategy profile x ∗
with the property that no single player can benefit from a unilateral deviation from xi∗ as long
as the other players do not change their own strategies.
The GNEP extends the classical NEP by assuming that each player’s strategy set can
depend on the rival players’ strategies x−i . So we will write Ki (x−i ) to indicate that we
might have a different closed convex set Ki for each different x−i . Analogously to the NEP
case, the aim of each player i is, given x−i , to choose a strategy xi ∈ Ki (x−i ) that solves the
problem
Minimize fi (xi , x−i )
123
subject to xi ∈ Ki (x−i ).
J Glob Optim
∗ ) such that,
A generalized Nash equilibrium (GNE) is a tuple of strategies x ∗ = (x1∗ , . . . , xN
for all i = 1, . . . , N ,
∗
∗
∗
fi (xi∗ , x−i
) ≤ fi (xi , x−i
) ∀xi ∈ Ki (x−i
).
The requirement that the feasible sets depend on the variables of players’ rivals is natural in
many applications (for example, the case in which the players share some common resource).
A GNEP is termed with shared constraints when the feasible sets Ki (x−i ) are defined by
Ki (x−i ) = {yi ∈ Ki | (yi , x−i ) ∈ Q},
where Ki is the nonempty closed and convex set of individual constraints of player i and Q
is a nonempty closed convex set in IRn representing the set of shared constraints, common
for all the players.
When the functions fi (·, x−i ) are convex and differentiable, we can define
N
F (x) = ∇xi fi (x) i=1
∗
and we see that x ∗ is a GNE if x ∗ ∈ K(x ∗ ) = N
i=1 Ki (x−i ) and
F (x ∗ )T (y − x ∗ ) ≥ 0 ∀y ∈ K(x ∗ ).
This problem is the quasi-variational inequality Problem QV I P (K; F ) (see, for instance,
[1,4,5,8,16–18,22,25,28,31] and the references quoted therein).
When the functions fi (·, x−i ) are convex and nondifferentiable, we define the NikaidoIsoda function
Ψ : IRn × IRn → IR
Ψ (x, y) =
N
fi (xi , x−i ) − fi (yi , x−i )
i=1
and we see that x ∗ is a GNE if x ∗ ∈ K(x ∗ ) =
N
i=1
∗ ) and
Ki (x−i
f (x ∗ , y) ≥ 0 ∀y ∈ K(x ∗ )
where f (x, y) = −Ψ (x, y) [2]. This problem is the quasi-equilibrium problem QE(K; f ).
We are interested in this work to develop efficient numerical methods for solving problems
QE(K; f ) and QV I (K; F ) and thus, in particular, for solving GNEPs.
Recently Fukushima [6], Taji [23] and Kubota et al. [13] proposed a class of gap functions for solving problem QV I (K; F ) (see also [15]). Unfortunately, contrary to the case of
variational inequalities, their gap functions were only directionally differentiable when the
function F is differentiable and the constraint set is given by convex differentiable inequalities. Furthermore, the gap functions being nonconvex, supplementary conditions must be
imposed to obtain that any stationary point of these functions is a global minimum. Such
conditions are known for problem V I (K; F ) but, to our knowledge, unknown for problem
QV I (K; F ). So in this paper we prefer to consider projection and extragradient methods
rather than methods based on gap functions for solving problems QEP (K; f ). More precisely, our aim is to generalize a class of double-projection methods for solving problem
V I (K; F ) to the case of equilibrium problems with a moving constraint set K(x).
After recalling in Sect. 2 several extragradient-type methods for solving variational
inequality problems and as well as their main properties, we propose in the next section
(Sect. 3) a very general algorithm inspired by these ideas for solving problem QE(K, f ).
The strategy is to reduce at each step the distance to the solution set. After that, conditions
are given on the data to force the convergence of this very general algorithm. In particular the
123
J Glob Optim
function f is supposed to be pseudomonotone and continuous and the multivalued function
K to be continuous. Let us mention that a similar general algorithm has been proposed in
[30], Section 4 for solving problem V I (K; F ).
In Sect. 4 we introduce two linesearches that give rise to two implementable algorithms
denoted Algorithms 1 and 2, respectively. Their convergence is established under mild
assumptions. In the next two sections (Sects. 5 and 6) we consider two variants of each
of the previous algorithms. The variants of Algorithm 1 are related to the classical extragradient method. When K(x) = K for every x ∈ X, one of them coincides with an algorithm
due to Tran et al. [24]. The variants of Algorithm 2 are related to Noor’s modified extragradient method [14]. When f (x, y) = F (x), y − x, one of them coincides with an algorithm
proposed in [31] for solving problem QV I (K; F ). Finally the last section of the paper is
devoted to preliminary numerical results.
2 Preliminaries
Let us consider for a moment the case of a variational inequality with a fixed constraint
set K(x) = K for all x ∈ X. It is well-known that this problem can be reformulated as a
fixed-point equation:
x − PK (x − λF (x)) = 0
(1)
where PK denotes the orthogonal projection from IRn onto K, and λ > 0 is a constant. The
corresponding fixed point algorithm: x k+1 = PK (x k − λF (x k )) is convergent to a solution
of problem V I (K; F ) under strong assumptions: F is Lipschitz continuous and strongly
monotone. To avoid that, the following modified fixed point equations have been introduced:
x − PK (x − λF (x̄)) = 0 or x − PK (x̄ − λF (x̄)) = 0
(2)
where x̄ = PK (x − λF (x)). When F is Lipschitz continuous, it can be proven (see [12,29])
that if x satisfies (2), then x satisfies (1) (and thus is a solution of problem V I P (K; F )) provided that the Lipschitz constant L is such that λ < 1/L and λ < min{3, 1/L2 }, respectively.
The first equation of (2) gives rise to the classical extragradient method [12] and its variants
[9,10]: Given x k ∈ K, x k+1 is obtained after two projections as follows:
k
y
= PK (x k − λF (x k ))
x k+1 = PK (x k − λF (y k )).
The second equation of (2) gives rise to the Noor’s modified extragradient iteration [14]:
k
y
= PK (x k − λF (x k ))
x k+1 = PK (y k − λF (y k )).
These two methods generate sequences converging to a solution of problem V I (K; F ) under
the assumption that F is pseudomonotone and Lipschitz continuous with a condition on the
Lipschitz constant.
A well-known strategy [26,27] to avoid the use of the Lipschitz constant is first to define
y k = PK (x k − λF (x k )) and then to find a direction d k such that the inequality
d k , x k − x ∗ ≥ μx k − y k 2 with μ > 0
123
(3)
J Glob Optim
holds for any solution x ∗ of problem V I P (K; F ). When y k = x k , the direction −d k is a
descent direction at x k for the distance function to K ∗ : 21 x − x ∗ 2 (x ∗ ∈ K ∗ ). For the
k
classical extragradient method an example of such a direction is given by d k = F α(zk ) where
zk = (1 − αk )x k + αk y k and αk = α mk with mk being the smallest nonnegative integer m
satisfying the inequality
F (x k − α m (x k − y k )), x k − y k ≥ c x k − y k 2
(4)
and α, c ∈ (0, 1). In fact, this vector zk gives rise to a hyperplane
H k = {x ∈ IRn | F (zk ), x − zk = 0}
(5)
which strictly separates x k from the solution set K ∗ of problem V I P (K; F ). The direction
d k satisfies (3) with μ = c and the next iterate x k+1 is given by
x k+1 = PK (x k − βk d k )
where βk > 0 is chosen such that x k+1 − x ∗ 2 < x k − x ∗ 2 for every solution x ∗ ∈ K ∗
(see [9,26] for more details). For example, the steplength βk can be chosen in such a way
that x k − βk d k be the orthogonal projection of x k onto H k . It is easy to see that this step is
given by
βk = αk
F (zk ), x k − zk .
F (zk )2
There is a similar strategy to avoid the use of the Lipschitz constant for the modified extragradient method. The parameter μ, the direction d k , the inequality (4) defining the linesearch,
and the hyperplane H k become
μ = 1 − c, d k = x k − y k +
F (zk )
,
αk
F (x k ) − F (x k − α m (x k − y k )), x k − y k ≤ c x k − y k 2 ,
(6)
Gk = {x ∈ IRn | d k , x k − x = x k − y k , x k − y k − F (x k ) + F (zk )}.
(7)
= PK
− βk
where, for example, the steplength βk
The next iterate is given by
can be chosen in such a way that x k − βk d k be the orthogonal projection of x k onto Gk .
It is easy to see that this step is given by
x k+1
βk =
(x k
dk)
x k − y k , x k − y k − F (x k ) + F (zk )
d k 2
(see [27,30] for more details).
The general algorithm and its six implementable versions proposed in this paper for solving problem QE(K; f ) are inspired by these ideas. First y k is computed as the solution of
the convex minimization problem
1
min f (x k , y) + y − x k 2 .
2
y∈K(x k )
This solution y k coincides with PK(x k ) (x k − F (x k )) when f (x, y) = F (x), y − x. In a second step, a direction d k is supposed to be computed satisfying (3) for any solution x ∗ ∈ K ∗
and in a third step a line search is performed along the opposite of this direction to reduce
the distance of x k to the solution set.
123
J Glob Optim
In order to prove the convergence of our algorithms we assume that f is defined and finite
on X × where ⊂ IRn is an open convex set containing X.
Another important assumption used in the convergence proof of our algorithms is the
continuity of the multivalued mapping K. Let us recall that
– K is upper semicontinuous (u.s.c) at x̄ ∈ X if
⎫
{x k } ⊂ X and x k → x̄ (k → ∞) ⎬
y k ∈ K(x k )
⎭
y k → ȳ (k → ∞)
⇒
ȳ ∈ K(x̄).
– K is lower semicontinuous (l.s.c) at x̄ ∈ X if x k ∈ X and x k → x̄, then for any ȳ ∈ K(x̄),
there exists a sequence {y k } with y k ∈ K(x k ), such that y k → ȳ (k → ∞).
– K is continuous on X if K is u.s.c. and l.s.c. at every point of X.
3 A general Algorithm
From now on the following assumption is supposed to be satisfied for problem (QEP):
Assumption (A)
(a) f (x, ·) is convex on for all x ∈ X, f is continuous on X × , and f (x, x) = 0 for
all x ∈ X.
(b) K is continuous on X and K(x) is a nonempty closed convex subset of X for all x ∈ X;
(c) x ∈ K(x) for all x ∈ X;
(d) S ∗ := {x ∈ S | f (x, y) ≥ 0, ∀y ∈ T } is nonempty, where S = ∩x∈X K(x) and
T = ∪x∈X K(x).
(e) f is pseudomonotone on X with respect to S ∗ , i.e.,
∀x ∗ ∈ S ∗ ∀y ∈ X f (y, x ∗ ) ≤ 0.
Assumption (d) has been used by Zhang et al. [31] to prove the convergence of their method
for solving problem QV I (K; F ). Since S ∗ is contained in the solution set of QE(K; f ),
this assumption can be seen as a generalization to problem QE(K; f ) of the nonemptyness
of the solution set of problem E(K; f ). Indeed, when K(x) = K for every x ∈ X, the sets
S, T , and K coincide and the set S ∗ is the solution set of EP (K; f ). However, for problem
QE(K; f ), this assumption is not easy to check. Let us also mention that assumption (c) is
true for the GNEP when there is a shared constraint between the players.
Our general algorithm can be expressed as follows:
General Algorithm
Let x 0 ∈ X, let μ ∈ (0, 1) and let γ ∈ (0, 2). Set k = 0.
Step 1 Compute y k = arg miny∈K(x k ) [f (x k , y) + 21 y − x k 2 ]. If y k = x k , Stop.
Step 2 Otherwise find d k such that d k , x k − x ∗ ≥ μx k − y k 2 > 0 for every x ∗ ∈ S ∗ .
Step 3 Compute x k (βk ) = PK(x k ) (x k − βk d k ) where βk is such that
x k (βk ) − x ∗ 2 ≤ x k − x ∗ 2 − γ (2 − γ ) μ2
x k − y k 4
d k 2
for every x ∗ ∈ S ∗ . Set x k+1 = x k (βk ), k := k + 1 and go to Step 1.
123
(8)
J Glob Optim
Comments
(1) When y k = x k , the vector −d k is a descent direction at x k for the function 21 x − x ∗ 2
for every x ∗ ∈ S ∗ . In particular, d k = 0 for every k.
(2) An example of βk satisfying (8) is given by
βk = γ μ
x k − y k 2
.
d k 2
Indeed for any x ∗ ∈ S ∗ , we have x ∗ ∈ K(x k ) and consequently, using the nonexpansiveness of the orthogonal projection and the property of d k , we obtain
x k (βk ) − x ∗ 2 ≤ x k − x ∗ 2 − 2βk d k , x k − x ∗ + βk2 d k 2
(9)
≤ x k − x ∗ 2 − 2μβk x k − y k 2 + βk2 d k 2 .
= x k − x ∗ 2 − γ (2 − γ ) μ2
x k − y k 4
.
d k 2
(3) When f (x, y) = F (x), y − x for every x, y ∈ X, Step 1 becomes: Compute y k =
PK(x k ) (x k − F (x k )).
First we give a characterization of y k computed from x k at Step 1 of the General Algorithm.
Proposition 1 For every y ∈ K(x k ), we have
f (x k , y) ≥ f (x k , y k ) + x k − y k , y − y k .
In particular, f (x k , y k ) + x k − y k 2 ≤ 0.
Proof The vector y k being a solution of a convex minimization problem, the optimality
conditions imply that there exists s k ∈ ∂f (x k , y k ) such that
0 ∈ s k + y k − x k + NK(x k ) (y k )
where NK(x k ) (y k ) denotes the normal cone to K(x k ) at y k . Hence, by definition of this cone,
we obtain that
∀y ∈ K(x k ) x k − y k − s k , y − y k ≤ 0.
(10)
On the other hand, since s k ∈ ∂f (x k , y k ), we can write
∀y ∈ K(x k ) f (x k , y) ≥ f (x k , y k ) + s k , y − y k .
(11)
Combining (10) and (11) and taking y = x k , we obtain the desired result because x k ∈ K(x k )
by Assumption (A)(c).
Now we justify the stopping criterion: y k = x k .
Proposition 2 If y k = x k , then x k is a solution of problem QE(K; f ).
Proof Since y k = x k and x k ∈ K(x k ), it follows from Proposition 1 that
∀y ∈ K(x k ) f (x k , y) ≥ f (x k , x k ) + x k − x k , y − x k = 0,
i.e., that x k is a solution of QE(K; f ).
123
J Glob Optim
Next we assume that x k = y k for every k and we prove that the sequence {x k } generated
by the General Algorithm is bounded.
Proposition 3 The sequence {x k } is bounded.
Proof Since, by construction (see Step 3 of the General Algorithm), {x k −x ∗ } is a decreasing sequence, we have for every k,
x k ≤ x k − x ∗ + x ∗ ≤ x 0 − x ∗ + x ∗ ,
and thus {x k } is bounded.
To prove the boundedness of the sequence {y k }, we need the next lemma.
Lemma 1 x k − y k ≤ g for every g ∈ ∂f (x k , x k ).
Proof Let g ∈ ∂f (x k , x k ). Then
f (x k , y k ) ≥ f (x k , x k ) + g, y k − x k = g, y k − x k .
Using successively Proposition 1, the previous inequality and the Cauchy-Schwarz inequality,
we obtain
x k − y k 2 ≤ −f (x k , y k ) ≤ −g, y k − x k ≤ g x k − y k ,
and thus x k − y k ≤ g.
The sequence {x k } being bounded, let x̄ be one of its limit points. Then there exists a
subsequence {x kj } converging to x̄. Thanks to Lemma 1 we can prove that the corresponding
sequence {y kj } is also bounded.
Proposition 4 The sequence {y kj } is bounded.
Proof By Lemma 1 it is sufficient to prove that there exists M > 0 such that
g ≤ M
for every g ∈ ∂f (x kj , x kj ) and all j.
Since x̄ ∈ , {x kj } ⊂ , f (x̄, ·) is finite on and since the sequence of convex functions
{f (x kj , ·)} converges pointwise on to the convex function f (x̄, ·), it follows from [20]
Theorem 24.5 that there exists j0 such that for all j ≥ j0 ,
∂f (x kj , x kj ) ⊆ ∂f (x̄, x̄) + B
where B denotes the closed Euclidean unit ball of IRn . Since B and ∂f (x̄, x̄) are bounded,
there exists M > 0 such that
g ≤ M
for every g ∈ ∂f (x kj , x kj ) and all j ≥ j0 .
Hence the sequence {y kj } is bounded.
Proposition 5 Let x̄ be a limit point of {x k }. Assume that x kj → x̄ and that x kj −y kj → 0
as j → ∞. Then x̄ is a solution of problem QE(K; f ).
123
J Glob Optim
Proof By assumption y kj = y kj − x kj + x kj → x̄. Since y kj ∈ K(x kj ) for every j and
since K is u.s.c. on X, we obtain that x̄ ∈ K(x̄).
Now let ȳ ∈ K(x̄). We have to prove that f (x̄, ȳ) ≥ 0. Since K is l.s.c. on X, there exists
a sequence {ȳ kj } such that
ȳ kj ∈ K(x kj ) for every j and ȳ kj → ȳ.
So, for every j , we have, by definition of y kj , that
1
1
f (x kj , y kj ) + x kj − y kj 2 ≤ f (x kj , ȳ kj ) + x kj − ȳ kj 2 .
2
2
Taking the limit as j → ∞ and remembering that f is continuous, we obtain
1
1
0 = f (x̄, x̄) + x̄ − x̄2 ≤ f (x̄, ȳ) + x̄ − ȳ2 .
2
2
(12)
But this implies that f (x̄, ȳ) ≥ 0. Indeed, the inequality (12) means that x̄ is a solution of
the convex minimization problem
1
2
min f (x̄, y) + y − x̄ .
y∈K(x̄)
2
Hence 0 ∈ ∂f (x̄, x̄) + NK(x̄) (x̄), that is, x̄ is a solution of min f (x̄, y) subject to y ∈ K(x̄).
Consequently f (x̄, ȳ) ≥ 0.
Finally we obtain the convergence of the whole sequence {x k } to a solution of problem
QE(K; f ) when the function f is strictly monotone at a limit point of {x k }. Let us recall
that f is strictly monotone at x ∈ X if for all y ∈ X, y = x, we have
f (x, y) + f (y, x) < 0.
Proposition 6 If, in addition to the assumptions of Proposition 5, the function f is strictly
monotone at x̄, then the whole sequence {x k } converges to x̄ as k → ∞. Furthermore x̄ is a
solution of problem QE(K; f ).
Proof Let x kj → x̄. By Proposition 5, x̄ is a solution of problem QE(K; f ).
(1) First we prove that x̄ ∈ S ∗ . Let x ∗ ∈ S ∗ . Then x ∗ ∈ ∩x∈X K(x) and f (x ∗ , y) ≥ 0 for
every y ∈ K(x) and every x ∈ X. Since x̄ ∈ K(x̄), we have f (x ∗ , x̄) ≥ 0.
On the other hand f (x̄, x ∗ ) = 0. Indeed, by Assumption (A)(e), we have that f (x̄, x ∗ ) ≤
0 and, since x̄ ∈ K ∗ and x ∗ ∈ K (x̄), we have that f (x̄, x ∗ ) ≥ 0. Consequently x ∗ = x̄.
Indeed, if x ∗ = x̄, we deduce from the strict monotonicity of f at x̄ that
f (x ∗ , x̄) = f (x ∗ , x̄) + f (x̄, x ∗ ) < 0,
which contradicts f (x ∗ , x̄) ≥ 0. Hence x̄ = x ∗ ∈ S ∗ .
(2) Next we prove that x k → x̄. Since x̄ = x ∗ ∈ S ∗ , it follows from Step 3 of the General
Algorithm that the sequence {x k − x̄} is nonincreasing and consequently converges
to some a ≥ 0. Since x kj → x̄, we deduce that the whole sequence x k − x̄ → 0,
that is, x k → x̄ as k → ∞.
123
J Glob Optim
The convergence of the General Algorithm is obtained under the assumption that x k −
→ 0 as k → ∞. Thanks to the inequality (8) the sequence {x k − x ∗ } is nonincreasing
and converges to some a ≥ 0, which implies that
yk x k − y k 4
→ 0 as k → ∞.
d k 2
Consequently x k − y k → 0 when the sequence {d k } is bounded. It remains to examine
the case when the sequence {d k } is unbounded.
4 A class of double-projection methods
In this section we give two examples of directions d k such that for all k,
d k , x k − x ∗ ≥ μx k − y k 2 > 0 for some μ ∈ (0, 1) and for every x ∗ ∈ S ∗ .
They are based on a linesearch performed between y k and x k . This linesearch has the property
that when the steplengths tend to zero, then the sequence {d k } is unbounded and x k −y k →
0. More precisely, Step 2 of the General Algorithm is replaced by one of the following linesearch procedures:
Linesearch 1: Given x k , y k as before and α, c ∈ (0, 1)
Find m the smallest nonnegative integer such that
f (zk,m , x k ) − f (zk,m , y k ) ≥ c x k − y k 2
where zk,m := (1 − α m )x k + α m y k .
k
Set αk = α m and zk = zk,m . Set d k = αg k where g k ∈ ∂f (zk , x k ).
When K(x) = K for every x ∈ X, this linesearch has been used by Tran et al. [24] for
solving problem E(K; f ). When f (x, y) = F (x), y − x for every x, y ∈ X, the inequality
k
satisfied by the linesearch coincides with (4) and the direction d k becomes equal to F α(zk ) .
Linesearch 2: Given x k , y k as before and α, c ∈ (0, 1)
Find m the smallest nonnegative integer such that
f (zk,m , x k ) − f (zk,m , y k ) + f (x k , y k ) ≥ −c x k − y k 2
where zk,m := (1 − α m )x k + α m y k .
k
Set αk = α m and zk = zk,m . Set d k = x k − y k + αg k where g k ∈ ∂f (zk , x k ).
When f (x, y) = F (x), y − x for every x, y ∈ X, the inequality satisfied by this
k
linesearch coincides with (6) and the direction d k becomes equal to x k − y k + F α(zk ) .
This linesearch has been used by Zhang et al. [31] for solving problem QV I (K; F ).
Let us also mention that Linesearch 2 is closely related to Linesearch 1. Indeed, from
Proposition 1, we have that
f (x k , y k ) + x k − y k 2 ≤ 0.
So the vector zk,m obtained from Linesearch 2 satisfies
f (zk,m , x k ) − f (zk,m , y k ) ≥ −f (x k , y k ) − cx k − y k 2 ≥ (1 − c) x k − y k 2 ,
that is, the inequality defining Linesearch 1 with c replaced by 1 − c.
123
J Glob Optim
Finally, let us also observe that when f (x, y) = F (x), y − x, only one projection is
used in our algorithms to generate zk . This differs from many algorithms for solving problem V I (K; F ) where a projection is required at each trial point PK(x k ) (x k − αk F (x k )) of a
backtracking procedure to obtain zk . This can lead to expensive computation (see [30] and
the references quoted therein).
First we prove that the two linesearches are finite when y k = x k .
Proposition 7 Assume y k = x k . Then the two Linesearches 1 and 2 give αk and zk after
finitely many iterations.
Proof Suppose that the linesearches are not finite. Then for all m ∈ IN, we have (for Linesearch 1):
f (zk,m , x k ) − f (zk,m , y k ) < c x k − y k 2 .
Since f (·, x) is continuous for all x ∈ , and zk,m → x k as m → ∞, we obtain
−f (x k , y k ) = f (x k , x k ) − f (x k , y k ) ≤ c x k − y k 2 .
On the other hand, by Proposition 1, we have that
f (x k , y k ) + x k − y k 2 ≤ 0.
Combining these last two inequalities yields
x k − y k 2 ≤ c x k − y k 2 .
Since c ∈ (0, 1), we deduce that y k = x k , which contradicts the assumption y k = x k .
Similarly, for Linesearch 2, we have for all m ∈ IN, that
f (zk,m , x k ) − f (zk,m , y k ) + f (x k , y k ) < −c x k − y k 2 .
Taking the limit as m → ∞ and noting that f (·, x) is continuous for all x ∈ , we obtain
0 = f (x k , x k ) − f (x k , y k ) + f (x k , y k ) ≤ −c x k − y k 2 .
But then y k = x k , which contradicts the assumption y k = x k .
Now our aim is to prove that the direction d k obtained from each linesearch satisfies the
property: d k , x k − x ∗ ≥ μx k − y k 2 for some μ ∈ (0, 1) and for every x ∗ ∈ S ∗ .
Proposition 8 For the two linesearches and for every g k ∈ ∂f (zk , x k ) and x ∗ ∈ S ∗ , we have
k
g
k
∗
, x − x ≥ f (zk , x k ) − f (zk , y k ).
αk
Proof Let g k ∈ ∂f (zk , x k ) and x ∗ ∈ S ∗ . Then
f (zk , x ∗ ) ≥ f (zk , x k ) + g k , x ∗ − x k .
Since f is pseudomonotone on X with respect to S ∗ , we have f (zk , x ∗ ) ≤ 0. So
g k , x k − x ∗ ≥ f (zk , x k ).
(13)
Since f (zk , ·) is convex, we can write, using the definition of zk ,
f (zk , zk ) ≤ (1 − αk )f (zk , x k ) + αk f (zk , y k ),
123
J Glob Optim
that is,
f (zk , x k ) ≥ αk [f (zk , x k ) − f (zk , y k )].
Combining (13) and (14) yields the announced result.
(14)
It follows immediately from Proposition 8 that if Linesearch 1 is used, then the required
k
property on d k = αg k :
d k , x k − x ∗ ≥ μ x k − y k 2
is satisfied for μ = c. In particular, when Linesearch 1 is used (that is, when y k = x k ), the
direction d k = 0 whatever g k ∈ ∂f (zk , x k ).
To get a similar property but for Linesearch 2, we need to prove the following proposition.
Proposition 9 For every k and every x ∗ ∈ S ∗ , we have
x k − y k , x k − x ∗ ≥ f (x k , y k ) + x k − y k 2 .
Proof Let x ∗ ∈ S ∗ . Since f is pseudomonotone on X with respect to S ∗ , and since x ∗ ∈
K(x k ), we can write, using Proposition 1, that
0 ≥ f (x k , x ∗ ) ≥ f (x k , y k ) + x k − y k , x ∗ − y k .
But then we have
x k − y k , y k − x k + x k − y k , x k − x ∗ ≥ f (x k , y k ),
that is, the thesis.
From Propositions 8 and 9, we deduce immediately that the required property on d k is
k
satisfied for the second linesearch with μ = 1 − c. Indeed for d k = x k − y k + αg k we have
d k , x k − x ∗ = x k − y k +
gk
k
αk , x
− x∗
≥ f (zk , x k ) − f (zk , y k ) + f (x k , y k ) + x k − y k 2
≥ (1 − c) x k − y k 2 .
In particular, when Linesearch 2 is used, the direction d k = 0 whatever g k ∈ ∂f (zk , x k ).
From now on we denote by Algorithms 1 and 2 the General Algorithm with Step 2 replaced
by Linesearches 1 and 2, respectively.
Algorithm 1
Let x 0 ∈ X, let α, c ∈ (0, 1), and let γ ∈ (0, 2). Set k = 0.
Step 1 Compute y k = arg miny∈K(x k ) [f (x k , y) + 21 y − x k 2 ]. If y k = x k , Stop.
Step 2 Otherwise find m the smallest nonnegative integer such that
f (zk,m , x k ) − f (zk,m , y k ) ≥ c x k − y k 2
where zk,m := (1 − α m )x k + α m y k .
Set αk = α m and zk = zk,m .
Step 3 Compute g k ∈ ∂f (zk , x k ) and x k+1 = PK(x k ) (x k − βk d k ) where
dk =
123
gk
αk
and βk = γ c
x k −y k 2
.
d k 2
Set k := k + 1 and go to Step 1.
J Glob Optim
Algorithm 2
Let x 0 ∈ X, let α, c ∈ (0, 1), and let γ ∈ (0, 2). Set k = 0.
Step 1 Compute y k = arg miny∈K(x k ) [f (x k , y) + 21 y − x k 2 ]. If y k = x k , Stop.
Step 2 Otherwise find m the smallest nonnegative integer such that
f (zk,m , x k ) − f (zk,m , y k ) + f (x k , y k ) ≥ −c x k − y k 2
where zk,m := (1 − α m )x k + α m y k .
Set αk = α m and zk = zk,m .
Step 3 Compute g k ∈ ∂f (zk , x k ) and x k+1 = PK(x k ) (x k − βk d k ) where
dk = xk − yk +
gk
αk
and βk = γ (1 − c) xd−y
k 2 . Set k := k + 1 and go to Step 1.
k
k 2
Now for Algorithms 1 and 2 we have the following boundedness properties:
Proposition 10 Let x̄ be a limit point of {x k } and assume that x kj → x̄. Then the sequences
{y kj }, {zkj }, and {g kj } are bounded.
Proof Since the sequences {x kj } and {y kj } are bounded (see Propositions 3 and 4), it follows
that the sequence {zkj } is also bounded because zkj belongs to the segment [x kj ; y kj ] for
every j . So a subsequence of {zkj }, again denoted {zkj }, converges to some z̄ ∈ X. Since
x̄ ∈ X ⊆ , {x kj } ⊆ X ⊆ , x kj → x̄, and the sequence of convex functions {f (zkj , ·)}
converges pointwise to the convex function f (z̄, ·), it follows from Theorem 24.5 of [20]
that there exists j0 such that for all j ≥ j0 ,
∂f (zkj , x kj ) ⊆ ∂f (z̄, x̄) + B
where B denotes the closed Euclidean unit ball of IRn . Since B and ∂f (z̄, x̄) are bounded,
the sequence {g kj } is also bounded.
In order to apply Proposition 5, we need to prove the next result.
Proposition 11 Let x kj → x̄. Then x kj − y kj → 0 as j → ∞.
Proof We examine two cases:
(1) inf j αkj > 0: The sequence {d kj } is bounded because the sequences {x kj }, {y kj }, and
{g kj } are bounded (Proposition 10) and
d kj =
g kj
αkj
(Linesearch 1) or d kj = x kj − y kj +
g kj
αkj
(Linesearch 2).
Since, from (8), xd−y
→ 0, we deduce that x kj − y kj → 0.
k 2
(2) inf j αkj = 0: Then αkj → 0 (for a subsequence). But this implies that αkj < 1 for j
k
k 4
large enough and that the linesearch conditions are not satisfied for
αkj
αkj k
x kj +
y j.
z̄kj = 1 −
α
α
αkj
α
. Let us denote
It is immediate that z̄kj → x̄. Now let us examine separately the two linesearches.
123
J Glob Optim
(a) If the second linesearch is used, we have
f (z̄kj , x kj ) − f (z̄kj , y kj ) + f (x kj , y kj ) < −c x kj − y kj 2 .
Let ȳ be a limit point of the bounded sequence {y kj }. Then, f being continuous, the
left-hand side of the previous inequality tends to
f (x̄, x̄) − f (x̄, ȳ) + f (x̄, ȳ) = 0.
Consequently, we obtain that x kj − y kj → 0.
(b) If the first linesearch is used, we have
f (z̄kj , x kj ) − f (z̄kj , y kj ) < c x kj − y kj 2 .
By definition of y kj we also have
x kj − y kj 2 ≤ −f (x kj , y kj ).
Let ȳ be a limit point of {y kj }. Then combining the two inequalities and taking the limit
as j → ∞, we obtain
f (x̄, x̄) − f (x̄, ȳ) ≤ −cf (x̄, ȳ),
which implies that f (x̄, ȳ) ≥ 0. So −f (x kj , y kj )→−f (x̄, ȳ) ≤ 0 and x kj −y kj →0.
Finally using successively Propositions 5, 6 and 11, we obtain the following convergence
result for Algorithms 1 and 2.
Theorem 1 Any limit point of the sequence {x k } generated by Algorithm 1 or Algorithm 2
is a solution of problem QE(K; f ). If f is strictly monotone at a limit point x̄ of {x k }, then
x k → x̄ as k → ∞.
When f (x, y) = F (x), y−x for all x, y ∈ X, our Algorithm 2 coincides with Algorithm
2 in [31]. So, as a by-product we find again the convergence results obtained in Theorems
5.1 and 5.2 of [31]. In the next section we consider other choices for βk .
5 Variants of Algorithm 1
In Algorithm 1 the next iterate was defined as x k+1 = x k (βk ) where for every β > 0, x k (β) =
k
PK(x k ) (x k − βd k ). The direction was equal to d k = αg k with g k ∈ ∂f (zk , x k ) and αk obtained
by using the first linesearch. Furthermore the step βk was chosen such that the following
inequality holds:
x k (βk ) − x ∗ 2 ≤ x k − x ∗ 2 − γ (2 − γ ) c2
x k − y k 4
.
d k 2
(15)
An example of such a step is βk = γ c xd−y
k 2 .
In this section we show that it is possible to choose other steps βk while keeping true
the inequality (15). So the convergence results obtained in the previous section remain valid
for these variants. These steps will give rise to better decreases on the distance between the
iterates and the set S ∗ .
k
123
k 2
J Glob Optim
By definition of g k ∈ ∂f (zk , x k ), we have
g k , x k − x ∗ ≥ f (zk , x k ) − f (zk , x ∗ )
where x ∗ is any element in S ∗ . Since f is pseudomonotone on X with respect to S ∗ , we
obtain that f (zk , x ∗ ) ≤ 0 and thus that
d k , x k − x ∗ = gk k
f (zk , x k )
, x − x∗ ≥
.
αk
αk
(16)
Using this inequality in (9) and taking
βk = γ
f (zk , x k )
αk ,
g k 2
(17)
we deduce that
x k+1 − x ∗ 2 ≤ x k − x ∗ 2 − 2
β2
βk
f (zk , x k ) + k2 g k 2
αk
αk
≤ x k − x ∗ 2 − γ (2 − γ )
f (zk , x k )2
.
g k 2
Now from the convexity of f (zk , ·) and the first linesearch, we obtain that
f (zk , x k ) ≥ αk (f (zk , x k ) − f (zk , y k )) ≥ αk cx k − y k 2 .
(18)
So if we use the new step βk given in (17), we can conclude that
x k+1 − x ∗ 2 ≤ x k − x ∗ 2 − γ (2 − γ ) αk2 c2
x k − y k 4
,
g k 2
and thus that inequality (15) holds.
Replacing βk by its new value in Step 3 of the General Algorithm, we obtain
f (zk , x k )
k
α
d
= PK(x k ) [x k − γ σk g k ]
x k+1 = PK(x k ) x k − γ
k
g k 2
where σk =
f (zk ,x k )
.
g k 2
Consequently, in that case, Algorithm 1 becomes
Algorithm 1a
Let x 0 ∈ X, let α, c ∈ (0, 1), and let γ ∈ (0, 2). Set k = 0.
Step 1 Compute y k = arg miny∈K(x k ) [f (x k , y) + 21 y − x k 2 ]. If y k = x k , Stop.
Step 2 Otherwise find m the smallest nonnegative integer such that
f (zk,m , x k ) − f (zk,m , y k ) ≥ c x k − y k 2
where zk,m := (1 − α m )x k + α m y k .
Set αk = α m and zk = zk,m .
Step 3 Compute g k ∈ ∂f (zk , x k ) and x k+1 = PK(x k ) (x k − γ σk g k ) where
σk =
f (zk ,x k )
.
g k 2
Set k := k + 1 and go to Step 1.
123
J Glob Optim
Comments
(1) When K(x) = K for every x ∈ X, Algorithm 1a coincides with Algorithm 2a of Tran
et al. [24]. Moreover, when f (x, y) = F (x), y − x for every x, y ∈ X, Algorithm 1a
corresponds to an extragradient algorithm due to Iusem and al. [9].
(2) Consider the hyperplane H k defined by
H k = x ∈ IRn | g k , x k − x = f (zk , x k )
and observe that when f (x, y) = F (x), y − x for every x, y ∈ X, the hyperplane H k
coincides with the one defined in (5). Now this hyperplane H k separates x k from S ∗ .
Indeed, from (18), it follows that f (zk , x k ) > 0 = g k , x k − x k and from (16), that
g k , x k − x ∗ ≥ f (zk , x k ) for every x ∗ ∈ S ∗ . Furthermore g k is the normal vector to
H k and since x k − σk g k ∈ H k , we can say that x k − σk g k is the orthogonal projection
of x k onto H k . So when γ = 1, the new iterate x k+1 is obtained by projecting onto
K(x k ) the projection of x k onto H k .
Since the set S ∗ ⊆ K ∗ is contained in K(x k ) ∩ H+k where
H+k = x ∈ IRn | g k , x k − x ≥ f (zk , x k ) ,
a variant of Algorithm 1a consists in replacing in Step 3 the iterate x k+1 = PK(x k ) (x k −
γ σk g k ) by x k+1 = PK(x k )∩H k (x k − γ σk g k ). Using the nonexpansiveness of PK(x k )∩H k
+
+
instead of the one of PK(x k ) , we immediately obtain that the inequality (15) holds. So the
convergence of the sequence {x k } is preserved for this variant. In the sequel we denote by
Algorithm 1b the Algorithm 1a where Step 3 has been replaced by
Step 3 Compute g k ∈ ∂f (zk , x k ) and x k+1 = PK(x k )∩H k (x k − γ σk g k ) where
+
σk =
f (zk ,x k )
g k 2
and H+k = {x ∈ IRn | g k , x k − x ≥ f (zk , x k )}.
Set k := k + 1 and go to Step 1.
When γ = 1 we have that x k − γ σk g k = PH k (x k ) and we can use the proof of Lemma 2.2
+
in [21] to show that
x k+1 = PK(x k )∩H k (x k ).
+
When γ = 1, it is also possible to give a geometric interpretation of Step 3 in Algorithms 1a
and 1b. In that purpose we recall the following property of the orthogonal projection onto a
convex set.
Lemma 2 [26] Let C be a nonempty closed convex subset in IRn . Then for any x ∈ IRn and
z ∈ C,
PC (x) − z2 ≤ x − z2 − PC (x) − x2 .
Using Lemma 2 with C = K(x k ), x = x k − βd k and z = x ∗ , we can write
x k (β) − x ∗ 2 ≤ x k − βd k − x ∗ 2 − x k (β) − x k + βd k 2 .
Developing the first term of the right-hand side of this inequality and using successively (9),
(16), and (17), we obtain
x k (β) − x ∗ 2 ≤ x k − x ∗ 2 + ϕ k (β)
123
(19)
J Glob Optim
where
ϕ k (β) = ϕk (β) − x k (β) − x k +
β k 2
g αk
with
ϕk (β) = −2
β
β2
f (zk , x k ) + 2 g k 2 .
αk
αk
Since ϕ k (β) ≤ ϕk (β), we have, in particular, that
x k (β) − x ∗ 2 ≤ x k − x ∗ 2 + ϕk (β).
(20)
f (zk ,x k )
αk = σk αk minimizes the right-hand side of (20). Since
g k 2
k
σk αk d ) = PK(x k ) (x k − σk g k ), it follows that the new iterate x k+1
It is easy to check that β1 =
x k (β1 ) = PK(x k ) (x k −
in Algorithm 1a is given by
x k+1 = x k (β1 ).
Now if we minimize the right-hand side of (19), it can be shown exactly as in [26] that the
function ϕ k (β) is convex and admits a minimum for a steplength β2 ≥ β1 . Computing an
explicit value for β2 seems difficult but it is possible, using a proof similar to the one of
Lemma 3.2 in [26], to show that
x k (β2 ) = PK(x k )∩H k (x k − β1 d k ) = PK(x k )∩H k (x k − σk g k ).
+
Hence the new iterate
x k+1
+
in Algorithm 1b is given by
x k+1 = x k (β2 ).
To end this section we summarize Algorithm 1b.
Algorithm 1b
Let x 0 ∈ X, let α, c ∈ (0, 1), and let γ ∈ (0, 2). Set k = 0.
Step 1 Compute y k = arg miny∈K(x k ) [f (x k , y) + 21 y − x k 2 ]. If y k = x k , Stop.
Step 2 Otherwise find m the smallest nonnegative integer such that
f (zk,m , x k ) − f (zk,m , y k ) ≥ c x k − y k 2
where zk,m := (1 − α m )x k + α m y k .
Set αk = α m and zk = zk,m .
Step 3 Compute g k ∈ ∂f (zk , x k ) and x k+1 = PK(x k )∩H k (x k − γ σk g k ) where
+
σk =
f (zk ,x k )
g k 2
and H+k = {x ∈ IRn | g k , x k − x ≥ f (zk , x k )}.
Set k := k + 1 and go to Step 1.
When γ = 1, we have seen that x k+1 = PK(x k )∩H k (x k ). Consequently, when K(x) = K
+
for every x ∈ X and f (x, y) = F (x), y −x, we find again the projection method introduced
by Solodov and al. for solving variational inequality problems [21] (see also Algorithm 3.1
in [26]).
123
J Glob Optim
6 Variants of Algorithm 2
In Algorithm 2 the next iterate was defined as x k+1 = x k (βk ) where for every β > 0, x k (β) =
k
PK(x k ) (x k − βd k ). The direction was equal to d k = x k − y k + αg k with g k ∈ ∂f (zk , x k ) and
αk obtained by using the second linesearch. Furthermore the step βk was chosen such that
the following inequality holds:
x k (βk ) − x ∗ 2 ≤ x k − x ∗ 2 − γ (2 − γ ) (1 − c)2
x k − y k 4
.
d k 2
(21)
An example of such a step is βk = γ (1 − c) xd−y
k 2 .
In this section we show that it is possible to choose other steps βk while keeping true
the inequality (21). So the convergence results obtained in Section 3 remain valid for these
variants. These steps will give rise to better decreases on the distance between the iterates
and the set S ∗ .
From the definition of d k , (16) and Proposition 9, we obtain that
k
d k , x k − x ∗ ≥
k 2
f (zk , x k )
+ f (x k , y k ) + x k − y k 2 .
αk
(22)
Using this inequality in (9) and taking
βk = γ σk where σk =
f (zk ,x k )
αk
+ f (x k , y k ) + x k − y k 2
d k 2
,
we deduce that
x k+1 − x ∗ 2 ≤ x k − x ∗ 2 − 2βk σk d k 2 + βk2 d k 2
≤ x k − x ∗ 2 − γ (2 − γ ) σk2 d k 2 .
Now we observe that
σk2 d k 2 ≥ (1 − c)2
x k − y k 4
> 0.
d k 2
(23)
Indeed, using successively the definition of σk , the second linesearch and the convexity of
f (zk , ·) (see the first inequality in (18)), we obtain
σk d k 2 =
≥
f (zk , x k )
+ f (x k , y k ) + x k − y k 2
αk
f (zk , x k )
+ f (zk , y k ) − f (zk , x k ) + (1 − c)x k − y k 2
αk
≥ (1 − c)x k − y k 2 .
(24)
Taking the square of both members of (24) we deduce the inequality (23) as well as the
estimation
x k+1 − x ∗ 2 ≤ x k − x ∗ 2 − γ (2 − γ ) (1 − c)2
x k − y k 4
.
d k 2
So inequality (21) is satisfied and the convergence of Algorithm 2 is preserved when the
stepsize βk is equal to γ σk .
123
J Glob Optim
With this stepsize βk , Algorithm 2 becomes
Algorithm 2a
Let x 0 ∈ X, let α, c ∈ (0, 1), and let γ ∈ (0, 2). Set k = 0.
Step 1 Compute y k = arg miny∈K(x k ) [f (x k , y) + 21 y − x k 2 ]. If y k = x k , Stop.
Step 2 Otherwise find m the smallest nonnegative integer such that
f (zk,m , x k ) − f (zk,m , y k ) + f (x k , y k ) ≥ −c x k − y k 2
where zk,m := (1 − α m )x k + α m y k .
Set αk = α m and zk = zk,m .
Step 3 Compute g k ∈ ∂f (zk , x k ) and x k+1 = PK(x k ) (x k − γ σk d k ) where
gk
and σk =
d =x −y +
αk
k
k
k
f (zk ,x k )
αk
+ f (x k , y k ) + x k − y k 2
d k 2
.
Set k := k + 1 and go to Step 1.
Comments
(1) When for every x, y ∈ X, K(x) = K and f (x, y) = F (x), y − x, Algorithm 2a
corresponds to Algorithm NVE-1 due to Wang et al. [27].
(2) Consider the hyperplane Gk defined by
f (zk , x k )
k
n
k
k
k
k
k
k 2
+ f (x , y ) + x − y G = x ∈ IR | d , x − x =
αk
and observe that when f (x, y) = F (x), y − x for every x, y ∈ X, this hyperplane
coincides with the one defined in (7). Now using (22) and (24), it is easy to prove that
Gk separates x k from S ∗ and that x k − σk d k is the orthogonal projection of x k onto
Gk . So when γ = 1, the new iterate x k+1 is obtained by projecting onto K(x k ) the
projection of x k onto Gk .
Since the set S ∗ ⊆ K ∗ is contained in K(x k ) ∩ Gk+ where
f (zk , x k )
+ f (x k , y k ) + x k − y k 2 ,
Gk+ = x ∈ IRn | d k , x k − x ≥
αk
a variant of Algorithm 2a consists in replacing in Step 3 the iterate x k+1 = PK(x k ) (x k −
γ σk d k ) by x k+1 = PK(x k )∩Gk (x k − γ σk d k ). Using the nonexpansiveness of PK(x k )∩Gk
+
+
instead of the one of PK(x k ) , we immediately obtain that the inequality (21) holds. So the
convergence of the sequence {x k } is preserved for this variant. In the sequel we denote by
Algorithm 2b the Algorithm 2a where Step 3 has been replaced by
Step 3 Compute g k ∈ ∂f (zk , x k ) and x k+1 = PK(x k )∩Gk (x k − γ σk d k ) where
+
dk = xk − yk +
gk
and σk =
αk
f (zk ,x k )
αk
+ f (x k , y k ) + x k − y k 2
d k 2
.
Set k := k + 1 and go to Step 1.
123
J Glob Optim
When γ = 1 we have that x k − γ σk g k = PGk (x k ) and we can use the proof of Lemma 2.2
+
in [21] to show that
x k+1 = PK(x k )∩Gk (x k ).
+
When γ = 1, it is also possible to give a geometric interpretation of Step 3 in Algorithms 2a
and 2b. Using (9), (22) and Lemma 2 as in the previous section, we obtain for any x ∗ ∈ S ∗
x k (β) − x ∗ 2 ≤ x k − x ∗ 2 − ϕ k (β)
(25)
where
ϕ k (β) = ϕk (β) − x k (β) − x k + βd k 2
with
ϕk (β) = −2β σk d k 2 + β 2 d k 2 .
Since ϕ k (β) ≤ ϕk (β), we have, in particular, that
x k (β) − x ∗ 2 ≤ x k − x ∗ 2 + ϕk (β).
(26)
It is easy to check that β1 = σk minimizes the right-hand side of (26). Since x k (β1 ) =
PK(x k ) (x k − β1 d k ) = PK(x k ) (x k − σk d k ), it follows that the iterate x k+1 in Algorithm 2a is
given by
x k+1 = x k (β1 ).
Now if we minimize the right-hand side of (25), it can be shown exactly as in [27] that the
function ϕ k (β) is convex and admits a minimum for a steplength β2 ≥ β1 . Computing an
explicit value for β2 seems difficult but it is possible, using a proof similar to the one of
Lemma 3.2 in [27], to show that
x k (β2 ) = PK(x k )∩Gk (x k − β1 d k ) = PK(x k )∩Gk (x k − σk d k )
+
+
Hence the iterate x k+1 in Algorithm 2b is given by
x k+1 = x k (β2 ).
Since ϕ k (β2 ) ≤ ϕ k (β1 ) ≤ ϕk (β1 ), Algorithm 2b should be faster than Algorithm 2a. To end
this section we give a description of Algorithm 2b.
Algorithm 2b
Let x 0 ∈ X, let α, c ∈ (0, 1), and let γ ∈ (0, 2). Set k = 0.
Step 1 Compute y k = arg miny∈K(x k ) [f (x k , y) + 21 y − x k 2 ]. If y k = x k , Stop.
Step 2 Otherwise find m the smallest nonnegative integer such that
f (zk,m , x k ) − f (zk,m , y k ) + f (x k , y k ) ≥ −c x k − y k 2
where zk,m := (1 − α m )x k + α m y k .
Set αk = α m and zk = zk,m .
Step 3 Compute g k ∈ ∂f (zk , x k ) and x k+1 = PK(x k )∩Gk (x k − γ σk d k ) where
+
dk = xk − yk +
123
gk
,
αk
σk =
f (zk ,x k )
αk
+ f (x k , y k ) + x k − y k 2
d k 2
J Glob Optim
and
Gk+ = {x ∈ IRn | d k , x k − x ≥
f (zk , x k )
+ f (x k , y k ) + x k − y k 2 }.
αk
Set k := k + 1 and go to Step 1.
When γ = 1, and for every x, y ∈ X, K(x) = K and f (x, y) = F (x), y − x, we find
again the projection method NVE-2 introduced by Wang et al. [27] for solving variational
inequality problems.
7 Numerical results
In order to compare the numerical behavior of the six algorithms presented in the previous sections we have implemented them in MATLAB and we have first solved an academic
quasi-equilibrium problem of size n = 5 and then three examples of generalized Nash games.
Example 1 The bifunction f of the quasi-equilibrium problem comes from the Cournot-Nash
equilibrium model considered in [24]. It is defined for each x, y ∈ IR5 , by
f (x, y) = P x + Qy + q, y − x
where q ∈ IR5 , and P and Q are two square matrices of order 5 such that Q is symmetric
positive semidefinite and Q − P is negative semidefinite. In our tests the vector q and the
matrices P and Q are chosen as follows:
⎤
⎡
⎤
⎡
⎤
⎡
1.6 1
0
0 0
3.1 2
0
0 0
1
⎢ 1 1.6 0
⎢ 2 3.6 0
⎢ −2 ⎥
0 0⎥
0 0⎥
⎥
⎢
⎥
⎢
⎥
⎢
⎢
⎥
⎢
⎥
⎢
0 1.5 1 0 ⎥
0 3.5 2 0 ⎥ ; Q = ⎢ 0
q = ⎢ −1 ⎥ ; P = ⎢ 0
⎥.
⎣ 0
⎣ 0
⎣ 2⎦
0
1 1.5 0 ⎦
0
2 3.3 0 ⎦
0
0
0
0 2
0
0
0
0 3
−1
Instead of the fixed constraint set used in [24] we consider the moving set K(x) =
1≤i≤5 Ki (x) where for each x ∈ IR5 and each i, the set Ki (x) is defined by
⎧
⎫
⎨
⎬
xj ≥ −1 .
Ki (x) = yi ∈ IR | yi +
⎩
⎭
1≤j ≤5,j =i
The numerical results obtained for this example are listed in Table 1 for Algorithms 1, 1a,
and 1b and in Table 2 for Algorithms 2, 2a, and 2b. Two different starting points are chosen
and for each test the number of iterations and the CPU time needed to get a solution are
reported.
Table 1 The results of Example 1 for Algorithms 1, 1a, and 1b
Starting point
Number of iterations
CPU (s)
Algorithm 1
Algorithm 1a
Algorithm 1b
Algorithm 1
Algorithm 1a
Algorithm 1b
(1, 3, 1, 1, 2)T
27
20
20
0.64
0.48
0.53
(0, 0, 0, 0, 0)T
20
15
15
0.59
0.54
0.31
123
J Glob Optim
Table 2 The results of Example 1 for Algorithms 2, 2a, and 2b
Starting point
Number of iterations
CPU (s)
Algorithm 2
Algorithm 2a
Algorithm 2b
Algorithm 2
Algorithm 2a
Algorithm 2b
(1, 3, 1, 1, 2)T
105
41
41
3.04
0.96
1.20
(0, 0, 0, 0, 0)T
59
32
32
1.42
0.65
0.53
Algorithm 1a
Algorithm 1b
Table 3 The results of Example 2 for Algorithms 1, 1a, and 1b
Starting point
Number of iterations
CPU (s)
Algorithm 1
Algorithm 1a
Algorithm 1b
289
145
128
9.23
4.78
2.42
3
2
2
0.26
0.20
0.15
(10, 10)T
255
120
120
8.43
3.70
2.57
(0, 10)T
259
193
195
8.20
6.10
3.76
(5, 5)T
254
119
119
8.21
3.81
2.40
(0, 0)T
(10, 0)T
Algorithm 1
In a second step we consider three examples of generalized Nash games. They are described
in the paper by Zhang et al. [31] (see also [8]) where they have been solved after having written
them under the form of quasi-variational inequalities. The algorithm used by these authors is
nothing else than our Algorithm 2 when f (x, y) = F (x), y − x. In this paper we consider
these three quasi-variational inequality problems and we solve them by using each of our
six algorithms with the equilibrium function f (x, y) = F (x), y − x corresponding to a
quasi-variational inequality problem. Doing so we can compare the numerical behavior of
our six algorithms with the Zhang and al. Algorithm 2 given in [31, p. 101].
Example 2 Consider a two-person game whose QVI formulation involves the function F =
(F1 , F2 ) and the multivalued mapping K defined for each x = (x1 , x2 ) ∈ IR2 by
F1 (x1 , x2 ) = 2x1 + 83 x2 − 34,
F2 (x1 , x2 ) = 2x2 + 54 x1 − 24.25,
K1 (x2 ) = {y1 ∈ IR | 0 ≤ y1 ≤ 10, y1 ≤ 15 − x2 },
K2 (x1 ) = {y2 ∈ IR | 0 ≤ y2 ≤ 10, y2 ≤ 15 − x1 }.
The set of solutions is composed of the point (5, 9)T and the line segment [(9, 6)T , (10, 5)T ].
The numerical results obtained for this example are listed in Tables 3 and 4 where different
starting points are chosen. The computed solution up to 10−6 is (5, 9)T for all the chosen
starting points except for the point (10, 0)T where the solution (10, 5)T is obtained.
Example 3 Consider an oligopolistic market in which five firms supply a homogeneous product in a noncooperative fashion. The function F = (F1 , . . . , F5 ) is defined for each x ∈ IR5 ,
by
(1/βi ) xi
5000 (1/η) xi
− 1 , i = 1, . . . , 5
+
Fi (x) = ci +
τi
Q
ηQ
123
J Glob Optim
Table 4 The results of Example 2 for Algorithms 2, 2a, and 2b
Starting point
Number of iterations
CPU (s)
Algorithm 2
Algorithm 2a
Algorithm 2b
Algorithm 2
Algorithm 2a
Algorithm 2b
(0, 0)T
236
187
188
7.75
6.34
4.32
(10, 0)T
(10, 10)T
297
245
246
9.37
8.12
5.18
(0, 10)T
188
146
145
6.01
4.85
2.78
(5, 5)T
297
244
245
9.35
7.95
5.10
Table 5 The results of Example 3 for Algorithms 1, 1a, and 1b
Starting point
Number of iterations
CPU (s)
Algorithm 1 Algorithm 1a Algorithm 1b Algorithm 1 Algorithm 1a Algorithm 1b
(10, 10, 10, 10, 10)T 36
24
24
1.31
0.82
0.60
(50, 50, 50, 50, 50)T 44
27
27
1.68
0.96
0.53
Table 6 The results of Example 3 for Algorithms 2, 2a, and 2b
Starting point
Number of iterations
CPU (s)
Algorithm 2 Algorithm 2a Algorithm 2b Algorithm 2 Algorithm 2a Algorithm 2b
(10, 10, 10, 10, 10)T
96
65
66
3.07
2.23
1.39
(50, 50, 50, 50, 50)T 110
70
71
3.54
2.35
1.28
where Q = 1≤i≤5 x i . The reader is referred to the paper by Zhang et al. [31] for more
details about this problem and the different values of the parameters ci , βi , τi and η. On the
other hand, the sets Ki (x−i ), i = 1, . . . , 5, are defined as follows
xj ≤ N }.
Ki (x−i ) = {xi ∈ IR | 1 ≤ xi ≤ 150, xi +
j =i
The numerical results obtained for this example are listed in Tables 5 and 6 for two different
starting points. The computed solution up to 10−6 is the same for the two starting points:
(36.9325, 41.8181, 43.7066, 42.6592, 39.1790).
Example 4 In this example, the set K2 (x1 ) in Example 2 is replaced by K2 (x1 ) = {y2 ∈
IR | 0 ≤ y2 ≤ 10}. The corresponding numerical results are given in Tables 7 and 8. As for
Example 2, the computed solution up to 10−6 is (5, 9)T for all the chosen starting points
except for the point (10, 0)T where the solution (10, 5)T is obtained.
For each example and for each test the algorithm is terminated when x k − y k < ε for an
iterate x k . For each of our computational experiments the following values of the parameters
have been used: ε = 10−6 , γ = 1.99, α = 0.5 and c = 0.5. Furthermore, for each test the
number of iterations and the corresponding CPU times in seconds are reported. The symbol
“” means that the number of iterations exceeds 1000.
123
J Glob Optim
Table 7 The results of Example 4 for Algorithms 1, 1a, and 1b
Starting point
Number of iterations
CPU (s)
Algorithm 1
Algorithm 1a
Algorithm 1b
Algorithm 1
Algorithm 1a
Algorithm 1b
(0, 0)T
282
201
201
9.14
6.26
4.25
(10, 0)T
155
207
130
5.06
6.59
2.62
(10, 10)T
218
177
177
6.65
5.71
3.87
(0, 10)T
258
195
195
8.21
6.10
4.03
(5, 5)T
281
123
123
8.70
3.93
2.45
Table 8 The results of Example 4 for Algorithms 2, 2a, and 2b
Starting point
Number of iterations
CPU (s)
Algorithm 2
Algorithm 2a
Algorithm 2b
Algorithm 2
Algorithm 2a
Algorithm 2b
(0, 0)T
298
246
246
9.45
7.81
5.26
(10, 0)T
305
248
247
9.68
7.79
5.21
(10, 10)T
198
168
168
6.23
5.28
3.42
(0, 10)T
190
146
146
5.98
4.64
2.84
(5, 5)T
289
237
237
9.18
7.64
4.76
From these preliminary results, it seems that the numerical behavior of Algorithms 1a, 1b,
and 2a, 2b is better than the one of Algorithms 1 and 2, respectively. Furthermore the number
of iterations used by Algorithms 1a and 1b to get a solution is very similar. The difference
between them lies in the CPU time which is systematically smaller for the second one. A
similar behavior can be observed for Algorithms 2a and 2b.
8 Conclusion
In this paper we have presented a very general algorithm for solving the class of quasi-equilibrium problems. Then we have studied and proven the convergence of six new implementable
versions of this algorithm based on the extragradient method and on two different linesearches. Only very preliminary numerical results have been reported. Many other test problems
should be considered and other choices of linesearches and parameter values should be studied to improve the performance of these algorithms. This could be the subject of future
researches.
Acknowledgments
The authors thank the two referees for their helpful suggestions for improving this paper.
References
1. Bensoussan, A.: Points de Nash dans le cas de fonctionnelles quadratiques et jeux différentiels linéaires
à N personnes. SIAM J. Control 12, 460–499 (1974)
2. Blum, E., Oettli, W.: From optimization and variational inequalities to equilibrium problems. Math.
Stud. 63, 123–145 (1994)
123
J Glob Optim
3. Chinchuluun, A., Pardalos, P.M., Migdalas, A., Pitsoulis, L. (eds.): Pareto Optimality, Game Theory and
Equilibria. Springer, New York (2008)
4. Facchinei, F., Kanzow, C.: Generalized Nash equilibrium problems. Ann. Oper. Res. 175, 177–211 (2010)
5. Facchinei, F., Pang, J.-S.: Finite-Dimensional Variational Inequalities and Complementarity Problems. Springer, New York (2003)
6. Fukushima, M.: A class of gap functions for quasi-variational inequality problems. J. Ind. Manag.
Optim. 3, 165–171 (2007)
7. Gianessi, F., Maugeri, A., Pardalos, P.M. (eds.): Equilibrium Problems: Nonsmooth Optimization and
Variational Inequality Models. Kluwer, New York (2002)
8. Harker, P.T.: Generalized Nash games and quasi-variational inequalities. Eur. J. Oper. Res. 54,
81–94 (1991)
9. Iusem, A.N., Svaiter, B.F.: A variant of Korpelevich’s method for variational inequalities with a new
search strategy. Optimization 42, 309–321 (1997)
10. Khobotov, E.N.: Modification of the extragradient method for solving variational inequalities and certain
optimization problems. USSR Comput. Math. Phys. 27, 120–127 (1987)
11. Konnov, I.V.: Equilibrium Models and Variational Inequalities. Mathematics in Science and Engineering. Elsevier, Amsterdam (2007)
12. Korpelevich, G.M.: The extragradient method for finding saddle points and other problems. Matekon 12, 747–756 (1976)
13. Kubota, K., Fukushima, M.: Gap function approach to the generalized Nash equilibrium problem.
J. Optim. Theory Appl. 144, 511–531 (2010)
14. Noor, M.A.: A modified extragradient method for general monotone variational inequalities. Comput.
Math. Appl. 38, 19–24 (1999)
15. Noor, M.A.: On merit functions for quasivariational inequalities. J. Math. Inequal. 1, 259–268 (2007)
16. Palomar, D.P., Eldar, Y.C. (eds.): Convex Optimization in Signal Processing and Communications.
Cambridge University Press, Cambridge (2010)
17. Pang, J.-S., Fukushima, M.: Quasi-variational inequalities, generalized Nash equilibria, and multi-leaderfollower games. Comput. Manag. Sci. 2, 21–56 (2005)
18. Pang, J.-S., Fukushima, M.: Quasi-variational inequalities, generalized Nash equilibria, and multi-leaderfollower games. Erratum. Comput. Manag. Sci. 6, 373–375 (2009)
19. Pardalos, P.M., Rassias, T.M., Khan, A.A. (eds.): Nonlinear Analysis and Variational Problems. Springer,
New York (2010)
20. Rockafellar, R.T.: Convex Analysis. Princeton University Press, Princeton (1970)
21. Solodov, M.V., Svaiter, B.F.: A new projection method for variational inequality problems. SIAM J.
Control Optim. 37, 765–776 (1999)
22. Smeers, Y., Oggioni, G., Allevi, E., Schaible, S.: Generalized Nash equilibrium and market coupling in
the European power system. EPRG Working Paper 1016, Cambridge Working Paper in Economics 1034
(2010)
23. Taji, K.: On gap functions for quasi-variational inequalities. Abstract and Applied Analysis 2008, Article
ID 531361, 7 pages (2008)
24. Tran, D.Q., Le Dung, M., Nguyen, V.H.: Extragradient algorithms extended to equilibrium problems.
Optimization 57, 749–776 (2008)
25. von Heusinger, A.: Numerical Methods for the Solution of the Generalized Nash Equilibrium Problem.
PhD Thesis, University of Wuerzburg, Germany (2009)
26. Wang, Y., Xiu, N., Wang, C.: Unified framework of extragradient-type methods for pseudomonotone
variational inequalities. J. Optim. Theory Appl. 111, 641–656 (2001)
27. Wang, Y., Xiu, N., Wang, C.: A new version of extragradient method for variational inequality problems. Comput. Math. Appl. 42, 969–979 (2001)
28. Wei, J.Y., Smeers, Y.: Spatial oligopolistic electricity models with Cournot generators and regulated
transmission prices. Oper. Res. 47, 102–112 (1999)
29. Xiu, N., Wang, Y., Zhang, X.: Modified fixed-point equations and related iterative methods for variational
inequalities. Comput. Math. Appl. 47, 913–920 (2004)
30. Xiu, N., Zhang, J.: Some recent advances in projection-type methods for variational inequalities.
J. Comput. Appl. Math. 152, 559–585 (2003)
31. Zhang, J., Qu, B., Xiu, N.: Some projection-like methods for the generalized Nash equilibria. Comput.
Optim. Appl. 45, 89–109 (2010)
123