 Research
 Open Access
 Published:
A new extragradient method for generalized variational inequality in Euclidean space
Fixed Point Theory and Applications volume 2013, Article number: 139 (2013)
Abstract
In this paper, we extend the extragradient projection method proposed in (Wang et al. in J. Optim. Theory Appl. 119:167183, 2003) for the classical variational inequalities to the generalized variational inequalities. For this algorithm, we first prove its expansion property of the generated sequence with respect to the starting point and then show that the existence of the solution to the problem can be verified through the behavior of the generated sequence. The global convergence of the method is also established under mild conditions.
MSC:90C30, 15A06.
1 Introduction
Let F(x) be a multivalued mapping from {R}^{n} into {2}^{{R}^{n}} with nonempty values, and let X be a nonempty closed convex set in {R}^{n}. The problem of generalized variational inequalities (GVI) [1, 2] is to find {x}^{\ast}\in X such that there exists {\omega}^{\ast}\in F({x}^{\ast}) satisfying
where \u3008\cdot ,\cdot \u3009 stands for the Euclidean inner product of vectors in {R}^{n}. The solution set of problem (1.1) is denoted by {X}^{\ast}. Certainly, the GVI reduces to the classical variational inequalities (VI) when F is a singlevalued mapping, which has been well studied in the past decades [3, 4].
For the GVI, theories and solution methods have been extensively considered in the literature [2, 5–12], and it is well known that the existence of solutions is an important topic for the GVI [1]. Generally, there are mainly two approaches to attack the solution existence problem of the GVI. One is an analytic approach which first reformulates the GVI as a wellstudied mathematical problem and then invokes an existence theorem for the latter problem [13]. The second is a constructive approach in which the existence can be verified by the behavior of the proposed method. The algorithm that is considered in this paper belongs to the second approach.
First, we give a short summary to the constructive approach on the existence theory for the VI. For this approach, the equivalence between the existence of solutions to the VI problem and the boundedness of the sequence generated by some modified extragradient methods was first established by Sun in [14]. Later, Wang et al. [4] established the same theory by a new type of extragradienttype method. Furthermore, the generated sequence possesses an expansion property with respect to the starting point and converges to a solution point if the solution set of the VI is nonempty. Now a question is posed naturally: as the GVI problem is an extension of the VI, can this theory be extended to the GVI? This constitutes the main motivation of the paper.
In this paper, inspired by the work in [4], we propose a new type of extragradient projection method for the problem GVI. We first establish the existence results for the GVI under pseudomonotonicity and continuity assumption of the underlying mapping F, and then show the global convergence of the proposed method.
The rest of this paper is organized as follows. In Section 2, we give some related concepts and conclusions. In Section 3, we present the description of the algorithm and establish some properties of the algorithm. The global convergence of the sequence is also established.
2 Preliminaries
For a nonempty closed convex set K\subset {R}^{n} and a vector x\in {R}^{n}, the orthogonal projection of x onto K, i.e.,
is denoted by {P}_{K}(x). In what follows, we state some wellknown properties of the projection operator which will be used in the sequel.
Lemma 2.1 [15]
Let K be a nonempty, closed and convex subset in {R}^{n}. Then, for any x,y\in {R}^{n} and z\in K, the following statements hold:

(i)
\u3008{P}_{K}(x)x,z{P}_{K}(x)\u3009\ge 0;

(ii)
{\parallel {P}_{K}(x){P}_{K}(y)\parallel}^{2}\le {\parallel xy\parallel}^{2}{\parallel {P}_{K}(x)x+y{P}_{K}(y)\parallel}^{2}.
Remark 2.1 In fact, (i) in Lemma 2.1 provides also a sufficient and necessary condition for a vector u\in K to be the projection of the vector x; i.e., u={P}_{K}(x) if and only if
Definition 2.1 Let K be a nonempty subset of {R}^{n}. A multivalued mapping F:K\to {2}^{{R}^{n}} is said to be

(i)
monotone if and only if
\u3008uv,xy\u3009\ge 0,\phantom{\rule{1em}{0ex}}\mathrm{\forall}x,y\in K,u\in F(x),v\in F(y); 
(ii)
pseudomonotone if and only if, for any x,y\in K, u\in F(x), v\in F(y),
\u3008u,yx\u3009\ge 0\phantom{\rule{1em}{0ex}}\u27f9\phantom{\rule{1em}{0ex}}\u3008v,yx\u3009\ge 0.
Now let us recall the definition of a continuous multivalued mapping F.
Definition 2.2 Assume that F:X\to {2}^{{R}^{n}} is a multivalued mapping, then

(i)
F is said to be upper semicontinuous at x\in X if for every open set V containing F(x), there is an open set U containing x such that F(y)\subset V for all y\in X\cap U;

(ii)
F is said to be lower semicontinuous at x\in X if given any sequence \{{x}^{k}\} converging to x and any y\in F(x), there exists a sequence {y}^{k}\in F({x}^{k}) that converges to y.
F is said to be continuous at x\in X if it is both upper semicontinuous and lower semicontinuous at x.
For the simplicity of our description, we list the assumptions needed in the sequel.
Assumption 2.1 Suppose that X is a nonempty closed convex set in {R}^{n}. The multivalued mapping F:X\to {2}^{{R}^{n}} is pseudomonotone and continuous on X with nonempty compact convex values.
3 Main results
For x\in {R}^{n}, \xi \in F(x), we first define the projection residue
It is well known that the projection residue is related intimately to the solution set {X}^{\ast}.
Proposition 3.1 For x\in X and \xi \in F(x), they solve problem (1.1) if and only if
The basic idea of the designed algorithm is as follows. At each step of the algorithm, compute the projection residue r({x}^{k},{\xi}^{k}) at iterate {x}^{k}. If it is a zero vector, then stop with {x}^{k} being a solution of the GVI; otherwise, find a trial point {y}^{k} by a backtracking search at {x}^{k} along the residue r({x}^{k},{\xi}^{k}), and the new iterate is obtained by projecting {x}^{0} onto the intersection of X with two halfspaces respectively associated with {y}^{k} and {x}^{k}. Repeat this process until the projection residue is a zero vector.
Algorithm 3.1 Choose \sigma ,\gamma \in (0,1), {x}^{0}\in X, k=0.
Step 1: Given the current iterate {x}^{k}\in X, if \parallel r({x}^{k},\xi )\parallel =0 for some \xi \in F({x}^{k}), stop; else take any {\xi}^{k}\in F({x}^{k}) and compute
Then, let
where {\eta}_{k}={\gamma}^{{m}_{k}}, with {m}_{k} being the smallest nonnegative integer m satisfying: \mathrm{\exists}{\zeta}^{k}\in F({x}^{k}{\gamma}^{m}r({x}^{k},{\xi}^{k})) such that
Step 2: Let {x}^{k+1}={P}_{{H}_{k}^{1}\cap {H}_{k}^{2}\cap X}({x}^{0}), where
Select k=k+1 and go to Step 1.
Now, we first discuss the feasibility of the stepsize rule (3.1).
Lemma 3.1 If {x}^{k} is not a solution of problem (1.1), then there exists the smallest nonnegative integer m satisfying (3.1).
Proof By the definition of r({x}^{k},{\xi}^{k}) and Lemma 2.1, we know that
which implies
Since \gamma \in (0,1), we get
By this and the fact that F is lower semicontinuous, there exists {\zeta}^{m}\in F({x}^{k}{\gamma}^{m}r({x}^{k},{\xi}^{k})) such that
So,
which implies the conclusion. □
The following lemma shows that the halfspace {H}_{k}^{1} in Algorithm 3.1 strictly separates {x}^{k} and the solution set {X}^{\ast} if {X}^{\ast} is nonempty.
Lemma 3.2 If {X}^{\ast}\ne \mathrm{\varnothing}, the halfspace {H}_{k}^{1} in Algorithm 3.1 separates the point {x}^{k} from the set {X}^{\ast}. Moreover,
Proof By the definition of r({x}^{k},{\xi}^{k}) and Algorithm 3.1, we know
which can be written
Then, by this and (3.1), we get
where {\zeta}^{k} is a vector in F({y}^{k}). So, by the definition of {H}_{k}^{1} and (3.3), we get {x}^{k}\notin {H}_{k}^{1}.
On the other hand, for any {x}^{\ast}\in {X}^{\ast} and x\in X, we have
Since F is pseudomonotone on X, we get
Let x={y}^{k} in (3.4), for any {\zeta}^{k}\in F({y}^{k}), we have
which implies {x}^{\ast}\in {H}_{k}^{1}. Moreover, it is easy to see that {X}^{\ast}\subseteq {H}_{k}^{1}\cap X, \mathrm{\forall}k\ge 0. □
The following lemma says that if the solution set is nonempty, then {X}^{\ast}\subseteq {H}_{k}^{1}\cap {H}_{k}^{2}\cap X and thus {H}_{k}^{1}\cap {H}_{k}^{2}\cap X is a nonempty set.
Lemma 3.3 If the solution set {X}^{\ast}\ne \mathrm{\varnothing}, then {X}^{\ast}\subseteq {H}_{k}^{1}\cap {H}_{k}^{2}\cap X for all k\ge 0 under Assumption 2.1.
Proof From the analysis above, it is sufficient to prove that {X}^{\ast}\subseteq {H}_{k}^{2} for all k\ge 0. The proof will be given by induction. Obviously, if k=0,
Now, suppose that
holds for k=l\ge 0. Then
For any {x}^{\ast}\in {X}^{\ast}, by Lemma 2.1 and the fact that
it holds that
Thus {X}^{\ast}\subseteq {H}_{l+1}^{2}. This shows that {X}^{\ast}\subseteq {H}_{k}^{2} for all k\ge 0, and the desired result follows. □
For the case that the solution set is empty, we have that {H}_{k}^{1}\cap {H}_{k}^{2}\cap X is also nonempty from the following lemma, which implies the feasibility of Algorithm 3.1.
Lemma 3.4 Suppose that {X}^{\ast}=\mathrm{\varnothing}, then {H}_{k}^{1}\cap {H}_{k}^{2}\cap X\ne \mathrm{\varnothing} for all k\ge 0 under Assumption 2.1.
Before proving Lemma 3.4, we present a fundamental existence result for the GVI problem defined over a compact convex region [16]. For the sake of completeness, we give the proof process.
Lemma 3.5 Let X\subseteq {R}^{n} be a nonempty bounded closed convex set and let the multivalued mapping F:X\to {2}^{{R}^{n}} be lower semicontinuous with nonempty closed convex values. Then the solution set {X}^{\ast} is nonempty.
Proof Since the multivalued mapping F is lower semicontinuous and has nonempty closed convex values, by Michael’s selection theorem (see, for instance, Theorem 24.1 in [17]), it admits a continuous selection; that is, there exists a continuous mapping G:X\to {R}^{n} such that G(x)\in F(x) for every x\in X. Since X is a nonempty bounded closed convex set, the VI(X,G), which consists of finding an x\in X such that
has a solution (see Lemma 3.1 in [18]), i.e., the solution set {X}^{\prime} of the problem VI(X,G) is nonempty. It follows from {X}^{\prime}\subseteq {X}^{\ast} that {X}^{\ast} is nonempty. □
Proof of Lemma 3.4 On the contrary, suppose that there exists {k}_{0}\ge 1 such that {H}_{{k}_{0}}^{1}\cap {H}_{{k}_{0}}^{2}\cap X=\mathrm{\varnothing}. Then there exists a positive number M such that
and
where
Let Y=X\cap B({x}^{0},2M) and consider the GVI(F,Y). From Lemma 3.5, we know that the solution set {Y}^{\ast} of GVI(F,Y) is nonempty. In order to avoid confusion with the sequence \{{H}_{k}^{1}\}, \{{H}_{k}^{2}\}, and \{{x}^{k}\}, we denote the three corresponding sequences by \{{\overline{H}}_{k}^{1}\}, \{{\overline{H}}_{k}^{2}\} and \{{\overline{x}}^{k}\}, respectively, when Algorithm 3.1 is applied to GVI(F,Y) with a starting point {x}^{0}. We claim that

(i)
the set has at least {k}_{0}+1 elements: {\overline{x}}^{0},{\overline{x}}^{1},\dots ,{\overline{x}}^{{k}_{0}};

(ii)
{\overline{x}}^{k}={x}^{k}, {\overline{H}}_{k}^{1}={H}_{k}^{1}, {\overline{H}}_{k}^{2}={H}_{k}^{2} for 0\le k\le {k}_{0};

(iii)
{x}^{{k}_{0}} is not a solution of GVI(F,Y).
Since {Y}^{\ast}\ne \mathrm{\varnothing}, using Lemma 3.3, we know that {\overline{H}}_{{k}_{0}}^{1}\cap {\overline{H}}_{{k}_{0}}^{2}\cap X\ne \mathrm{\varnothing}, so {H}_{{k}_{0}}^{1}\cap {H}_{{k}_{0}}^{2}\cap X\ne \mathrm{\varnothing}, which contradicts the supposition that {H}_{{k}_{0}}^{1}\cap {H}_{{k}_{0}}^{2}\cap X=\mathrm{\varnothing}. □
From Lemma 3.4, if the solution set of problem (1.1) is empty, then Algorithm 3.1 generates an infinite sequence. More generally, we have the following conclusion.
Theorem 3.1 Suppose that Assumption 2.1 holds. Assume that Algorithm 3.1 generates an infinite sequence \{{x}^{k}\}. If the solution set {X}^{\ast} is nonempty, then the sequence \{{x}^{k}\} is bounded and all its cluster points belong to the solution set. Otherwise,
if the solution set {X}^{\ast} is empty.
Proof First, we suppose that the solution set is nonempty. Since
by Lemma 3.3 and the definition of the projection, it holds that
for any {x}^{\ast}\in {X}^{\ast}. So, \{{x}^{k}\} is a bounded sequence.
Since {x}^{k+1}\in {H}_{k}^{2}, it is obvious that
from the definition of the projection operator. For {x}^{k}, since
it holds that {x}^{k}={P}_{{H}_{k}^{2}}({x}^{0}) from Remark 2.1. Thus, using Lemma 2.1, one has
i.e.,
which can be written as
Thus, the sequence \{\parallel {x}^{k}{x}^{0}\parallel \} is nondecreasing and bounded, and hence convergent, which implies that
On the other hand, by {x}^{k+1}\in {H}_{k}^{1}, we get
Since
and by (3.6) we have
which implies
Using the CauchySchwarz inequality and (3.1), we obtain
Since F is continuous with compact values, Proposition 3.11 in [19] implies that \{F({y}^{k}):k\in N\} is a bounded set, and so the sequence \{{\zeta}^{k}:{\zeta}^{k}\in F({y}^{k})\} is bounded. By (3.5) and (3.7), it follows that
For any convergent sequence \{{x}^{{k}_{j}}\} of \{{x}^{k}\}, its limit is denoted by \overline{x}, i.e.,
Without loss of generality, suppose that \{{\eta}_{{k}_{j}}\} has a limit. Then
or
For the first case, by the choice of {\eta}_{{k}_{j}} in Algorithm 3.1, we know that
for all \zeta \in F({x}^{{k}_{j}}\frac{{\eta}_{{k}_{j}}}{\gamma}r({x}^{{k}_{j}},{\xi}^{{k}_{j}})).
Since
and F is lower semicontinuous on X, \mathrm{\exists}{\zeta}^{{k}_{j}}\in F({x}^{{k}_{j}}\frac{{\eta}_{{k}_{j}}}{\gamma}r({x}^{{k}_{j}},{\xi}^{{k}_{j}})) such that
where ξ is a vector in F(\overline{x}). So, by (3.8) we obtain
Using the similar arguments of (3.2), we have
Combining this and (3.9), we know that r(\overline{x},\xi )=0 and thus \overline{x} is a solution of problem (1.1).
For the second case
it is easy to see that the limit point \overline{x} of {x}^{{k}_{j}} is a solution of problem (1.1).
Now, we consider the case that the solution set is empty. Since the inequality
also holds in this case, the sequence \{\parallel {x}^{k}{x}^{0}\parallel \} is still nondecreasing. Next, we claim that
Otherwise, \{\parallel {x}^{k}{x}^{0}\parallel \} is a bounded sequence. A similar discussion as above would lead to the conclusion that every cluster point of \{{x}^{k}\} is a solution of problem (1.1), which contradicts the emptiness of the solution set. □
Theorem 3.2 Under the assumption of Theorem 3.1, if the solution set {X}^{\ast} is nonempty, then the sequence \{{x}^{k}\} globally converges to a solution {x}^{\ast} such that {x}^{\ast}={P}_{{X}^{\ast}}({x}^{0}); otherwise, {lim}_{k\to +\mathrm{\infty}}\parallel {x}^{k}{x}^{0}\parallel =+\mathrm{\infty}. That is, the solution set of problem (1.1) is empty if and only if the generated sequence diverges to infinity.
Proof First, we suppose that the solution set is nonempty. From Theorem 3.1, we know that the sequence \{{x}^{k}\} is bounded and that every cluster point {x}^{\ast} of \{{x}^{k}\} is a solution of problem (1.1). Suppose that the subsequence \{{x}^{{k}_{j}}\} converges to {x}^{\ast}, i.e.,
Let \overline{x}={P}_{{X}^{\ast}}({x}^{0}). Since \overline{x}\in {X}^{\ast}, by Lemma 3.3 we have
for all j. So, by the iterative sequence of Algorithm 3.1, we have
Thus,
Letting j\to \mathrm{\infty}, we have
where the last inequality is due to Lemma 2.1 and the fact that \overline{x}={P}_{{X}^{\ast}}({x}^{0}) and {x}^{\ast}\in {X}^{\ast}. So,
Thus the sequence \{{x}^{k}\} has a unique cluster point {P}_{{X}^{\ast}}({x}^{0}), which shows the global convergence of \{{x}^{k}\}.
For the case that the solution set is empty, the conclusion can be obtained directly from Theorem 3.1. □
4 Discussion
Certainly, the proposed extragradient method for the GVI in this paper has a good theoretical property in theory, as the generated sequence not only has an expansion property w.r.t. the starting point, but also the existence of the solution to the problem can be verified through the behavior of the generated sequence. However, the proposed algorithm is not easy to be realized in practice as the termination criterion is not easy to execute. This is an interesting topic for further research.
References
Facchinei F, Pang JS: FiniteDimensional Variational Inequalities and Complementarity Problems. Springer, New York; 2003.
Xia FQ, Huang NJ: A projectionproximal point algorithm for solving generalized variational inequalities. J. Optim. Theory Appl. 2011, 150: 98–117. 10.1007/s1095701198253
Wang Y, Xiu N, Wang C: Unified framework of extragradienttype methods for pseudomonotone variational inequalities. J. Optim. Theory Appl. 2001, 111: 641–656. 10.1023/A:1012606212823
Wang Y, Xiu N, Zhang J: Modified extragradient method for variational inequalities and verification of solution existence. J. Optim. Theory Appl. 2003, 119: 167–183.
Auslender A, Teboulle M: Lagrangian duality and related multiplier methods for variational inequality problems. SIAM J. Optim. 2000, 10: 1097–1115. 10.1137/S1052623499352656
Censor Y, Gibali A, Reich S: The subgradient extragradient method for solving variational inequalities in Hilbert space. J. Optim. Theory Appl. 2011, 148: 318–335. 10.1007/s1095701097573
Fang SC, Peterson EL: Generalized variational inequalities. J. Optim. Theory Appl. 1982, 38: 363–383. 10.1007/BF00935344
Fand C, He Y: A double projection algorithm for multivalued variational inequalities and a unified framework of the method. Appl. Math. Comput. 2001, 217: 9543–9551.
He Y: Stable pseudomonotone variational inequality in reflexive Banach spaces. J. Math. Anal. Appl. 2007, 330: 352–363. 10.1016/j.jmaa.2006.07.063
Li F, He Y: An algorithm for generalized variational inequality with pseudomonotone mapping. J. Comput. Appl. Math. 2009, 228: 212–218. 10.1016/j.cam.2008.09.014
Saigal R: Extension of the generalized complementarity problem. Math. Oper. Res. 1976, 1: 260–266. 10.1287/moor.1.3.260
Salmon G, Strodiot JJ, Nguyen VH: A bundle method for solving variational inequalities. SIAM J. Optim. 2003, 14: 869–893.
He Y: The Tikhonov regularization method for setvalued variational inequalities. Abstr. Appl. Anal. 2012., 2012: Article ID 172061. doi:10.1155/2012/172061
Sun D: A class of iterative methods for solving nonlinear projection equations. J. Optim. Theory Appl. 1996, 91: 123–140. 10.1007/BF02192286
Polyak BT: Introduction to Optimization. Optimization Software, New York; 1987.
Fang, C, He, Y: An extragradient method for generalized variational inequality. J. Optim. Theory Appl. (2012, submitted)
Deimling K: Nonlinear Functional Analysis. Springer, Berlin; 1985.
Hartman P, Stampacchia G: On some nonlinear elliptic differential functional equations. Acta Math. 1966, 115: 271–310. 10.1007/BF02392210
Aubin JP, Ekeland I: Applied Nonlinear Analysis. Wiley, New York; 1984.
Acknowledgements
The author would like to thank two anonymous reviewers for their insightful comments and constructive suggestions, and Professor Wang Yiju for his careful reading, which helped improve the presentation of the paper. This work was supported by the Natural Science Foundation of China (Grant Nos. 11171180, 11101303).
Author information
Authors and Affiliations
Corresponding author
Additional information
Competing interests
The author declares that they have no competing interests.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Chen, H. A new extragradient method for generalized variational inequality in Euclidean space. Fixed Point Theory Appl 2013, 139 (2013). https://doi.org/10.1186/168718122013139
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/168718122013139
Keywords
 generalized variational inequalities
 extragradient method
 solution existence
 multivalued mapping