# A new extragradient method for generalized variational inequality in Euclidean space

- Haibin Chen
^{1}Email author

**2013**:139

https://doi.org/10.1186/1687-1812-2013-139

© Chen; licensee Springer. 2013

**Received: **2 March 2013

**Accepted: **13 May 2013

**Published: **29 May 2013

## Abstract

In this paper, we extend the extragradient projection method proposed in (Wang *et al.* in J. Optim. Theory Appl. 119:167-183, 2003) for the classical variational inequalities to the generalized variational inequalities. For this algorithm, we first prove its expansion property of the generated sequence with respect to the starting point and then show that the existence of the solution to the problem can be verified through the behavior of the generated sequence. The global convergence of the method is also established under mild conditions.

**MSC:**90C30, 15A06.

## Keywords

## 1 Introduction

*X*be a nonempty closed convex set in ${R}^{n}$. The problem of generalized variational inequalities (GVI) [1, 2] is to find ${x}^{\ast}\in X$ such that there exists ${\omega}^{\ast}\in F({x}^{\ast})$ satisfying

where $\u3008\cdot ,\cdot \u3009$ stands for the Euclidean inner product of vectors in ${R}^{n}$. The solution set of problem (1.1) is denoted by ${X}^{\ast}$. Certainly, the GVI reduces to the classical variational inequalities (VI) when *F* is a single-valued mapping, which has been well studied in the past decades [3, 4].

For the GVI, theories and solution methods have been extensively considered in the literature [2, 5–12], and it is well known that the existence of solutions is an important topic for the GVI [1]. Generally, there are mainly two approaches to attack the solution existence problem of the GVI. One is an analytic approach which first reformulates the GVI as a well-studied mathematical problem and then invokes an existence theorem for the latter problem [13]. The second is a constructive approach in which the existence can be verified by the behavior of the proposed method. The algorithm that is considered in this paper belongs to the second approach.

First, we give a short summary to the constructive approach on the existence theory for the VI. For this approach, the equivalence between the existence of solutions to the VI problem and the boundedness of the sequence generated by some modified extragradient methods was first established by Sun in [14]. Later, Wang *et al.* [4] established the same theory by a new type of extragradient-type method. Furthermore, the generated sequence possesses an expansion property with respect to the starting point and converges to a solution point if the solution set of the VI is nonempty. Now a question is posed naturally: as the GVI problem is an extension of the VI, can this theory be extended to the GVI? This constitutes the main motivation of the paper.

In this paper, inspired by the work in [4], we propose a new type of extragradient projection method for the problem GVI. We first establish the existence results for the GVI under pseudomonotonicity and continuity assumption of the underlying mapping *F*, and then show the global convergence of the proposed method.

The rest of this paper is organized as follows. In Section 2, we give some related concepts and conclusions. In Section 3, we present the description of the algorithm and establish some properties of the algorithm. The global convergence of the sequence is also established.

## 2 Preliminaries

*x*onto

*K*,

*i.e.*,

is denoted by ${P}_{K}(x)$. In what follows, we state some well-known properties of the projection operator which will be used in the sequel.

**Lemma 2.1** [15]

*Let*

*K*

*be a nonempty*,

*closed and convex subset in*${R}^{n}$.

*Then*,

*for any*$x,y\in {R}^{n}$

*and*$z\in K$,

*the following statements hold*:

- (i)
$\u3008{P}_{K}(x)-x,z-{P}_{K}(x)\u3009\ge 0$;

- (ii)
${\parallel {P}_{K}(x)-{P}_{K}(y)\parallel}^{2}\le {\parallel x-y\parallel}^{2}-{\parallel {P}_{K}(x)-x+y-{P}_{K}(y)\parallel}^{2}$.

**Remark 2.1**In fact, (i) in Lemma 2.1 provides also a sufficient and necessary condition for a vector $u\in K$ to be the projection of the vector

*x*;

*i.e.*, $u={P}_{K}(x)$ if and only if

**Definition 2.1**Let

*K*be a nonempty subset of ${R}^{n}$. A multi-valued mapping $F:K\to {2}^{{R}^{n}}$ is said to be

- (i)monotone if and only if$\u3008u-v,x-y\u3009\ge 0,\phantom{\rule{1em}{0ex}}\mathrm{\forall}x,y\in K,u\in F(x),v\in F(y);$
- (ii)pseudomonotone if and only if, for any $x,y\in K$, $u\in F(x)$, $v\in F(y)$,$\u3008u,y-x\u3009\ge 0\phantom{\rule{1em}{0ex}}\u27f9\phantom{\rule{1em}{0ex}}\u3008v,y-x\u3009\ge 0.$

Now let us recall the definition of a continuous multi-valued mapping *F*.

**Definition 2.2**Assume that $F:X\to {2}^{{R}^{n}}$ is a multi-valued mapping, then

- (i)
*F*is said to be upper semicontinuous at $x\in X$ if for every open set*V*containing $F(x)$, there is an open set*U*containing*x*such that $F(y)\subset V$ for all $y\in X\cap U$; - (ii)
*F*is said to be lower semicontinuous at $x\in X$ if given any sequence $\{{x}^{k}\}$ converging to*x*and any $y\in F(x)$, there exists a sequence ${y}^{k}\in F({x}^{k})$ that converges to*y*.

*F* is said to be continuous at $x\in X$ if it is both upper semicontinuous and lower semicontinuous at *x*.

For the simplicity of our description, we list the assumptions needed in the sequel.

**Assumption 2.1** Suppose that *X* is a nonempty closed convex set in ${R}^{n}$. The multi-valued mapping $F:X\to {2}^{{R}^{n}}$ is pseudomonotone and continuous on *X* with nonempty compact convex values.

## 3 Main results

It is well known that the projection residue is related intimately to the solution set ${X}^{\ast}$.

**Proposition 3.1**

*For*$x\in X$

*and*$\xi \in F(x)$,

*they solve problem*(1.1)

*if and only if*

The basic idea of the designed algorithm is as follows. At each step of the algorithm, compute the projection residue $r({x}^{k},{\xi}^{k})$ at iterate ${x}^{k}$. If it is a zero vector, then stop with ${x}^{k}$ being a solution of the GVI; otherwise, find a trial point ${y}^{k}$ by a back-tracking search at ${x}^{k}$ along the residue $r({x}^{k},{\xi}^{k})$, and the new iterate is obtained by projecting ${x}^{0}$ onto the intersection of *X* with two halfspaces respectively associated with ${y}^{k}$ and ${x}^{k}$. Repeat this process until the projection residue is a zero vector.

**Algorithm 3.1** Choose $\sigma ,\gamma \in (0,1)$, ${x}^{0}\in X$, $k=0$.

*m*satisfying: $\mathrm{\exists}{\zeta}^{k}\in F({x}^{k}-{\gamma}^{m}r({x}^{k},{\xi}^{k}))$ such that

Select $k=k+1$ and go to Step 1.

Now, we first discuss the feasibility of the stepsize rule (3.1).

**Lemma 3.1** *If* ${x}^{k}$ *is not a solution of problem* (1.1), *then there exists the smallest nonnegative integer* *m* *satisfying* (3.1).

*Proof*By the definition of $r({x}^{k},{\xi}^{k})$ and Lemma 2.1, we know that

*F*is lower semicontinuous, there exists ${\zeta}^{m}\in F({x}^{k}-{\gamma}^{m}r({x}^{k},{\xi}^{k}))$ such that

which implies the conclusion. □

The following lemma shows that the halfspace ${H}_{k}^{1}$ in Algorithm 3.1 strictly separates ${x}^{k}$ and the solution set ${X}^{\ast}$ if ${X}^{\ast}$ is nonempty.

**Lemma 3.2**

*If*${X}^{\ast}\ne \mathrm{\varnothing}$,

*the halfspace*${H}_{k}^{1}$

*in Algorithm*3.1

*separates the point*${x}^{k}$

*from the set*${X}^{\ast}$.

*Moreover*,

*Proof*By the definition of $r({x}^{k},{\xi}^{k})$ and Algorithm 3.1, we know

where ${\zeta}^{k}$ is a vector in $F({y}^{k})$. So, by the definition of ${H}_{k}^{1}$ and (3.3), we get ${x}^{k}\notin {H}_{k}^{1}$.

*F*is pseudomonotone on

*X*, we get

which implies ${x}^{\ast}\in {H}_{k}^{1}$. Moreover, it is easy to see that ${X}^{\ast}\subseteq {H}_{k}^{1}\cap X$, $\mathrm{\forall}k\ge 0$. □

The following lemma says that if the solution set is nonempty, then ${X}^{\ast}\subseteq {H}_{k}^{1}\cap {H}_{k}^{2}\cap X$ and thus ${H}_{k}^{1}\cap {H}_{k}^{2}\cap X$ is a nonempty set.

**Lemma 3.3** *If the solution set* ${X}^{\ast}\ne \mathrm{\varnothing}$, *then* ${X}^{\ast}\subseteq {H}_{k}^{1}\cap {H}_{k}^{2}\cap X$ *for all* $k\ge 0$ *under Assumption * 2.1.

*Proof*From the analysis above, it is sufficient to prove that ${X}^{\ast}\subseteq {H}_{k}^{2}$ for all $k\ge 0$. The proof will be given by induction. Obviously, if $k=0$,

Thus ${X}^{\ast}\subseteq {H}_{l+1}^{2}$. This shows that ${X}^{\ast}\subseteq {H}_{k}^{2}$ for all $k\ge 0$, and the desired result follows. □

For the case that the solution set is empty, we have that ${H}_{k}^{1}\cap {H}_{k}^{2}\cap X$ is also nonempty from the following lemma, which implies the feasibility of Algorithm 3.1.

**Lemma 3.4** *Suppose that* ${X}^{\ast}=\mathrm{\varnothing}$, *then* ${H}_{k}^{1}\cap {H}_{k}^{2}\cap X\ne \mathrm{\varnothing}$ *for all* $k\ge 0$ *under Assumption * 2.1.

Before proving Lemma 3.4, we present a fundamental existence result for the GVI problem defined over a compact convex region [16]. For the sake of completeness, we give the proof process.

**Lemma 3.5** *Let* $X\subseteq {R}^{n}$ *be a nonempty bounded closed convex set and let the multi*-*valued mapping* $F:X\to {2}^{{R}^{n}}$ *be lower semicontinuous with nonempty closed convex values*. *Then the solution set* ${X}^{\ast}$ *is nonempty*.

*Proof*Since the multi-valued mapping

*F*is lower semicontinuous and has nonempty closed convex values, by Michael’s selection theorem (see, for instance, Theorem 24.1 in [17]), it admits a continuous selection; that is, there exists a continuous mapping $G:X\to {R}^{n}$ such that $G(x)\in F(x)$ for every $x\in X$. Since

*X*is a nonempty bounded closed convex set, the $VI(X,G)$, which consists of finding an $x\in X$ such that

has a solution (see Lemma 3.1 in [18]), *i.e.*, the solution set ${X}^{\prime}$ of the problem $VI(X,G)$ is nonempty. It follows from ${X}^{\prime}\subseteq {X}^{\ast}$ that ${X}^{\ast}$ is nonempty. □

*Proof of Lemma 3.4*On the contrary, suppose that there exists ${k}_{0}\ge 1$ such that ${H}_{{k}_{0}}^{1}\cap {H}_{{k}_{0}}^{2}\cap X=\mathrm{\varnothing}$. Then there exists a positive number

*M*such that

- (i)
the set has at least ${k}_{0}+1$ elements: ${\overline{x}}^{0},{\overline{x}}^{1},\dots ,{\overline{x}}^{{k}_{0}}$;

- (ii)
${\overline{x}}^{k}={x}^{k}$, ${\overline{H}}_{k}^{1}={H}_{k}^{1}$, ${\overline{H}}_{k}^{2}={H}_{k}^{2}$ for $0\le k\le {k}_{0}$;

- (iii)
${x}^{{k}_{0}}$ is not a solution of $GVI(F,Y)$.

Since ${Y}^{\ast}\ne \mathrm{\varnothing}$, using Lemma 3.3, we know that ${\overline{H}}_{{k}_{0}}^{1}\cap {\overline{H}}_{{k}_{0}}^{2}\cap X\ne \mathrm{\varnothing}$, so ${H}_{{k}_{0}}^{1}\cap {H}_{{k}_{0}}^{2}\cap X\ne \mathrm{\varnothing}$, which contradicts the supposition that ${H}_{{k}_{0}}^{1}\cap {H}_{{k}_{0}}^{2}\cap X=\mathrm{\varnothing}$. □

From Lemma 3.4, if the solution set of problem (1.1) is empty, then Algorithm 3.1 generates an infinite sequence. More generally, we have the following conclusion.

**Theorem 3.1**

*Suppose that Assumption*2.1

*holds*.

*Assume that Algorithm*3.1

*generates an infinite sequence*$\{{x}^{k}\}$.

*If the solution set*${X}^{\ast}$

*is nonempty*,

*then the sequence*$\{{x}^{k}\}$

*is bounded and all its cluster points belong to the solution set*.

*Otherwise*,

*if the solution set* ${X}^{\ast}$ *is empty*.

*Proof*First, we suppose that the solution set is nonempty. Since

for any ${x}^{\ast}\in {X}^{\ast}$. So, $\{{x}^{k}\}$ is a bounded sequence.

*i.e.*,

*F*is continuous with compact values, Proposition 3.11 in [19] implies that $\{F({y}^{k}):k\in N\}$ is a bounded set, and so the sequence $\{{\zeta}^{k}:{\zeta}^{k}\in F({y}^{k})\}$ is bounded. By (3.5) and (3.7), it follows that

*i.e.*,

for all $\zeta \in F({x}^{{k}_{j}}-\frac{{\eta}_{{k}_{j}}}{\gamma}r({x}^{{k}_{j}},{\xi}^{{k}_{j}}))$.

*F*is lower semicontinuous on

*X*, $\mathrm{\exists}{\zeta}^{{k}_{j}}\in F({x}^{{k}_{j}}-\frac{{\eta}_{{k}_{j}}}{\gamma}r({x}^{{k}_{j}},{\xi}^{{k}_{j}}))$ such that

*ξ*is a vector in $F(\overline{x})$. So, by (3.8) we obtain

Combining this and (3.9), we know that $r(\overline{x},\xi )=0$ and thus $\overline{x}$ is a solution of problem (1.1).

it is easy to see that the limit point $\overline{x}$ of ${x}^{{k}_{j}}$ is a solution of problem (1.1).

Otherwise, $\{\parallel {x}^{k}-{x}^{0}\parallel \}$ is a bounded sequence. A similar discussion as above would lead to the conclusion that every cluster point of $\{{x}^{k}\}$ is a solution of problem (1.1), which contradicts the emptiness of the solution set. □

**Theorem 3.2** *Under the assumption of Theorem * 3.1, *if the solution set* ${X}^{\ast}$ *is nonempty*, *then the sequence* $\{{x}^{k}\}$ *globally converges to a solution* ${x}^{\ast}$ *such that* ${x}^{\ast}={P}_{{X}^{\ast}}({x}^{0})$; *otherwise*, ${lim}_{k\to +\mathrm{\infty}}\parallel {x}^{k}-{x}^{0}\parallel =+\mathrm{\infty}$. *That is*, *the solution set of problem* (1.1) *is empty if and only if the generated sequence diverges to infinity*.

*Proof*First, we suppose that the solution set is nonempty. From Theorem 3.1, we know that the sequence $\{{x}^{k}\}$ is bounded and that every cluster point ${x}^{\ast}$ of $\{{x}^{k}\}$ is a solution of problem (1.1). Suppose that the subsequence $\{{x}^{{k}_{j}}\}$ converges to ${x}^{\ast}$,

*i.e.*,

*j*. So, by the iterative sequence of Algorithm 3.1, we have

Thus the sequence $\{{x}^{k}\}$ has a unique cluster point ${P}_{{X}^{\ast}}({x}^{0})$, which shows the global convergence of $\{{x}^{k}\}$.

For the case that the solution set is empty, the conclusion can be obtained directly from Theorem 3.1. □

## 4 Discussion

Certainly, the proposed extragradient method for the GVI in this paper has a good theoretical property in theory, as the generated sequence not only has an expansion property w.r.t. the starting point, but also the existence of the solution to the problem can be verified through the behavior of the generated sequence. However, the proposed algorithm is not easy to be realized in practice as the termination criterion is not easy to execute. This is an interesting topic for further research.

## Declarations

### Acknowledgements

The author would like to thank two anonymous reviewers for their insightful comments and constructive suggestions, and Professor Wang Yiju for his careful reading, which helped improve the presentation of the paper. This work was supported by the Natural Science Foundation of China (Grant Nos. 11171180, 11101303).

## Authors’ Affiliations

## References

- Facchinei F, Pang JS:
*Finite-Dimensional Variational Inequalities and Complementarity Problems*. Springer, New York; 2003.Google Scholar - Xia FQ, Huang NJ: A projection-proximal point algorithm for solving generalized variational inequalities.
*J. Optim. Theory Appl.*2011, 150: 98–117. 10.1007/s10957-011-9825-3MathSciNetView ArticleGoogle Scholar - Wang Y, Xiu N, Wang C: Unified framework of extragradient-type methods for pseudomonotone variational inequalities.
*J. Optim. Theory Appl.*2001, 111: 641–656. 10.1023/A:1012606212823MathSciNetView ArticleGoogle Scholar - Wang Y, Xiu N, Zhang J: Modified extragradient method for variational inequalities and verification of solution existence.
*J. Optim. Theory Appl.*2003, 119: 167–183.MathSciNetView ArticleGoogle Scholar - Auslender A, Teboulle M: Lagrangian duality and related multiplier methods for variational inequality problems.
*SIAM J. Optim.*2000, 10: 1097–1115. 10.1137/S1052623499352656MathSciNetView ArticleGoogle Scholar - Censor Y, Gibali A, Reich S: The subgradient extragradient method for solving variational inequalities in Hilbert space.
*J. Optim. Theory Appl.*2011, 148: 318–335. 10.1007/s10957-010-9757-3MathSciNetView ArticleGoogle Scholar - Fang SC, Peterson EL: Generalized variational inequalities.
*J. Optim. Theory Appl.*1982, 38: 363–383. 10.1007/BF00935344MathSciNetView ArticleGoogle Scholar - Fand C, He Y: A double projection algorithm for multi-valued variational inequalities and a unified framework of the method.
*Appl. Math. Comput.*2001, 217: 9543–9551.Google Scholar - He Y: Stable pseudomonotone variational inequality in reflexive Banach spaces.
*J. Math. Anal. Appl.*2007, 330: 352–363. 10.1016/j.jmaa.2006.07.063MathSciNetView ArticleGoogle Scholar - Li F, He Y: An algorithm for generalized variational inequality with pseudomonotone mapping.
*J. Comput. Appl. Math.*2009, 228: 212–218. 10.1016/j.cam.2008.09.014MathSciNetView ArticleGoogle Scholar - Saigal R: Extension of the generalized complementarity problem.
*Math. Oper. Res.*1976, 1: 260–266. 10.1287/moor.1.3.260MathSciNetView ArticleGoogle Scholar - Salmon G, Strodiot JJ, Nguyen VH: A bundle method for solving variational inequalities.
*SIAM J. Optim.*2003, 14: 869–893.MathSciNetView ArticleGoogle Scholar - He Y: The Tikhonov regularization method for set-valued variational inequalities.
*Abstr. Appl. Anal.*2012., 2012: Article ID 172061. doi:10.1155/2012/172061Google Scholar - Sun D: A class of iterative methods for solving nonlinear projection equations.
*J. Optim. Theory Appl.*1996, 91: 123–140. 10.1007/BF02192286MathSciNetView ArticleGoogle Scholar - Polyak BT:
*Introduction to Optimization*. Optimization Software, New York; 1987.Google Scholar - Fang, C, He, Y: An extragradient method for generalized variational inequality. J. Optim. Theory Appl. (2012, submitted)Google Scholar
- Deimling K:
*Nonlinear Functional Analysis*. Springer, Berlin; 1985.View ArticleGoogle Scholar - Hartman P, Stampacchia G: On some nonlinear elliptic differential functional equations.
*Acta Math.*1966, 115: 271–310. 10.1007/BF02392210MathSciNetView ArticleGoogle Scholar - Aubin JP, Ekeland I:
*Applied Nonlinear Analysis*. Wiley, New York; 1984.Google Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.