Research | Open | Published:

# Composite iterative schemes for maximal monotone operators in reflexive Banach spaces

*Fixed Point Theory and Applications***volume 2011**, Article number: 7 (2011)

## Abstract

In this article, we introduce composite iterative schemes for finding a zero point of a finite family of maximal monotone operators in a reflexive Banach space. Then, we prove strong convergence theorems by using a shrinking projection method. Moreover, we also apply our results to a system of convex minimization problems in reflexive Banach spaces.

**AMS Subject Classification**: 47H09, 47H10

## Introduction

Let *E* be a real Banach space and *C* a nonempty subset of *E*. Let *E** be the dual space of *E*. We denote the value of *x** ∈ *E** at *x 2 E* by 〈*x**, *x*〉. Let *A* : *E* → 2^{E*} be a set-valued mapping. We denote dom *A* by *domain* of *A*, that is, dom *A* = {*x* ∈ *E* : *Ax* ≠ ∅}and also denote *G*(*A*) by the *graph* of *A*, that is, *G*(*A*) = *f* (*x*, *x**) ∈ *E* × *E** : *x** ∈ *Ax*}. A set-valued mapping *A* is said to be *monotone* if 〈*x** - *y**, *x - y*〉 ≥ 0 whenever (*x*, *x**); (*y*, *y**) ∈ *G*(*A*). It is said to be *maximal monotone* if its graph is not contained in the graph of any other monotone operator on *E*. It is known that if *A* is maximal monotone, then the set *A*^{-1}(0*) = {*z* ∈ *E* : 0* ∈ *Az*} is closed and convex.

The problem of finding zero points for maximal monotone operators plays an important role in optimizations. This is because it can be reduced to a convex minimization problem and a variational inequality problem. Many authors have studied the convergence of such problems in several settings, (see [1–6]). Initiated by Martinet [7], in a Hilbert space, Rockafellar [8] introduced the following iterative schemes:

where {*λ*_{
n
}} ⊂ (0, ∞) and *J*_{
λ
}is the resolvent of *A* defined by *J*_{
λ
} = (*I* + *λA*)^{-1} for all *λ >* 0, and *A* is a maximal monotone operator on *E*. Such an algorithm is called the *proximal point algorithm*. He proved that the sequence {*x*_{
n
} } generated by (1.1) converges weakly to an element in *A*^{-1} (0) provided lim inf_{n→∞}*λ*_{
n
}> 0. Later, Kamimura and Takahashi [9] introduced the following iteration in a Hilbert space:

where {*α*_{
n
} } ⊂ [0, 1] and {*λ*_{
n
} } ⊂ (0, *∞*). The weak convergence theorems are also established in a Hilbert space under suitable conditions imposed on {*α*_{
n
} } and {*λ*_{
n
} }.

In 2005, Kohsaka and Takahashi [10] studied the above iteration process in a more general setting, reflexive Banach spaces. In fact, those authors proposed the following algorithm:

where {*α*_{
n
} }⊂ [0, 1], {*λ*_{
n
} } ⊂ (0, *∞*), *f* : *E* → ℝ is a Bregman function and *J*_{
λ
} = (∇*f* + *λA*) ^{-1} ∇*f* for all *λ* > 0. They also proved a weak convergence theorem of the proposed algorithm.

Very recently, in 2010, Reich and Sabach [11] proposed an algorithm for finding a zero point of maximal monotone operators *A*_{
i
} : *E* → 2^{E*} (*i* = 1, 2,..., *N*) in a general reflexive Banach space *E* as follows:

where , is an error sequence in *E* with and the Bregman projection with respect to *f* from *E* onto a closed and convex subset *K* of *E*. Those authors showed that the sequence {*x*_{
n
} } defined by (1.4) converges strongly to a common element in under some mild conditions.

Motivated by the previous ones, we first introduce a composite iterative scheme which is different from (1.4) for finding a zero point of maximal monotone operators *A*_{
i
} : *E* → 2^{E*} (*i* = 1, 2,..., *N*) in reflexive Banach spaces. Using the shrinking projection technique, introduced by Takahashi et al. [12], we then prove that a sequence generated by the proposed algorithm converges strongly to an element in under some appropriate control conditions. Finally, we also apply our result to a system of convex minimization problems.

## Preliminaries and lemmas

Let *E* be a real reflexive Banach space with a norm ||·|| and *E** be the dual space of *E*. Throughout this article, *f* : *E* → (*-∞*, +∞] is a proper, lower semi-continuous, and convex function, and the Fenchel conjugate of *f* is the function *f**: *E** → (*-∞*, +∞] defined by

We denote by dom *f* the domain of *f*, that is, the set {*x* ∈ *E* : *f*(*x*) *<* +∞). For any *x* ∈ int dom *f* and *y* ∈ *E*, the *right-hand derivative* of *f* at *x* in the direction *y* is defined by

The function *f* is said to be *Gâteaux differentiable* at *x* exists for any *y*. In this case, *f*^{o}(*x*, *y*) coincides with ∇*f* (*x*), the value of the *gradient* ∇*f* of *f* at *x*. The function *f* is said to be *Gâteaux differentiable* if it is Gâteaux differentiable for any *x* ∈ int dom *f*. The function *f* is said to be *Fréchet differentiable at x* if this limit is attained uniformly in ||*y*|| = 1. Finally, *f* is said to be *uniformly Fréchet differentiable* on a subset *C* of *E* if the limit is attained uniformly for *x* ∈ *C* and ||*y*|| = 1.

Let *E* be a reflexive Banach space. The Legendre function is defined from a general Banach space *E* into (-∞, +∞] (see [13]). According to [13], the function *f* is *Legendre* if and only if it satisfies the following conditions:

(L1) The interior of the domain of *f* (denoted by int dom *f* ) is nonempty, *f** is Gâteaux differentiable on int dom *f*, and dom ∇*f* = int dom *f* ;

(L2) The interior of the domain *f**(denoted by int dom *f**) is nonempty, *f** is Gâteaux differentiable on int dom *f**, and dom ∇*f** = int dom *f**.

Since *E* is reflexive, we always have (∂*f*)^{-1} = ∂*f** (see [14]). This fact, when combined with the conditions (L1) and (L2), implies the following equalities [15]:

Also, the conditions (L1) and (L2), in conjunction with [13], imply that the functions *f* and *f** are strictly convex on the interior of their respective domains. Several interesting examples of the Legendre functions are presented in [13, 16]. Especially, the functions with *s* ∈ (1, *∞*) are Legendre, where the Banach space *E* is smooth and strictly convex and, in particular, a Hilbert space. Throughout this article, we assume that the convex function *f* : *E* → (*∞*, +∞] is Legendre.

**Lemma 2.1**. [17]*If f* : *E* → ℝ *is uniformly Fréchet differentiable and bounded on bounded subsets of E, then* ∇ *f is uniformly continuous on bounded subsets of E from the strong topology of E to the strong topology of E**.

Let *f* : *E* → (*-∞*, +∞] be a convex and Gâteaux differentiable function. The function *D*_{
f
} : dom *f* × int dom *f* → [0, +∞) is defined as follows:

is called the *Bregman distance* with respect to *f*[18].

Recall that the *Bregman projection*[19] of *x* ∈ int dom *f* onto the nonempty, closed, and convex set *C* ⊂ dom *f* is necessarily the unique vector satisfying

Let *f* : *E* → (*-∞*, +∞] be a convex and Gâteaux differentiable function. The function *f* is said to be *totally convex* at *x* ∈ int dom *f* if its modulus of total convexity at *x*, that is, the function *ν*_{
f
} : int dom *f* × [0, +∞) → [0, +∞] defined by

is positive, whenever *t >* 0. The function *f* is said to be *totally convex* when it is totally convex at every point *x* ∈ int dom *f*. In addition, the function *f* is said to be *totally convex on bounded sets* if *ν*_{
f
} (*B*, *t*) is positive for any nonempty bounded subset *B* of *E* and *t >* 0, where the modulus of total convexity of the function *f* on the set *B* is the function *ν*_{
f
} : int dom *f* × [0, +∞) → [0, +∞] defined by

Let *C* be a nonempty, closed, and convex subset of *E*. Let *f* : *E* → ℝ be a Gâteaux differentiable and totally convex function and let *x* ∈ *E*. It is known from [20] that if and only if 〈∇*f* (*x*) - ∇*f*(*z*), *y - z*〉 ≤ 0 for all *y* ∈ *C*. We also have

Recall that the function *f* is said to be *sequentially consistent*[20] if, for any two sequences, {*x*_{
n
} } and {*y*_{
n
} }, in *E* such that the first is bounded:

The following lemmas were proved by Reich and Sabach [11].

**Lemma 2.2**. [11]*Let f* : *E* → ℝ *be a Gâteaux differentiable and totally convex function. If x*_{0} ∈ *E and the sequence**is bounded, then the sequence**is also bounded*.

We know that the resolvent of *A*, denoted by , is defined as follows [21]:

It is known that , and is single-valued (see [21]). If *f* is a Legendre function which is bounded, uniformly Fréchet differentiable on bounded, subsets of *E*, then (see [22]). The *Yosida approximation Aλ* : *E* → *E*, *λ* > 0, is also defined by

for all *x* ∈ *E*. From Proposition 2.7 in [11], we know that and 0* ∈ *Ax* if and only if 0* ∈ *A*_{
λ
}*x* for all *x* ∈ *E* and *λ* > 0.

**Lemma 2.3**. [11]*Let A* : *E* → 2^{E*}*be a maximal monotone operator such that A*^{-1}(0*) ≠ ∅. *Then*,

*for all λ* > 0, *p* ∈ *A*^{-1}(0*) *and x* ∈ *E*.

## Strong convergence theorems

Now, in this section, we prove our main results of this article.

**Theorem 3.1**. *Let E be a real reflexive Banach space and f* : *E* → ℝ *a Legendre function which is bounded, uniformly Fréchet differentiable and totally convex on bounded subsets of E. Let A*_{
i
}: *E* → 2^{E*} (*i* = 1, 2,..., *N*) *be maximal monotone operators such that*. *Let**be such that* lim_{n→ ∞}*e*_{
n
}= 0. *Define a sequence**in E as follows:*

*If**for each i* = 1, 2,..., *N, then the sequence* {*x*_{
n
}} *converges strongly to a point*

*Proof*. We divide our proof into six steps as follows:

**Step 1**. *F* ⊂ *C*_{
n
} for all *n* ≥ 1.

Since is closed and convex for each *i* = 1, 2,..., *N*, we get that is a nonempty, closed and convex subset of *E*. It is easy to see that *C*_{
n
} is closed and convex for all *n* ≥ 1. Indeed, for each *z* ∈ *C*_{
n
} , it follows that *D*_{
f
} (*z*, *y*_{
n
} ) ≤ *D*_{
f
} (*z*, *x*_{
n
} + *e*_{
n
} ) is equivalent to

This shows that *C*_{
n
} is closed and convex for all *n* ≥ 1. It is obvious that *F* ⊂ *C* 1 = *E*.

Now, suppose that *F* ⊂ *C*_{
k
} for some . For any *p* ∈ *F*, by Lemma 2.3, we have

This implies that *F* ⊂ *C*_{k+1}. By induction, we can conclude that *F* ⊂ *C*_{
n
}for all *n* ≥ 1.

**Step 2**. lim_{n→∞}*D*_{
f
}(*x*_{
n
}, *x*_{0}) exists.

From and we have

By (2.1), for any *p* ∈ *F* ⊂ *C*_{
n
}, we have

Combining (3.3) and (3.4), we know that lim_{n→ ∞}*D*_{
f
}(*x*_{
n
}, *x*_{1}) exists.

**Step 3**. lim_{n→ ∞}||∇*f*(*y*_{
n
}) - ∇*f*(*x*_{
n
}+ *e*_{
n
})|| = 0

Since for *m* > *n* ≥ 1, by (2.1), it follows that

Letting *m*, *n* → *∞*, we have *D*_{
f
}(*x*_{
m
}, *x*_{
n
} ) → 0. Since *f* is totally convex on bounded subsets of *E*, *f* is sequentially consistent by Butnariu and Resmerita [20]. It follows that *||x*_{
m
} *- x*_{
n
}*||* → 0 as *m*, *n* → *∞*. Therefore, {*x*_{
n
} } is a Cauchy sequence. By the completeness of the space *E*, we can assume that *x*_{
n
} → *q* ∈ *E* as *n* → ∞. In particular, we obtain

Since *e*_{
n
} → 0, we also obtain

Since

We know from [23] that, if *f* is bounded on bounded subsets of *E*, then ∇*f* is also bounded on bounded subsets of *E*. Moreover, if *f* is uniformly Fréchet differentiable on bounded subsets of *E*, then *f* is uniformly continuous on bounded subsets of *E* (see [24]). Using (3.5), we have

Also, we have

and hence,

and, since *e*_{
n
} → 0,

Since *f* is uniformly Fréchet differentiable on bounded subsets of *E*, ∇*f* is norm-to-norm uniformly continuous on bounded subsets of *E* by Lemma 2.1. Hence, we have

Step 4. .

Denote for each *i* ∈ {1, 2,..., *N*} and for each *n* ≥ 1. We note that for each *n* ≥ 1. For any *p* ∈ *F*, by (3.2), it follows that

Since , by Lemma 2.3 and (3.8), it follows that

From (3.6) and (3.7), we get that . Since *f* is sequentially consistent,

Thus, from (3.6) and (3.9), it follows that

and hence,

Again, since , by Lemma 2.3 and (3.8), we know that

From (3.10) and (3.11), we have

Since *f* is sequentially consistent, it follows that

From (3.10) and (3.12), we have

and hence,

In a similar way, we can show that

, and

Hence, we can conclude that

for each *i* = 1,2,..., *N*.

Step 5.

For each *i* = 1, 2,..., *N*, we note that and so

From (3.13) and , we have

We note that for each *i* = 1, 2,..., *N*. If (*w*, *w**) ∈ *G*(*Ai*) for each *i* = 1, 2,..., *N* , then it follows from the monotonicity of *A*_{
i
} that

Since *x*_{
n
} → *q* and *e*_{
n
} → 0, *x*_{
n
} + *e*_{
n
} → *q*. Therefore, for each *i* = 1, 2,..., *N*. Thus, from (3.14), we have

By the maximality of *A*_{
i
} , we have for each *i* = 1, 2,..., *N*. Hence, .

**Step 6**. .

From , we have

Since *F* ⊂ *C*_{
n
} , we also have

Letting *n* → ∞ in (3.15), we obtain

Hence, we have . This completes the proof.

As a direct consequence of Theorem 3.1, we also obtain the following result concerning a system of convex minimization problems in reflexive Banach spaces:

**Theorem 3.2**. *Let E be a real reflexive Banach space and f* : *E* → ℝ *a Legendre function which is bounded, uniformly Fréchet differentiable, and totally convex on bounded subsets of E. Let g*_{
i
} : *E* → (- ∞, ∞] (*i* = 1, 2,..., *N*) *be proper lower semi-continuous convex functions such that* *. Let**be a sequence in E such that* lim_{n→ ∞}*e*_{
n
}= 0. *Define a sequence**in E as follows:*

*If**for each i* = 1, 2,..., *N, then the sequence* {*x*_{
n
} } *converges strongly to a point*.

*Proof*. By Rockafellar's theorem [25, 26], ∂*g*_{
i
} are maximal monotone operators for each *i* = 1, 2,..., *N*. Let *λ*^{i} *>* 0 for each *i* = 1, 2,..., *N*. Then if and only if

which is equivalent to

Using Theorem 3.1, we can complete the proof.

**Remark 3.3**. By means of the composite iterative scheme together with the shrinking projection method, we can construct the proximal point algorithms for finding a common element in the set . Moreover, our algorithm is different from that of Reich and Sabach [11] which is based on a finite intersection of sets.

**Remark 3.4**. Theorems 3.1 and 3.2 also hold in a uniformly convex and uniformly smooth Banach space with the generalized duality mapping.

## References

- 1.
Brézis H, Lions PL:

**Produits infinis de résolvantes.***Israel J Math*1978,**29:**329–345. 10.1007/BF02761171 - 2.
Burachik RS, Iusem AN:

**A generalized proximal point algorithm for the variational inequality problem in a Hilbert space.***SIAM J Optim*1998,**8:**197–216. 10.1137/S1052623495286302 - 3.
Güler O:

**On the convergence of the proximal point algorithm for convex minimization.***SIAM J Control Optim*1991,**29:**403–419. 10.1137/0329022 - 4.
Passty GB:

**Ergodic convergence to a zero of the sum of monotone operators in Hilbert space.***J Math Anal Appl*1979,**72:**383–390. 10.1016/0022-247X(79)90234-8 - 5.
Reich S, Sabach S:

**A projection method for solving nonlinear problems in reflexive Banach spaces.***J Fixed Point Theory Appl* - 6.
Solodov MV, Svaiter BF:

**Forcing strong convergence of proximal point iterations in a Hilbert space.***Math Program*2000,**87:**189–202. - 7.
Martinet B:

**Régularisation d'inéquations variationelles par approximations successives.***Rev Francaise d'Informatique et de Recherche Opérationelle*1970,**4:**154–159. - 8.
Rockafellar RT:

**Monotone operators and the proximal point algorithm.***SIAM J Control Optim*1976,**14:**877–898. 10.1137/0314056 - 9.
Kamimura S, Takahashi W:

**Approximating solutions of maximal monotone operators in Hilbert spaces.***J Approx Theory*2000,**106:**226–240. 10.1006/jath.2000.3493 - 10.
Kohsaka F, Takahashi W:

**Proximal point algorithms with Bregman functions in Banach spaces.***J Nonlinear Convex Anal*2005,**6:**505–523. - 11.
Reich S, Sabach S:

**Two strong convergence theorems for a proximal method in reflexive Banach spaces.***Numer Funct Anal Optim*2010,**31:**22–44. 10.1080/01630560903499852 - 12.
Takahashi W, Takeuchi Y, Kubota R:

**Strong convergence theorems by hybrid methods for families of nonexpansive mappings in Hilbert spaces.***J Math Anal Appl*2008,**341:**276–286. 10.1016/j.jmaa.2007.09.062 - 13.
Bauschke HH, Borwein JM, Combettes PL:

**Essential smoothness, essential strict convexity, and Legendre functions in Banach spaces.***Commun Contemp Math*2001,**3:**615–647. 10.1142/S0219199701000524 - 14.
Bonnans JF, Shapiro A:

*Perturbation Analysis of Optimization Problems.*Springer Verlag, New York; 2000. - 15.
Reich S, Sabach S:

**Two strong convergence theorems for Bregman strongly nonexpansive operators in reflexive Banach spaces.***Nonlinear Anal*2010,**73:**122–135. 10.1016/j.na.2010.03.005 - 16.
Bauschke HH, Borwein JM:

**Legendre functions and the method of random Bregman projections.***J Convex Anal*1997,**4:**27–67. - 17.
Reich S, Sabach S:

**A strong convergence theorem for a proximal-type algorithm in reflexive Banach spaces.***J Nonlinear Convex Anal*2009,**10:**471–485. - 18.
Censor Y, Lent A:

**An iterative row-action method for interval convex programming.***J Optim Theory Appl*1981,**34:**321–353. 10.1007/BF00934676 - 19.
Bregman LM:

**The relaxation method for finding the common point of convex sets and its application to the solution of problems in convex programming.***USSR Comput Math Math Phys*1967,**7:**200–217. - 20.
Butnariu D, Resmerita E:

**Bregman distances, totally convex functions and a method for solving operator equations in Banach spaces.***Abstr Appl Anal*2006,**2006:**1–39. (Art ID 84919) - 21.
Bauschke HH, Borwein JM, Combettes PL:

**Bregman monotone optimization algorithms.***SIAM J Control Optim*2003,**42:**596–636. 10.1137/S0363012902407120 - 22.
Reich S, Sabach S:

**Existence and approximation of fixed points of Bregman firmly nonexpansive mappings in reflexive Banach spaces.**In*Fixed-Point Algorithms for Inverse Problems in Science and Engineering*.*Volume 49*. Edited by: Bauschke HH et al. Springer, New York; 2011:301–316. 10.1007/978-1-4419-9569-8_15 - 23.
Butnariu D, Iusem AN:

*Totally Convex Functions for Fixed Points Computation and Infinite Dimensional Optimization.*Kluwer Academic Publishers, Dordrecht; 2000. - 24.
Ambrosetti A, Prodi G:

*A Primer of Nonlinear Analysis.*Cambridge University Press, Cambridge; 1993. - 25.
Rockafellar RT:

**Characterization of the subdifferentials of convex functions.***Pac J Math*1966,**17:**497–510. - 26.
Rockafellar RT:

**On the maximal monotonicity of subdifferential mappings.***Pac J Math*1970,**33:**209–216.

## Acknowledgements

P. Cholamjiak would like to thank the Royal Golden Jubilee Project, the Thailand Research Fund; Y.J. Cho was supported by the Korea Research Foundation Grant funded by the Korean Government (KRF-2008-313-C00050); and S. Suantai was supported by the Thailand Research Fund, Thailand. Especially, P. Cholamjiak is grateful for the hospitality extended by Professor Sang Keun Lee, Chairman, and others in the Department of Mathematics Education, Gyeongsang National University, during his 3-month stay.

## Author information

## Additional information

### Competing interests

The authors declare that they have no competing interests.

### Authors' contributions

PC designed of the study, performed the nonlinear and convex analysis and also wrote the article. YJC participated in the design of the study, carried out the materials and helped to check the manuscript. SS conceived of the study, participated in its design and also helped to draft the manuscript. All authors read and approved the final manuscript.

## Rights and permissions

## About this article

#### Received

#### Accepted

#### Published

#### DOI

### Keywords

- Maximal monotone operator
- Shrinking projection method
- Proximal point algorithm
- Bregman projection
- Totally convex function
- Legendre function