 Research
 Open Access
 Published:
Acceleration of the Halpern algorithm to search for a fixed point of a nonexpansive mapping
Fixed Point Theory and Applications volume 2014, Article number: 202 (2014)
Abstract
This paper presents an algorithm to accelerate the Halpern fixed point algorithm in a real Hilbert space. To this goal, we first apply the Halpern algorithm to the smooth convex minimization problem, which is an example of a fixed point problem for a nonexpansive mapping, and indicate that the Halpern algorithm is based on the steepest descent method for solving the minimization problem. Next, we formulate a novel fixed point algorithm using the ideas of conjugate gradient methods that can accelerate the steepest descent method. We show that, under certain assumptions, our algorithm strongly converges to a fixed point of a nonexpansive mapping. We numerically compare our algorithm with the Halpern algorithm and show that it dramatically reduces the running time and iterations needed to find a fixed point compared with that algorithm.
MSC:47H10, 65K05, 90C25.
1 Introduction
Fixed point problems for nonexpansive mappings [[1], Chapter 4], [[2], Chapter 3], [[3], Chapter 1], [[4], Chapter 3] have been investigated in many practical applications, and they include convex feasibility problems [5], [[1], Example 5.21], convex optimization problems [[1], Corollary 17.5], problems of finding the zeros of monotone operators [[1], Proposition 23.38], and monotone variational inequalities [[1], Subchapter 25.5].
Fixed point problems can be solved by using useful fixed point algorithms, such as the Krasnosel’skiĭMann algorithm [[1], Subchapter 5.2], [[6], Subchapter 1.2], [7, 8], the Halpern algorithm [[6], Subchapter 1.2], [9, 10], and the hybrid method [11]. Meanwhile, to make practical systems and networks (see, e.g., [12–15] and references therein) stable and reliable, the fixed point has to be found at a faster pace. That is, we need a new algorithm that approximates the fixed point faster than the conventional ones. In this paper, we focus on the Halpern algorithm and present an algorithm to accelerate the search for a fixed point of a nonexpansive mapping.
To achieve the main objective of this paper, we first apply the Halpern algorithm to the smooth convex minimization problem, which is an example of a fixed point problem for a nonexpansive mapping, and indicate that the Halpern algorithm is based on the steepest descent method [[16], Subchapter 3.3] for solving the minimization problem.
A number of iterative methods [[16], Chapters 519] have been proposed to accelerate the steepest descent method. In particular, conjugate gradient methods [[16], Chapter 5] have been widely used as an efficient accelerated version of most gradient methods. Here, we focus on the conjugate gradient methods and devise an algorithm blending the conjugate gradient methods with the Halpern algorithm.
Our main contribution is to propose a novel algorithm for finding a fixed point of a nonexpansive mapping, for which we use the ideas of accelerated conjugate gradient methods for optimization over the fixed point set [17, 18], and prove that the algorithm converges to some fixed point in the sense of the strong topology of a real Hilbert space. To demonstrate the effectiveness and fast convergence of our algorithm, we numerically compare our algorithm with the Halpern algorithm. Numerical results show that it dramatically reduces the running time and iterations needed to find a fixed point compared with that algorithm.
This paper is organized as follows. Section 2 gives the mathematical preliminaries. Section 3 devises the acceleration algorithm for solving fixed point problems and presents its convergence analysis. Section 4 applies the proposed and conventional algorithms to a concrete fixed point problem and provides numerical examples comparing them.
2 Mathematical preliminaries
Let H be a real Hilbert space with the inner product $\u3008\cdot ,\cdot \u3009$ and its induced norm $\parallel \cdot \parallel $, and let ℕ be the set of all positive integers including zero.
2.1 Fixed point problem
Suppose that $C\subset H$ is nonempty, closed, and convex. A mapping $T:C\to C$ is said to be nonexpansive [[1], Definition 4.1(ii)], [[2], (3.2)], [[3], Subchapter 1.1], [[4], Subchapter 3.1] if
The fixed point set of $T:C\to C$ is denoted by
The metric projection onto C [[1], Subchapter 4.2, Chapter 28] is denoted by ${P}_{C}$. It is defined by ${P}_{C}(x)\in C$ and $\parallel x{P}_{C}(x)\parallel ={inf}_{y\in C}\parallel xy\parallel $ ($x\in H$). ${P}_{C}$ is nonexpansive with $Fix({P}_{C})=C$ [[1], Proposition 4.8, (4.8)].
Proposition 2.1 Suppose that $C\subset H$ is nonempty, closed, and convex, $T:C\to C$ is nonexpansive, and $x\in H$. Then

(i)
[[1], Corollary 4.15], [[2], Lemma 3.4], [[3], Proposition 5.3], [[4], Theorem 3.1.6] $Fix(T)$ is closed and convex.

(ii)
[[1], Theorem 3.14] $\stackrel{\u02c6}{x}={P}_{C}(x)$ if and only if $\u3008x\stackrel{\u02c6}{x},y\stackrel{\u02c6}{x}\u3009\le 0$ ($y\in C$).
Proposition 2.1(i) guarantees that if $Fix(T)\ne \mathrm{\varnothing}$, ${P}_{Fix(T)}(x)$ exists for all $x\in H$.
This paper discusses the following fixed point problem.
Problem 2.1 Suppose that $T:H\to H$ is nonexpansive with $Fix(T)\ne \mathrm{\varnothing}$. Then
2.2 The Halpern algorithm and our algorithm
The Halpern algorithm generates the sequence ${({x}_{n})}_{n\in \mathbb{N}}$ [[6], Subchapter 1.2] [9, 10] as follows: given ${x}_{0}\in H$ and ${({\alpha}_{n})}_{n\in \mathbb{N}}$ satisfying ${lim}_{n\to \mathrm{\infty}}{\alpha}_{n}=0$, ${\sum}_{n=0}^{\mathrm{\infty}}{\alpha}_{n}=\mathrm{\infty}$, and ${\sum}_{n=0}^{\mathrm{\infty}}{\alpha}_{n+1}{\alpha}_{n}<\mathrm{\infty}$,
Algorithm (1) strongly converges to ${P}_{Fix(T)}({x}_{0})$ ($\in Fix(T)$) [[6], Theorem 6.17], [9, 10].
Here, we shall discuss Problem 2.1 when $Fix(T)$ is the set of all minimizers of a convex, continuously Fréchet differentiable functional f over H and see that algorithm (1) is based on the steepest descent method [[16], Subchapter 3.3] to minimize f over H. Suppose that the gradient of f, denoted by ∇f, is Lipschitz continuous with a constant $L>0$ and define ${T}^{f}:H\to H$ by
where $\alpha \in (0,2/L]$ and $I:H\to H$ stands for the identity mapping. Accordingly, ${T}^{f}$ satisfies the nonexpansivity condition (see, e.g., [[12], Proposition 2.3]) and
Therefore, algorithm (1) with $T:={T}^{f}$ can be expressed as follows.
This implies that algorithm (3) uses the steepest descent direction [[16], Subchapter 3.3] ${d}_{n+1}^{f,\mathrm{SDD}}:=\mathrm{\nabla}f({x}_{n})$ of f at ${x}_{n}$, and hence algorithm (3) is based on the steepest descent method.
Meanwhile, conjugate gradient methods [[16], Chapter 5] are popular acceleration methods of the steepest descent method. The conjugate gradient direction of f at ${x}_{n}$ ($n\in \mathbb{N}$) is ${d}_{n+1}^{f,\mathrm{CGD}}:=\mathrm{\nabla}f({x}_{n})+{\beta}_{n}{d}_{n}^{f,\mathrm{CGD}}$, where ${d}_{0}^{f,\mathrm{CGD}}:=\mathrm{\nabla}f({x}_{0})$ and ${({\beta}_{n})}_{n\in \mathbb{N}}\subset (0,\mathrm{\infty})$, which, together with (2), implies that
Therefore, by replacing ${d}_{n+1}^{f}:=\mathrm{\nabla}f({x}_{n})$ in algorithm (3) with ${d}_{n+1}^{f,\mathrm{CGD}}$ defined by (4), we can formulate a novel algorithm for solving Problem 2.1.
Before presenting the algorithm, we provide the following lemmas which are used to prove the main theorem.
Proposition 2.2 [[6], Lemmas 1.2 and 1.3]
Let ${({a}_{n})}_{n\in \mathbb{N}},{({b}_{n})}_{n\in \mathbb{N}},{({c}_{n})}_{n\in \mathbb{N}},{({\alpha}_{n})}_{n\in \mathbb{N}}\subset (0,\mathrm{\infty})$ be sequences with ${a}_{n+1}\le (1{\alpha}_{n}){a}_{n}+{\alpha}_{n}{b}_{n}+{c}_{n}$ ($n\in \mathbb{N}$). Suppose that ${\sum}_{n=0}^{\mathrm{\infty}}{\alpha}_{n}=\mathrm{\infty}$, ${lim\hspace{0.17em}sup}_{n\to \mathrm{\infty}}{b}_{n}\le 0$, and ${\sum}_{n=0}^{\mathrm{\infty}}{c}_{n}<\mathrm{\infty}$. Then ${lim}_{n\to \mathrm{\infty}}{a}_{n}=0$.
Proposition 2.3 [[19], Lemma 1]
Suppose that ${({x}_{n})}_{n\in \mathbb{N}}\subset H$ weakly converges to $x\in H$ and $y\ne x$. Then ${lim\hspace{0.17em}inf}_{n\to \mathrm{\infty}}\parallel {x}_{n}x\parallel <{lim\hspace{0.17em}inf}_{n\to \mathrm{\infty}}\parallel {x}_{n}y\parallel $.
3 Acceleration of the Halpern algorithm
We present the following algorithm.
Algorithm 3.1
Step 0. Choose $\mu \in (0,1]$, $\alpha >0$, and ${x}_{0}\in H$ arbitrarily, and set ${({\alpha}_{n})}_{n\in \mathbb{N}}\subset (0,1)$, ${({\beta}_{n})}_{n\in \mathbb{N}}\subset [0,\mathrm{\infty})$. Compute ${d}_{0}:=(T({x}_{0}){x}_{0})/\alpha $.
Step 1. Given ${x}_{n},{d}_{n}\in H$, compute ${d}_{n+1}\in H$ by
Compute ${x}_{n+1}\in H$ as follows.
Put $n:=n+1$, and go to Step 1.
We can check that Algorithm 3.1 coincides with the Halpern algorithm (1) when ${\beta}_{n}:=0$ ($n\in \mathbb{N}$) and $\mu :=1$.
This section makes the following assumptions.
Assumption 3.1 The sequences ${({\alpha}_{n})}_{n\in \mathbb{N}}$ and ${({\beta}_{n})}_{n\in \mathbb{N}}$ satisfy^{a}
Moreover, ${({x}_{n})}_{n\in \mathbb{N}}$ in Algorithm 3.1 satisfies
Let us do a convergence analysis of Algorithm 3.1.
Theorem 3.1 Under Assumption 3.1, the sequence ${({x}_{n})}_{n\in \mathbb{N}}$ generated by Algorithm 3.1 strongly converges to ${P}_{Fix(T)}({x}_{0})$.
Algorithm 3.1 when ${\beta}_{n}:=0$ ($n\in \mathbb{N}$) is the Halpern algorithm defined by
Hence, the nonexpansivity of T ensures that, for all $x\in Fix(T)$ and for all $n\in \mathbb{N}$,
Suppose that $n:=0$. From (5), we have $\parallel {x}_{1}x\parallel \le \mu {\alpha}_{0}\parallel {x}_{0}x\parallel +(1\mu {\alpha}_{0})\parallel {x}_{0}x\parallel =\parallel {x}_{0}x\parallel $. Assume that $\parallel {x}_{m}x\parallel \le \parallel {x}_{0}x\parallel $ for some $m\in \mathbb{N}$. Then (5) implies that $\parallel {x}_{m+1}x\parallel \le \mu {\alpha}_{m}\parallel {x}_{0}x\parallel +(1\mu {\alpha}_{m})\parallel {x}_{m}x\parallel \le \mu {\alpha}_{m}\parallel {x}_{0}x\parallel +(1\mu {\alpha}_{m})\parallel {x}_{0}x\parallel =\parallel {x}_{0}x\parallel $. Hence, induction guarantees that
Therefore, we find that ${({x}_{n})}_{n\in \mathbb{N}}$ is bounded. Moreover, since the nonexpansivity of T ensures that ${(T({x}_{n}))}_{n\in \mathbb{N}}$ is also bounded, (C5) holds. Accordingly, Theorem 3.1 says that if ${({\alpha}_{n})}_{n\in \mathbb{N}}$ satisfies (C1)(C3), Algorithm 3.1 when ${\beta}_{n}:=0$ ($n\in \mathbb{N}$) (i.e., the Halpern algorithm) strongly converges to ${P}_{Fix(T)}({x}_{0})$. This means that Theorem 3.1 is a generalization of the convergence analysis of the Halpern algorithm.
3.1 Proof of Theorem 3.1
We first show the following lemma.
Lemma 3.1 Suppose that Assumption 3.1 holds. Then ${({d}_{n})}_{n\in \mathbb{N}}$, ${({x}_{n})}_{n\in \mathbb{N}}$, and ${({y}_{n})}_{n\in \mathbb{N}}$ are bounded.
Proof We have from (C1) and (C4) that ${lim}_{n\to \mathrm{\infty}}{\beta}_{n}=0$. Accordingly, there exists ${n}_{0}\in \mathbb{N}$ such that ${\beta}_{n}\le 1/2$ for all $n\ge {n}_{0}$. Define ${M}_{1}:=max\{\parallel {d}_{{n}_{0}}\parallel ,(2/\alpha ){sup}_{n\in \mathbb{N}}\parallel T({x}_{n}){x}_{n}\parallel \}$. Then (C5) implies that ${M}_{1}<\mathrm{\infty}$. Assume that $\parallel {d}_{n}\parallel \le {M}_{1}$ for some $n\ge {n}_{0}$. The triangle inequality ensures that
which means that $\parallel {d}_{n}\parallel \le {M}_{1}$ for all $n\ge {n}_{0}$, i.e., ${({d}_{n})}_{n\in \mathbb{N}}$ is bounded.
The definition of ${y}_{n}$ ($n\in \mathbb{N}$) implies that
The nonexpansivity of T and (6) imply that, for all $x\in Fix(T)$ and for all $n\ge {n}_{0}$,
Therefore, we find that, for all $x\in Fix(T)$ and for all $n\ge {n}_{0}$,
which, together with (C4) and ${\alpha}_{n}<1$ ($n\in \mathbb{N}$), means that, for all $x\in Fix(T)$ and for all $n\ge {n}_{0}$,
Induction guarantees that, for all $x\in Fix(T)$ and for all $n\ge {n}_{0}$,
Therefore, ${({x}_{n})}_{n\in \mathbb{N}}$ is bounded.
The definition of ${y}_{n}$ ($n\in \mathbb{N}$) and the boundedness of ${({x}_{n})}_{n\in \mathbb{N}}$ and ${({d}_{n})}_{n\in \mathbb{N}}$ imply that ${({y}_{n})}_{n\in \mathbb{N}}$ is also bounded. This completes the proof. □
Lemma 3.2 Suppose that Assumption 3.1 holds. Then

(i)
${lim}_{n\to \mathrm{\infty}}\parallel {x}_{n+1}{x}_{n}\parallel =0$.

(ii)
${lim}_{n\to \mathrm{\infty}}\parallel {x}_{n}T({x}_{n})\parallel =0$.

(iii)
${lim\hspace{0.17em}sup}_{n\to \mathrm{\infty}}\u3008{x}_{0}{x}^{\star},{y}_{n}{x}^{\star}\u3009\le 0$, where ${x}^{\star}:={P}_{Fix(T)}({x}_{0})$.
Proof (i) Equation (6), the triangle inequality, and the nonexpansivity of T imply that, for all $n\in \mathbb{N}$,
which, together with $\parallel {d}_{n}\parallel \le {M}_{1}$ ($n\ge {n}_{0}$) and (C4), implies that, for all $n\ge {n}_{0}$,
On the other hand, from ${\alpha}_{n}\le {\alpha}_{n+1}{\alpha}_{n}+{\alpha}_{n+1}$ and ${\alpha}_{n}<1$ ($n\in \mathbb{N}$), we have that, for all $n\in \mathbb{N}$,
We also find that, for all $n\in \mathbb{N}\mathrm{\setminus}\{0\}$,
where ${M}_{2}:={sup}_{n\in \mathbb{N}}\mu (\parallel {x}_{0}\parallel +\parallel {y}_{n}\parallel )<\mathrm{\infty}$. Hence, (7) and (8) ensure that, for all $n\ge {n}_{0}$,
Proposition 2.2, (C1), (C2), and (C3) lead us to
(ii) From $\parallel {x}_{n+1}{y}_{n}\parallel =\mu {\alpha}_{n}\parallel {x}_{0}{y}_{n}\parallel \le {M}_{2}{\alpha}_{n}$ ($n\in \mathbb{N}$), (C1) means that ${lim}_{n\to \mathrm{\infty}}\parallel {x}_{n+1}{y}_{n}\parallel =0$. Since the triangle inequality ensures that $\parallel {y}_{n}{x}_{n}\parallel \le \parallel {y}_{n}{x}_{n+1}\parallel +\parallel {x}_{n+1}{x}_{n}\parallel $ ($n\in \mathbb{N}$), we find from (9) that
From the definition of ${d}_{n+1}$ ($n\in \mathbb{N}$), we have, for all $n\ge {n}_{0}$,
Since Equation (10) and ${lim}_{n\to \mathrm{\infty}}{\beta}_{n}=0$ guarantee that the righthand side of the above inequality converges to 0, we find that

(iii)
From the limit superior of ${(\u3008{x}_{0}{x}^{\star},{y}_{n}{x}^{\star}\u3009)}_{n\in \mathbb{N}}$, there exists ${({y}_{{n}_{k}})}_{k\in \mathbb{N}}$ ($\subset {({y}_{n})}_{n\in \mathbb{N}}$) such that
$$\underset{n\to \mathrm{\infty}}{lim\hspace{0.17em}sup}\u3008{x}_{0}{x}^{\star},{y}_{n}{x}^{\star}\u3009=\underset{k\to \mathrm{\infty}}{lim}\u3008{x}_{0}{x}^{\star},{y}_{{n}_{k}}{x}^{\star}\u3009.$$(12)
Moreover, since ${({y}_{{n}_{k}})}_{k\in \mathbb{N}}$ is bounded, there exists ${({y}_{{n}_{{k}_{i}}})}_{i\in \mathbb{N}}$ ($\subset {({y}_{{n}_{k}})}_{k\in \mathbb{N}}$) which weakly converges to some point $\stackrel{\u02c6}{y}$ (∈H). Equation (10) guarantees that ${({x}_{{n}_{{k}_{i}}})}_{i\in \mathbb{N}}$ weakly converges to $\stackrel{\u02c6}{y}$.
We shall show that $\stackrel{\u02c6}{y}\in Fix(T)$. Assume that $\stackrel{\u02c6}{y}\notin Fix(T)$, i.e., $\stackrel{\u02c6}{y}\ne T(\stackrel{\u02c6}{y})$. Proposition 2.3, (11), and the nonexpansivity of T ensure that
This is a contradiction. Hence, $\stackrel{\u02c6}{y}\in Fix(T)$. Hence, (12) and Proposition 2.1(ii) guarantee that
This completes the proof. □
Now, we are in a position to prove Theorem 3.1.
Proof of Theorem 3.1 The inequality ${\parallel x+y\parallel}^{2}\le {\parallel x\parallel}^{2}+2\u3008y,x+y\u3009$ ($x,y\in H$), (6), and the nonexpansivity of T imply that, for all $n\in \mathbb{N}$,
where ${\beta}_{n}\le {\alpha}_{n}^{2}$ ($n\in \mathbb{N}$) and ${M}_{3}:={sup}_{n\in \mathbb{N}}2\alpha \u3008{y}_{n}{x}^{\star},{d}_{n}\u3009<\mathrm{\infty}$. We thus have that, for all $n\in \mathbb{N}$,
Proposition 2.2, (C1), (C2), and Lemma 3.2(iii) lead one to deduce that
This guarantees that ${({x}_{n})}_{n\in \mathbb{N}}$ generated by Algorithm 3.1 strongly converges to ${x}^{\star}:={P}_{Fix(T)}({x}_{0})$. □
Suppose that $Fix(T)$ is bounded. Then we can set a bounded, closed convex set C ($\supset Fix(T)$) such that ${P}_{C}$ can be computed within a finite number of arithmetic operations (e.g., C is a closed ball with a large enough radius). Hence, we can compute
instead of ${x}_{n+1}$ in Algorithm 3.1. From ${({x}_{n})}_{n\in \mathbb{N}}\subset C$, the boundedness of C means that ${({x}_{n})}_{n\in \mathbb{N}}$ is bounded. The nonexpansivity of T guarantees that $\parallel T({x}_{n})T(x)\parallel \le \parallel {x}_{n}x\parallel $ ($x\in Fix(T)$), which means that ${(T({x}_{n}))}_{n\in \mathbb{N}}$ is bounded. Therefore, (C5) holds. We can prove that Algorithm 3.1 with (13) strongly converges to a point in $Fix(T)$ by referring to the proof of Theorem 3.1.
Let us consider the case where $Fix(T)$ is unbounded. In this case, we cannot choose a bounded C satisfying $Fix(T)\subset C$. Although we can execute Algorithm 3.1, we need to verify the boundedness of ${(T({x}_{n}){x}_{n})}_{n\in \mathbb{N}}$. Instead, we can apply the Halpern algorithm (1) to this case without any problem. However, the Halpern algorithm would converge slowly because it is based on the steepest descent method (see Section 1). Hence, in this case, it would be desirable to execute not only the Halpern algorithm but also Algorithm 3.1.
4 Numerical examples and conclusion
Let us apply the Halpern algorithm (1) and Algorithm 3.1 to the following convex feasibility problem [5], [[1], Example 5.21].
Problem 4.1 Given a nonempty, closed convex set ${C}_{i}\subset {\mathbb{R}}^{N}$ ($i=0,1,\dots ,m$),
where one assumes that $C\ne \mathrm{\varnothing}$.
Define a mapping $T:{\mathbb{R}}^{N}\to {\mathbb{R}}^{N}$ by
where ${P}_{i}:={P}_{{C}_{i}}$ ($i=0,1,\dots ,m$) stands for the metric projection onto ${C}_{i}$. Since ${P}_{i}$ ($i=0,1,\dots ,m$) is nonexpansive, T defined by (14) is also nonexpansive. Moreover, we find that
Therefore, Problem 4.1 coincides with Problem 2.1 with T defined by (14).
The experiment used an Apple Macbook Air with a 1.30GHz Intel(R) Core(TM) i54250U CPU and 4GB DDR3 memory. The Halpern algorithm (1) and Algorithm 3.1 were written in Java. The operating system of the computer was Mac OSX version 10.8.5.
We set $\alpha :=1$, $\mu :=1/{10}^{5}$, ${\alpha}_{n}:=1/(n+1)$ ($n\in \mathbb{N}$), and ${\beta}_{n}:=1/{(n+1)}^{2}$ ($n\in \mathbb{N}$) in Algorithm 3.1 and compared Algorithm 3.1 with the Halpern algorithm (1) with ${\alpha}_{n}:=\mu /(n+1)$ ($n\in \mathbb{N}$). In the experiment, we set ${C}_{i}$ ($i=0,1,\dots ,m$) as a closed ball with center ${c}_{i}\in {\mathbb{R}}^{N}$ and radius ${r}_{i}>0$. Thus, ${P}_{i}$ ($i=0,1,\dots ,m$) can be computed with
or ${P}_{i}(x):=x$ if $\parallel {c}_{i}x\parallel \le {r}_{i}$.
We set $N:=100$, $m:=3$, ${r}_{i}:=1$ ($i=0,1,2,3$), and ${c}_{0}:=0$. The experiment used random vectors ${c}_{i}\in {(1/\sqrt{N},1/\sqrt{N})}^{N}$ ($i=1,2,3$) generated by the java.util.Random class so as to satisfy $C\ne \mathrm{\varnothing}$. We also used the java.util.Random class to set a random initial point in ${(16,16)}^{N}$.
Figure 1 describes the behaviors of $\parallel T({x}_{n}){x}_{n}\parallel $ for the Halpern algorithm (1) and Algorithm 3.1 (proposed). The xaxis and yaxis represent the elapsed time and value of $\parallel T({x}_{n}){x}_{n}\parallel $. The results show that compared with the Halpern algorithm, Algorithm 3.1 dramatically reduces the time required to satisfy $\parallel T({x}_{n}){x}_{n}\parallel <{10}^{6}$. We found that the Halpern algorithm took 850 iterations to satisfy $\parallel T({x}_{n}){x}_{n}\parallel <{10}^{6}$, whereas Algorithm 3.1 took only six.
This paper presented an algorithm to accelerate the Halpern algorithm for finding a fixed point of a nonexpansive mapping on a real Hilbert space and its convergence analysis. The convergence analysis guarantees that the proposed algorithm strongly converges to a fixed point of a nonexpansive mapping under certain assumptions. We numerically compared the abilities of the proposed and Halpern algorithms in solving a concrete fixed point problem. The results showed that the proposed algorithm performs better than the Halpern algorithm.
Endnote
Examples of ${({\alpha}_{n})}_{n\in \mathbb{N}}$ and ${({\beta}_{n})}_{n\in \mathbb{N}}$ satisfying (C1)(C4) are ${\alpha}_{n}:=1/{(n+1)}^{a}$ and ${\beta}_{n}:=1/{(n+1)}^{2a}$ ($n\in \mathbb{N}$), where $a\in (0,1]$.
References
 1.
Bauschke HH, Combettes PL: Convex Analysis and Monotone Operator Theory in Hilbert Spaces. Springer, Berlin; 2011.
 2.
Goebel K, Kirk WA Cambridge Studies in Advanced Mathematics. In Topics in Metric Fixed Point Theory. Cambridge University Press, Cambridge; 1990.
 3.
Goebel K, Reich S: Uniform Convexity, Hyperbolic Geometry, and Nonexpansive Mappings. Dekker, New York; 1984.
 4.
Takahashi W: Nonlinear Functional Analysis. Yokohama Publishers, Yokohama; 2000.
 5.
Bauschke HH, Borwein JM: On projection algorithms for solving convex feasibility problems. SIAM Rev. 1996, 38(3):367–426. 10.1137/S0036144593251710
 6.
Berinde V: Iterative Approximation of Fixed Points. Springer, Berlin; 2007.
 7.
Krasnosel’skiĭ MA: Two remarks on the method of successive approximations. Usp. Mat. Nauk 1955, 10: 123–127.
 8.
Mann WR: Mean value methods in iteration. Proc. Am. Math. Soc. 1953, 4: 506–510. 10.1090/S00029939195300548463
 9.
Halpern B: Fixed points of nonexpansive maps. Bull. Am. Math. Soc. 1967, 73: 957–961. 10.1090/S000299041967118640
 10.
Wittmann R: Approximation of fixed points of nonexpansive mappings. Arch. Math. 1992, 58(5):486–491. 10.1007/BF01190119
 11.
Nakajo K, Takahashi W: Strong convergence theorems for nonexpansive mappings and nonexpansive semigroups. J. Math. Anal. Appl. 2003, 279: 372–379. 10.1016/S0022247X(02)004584
 12.
Iiduka H: Iterative algorithm for solving triplehierarchical constrained optimization problem. J. Optim. Theory Appl. 2011, 148: 580–592. 10.1007/s109570109769z
 13.
Iiduka H: Fixed point optimization algorithm and its application to power control in CDMA data networks. Math. Program. 2012, 133: 227–242. 10.1007/s101070100427x
 14.
Iiduka H: Iterative algorithm for triplehierarchical constrained nonconvex optimization problem and its application to network bandwidth allocation. SIAM J. Optim. 2012, 22(3):862–878. 10.1137/110849456
 15.
Iiduka H: Fixed point optimization algorithms for distributed optimization in networked systems. SIAM J. Optim. 2013, 23: 1–26. 10.1137/120866877
 16.
Nocedal J, Wright SJ Springer Series in Operations Research and Financial Engineering. In Numerical Optimization. 2nd edition. Springer, Berlin; 2006.
 17.
Iiduka H: Acceleration method for convex optimization over the fixed point set of a nonexpansive mapping. Math. Program. 2014. 10.1007/s1010701307411
 18.
Iiduka H, Yamada I: A use of conjugate gradient direction for the convex optimization problem over the fixed point set of a nonexpansive mapping. SIAM J. Optim. 2009, 19(4):1881–1893. 10.1137/070702497
 19.
Opial Z: Weak convergence of the sequence of successive approximation for nonexpansive mappings. Bull. Am. Math. Soc. 1967, 73: 591–597. 10.1090/S000299041967117610
Acknowledgements
We are sincerely grateful to the associate editor LaiJiu Lin and the two anonymous reviewers for helping us improve the original manuscript.
Author information
Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
All authors contributed equally to the writing of this paper. All authors read and approved the final manuscript.
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0), which permits use, duplication, adaptation, distribution, and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Sakurai, K., Iiduka, H. Acceleration of the Halpern algorithm to search for a fixed point of a nonexpansive mapping. Fixed Point Theory Appl 2014, 202 (2014). https://doi.org/10.1186/168718122014202
Received:
Accepted:
Published:
Keywords
 conjugate gradient method
 fixed point
 Halpern algorithm
 nonexpansive mapping
 smooth convex optimization
 steepest descent method