Skip to content
• Research
• Open Access

# Higher-order Lipschitz mappings

Fixed Point Theory and Applications20152015:88

https://doi.org/10.1186/s13663-015-0334-1

• Received: 10 March 2015
• Accepted: 18 May 2015
• Published:

## Abstract

We study self-mappings on complete metric spaces, which we refer to as higher-order Lipschitz mappings. These mappings generalise Lipschitz mappings, the latter which are equivalent to first-order Lipschitz mappings studied in this paper. The main result of this paper is to extend the Banach fixed point theorem (and an often-cited generalisation) to higher-order contraction mappings. We also present results on the problem of local Lipschitzity of these higher-order Lipschitz mappings.

## Keywords

• metric space
• fixed point
• Lipschitz mapping
• higher-order Lipschitz mapping
• local Lipschitzity
• stable and unstable polynomials
• Baire category theorem
• Perron-Frobenius theorem

• 47H09
• 47H10

## 1 Introduction

Let $$(\mathcal{X},d)$$ be a complete metric space and let $$T:\mathcal{X}\to\mathcal{X}$$ be a Lipschitz mapping, that is, $$d(Ty,Tx)\leq cd(y,x)$$ for all $$x,y \in\mathcal{X}$$ where $$c\ge0$$. When $$0\le c<1$$, then T is referred to as a contraction mapping and when $$c=1$$, then T is referred to as a non-expansive mapping. In this paper, we consider the following generalisation of Lipschitz mappings:

### Definition 1.1

(Higher-order Lipschitz mapping)

A mapping $$T:\mathcal{X}\to\mathcal{X}$$ on a metric space $$(\mathcal{X},d)$$ is an rth-order Lipschitz mapping if
$$d\bigl(T^{r}y,T^{r}x\bigr)\leq\sum _{k=0}^{r-1}c_{k}d\bigl(T^{k}y,T^{k}x \bigr) \quad \forall x,y\in \mathcal{X},$$
(1)
where r is a natural number and $$c_{k}$$, for all $$0\le k\le r-1$$, are non-negative real numbers.
An example is when $$\mathcal{X}$$ is a finite dimensional vector space and $$T:\mathcal{X}\to\mathcal{X}$$ is a matrix. Indeed, by the Cayley-Hamilton theorem [1, 2], T in this case satisfies an identity
$$f(T)=T^{r}-a_{r-1}T^{r-1}-\cdots-a_{0}=0,$$
where $$f(z)=z^{r}-a_{r-1}z^{r-1}-\cdots-a_{0}$$ denotes the characteristic polynomial of T. It follows that we have the identity
$$T^{r}y-T^{r}x=a_{r-1}\bigl(T^{r-1}y-T^{r-1}x \bigr)+\cdots+a_{0}(y-x),$$
which upon taking norms and using the triangle inequality gives us
$$\bigl\| T^{r}y-T^{r}x\bigr\| \le c_{r-1} \bigl\| T^{r-1}y-T^{r-1}x\bigr\| +\cdots+c_{0}\|y-x\|,$$
(2)
where $$c_{k}:=|a_{k}|$$. Now, as in the first-order case, we classify higher-order Lipschitz mappings into three cases, thus
• T is an rth-order contraction mapping if the polynomial $$p(z):=z^{r}-\sum_{k=0}^{r-1}c_{k}z^{k}$$ is stable, that is, $$|\lambda|<1$$ if $$p(\lambda)=0$$.

• T is an rth-order non-expansive mapping if the polynomial $$p(z):=z^{r}-\sum_{k=0}^{r-1}c_{k}z^{k}$$ is tamely unstable, that is, there exists at least a magnitude-wise dominating root $$\lambda\in\mathbb{C}$$ such that $$p(\lambda)=0$$ and $$|\lambda|=1$$.

• T is an rth-order expansive (Lipschitz) mapping if the polynomial $$p(z):=z^{r}-\sum_{k=0}^{r-1}c_{k}z^{k}$$ is wildly unstable, that is, there exists $$\lambda\in\mathbb{C}$$ such that $$|\lambda|>1$$ and $$p(\lambda)=0$$.

Here, $$\mathbb{C}$$ denotes the field of complex numbers. In Section 2, we give equivalent classification based on the coefficients $$c_{k}$$.

In the following subsections, we review the pertinent results on the fixed point theory of Lipschitz mappings and show the relationship with the fixed point theory of the higher-order counterparts as introduced above.

### 1.1 Fixed point theory of contraction mappings in metric spaces

The basic result of metric fixed point theory is the Banach  fixed point theorem (or the contraction mapping theorem).

### Theorem 1.2

(Banach fixed point)

Let $$(\mathcal{X},d)$$ be a complete metric space and let $$T:\mathcal{X}\to\mathcal{X}$$ be a contraction mapping. Then T has a unique fixed point given by the limit of Picard iterates $$x_{n+1}:=Tx_{n}$$.

Theorem 1.2 is particularly useful in the demonstration of existence and uniqueness of solutions to certain problems in analysis and economics (see ). A survey of various extensions of Theorem 1.2 can be found in ; we highlight the important results related to those demonstrated in this paper. First, the higher-order contraction case when $$r>1$$ and $$c_{k}=0$$ for all $$k\geq1$$ is an often-cited generalisation in many texts on Theorem 1.2; this is the case when $$T^{r}$$, but not $$T^{k}$$ for all $$k< r$$, is a contraction mapping; that is:

### Theorem 1.3

Let $$(\mathcal{X},d)$$ be a complete metric space and $$T:\mathcal{X}\to\mathcal{X}$$ a mapping such that $$T^{r}$$ is a contraction for some $$r>1$$. Then T has a unique fixed point given by the limit of Picard iterates $$x_{n+1}:=Tx_{n}$$.

In Section 3, we demonstrate that the conclusions of Theorems 1.2 and 1.3 extend to all higher-order contraction mappings. Both (first-order) contraction mappings and the rth-order contraction mappings defined in Theorem 1.3 are special cases of the now-proven generalised Banach contraction conjecture (see Jachymski , Merryfield-Stein  and Arvanitakis ).

### Theorem 1.4

(Generalised Banach contraction theorem)

Let $$(\mathcal{X},d)$$ be a complete metric space and let $$T:\mathcal{X}\to\mathcal{X}$$ be a mapping such that
$$\mathop{\min_{1\leq i\leq r}}_{y\neq x}\frac{d(T^{i}y,T^{i}x)}{d(y,x)}\leq c$$
for some natural number r and real number $$c\in(0,1)$$. Then T has a unique fixed point.
In the present paper, we demonstrate the closely related result that an rth-order contraction mapping satisfies the minimising inequality
$$\mathop{\min_{1\leq i\leq r}}_{T^{i-1}y\ne T^{i-1}x}\frac {d(T^{i}y,T^{i}x)}{d(T^{i-1}y,T^{i-1}x)}\leq c$$
(3)
for some real number $$c\in(0,1)$$. Indeed inequality (3) also holds true for higher-order non-expansive and higher-order expansive mappings, respectively, for some $$c=1$$ and $$c>1$$.

Now an early continuous mapping generalisation of the Banach fixed point theorem is the following result due to Caccioppoli :

### Theorem 1.5

Let $$(\mathcal{X},d)$$ be a complete metric space and let $$T:\mathcal{X}\to\mathcal{X}$$ be a mapping such that
$$d\bigl(T^{n}y,T^{n}x\bigr)\le q_{n}d(y,x),$$
(4)
where $$\{q_{n}\}_{n\ge1}$$ is a summable non-negative sequence independent of $$\mathcal{X}$$. Then T has a unique fixed point given by the limit of the Picard iterates $$x_{n+1}:=Tx_{n}$$.
Whereas higher-order contraction mappings do not generally satisfy the hypothesis of Theorem 1.5 (note that the mappings satisfying Theorem 1.5 are necessarily uniformly continuous), we demonstrate in Section 4 that when the additional requirement of continuity is imposed on a higher-order contraction mapping, the inequality (4) holds locally in the sense that for every given $$x_{0}\in\mathcal{X}$$, there exists an (open) subset $$\mathcal{S}\subset\mathcal{X}$$ (depending on, but not necessarily a neighbourhood of, $$x_{0}$$) such that for all $$x\in\mathcal{S}$$
$$d\bigl(T^{n}x,T^{n}x_{0}\bigr)\le Mc^{n}d(x,x_{0})$$
for some constants $$M\ge1$$ and $$c\in(0,1)$$. Indeed the same is true for higher-order non-expansive and higher-order expansive mappings, respectively, but with the constant $$c=1$$ and $$c>1$$.
In general, higher-order Lipschitz mappings are not reducible to lower-order Lipschitz mappings within the same metric space $$(\mathcal {X},d)$$. Whereas Lipschitz mappings are (uniformly) continuous mappings, this need not be the case for general higher-order Lipschitz mappings; they may not even be continuous: for instance, the function $$T:\mathbb{R}\to\mathbb{R}$$ given by
$$Tx:=\textstyle\begin{cases} 1 & \mbox{if }x< 0,\\ 0 & \mbox{if }x\ge0, \end{cases}$$
with the metric induced by the usual absolute value on $$\mathbb{R}$$, is discontinuous at $$x=0$$ but we observe that $$T^{2}x=0$$ and so
$$\bigl|T^{2}y-T^{2}x\bigr|=0\le c|y-x|$$
for any $$c\ge0$$, thus making T a second-order Lipschitz (indeed, second-order contraction) mapping. The immediate cases for which a higher-order Lipschitz mapping may be of lower-order is when it is actually of lower-order on $$\mathcal{X}$$ itself (for instance, matrices, which are actually (first-order) Lipschitz mappings but satisfy (2) and also when it is of lower-order on $$T(\mathcal{X})$$; an example in the latter case is the mapping $$T:\mathbb{R}\to\mathbb{R}$$ given by
$$Tx:=\textstyle\begin{cases} 1-x & \mbox{if }x< 0,\\ x & \mbox{if }x\ge0, \end{cases}$$
with the metric induced by the usual absolute value on $$\mathbb{R}$$. Obviously T is discontinuous at $$x=0$$ and noting that $$T^{2}=T$$, then $$|T^{2}y-T^{2}x|=|Ty-Tx|$$; in other words, T is second-order non-expansive mapping but at the same time it is (first-order) non-expansive on the metric subspace $$(T(\mathbb{R}),|\cdot|)$$. We shall refer to T when it is actually of lower-order on $$\mathcal {X}$$ or $$T(\mathcal{X})$$ as a trivial mapping.

### 1.2 Fixed point theory of non-contraction mappings in Banach spaces

Now, for non-contraction mappings, complete metric spaces are in general not sufficient to guarantee the existence or uniqueness fixed points; in this regard, usually, compactness and/or convexity of subsets of normed linear spaces is required. Some noteworthy results are as follows.

### Theorem 1.6

(Edelstein )

Let T be a contractive mapping on a compact metric space, that is, $$d(Ty,Tx)\le d(y,x)$$ with equality only if $$x=y$$. Then T has a unique fixed point given by the limit of Picard iteration $$x_{n+1}:=Tx_{n}$$.

### Theorem 1.7

(Kirk )

Let T be a non-expansive self-mapping on a weakly compact convex subset $$\mathcal{C}$$ of a Banach space with normal structure - that is, for any bounded non-empty convex subset $$\mathcal{K}\subset\mathcal{C}$$ there exists a point $$x_{0}\in\mathcal {K}$$ such that $$\sup_{x\in\mathcal{K}}\|x_{0}-x\|<\operatorname{diam}(\mathcal {K}):=\sup_{x,y\in\mathcal{K}}\|x-y\|$$. Then T has a fixed point.

### Theorem 1.8

(Schauder1 )

Let T be a Lipschitz self-mapping on a compact convex subset of a Banach space. Then T has a fixed point.

In the present paper, we do not investigate the fixed point theory of higher-order Lipschitz mappings under the hypotheses in Theorems 1.6, 1.7 and 1.8 above. These are deferred to a sequel to this paper.

## 2 Preliminaries

First of all, we recall the following definitions:

A nowhere dense subset of a topological space is a set whose closure has an empty interior in the topological space; that is, it contains no open neighbourhood of its elements in the topological space.

A real matrix is non-negative if all its entries are non-negative real numbers; it is positive if all the entries are positive real numbers.

A non-negative matrix A is irreducible if for every pair of indices i, j, there exists a natural number n such that $$(\mathbf{A}^{n})_{ij}>0$$.

A real non-negative matrix A is primitive if there exists an integer $$n\ge1$$ such that $$\mathbf{A}^{n}$$ is positive; thus, a primitive matrix is irreducible.

A polynomial $$f(z)$$ is non-degenerate if whenever $$\alpha\neq \beta$$ but $$f(\alpha)=f(\beta)=0$$, then $$\alpha\neq\zeta\beta$$, where ζ is a root of unity.

An rth-order linear recurrence sequence $$S_{n}$$, satisfying the recursive equation $$S_{n+r}=\sum_{k=0}^{r-1}c_{k}S_{n+k}$$, is non-degenerate if the associated characteristic polynomial $$p(z)=z^{r}-\sum_{k=0}^{r-1}c_{k}z^{k}$$ is non-degenerate.

The following useful results are necessary for the proof of our main results. As used below and elsewhere, we employ the Kronecker delta symbol $$\delta_{jk}$$, which equals 1 when $$j=k$$ and equals 0 if otherwise.

### Theorem 2.1

Let $$S_{n}$$ be an rth-order linear recurrence sequence satisfying the recursion $$S_{n+r}=\sum_{k=0}^{r-1}c_{k}S_{n+k}$$. Let $$p(z)=z^{r}-\sum_{k=0}^{r-1}c_{k}z^{k}$$ be the associated characteristic polynomial having distinct (complex) roots $$\lambda_{1},\lambda_{2},\ldots$$ with respective multiplicities $$\mu_{1},\mu_{2},\ldots$$ - thus $$\sum_{i}\mu_{i}=r$$. Then $$S_{n}$$ has an explicit form
$$S_{n}=\sum_{i}p_{i}(n) \lambda_{i}^{n},$$
where $$p_{i}(n)$$ is a polynomial of degree $$\mu_{i}-1$$. Also, $$S_{n}$$ has an implicit form
$$S_{n}=\sum_{j=0}^{r-1}I_{j}(n)S_{j},$$
where $$I_{j}(n)$$ - an impulse-response sequence of $$S_{n}$$ - is also an rth-order linear recurrence sequence satisfying the same recursion as $$S_{n}$$ with initial values $$I_{j}(k)=\delta_{jk}$$ for all $$0\leq j,k\leq r-1$$.

### Proof

See for instance , Sections 1.1.4 through 1.1.6. □

The following corollary, which follows straightforwardly from the above theorem, is what is most useful for our purposes here.

### Corollary 2.2

Using the notation of Theorem  2.1, define $$\lambda:=\max |\lambda_{i} |$$ and $$\mu:=\max\mu_{i}$$. Then $$|S_{n}|\leq K\lambda^{n}n^{\mu -1}$$ for some absolute constant $$K>0$$ independent of n, λ, μ; moreover, if $$\lambda<1$$, then $$\lim_{n\to\infty}S_{n} =0$$.

### Theorem 2.3

(Skolem-Mahler-Lech )

The set of zeroes of a linear recurrence sequence $$S_{n}$$ over a field of characteristic zero comprises a finite set together with a finite number of arithmetic progressions. If $$S_{n}$$ is non-degenerate, then the set of zeroes is finite.

### Remark

The set referred to is the set of indices n for which $$S_{n} = 0$$ and in either case it may be empty.

### Theorem 2.4

(Baire category )

Let $$\mathcal{X}$$ be a complete metric space. Then $$\mathcal{X}$$ is not the countable union of nowhere dense closed sets.

### Theorem 2.5

(Perron-Frobenius [21, 22])

Let A be an irreducible non-negative $$r\times r$$ matrix with spectral radius $$\rho(\mathbf{A})$$. Then the following statements hold:
1. 1.

$$\rho(\mathbf{A})$$ is an eigenvalue of A and it is uniquely dominating if A is primitive;

2. 2.

$$\min_{i}\sum_{j}\mathbf{A}_{ij}\leq\rho(\mathbf{A})\leq\max_{i}\sum_{j}\mathbf{A}_{ij}$$;

3. 3.

Collatz-Wielandt formula: Let $$\mathbf{N}:=\{\mathbf{v}=\{v_{j}\geq0\} _{j=1}^{r}:\exists i,v_{i}\neq0 \}$$. Then $$\rho(\mathbf{A})=\max_{\mathbf{v}\in\mathbf{N}}\min_{1\leq i\leq r,v_{i}\neq0}\frac{1}{v_{i}}(\mathbf{A}\mathbf{v})_{i}$$.

### Remark

The Perron-Frobenius theorem is more general than this but this suffices for our purposes here. By a uniquely dominating eigenvalue, we imply one which is a unique maximum in absolute value.

### Theorem 2.6

(Keilson-Styan inequality )

Let A be a non-negative $$r\times r$$ matrix with spectral radius $$\rho(\mathbf{A})$$. Then $$\det(tI-\mathbf{A})\le t^{r}-\rho (\mathbf{A})^{r}$$ for all $$t\ge\rho(\mathbf{A})$$.

### Theorem 2.7

(Rouché)

Let $$g(z)=z^{r}$$ and $$h(z)=\sum_{k=0}^{r-1}a_{k}z^{k}$$ be complex-valued polynomials such that $$g(R)>\sum_{k=0}^{r-1}|a_{k}|R^{k}$$ for a real number $$R >0$$, then the polynomial $$f(z):=g(z)-h(z)$$ has all its roots lying strictly inside the circle $$|z|= R$$.

### Proof

This follows immediately from a more general theorem of Rouché, a proof of which can be found in Titchmarsh . □

### Theorem 2.8

(Bolzano’s intermediate value )

If a continuous real function defined on an interval is sometimes positive and sometimes negative, then it must be 0 at some point in the interval.

### Theorem 2.9

(Descartes’ rule of signs)

Let $$f(z):=\sum_{k=0}^{r}a_{k}z^{k}$$ be an rth degree polynomial over the real numbers $$a_{k}$$. Then the number of positive real roots of f is bounded above by the number of sign changes of the coefficients $$a_{k}$$ as one proceeds from $$k=0$$ to $$k=r$$ (ignoring zero coefficients).

### Proof

See for instance . □

### Proposition 2.10

Let $$p(z)=z^{r}-\sum_{k=0}^{r-1}c_{k}z^{k}$$, where $$c_{k}\geq0$$, be a polynomial.
1. (i)

If p is stable, then $$0< p(1)\leq1$$ and there exists $$\lambda\in [0,1)$$, which is unique and positive if $$c_{0}\neq0$$, such that $$p(\lambda)=0$$.

2. (ii)

If p is tamely unstable, then $$p(1)=0$$ and 1 is the only positive root of p.

3. (iii)

If p is wildly unstable, then $$p(1)< 0$$ and there exists a unique positive $$\lambda>1$$ such that $$p(\lambda)=0$$.

### Proof

First of all, that $$p(1)\leq1$$ follows since $$c_{k}\geq0$$.

For (i): Suppose to the contrary that $$p(1)\leq0$$. Now we note that for real numbers t, we have $$\lim_{t\to\infty}p(t)=\infty$$ and as such there exists $$t_{1}\geq1$$ such that $$p(t_{1})\geq0$$. Given that p is a continuous function (on the whole of the real line), then by Bolzano’s intermediate value theorem (Theorem 2.8), there exists $$t_{0}\in[1,t_{1}]$$ such that $$p(t_{0})=0$$, contradicting the fact that by assumption we should rather have $$t_{0}<1$$. Finally since $$p(0)=-c_{0}\leq0$$ and $$p(1)>0$$ then (by Bolzano’s intermediate value theorem again) there exists $$\lambda\in[0,1)$$ such that $$p(\lambda )=0$$. The uniqueness of λ when $$c_{0}\neq0$$, in which case λ cannot be 0, follows from Descartes’ rule of signs (Theorem 2.9).

For (ii): If $$1-p(1)=\sum_{k=0}^{r-1}c_{k} < 1$$, then from Rouché’s theorem (Theorem 2.7) all roots of $$p(z)$$ would be strictly less than 1 in absolute value; hence $$p(1)\leq0$$. Now suppose to the contrary that $$p(1) < 0$$; hence there exists $$t_{1}> 1$$ such that $$p(t_{1})\geq0$$ and so by Bolzano’s intermediate value theorem there exists $$t_{0}\in(1,t_{1}]$$ such that $$p(t_{0})=0$$, contradicting the fact that by assumption we should rather have $$t_{0}\leq1$$; thus $$p(1)=0$$. That 1 is the only positive root follows from Descartes’ rule of signs.

For (iii): Supposing to the contrary that $$p(1)\in[0,1]$$, then we would have $$1-p(1)\in[0,1]$$. But that would imply from Rouché’s theorem that all roots of $$p(z)$$ are at most 1 in absolute value, which is a contradiction. Finally, since $$p(1)<0$$ and $$\lim_{t\to \infty}p(t)=\infty$$, by Bolzano’s intermediate value theorem, there exists $$\lambda>1$$ such that $$p(\lambda)=0$$, the uniqueness of which follows from Descartes’ rule of signs. □

### Corollary 2.11

Let T be a second-order Lipschitz mapping on $$(\mathcal{X},d)$$. Then the inequality (1) takes the form
$$d\bigl(T^{2}y,T^{2}x\bigr)\leq\bigl(\lambda- \lambda'\bigr)d(Ty,Tx)+\lambda\lambda' d(y,x),$$
where $$0\leq\lambda'\leq\lambda$$.

### Proof

By Proposition 2.10, one root of the polynomial $$p(z)=z^{2}-c_{1}z-c_{0}$$ is real and non-negative, λ say; hence given that $$c_{0}\geq0$$ then the other root must be real and non-positive, $$-\lambda'$$ say, where $$\lambda'\geq0$$. We can therefore factor $$p(z)$$ as $$p(z)=(z-\lambda)(z+\lambda')=z^{2}-(\lambda -\lambda')z-\lambda\lambda'$$ and given that $$c_{1}=\lambda-\lambda'\geq 0$$, the conclusion follows. □

### Lemma 2.12

Let $$p(z):=z^{r}-\sum_{k=0}^{r-1}c_{k}z^{k}$$, where $$c_{k}\geq0$$, be a polynomial with unique positive root λ as given by Proposition  2.10. Then
1. 1.

λ dominates all other roots of $$p(z)$$; furthermore, λ is a uniquely dominating root if $$p(z)$$ is non-degenerate.

2. 2.

$$\lambda\in[1-p(1),(1-p(1))^{1/r}]$$, $$\lambda=1$$ and $$\lambda\in (1,1-p(1)]$$ if p is stable, tamely unstable and wildly unstable, respectively.

### Proof

We observe that the (companion) non-negative matrix
$$\mathbf{C}:= \begin{pmatrix} 0&1;&0;&\ldots&0;\\ 0&0;&1;&\ldots&0;\\ \vdots&\vdots&\vdots&\ddots&\vdots\\ 0&0;&0;&\ldots&1;\\ c_{0}&c_;{1}&c_;{2}&\ldots&c_;{r-1} \end{pmatrix}$$
(5)
has the characteristic polynomial $$p(z):=z^{r}-\sum_{k=0}^{r-1}c_{k}z^{k}$$. Now let $$I_{j}(n)$$ be the rth-order impulse-response sequence with characteristic polynomial $$p(z)$$ and initial values $$I_{j}(k)=\delta _{jk}$$ for all $$0\leq j,k\leq r-1$$. If we define the column vector $$\mathbf{v}(n):=[I_{j}(n),I_{j}(n+1),\ldots,I_{j}(n+r-1)]^{\mathbf{T}}$$ then $$\mathbf{v}(n+1)=\mathbf{C}\mathbf{v}(n)$$ and since $$\mathbf{v}(0)=[\delta_{j,0},\delta_{j,1},\ldots,\delta_{j,r-1}]^{\mathbf{T}}$$, then it follows by induction that
$$I_{j}(n+k)=\bigl(\mathbf{v}(n)\bigr)_{k+1}=\bigl(\mathbf{C} \mathbf{v}(n-1)\bigr)_{k+1}=\cdots =\bigl(\mathbf{C}^{n} \mathbf{v}(0)\bigr)_{k+1}=\bigl(\mathbf{C}^{n} \bigr)_{k+1,j+1}$$
for all $$n\geq0$$ and $$0\leq j,k\leq r-1$$. Given that $$\mathbf {C}^{n}\geq0$$ for all $$n\geq0$$, then C would be irreducible if for any given index $$j \in[0,r-1]$$ there exists a sufficiently large natural number n such that $$I_{j}(n)>0$$; but if this were not the case, then $$I_{j}(n)=0$$ for all sufficiently large n and so by the Skolem-Mahler-Lech theorem (Theorem 2.3) there would exist a natural m such that $$I_{j}(k+lm)=0$$ for all $$0\leq k\leq m-1$$ and all $$l\geq0$$, contradicting the fact that $$I_{j}(j)=1\neq0$$. Similarly, when $$I_{j}(n)$$ is non-degenerate, that is, when $$p(z)$$ is a non-degenerate polynomial, then by the Skolem-Mahler-Lech theorem, there would exist a sufficiently large natural number $$n_{j}$$ such that $$I_{j}(n)>0$$ for all $$n\geq n_{j}$$; consequently, the matrix $$\mathbf{C}^{\max_{j}\{n_{j}\}}$$ would be positive and so it follows that C is primitive when $$p(z)$$ is non-degenerate. This gives the first part of the lemma.

Now from Proposition 2.10, λ is the unique positive root of $$p(z)$$; consequently it follows from the first part of the Perron-Frobenius theorem (Theorem 2.5) that $$\rho(\mathbf {C})=\lambda$$ and the conclusion of the first part of the lemma follows immediately. Finally we observe that $$\min_{i}\sum_{j}\mathbf {C}_{ij}=\min\{1,\sum_{k=0}^{r-1}c_{k}\}=\min\{1,1-p(1)\}$$ and similarly we have $$\max_{i}\sum_{j}\mathbf{C}_{ij}=\max\{1,\sum_{k=0}^{r-1}c_{k}\}=\max \{1,1-p(1)\}$$; but by Proposition 2.10, we have $$p(1)>0$$, $$p(1)=0$$ and $$p(1)<0$$ if T is rth-order contraction, non-expansive and expansive, respectively; hence given that $$\rho (\mathbf{C})=\lambda$$, then from the second part of the Perron-Frobenius theorem plus the fact that $$\lambda^{r}\le1-p(1)$$ when $$\lambda\le1$$ (the Keilson-Styan inequality, Theorem 2.6 with $$t=1$$), the second part of the lemma follows. □

### Proposition 2.13

For all $$\lambda\in[0,1)$$, $$\mu\geq1$$ and integers $$m\geq0$$,
$$\sum_{k=0}^{\infty}\lambda^{m+k}(m+k)^{\mu-1}< L\lambda^{m}(m+1)^{\mu-1}$$
for some absolute constant L dependent only on λ, μ.

### Proof

First note that the polylogarithm function $$Li_{1-\mu}(\lambda):=\sum_{k=1}^{\infty}\lambda^{k}k^{\mu-1}$$ converges (via Cauchy’s root test, say) for all $$\lambda\in[0,1)$$. Now if $$m\in\{0,1\}$$ then
$$\sum_{k=0}^{\infty}\lambda^{m+k}(m+k)^{\mu-1}= \sum_{k=1}^{\infty}\lambda ^{k}k^{\mu-1}=Li_{1-\mu}( \lambda),$$
from which the proposition follows in that case by setting, for instance, $$L:=(1+Li_{1-\mu}(\lambda))\max\{1,2^{1-\mu}/\lambda\}$$. Now when $$m\geq2$$ and using the fact that $$mk\geq m+k$$ when also $$k\geq2$$, then we have
\begin{aligned} \sum_{k=0}^{\infty}\lambda^{m+k}(m+k)^{\mu-1} =& \lambda^{m}\bigl(m^{\mu-1}+\lambda (m+1)^{\mu-1}\bigr)+ \sum_{k=2}^{\infty}\lambda^{m+k}(m+k)^{\mu-1} \\ \leq&\lambda^{m}\bigl(m^{\mu-1}+\lambda(m+1)^{\mu-1} \bigr)+\sum_{k=2}^{\infty}\lambda ^{m+k}(mk)^{\mu-1} \\ =&\lambda^{m}\bigl(m^{\mu-1}+\lambda(m+1)^{\mu-1}\bigr)+ \lambda^{m}m^{\mu-1}\sum_{k=2}^{\infty}\lambda^{k}k^{\mu-1} \\ =&\lambda^{m}\bigl(m^{\mu-1}+\lambda(m+1)^{\mu-1}\bigr)+ \lambda^{m}m^{\mu-1}\bigl(-\lambda +Li_{1-\mu}(\lambda) \bigr) \\ < &\lambda^{m}(m+1)^{\mu-1}\bigl(1+\lambda-\lambda+Li_{1-\mu}( \lambda)\bigr) \\ =&\lambda^{m}(m+1)^{\mu-1}\bigl(1+Li_{1-\mu}(\lambda) \bigr). \end{aligned}
Setting $$L:=(1+Li_{1-\mu}(\lambda))\max\{1,2^{1-\mu}/\lambda\}$$ as before, the proposition follows. □

### Proposition 2.14

Let $$T:\mathcal{X}\to\mathcal{Y}$$ be a mapping between metric spaces $$(\mathcal{X},d_{\mathcal{X}})$$ and $$(\mathcal {Y},d_{\mathcal{Y}})$$. Then T is continuous at $$x^{*}\in\mathcal{X}$$ if and only if for every sequence $$\{x_{n}\}_{n\geq0}$$ converging to $$x^{*}$$ we find that the sequence $$\{Tx_{n}\}_{n\geq0}$$ is Cauchy.

### Proof

Indeed if T is continuous at $$x^{*}$$, then the conclusion of the proposition holds true, hence it suffices for us to show that $$\lim_{n\to\infty}Tx_{n}=Tx^{*}$$ for every sequence $$\{x_{n}\}_{n\geq0}$$ converging to $$x^{*}$$. Suppose to the contrary that for some sequence $$\{x_{n}\}_{n\geq0}$$ converging to $$x^{*}$$ we have $$\lim_{n\to\infty }Tx_{n}\neq Tx^{*}$$ and we define a sequence $$\{u_{n}\}_{n\geq0}$$ by $$u_{2n}=x^{*}$$ and $$u_{2n+1}=x_{n}$$. Clearly $$u_{n}$$ converges to $$x^{*}$$ but since we have
$$\lim_{n\to\infty}d_{\mathcal{Y}}(Tu_{2n+1},Tu_{2n})= \lim_{n\to\infty }d_{\mathcal{Y}}\bigl(Tx_{n},Tx^{*}\bigr)\neq0$$
we find that $$\{Tu_{n}\}_{n\geq0}$$ is not Cauchy, which is a contradiction. Hence $$\lim_{n\to\infty}Tx_{n}=Tx^{*}$$ for any sequence $$\{x_{n}\}_{n\geq0}$$ converging to $$x^{*}$$ and so T is continuous at $$x^{*}$$. □

## 3 Main result

We now prove our main results. We begin with a direct proof of the fixed point theorem for higher-order contraction mappings; thereafter, we provide a re-metrisation argument that relates higher-order Lipschitz mappings to (first-order) Lipschitz mappings. This re-metrisation of the original metric space does not necessarily result in a complete metric space even if the former is complete; consequently, we shall need a completion of the re-metrised space and an extension of the higher-order Lipschitz mapping into the complete re-metrised space.

### 3.1 Higher-order contraction mappings

The main result of this subsection is as follows, which extends the conclusion of the Banach fixed point theorem (Theorem 1.2) to higher-order contraction mappings:

### Theorem 3.1

(Higher-order contraction mapping theorem)

Let $$(\mathcal{X},d)$$ be a complete metric space and let $$T:\mathcal{X}\to\mathcal{X}$$ be an rth-order contraction mapping. Then T has a unique fixed point and $$\lim_{n\to\infty}T^{n}x$$ converges to this fixed point for arbitrary $$x\in\mathcal{X}$$.

We accomplish the proof below, but we require the following auxiliary lemma.

### Lemma 3.2

The sequence $$\{T^{n}x\}_{n\geq0}$$ is Cauchy for all $$x\in\mathcal{X}$$; moreover, $$\lim_{n\to\infty}T^{n}y=\lim_{n\to\infty}T^{n}x$$ for all $$x,y\in\mathcal{X}$$.

### Proof

For all $$x:=x_{0},y:=y_{0}\in\mathcal{X}$$ let $$x_{n}:=T^{n}x$$ and $$y_{n}:=T^{n}y$$. Then from the inequality (1), we have
$$d(y_{m+r},x_{m+r})\leq\sum _{k=0}^{r-1}c_{k}d(y_{m+k},x_{m+k})$$
(6)
for all $$m\geq0$$. Now we prove by induction that
$$d(y_{n},x_{n})\leq\sum _{j=0}^{r-1}I_{j}(n)d(y_{j},x_{j}),$$
(7)
where $$I_{j}(n)$$ satisfies $$I_{j}(n+r)=\sum_{k=0}^{r-1}c_{k}I_{j}(n+k)$$ and $$I_{j}(k)=\delta_{jk}$$ for all $$0\leq k\leq r-1$$ (thus, $$I_{j}(n)$$ is an impulse-response sequence). Indeed inequality (7) already holds with equality when $$0\leq n:=k\leq r-1$$, which serves as our base cases for the induction; thus for some non-negative integer m, suppose it holds for $$n=m,m+1,\ldots,m+r-1$$. Then from inequality (6) we have
\begin{aligned} d(y_{m+r},x_{m+r}) \leq&\sum_{k=0}^{r-1}c_{k}d(y_{m+k},x_{m+k}) \\ \leq&\sum_{k=0}^{r-1}c_{k}\sum _{j=0}^{r-1}I_{j}(m+k)d(y_{j},x_{j}) \\ =&\sum_{j=0}^{r-1}d(y_{j},x_{j}) \sum_{k=0}^{r-1}c_{k}I_{j}(m+k) \\ =&\sum_{j=0}^{r-1}d(y_{j},x_{j})I_{j}(m+r). \end{aligned}
Thus inequality (7) holds for $$n=m+r$$ as well and so it holds for all $$n\geq0$$. From Corollary 2.2, $$I_{j}(n)\to 0$$ as $$n\to\infty$$ and so from inequality (7) $$d(y_{n},x_{n})\to0$$ as $$n\to\infty$$; hence from the continuity of d, we have $$d(\lim_{n\to\infty}y_{n},\lim_{n\to\infty}x_{n})=0$$ and thus $$\lim_{n\to\infty}T^{n}y=\lim_{n\to\infty}T^{n}x$$, assuming the limit exists, which we show next.
Henceforward we make the substitution $$y=Tx$$, thus $$x_{n+1}=Tx_{n}$$. We show that $$\{x_{n}\}_{n\geq0}$$ is Cauchy but this is trivial if $$x_{1}=x_{0}$$, that is, if $$x_{0}$$ is a fixed point; hence we assume that $$x_{1}\neq x_{0}$$. Now let $$n>m$$ and letting λ and μ denote, respectively, the maximum in absolute value and multiplicity of the roots of the polynomial $$p(z)=z^{r}-\sum_{k=0}^{r-1}c_{k}z^{k}$$, then via the triangle inequality of d, inequality (7), Corollary 2.2 and Proposition 2.13, we have
\begin{aligned} d(x_{n},x_{m}) \leq&\sum_{k=0}^{n-m-1}d(x_{m+k+1},x_{m+k}) \\ \leq&\sum_{k=0}^{n-m-1}\sum _{j=0}^{r-1}I_{j}(m+k)d(x_{j+1},x_{j}) \\ =&\sum_{j=0}^{r-1}d(x_{j+1},x_{j}) \sum_{k=0}^{n-m-1}I_{j}(m+k) \\ \leq& K\sum_{j=0}^{r-1}d(x_{j+1},x_{j}) \sum_{k=0}^{n-m-1}\lambda ^{m+k}(m+k)^{\mu-1} \\ < &K;\sum_{j=0}^{r-1}d(x_{j+1},x_{j}) \sum_{k=0}^{\infty}\lambda ^{m+k}(m+k)^{\mu-1} \\ < & KL\lambda^{m}(m+1)^{\mu-1}\sum_{j=0}^{r-1}d(x_{j+1},x_{j}). \end{aligned}
Note that $$\sum_{j=0}^{r-1}d(x_{j+1},x_{j})\neq0$$ since by assumption $$x_{1}\neq x_{0}$$. Now given that $$0\leq\lambda<1$$, it follows that $$\lambda^{m}(m+1)^{\mu-1}\to0$$ as $$m\to\infty$$, hence for any arbitrary $$\varepsilon>0$$, we can find $$N(\varepsilon)$$ large enough such that
$$\lambda^{m}(m+1)^{\mu-1}< \frac{\varepsilon}{KL\sum_{j=0}^{r-1}d(x_{j+1},x_{j})} \quad\forall m\geq N( \varepsilon),$$
which implies that $$d(x_{n},x_{m})<\varepsilon$$ for all $$n>m\geq N(\varepsilon)$$ and thus the sequence $$\{x_{n}\}_{n\geq0}$$ is indeed Cauchy. □

### Proof of Theorem 3.1

Indeed by Lemma 3.2 the sequence $$\{x_{n}:=T^{n}x\}_{n\geq0}$$ is Cauchy for any arbitrary $$x\in\mathcal{X}$$ and is therefore convergent, say $$\lim_{n\to\infty}T^{n}x= \lim_{n\to\infty}Tx_{n}= x^{*}\in \mathcal{X}$$. Consequently $$\lim_{n\to\infty}T^{n}x^{*}=x^{*}$$ and so the set $$S(x^{*}):=\{T^{n}x^{*}\}_{n\geq0}$$ is a closed subset of $$\mathcal {X}$$; indeed $$(S(x^{*}),d)$$ is a complete metric subspace of $$\mathcal {X}$$ such that $$T(S(x^{*}))\subseteq S(x^{*})$$. But for every sequence $$\{s_{n}\}_{n\geq0}\subseteq S(x^{*})$$ convergent to $$x^{*}$$, it follows from Lemma 3.2 that the sequence $$\{Ts_{n}\}_{n\geq0}$$ is Cauchy; hence given that $$x^{*}\in S(x^{*})$$, then by Proposition 2.14 it follows that T is continuous at $$x^{*}$$ in $$S(x^{*})$$ and consequently $$Tx^{*}=x^{*}$$. Now to see that $$x^{*}$$ is a unique fixed point, observe that if $$Ty^{*}=y^{*}$$ for some $$y^{*}\neq x^{*}$$ then
\begin{aligned} 0 < &d;\bigl(y^{*},x^{*}\bigr)=d\bigl(T^{r}y^{*},T^{r}x^{*}\bigr) \\ \leq&\sum_{k=0}^{r-1}c_{k}d \bigl(T^{k}y^{*},T^{k}x^{*}\bigr) \\ =&\sum_{k=0}^{r-1}c_{k}d \bigl(y^{*},x^{*}\bigr) \\ =&\bigl(1-p(1)\bigr)d\bigl(y^{*},x^{*}\bigr). \end{aligned}
But by Proposition 2.10, we have $$0< p(1)\leq1$$, which leads to the contradiction that $$0< d(y^{*},x^{*})\leq (1-p(1))d(y^{*},x^{*})< d(y^{*},x^{*})$$. Equivalently, in a more straightforward fashion, if $$y^{*}\neq x^{*}$$ is also a fixed point, then we get the contradiction (via Lemma 3.2) that $$y^{*}=\lim_{n\to \infty}T^{n}y^{*}=\lim_{n\to\infty}T^{n}x^{*}=x^{*}$$. This completes the proof of Theorem 3.1. □

### 3.2 The general case

Now we consider the general case of higher-order Lipschitz mappings. First of all, it is worthy of note the theorem of Bessaga  states that whenever $$\mathcal{X}$$ is an arbitrary set with a self-map T satisfying the property that each iterate $$T^{n}$$ has a unique fixed point, then for each $$c\in(0,1)$$, there exists a metric $$d_{c}$$ on $$\mathcal{X}$$ such that $$(\mathcal{X},d_{c})$$ is a complete metric space and T is a contraction mapping on $$(\mathcal {X},d_{c})$$. Thus in light of Theorem 3.1 demonstrated in the previous subsection, we are motivated to consider a re-metrisation of the space $$(\mathcal{X},d)$$ over which a higher-order Lipschitz mapping is defined.

Now let T be an rth-order Lipschitz mapping on a complete metric space $$(\mathcal{X},d)$$ as given in (1) and let λ be the unique positive root of the polynomial $$p(z)=z^{r}-\sum_{k=0}^{r-1}c_{k}z^{k}$$ as guaranteed by Proposition 2.10, in particular we assume that $$p(0)\ne0$$. Define a new metric on the space $$\mathcal{X}$$ as follows:
$$D(y,x)=\sum_{k=0}^{r-1}b_{k}d \bigl(T^{k}y,T^{k}x\bigr) , \quad\mbox{where }b_{k}=\sum _{j=0}^{k}c_{j} \lambda^{j-k-1}.$$
(8)
That D is a metric on $$\mathcal{X}$$ is straightforward. Indeed, D is non-negative since d and $$b_{k}$$ are non-negative and it is sub-additive since d is sub-additive; furthermore, $$D(y,x)=0$$ if and only if $$y=x$$ since $$b_{k}\ne0$$ (because, by assumption, $$c_{0}\ne0$$) and finally $$D(y,x)=D(x,y)$$.

Now we have the following lemma.

### Lemma 3.3

Let $$(\mathcal{X},d)$$ be a (not necessarily complete) metric space and let $$T:\mathcal{X}\to\mathcal{X}$$ be an rth-order Lipschitz mapping. Let D be the new metric defined in (8). Then
$$D(Ty,Tx)\le\lambda D(y,x).$$
Moreover, a sequence $$\{x_{n}\}_{n\ge1}\subset(\mathcal{X},D)$$ is Cauchy in $$(\mathcal{X},D)$$ if and only if the sequence $$\{T^{k}x_{n}\} _{n\ge1}\subset(\mathcal{X},d)$$ is Cauchy in $$(\mathcal{X},d)$$ for all $$0\le k\le r-1$$.

First, we prove the following recurrence relation for the constants $$b_{k}$$ in (8).

### Proposition 3.4

Let $$b_{k}$$ be as defined in (8). Then
\begin{aligned}& b_{0}=\lambda^{-1}c_{0} ,\quad\quad b_{r-1}=1, \\& b_{k}=\lambda^{-1}(b_{k-1}+c_{k}) ,\quad 1 \le k\le r-1. \end{aligned}

### Proof

Obviously, $$b_{0}=\sum_{j=0}^{0}c_{j}\lambda^{j-k-1}=c_{0}\lambda^{-1}$$. Then also
$$b_{r-1}=\sum_{j=0}^{r-1}c_{j} \lambda^{j-r}=\lambda^{-r}\sum_{j=0}^{r-1}c_{j} \lambda^{j}= \lambda^{-r}\bigl( \lambda^{r}-p( \lambda)\bigr)=1.$$
Finally,
$$b_{k}=\sum_{j=0}^{k}c_{j} \lambda^{j-k-1}=\lambda^{-1}\sum_{j=0}^{k-1}c_{j} \lambda^{j-k}+c_{k}\lambda^{-1}=\lambda^{-1}(b_{k-1}+c_{k}),$$
which completes the proof. □

### Proof of Lemma 3.3

Via Proposition 3.4, we have
\begin{aligned} D(Ty,Tx) =&\sum_{k=0}^{r-1}b_{k}d \bigl(T^{k+1}y,T^{k+1}x\bigr) \\ =&\sum_{k=1}^{r}b_{k-1}d \bigl(T^{k}y,T^{k}x\bigr) \\ =&b_;{r-1}d\bigl(T^{r}y,T^{r}x\bigr)+\sum _{k=1}^{r-1}b_{k-1}d\bigl(T^{k}y,T^{k}x \bigr) \\ \le&\sum_{k=0}^{r-1}c_{k}d \bigl(T^{k}y,T^{k}x\bigr)+\sum_{k=1}^{r-1}b_{k-1}d \bigl(T^{k}y,T^{k}x\bigr) \\ =&c_;{0}d(y,x)+\sum_{k=1}^{r-1}(b_{k-1}+c_{k})d \bigl(T^{k}y,T^{k}x\bigr) \\ =&\lambda b_{0}d(y,x)+\sum_{k=1}^{r-1} \lambda b_{k}d\bigl(T^{k}y,T^{k}x\bigr) \\ =&\lambda\sum_{k=0}^{r-1}b_{k}d \bigl(T^{k}y,T^{k}x\bigr) \\ =&\lambda D(y,x). \end{aligned}
Finally, since $$b_{k}\ne0$$ for all $$0\le k\le r-1$$, then we note from (8) that if $$\lim_{n\ge m\to\infty}D(x_{n},x_{m})=0$$, then likewise $$\lim_{n\ge m\to\infty}d(T^{k}x_{n},T^{k}x_{m})=0$$ for all $$0\le k\le r-1$$; similarly, if $$\lim_{n\ge m\to\infty }d(T^{k}x_{n},T^{k}x_{m})=0$$ for all $$0\le k\le r-1$$, then likewise $$\lim_{n\ge m\to\infty}D(x_{n},x_{m})=0$$, which completes the proof. □
We note in the first place that Lemma 3.3 does not imply that T is uniformly continuous (or even continuous) in $$(\mathcal{X},d)$$ as was noted in the introduction; rather, T is Lipschitz continuous (and therefore uniformly continuous) in $$(\mathcal {X},D)$$. Secondly, when $$\lambda<1$$, then the Banach fixed point theorem (Theorem 1.2) cannot be applied to assert that T has a fixed point in $$(\mathcal{X},D)$$ unless T is continuous in $$(\mathcal{X},d)$$; however, the following theorem remedies the case when T is discontinuous on $$(\mathcal{X},d)$$. To proceed, let $$(\overline{\mathcal{X}},\overline{D})$$ be the canonical completion of the metric space $$(\mathcal{X},D)$$; that is,
$$\overline{D}\bigl([y_{n}],[x_{n}]\bigr):=\lim _{n\to\infty}D(y_{n},x_{n}),$$
where $$\{y_{n}\}_{n\ge1}$$, $$\{x_{n}\}_{n\ge1}$$ are Cauchy sequences in $$(\mathcal{X},D)$$ and $$[x_{n}]$$ denotes the equivalence class of $$\{ x_{n}\}_{n\ge1}$$ in $$(\mathcal{X},D)$$, where $$\{y_{n}\}_{n\ge1}$$ is equivalent to $$\{x_{n}\}_{n\ge1}$$ if $$\lim_{n\to\infty}D(y_{n},x_{n})=0$$.

### Theorem 3.5

Define the mapping,
$$\overline{T}:\overline{\mathcal{X}}\to\overline{\mathcal {X}},\quad\quad [x_{n}] \mapsto[Tx_{n}].$$
Then we have
$$\overline{D}\bigl(\overline{T}[y_{n}],\overline{T}[x_{n}] \bigr)\le\lambda\overline {D}\bigl([y_{n}],[x_{n}]\bigr).$$
In particular, if $$(\mathcal{X},d)$$ is complete, then T has a fixed point in $$(\mathcal{X},d)$$ if and only if $$\overline{T}$$ has a fixed point in $$(\overline{\mathcal{X}},\overline{D})$$.

### Proof

Since $$\{x_{n}\}_{n\ge1}$$ is Cauchy in $$(\mathcal{X},D)$$ then, by Lemma 3.3,
$$D(Tx_{n},Tx_{m})\le\lambda D(x_{n},x_{m})$$
and so $$\{Tx_{n}\}_{n\ge1}$$ is Cauchy in $$(\mathcal{X},D)$$; thus $$\overline{T}$$ is well defined. Now given Cauchy sequences $$\{y_{n}\} _{n\ge1}$$, $$\{x_{n}\}_{n\ge1}$$ in $$(\mathcal{X},D)$$, then we have
\begin{aligned} \overline{D}\bigl(\overline{T}[y_{n}],\overline{T}[x_{n}] \bigr) =&\overline {D}\bigl([Ty_{n}],[Tx_{n}]\bigr) \\ =&\lim_{n\to\infty}D(Ty_{n},Tx_{n}) \\ \le&\lambda\lim_{n\to\infty}D(y_{n},x_{n}) \\ =&\lambda\overline{D}\bigl([y_{n}],[x_{n}]\bigr). \end{aligned}
Finally, if $$x=Tx$$ in $$(\mathcal{X},d)$$, then let $$[x]$$ be the equivalence class of the constant sequence $$\{x,x,x,\ldots\}\in (\overline{\mathcal{X}},\overline{D})$$. Then
$$\overline{T}[x]=[Tx]=[x] .$$
On the other hand, if $$[x_{n}]=\overline{T}[x_{n}]=[Tx_{n}]$$ in $$(\overline {\mathcal{X}},\overline{D})$$, then by Lemma 3.3, T is continuous at $$x:=\lim_{n\to\infty}x_{n}$$ in $$(\mathcal{X},d)$$, hence
$$Tx=\lim_{n\to\infty}Tx_{n}=\lim_{n\to\infty}x_{n}=x,$$
which completes the proof. □

## 4 Local Lipschitzity

If $$r>1$$, a natural question that arises is whether there are pairs $$x_{0},y_{0}\in\mathcal{X}$$ and a constant $$c>0$$ such that $$d(T^{n}y_{0},T^{n}x_{0})\leq c^{n}d(y_{0},x_{0})$$ for all $$n\geq0$$; such a pair is therefore Lipschitzian under T, in the sense that this would be the case if T were a Lipschitz mapping with Lipschitz constant c. In this section, motivated by the existence of the unique positive real root λ of the polynomial $$p(z)$$ (in Definition 1.1) when $$p(0)\neq0$$, we show that near-local Lipschitzity of open subsets exists with respect to an arbitrary point: that is, given $$x_{0}\in\mathcal{X}$$ there exists an open subset $$\mathcal{S}\subset\mathcal{X}$$ and positive real number $$m_{0}\geq1$$ such that $$d(T^{n}x,T^{n}x_{0})\leq m_{0}\lambda^{n}d(x,x_{0})$$ for all $$x\in\mathcal{S}$$. We also prove the closely related result that there exists an open subset $$\mathcal{S}$$ of $$\mathcal{X}^{2}$$ and real number $$m_{0}\geq1$$ such that $$d(T^{n}y,T^{n}x)\leq m_{0}\lambda ^{n}d(y,x)$$ for all $$(x,y)\in\mathcal{S}$$. Though plausible, we are not able to determine whether or when local Lipschitzity can occur (non-trivially) in either case, that is, whether or when $$m_{0}$$ can take the value 1.

Unless otherwise mentioned, we assume $$p(0)\neq0$$ throughout the remaining part of this subsection.

### Proposition 4.1

Let $$(\mathcal{X},d)$$ be a complete metric space and let T be an rth-order Lipschitz mapping on $$\mathcal{X}$$. For every pair $$x\neq y\in\mathcal{X}$$ define
$$M:=M(y,x)=\max_{0\leq k\leq r-1}\lambda^{-k}\frac{d(T^{k}y,T^{k}x)}{d(y,x)} .$$
Then
$$M=\max_{n\geq0}\lambda^{-n}\frac{d(T^{n}y,T^{n}x)}{d(y,x)}.$$

### Proof

We prove by induction; by definition the conclusion holds up to $$n=r-1$$, hence assuming the conclusion holds up to $$n+r-1$$, we have
\begin{aligned} d\bigl(T^{n+r}y,T^{n+r}x\bigr) \leq&\sum _{k=0}^{r-1}c_{k}d\bigl(T^{n+k}y,T^{n+k}x \bigr) \\ \leq& M\sum_{k=0}^{r-1}c_{k} \lambda^{n+k} d(Ty,Tx) \\ =&M;\lambda^{n}d(Ty,Tx)\sum_{k=0}^{r-1}c_{k} \lambda^{k} \\ =&M;\lambda^{n}d(Ty,Tx) \bigl(\lambda^{r}-p(\lambda)\bigr) \\ =&M;\lambda^{n+r}d(Ty,Tx), \end{aligned}
and so the conclusion holds for $$n+r$$ as well and thus it holds for all n. □

### Remark

Obviously, $$M(y,x)\geq1$$ for every $$x,y\in\mathcal {X}$$ so if $$M(y,x)$$ can be 1 for some pair x, y then T is essentially locally Lipschitz on the pair x, y. We are not able to establish whether or when local Lipschitzity can occur, but the next results give near misses.

### Theorem 4.2

Let T be an rth-order Lipschitz mapping on a complete metric space $$(\mathcal{X},d)$$. Then for all $$x,y\in\mathcal{X}$$
$$\mathop{\min_{1\leq i\leq r}}_{d(T^{i-1}y,T^{i-1}x)\neq0} \frac {d(T^{i}y,T^{i}x)}{d(T^{i-1}y,T^{i-1}x)}\leq\lambda.$$
In particular, $$\min_{x\neq y}\frac{d(Ty,Tx)}{d(y,x)}\leq\lambda$$.

### Proof

Define the column vector $$\mathbf{v}:=[d(y,x),\dots ,d(T^{r-1}y,T^{r-1}x)]^{\mathbf{T}}$$ and let C be the companion non-negative $$r\times r$$ matrix defined in (5). Then
$$\mathbf{C}\mathbf{v}= \begin{pmatrix} d(Ty,Tx)\\ d(T^{2}y,T^{2}x)\\ \vdots\\ \sum_{k=0}^{r-1}c_{k}d(T^{r}y,T^{r}x) \end{pmatrix}.$$
Now by definition of T in Definition 1.1 we have $$(\mathbf{C}\mathbf{v})_{r}\geq d(T^{r}y,T^{r}x)$$, consequently we have the inequality
$$\mathop{\min_{1\leq i\leq r}}_{v_{i}\neq0}\frac{(\mathbf{C}\mathbf {v})_{i}}{\mathbf{v}_{i}}\geq \mathop{\min _{1\leq i\leq r}}_{ d(T^{i-1}y,T^{i-1}x)\neq0} \frac{d(T^{i}y,T^{i}x)}{d(T^{i-1}y,T^{i-1}x)} .$$
(9)
But by the Collatz-Wielandt formula of the Perron-Frobenius theorem (Theorem 2.5), the left-hand side of (9) attains the maximum value of $$\rho(\mathbf{C})$$, which equals λ from the proof of Lemma 2.12. The conclusion of the theorem then follows immediately. □

### Remark

We note that if the inequality in Theorem 4.2 is uniform for all $$1\leq i\leq r$$ for some $$x,y\in \mathcal{X}$$, then the maximum bound $$M(x,y)$$ as defined in Proposition 4.1 would be exactly equal to 1.

### Theorem 4.3

Let T be a continuous higher-order Lipschitz mapping on a complete metric space $$( \mathcal{X},d )$$. Then for every $$x_{0}\in\mathcal{X}$$ there exist an open set $$\mathcal{S}\subset\mathcal{X}$$ and a natural number $$m_{0}$$ such that
$$d\bigl(T^{n}x,T^{n}x_{0}\bigr)\leq m_{0} \lambda^{n}d(x,x_{0})$$
for all $$x\in\mathcal{S}$$ and $$n\geq0$$.

### Proof

By Proposition 4.1, there exists a bound $$M(x_{0},x)$$ for each $$x,x_{0}\in\mathcal{X}$$ such that $$d(T^{n}x_{0},T^{n}x)\leq M(x_{0},x)\lambda^{n}d(x_{0},x)$$ for all $$n\geq0$$. For all natural numbers m define the sets
$$\mathcal{X}_{m}:=\bigl\{ x\in\mathcal{X}:M(x_{0},x)\leq m \bigr\} .$$
Obviously if $$\{y_{k}\}_{k\geq1}\subset\mathcal{X}$$ is a sequence converging to y such that $$M(x_{0},y_{k})\leq m$$, then via the continuity of the metric d and of T we have
\begin{aligned} d\bigl(T^{n}x_{0},T^{n}y\bigr) =&\lim _{k\to\infty}d\bigl(T^{n}x_{0},T^{n}y_{k} \bigr) \\ \leq&\lim_{k\to\infty}M(x_{0},y_{k}) \lambda^{n}d(x_{0},y_{k}) \\ \leq& m\lim_{k\to\infty}\lambda^{n}d(x_{0},y_{k}) \\ =&m;\lambda^{n}d(x_{0},y) \end{aligned}
and as such the sets $$\mathcal{X}_{m}$$ are all closed. But every $$x\in \mathcal{X}$$ is contained in some $$\mathcal{X}_{m}$$ and so we have $$\mathcal{X}=\bigcup_{m\geq1}\mathcal{X}_{m}$$. Thus by the Baire category theorem (Theorem 2.4), there exist $$m_{0}\geq1$$ and an open set $$S\in\mathcal{X}_{m_{0}}$$ such that $$M(x_{0},x)\leq m_{0}$$ for all $$x\in\mathcal{S}$$. □

### Theorem 4.4

Let T be a continuous higher-order Lipschitz mapping on a complete metric space $$( \mathcal{X},d )$$. Then there exist an open set $$\mathcal{S}\subset\mathcal{X}^{2}$$ and a natural number $$m_{0}$$ such that
$$d\bigl(T^{n}y,T^{n}x\bigr)\leq m_{0} \lambda^{n}d(y,x)$$
for all $$(x,y)\in\mathcal{S}$$ and $$n\geq0$$.

### Proof

The proof is similar to that given for Theorem 4.3. Here we define the sets
$$\mathcal{X}_{m} :=\bigl\{ (x,y)\in\mathcal{X}^{2}:M(y,x)\leq m \bigr\} \quad \forall m\in\{ 1,2,3,\ldots\},$$
where $$M(y,x)$$ is the maximum bound defined in Proposition 4.1. With respect to a chosen metric on $$\mathcal{X}^{2}$$ which agrees with the product topology on $$\mathcal{X}^{2}$$, then $$\mathcal{X}^{2}$$ is also a complete metric space. Furthermore, if $$\{ (x_{k},y_{k})\}_{k\geq1}\subset\mathcal{X}^{2}$$ converges to $$( x,y )$$ say, such that $$M(y_{k},x_{k})\leq m$$ then via the continuity of the metric d and of T we have
\begin{aligned} d\bigl(T^{n}y,T^{n}x\bigr) =&\lim_{k\to\infty}d \bigl(T^{n}y_{k},T^{n}x_{k}\bigr) \\ \leq&\lim_{k\to\infty}M(y_{k},x_{k}) \lambda^{n}d(y_{k},x_{k}) \\ \leq& m\lim_{k\to\infty}\lambda^{n}d(y_{k},x_{k}) \\ =&m;\lambda^{n}d(y,x) \end{aligned}
and so the sets $$\mathcal{X}_{m}$$ are closed. Since every pair $$( x,y)\in\mathcal{X}^{2}$$ is contained in some $$\mathcal{X}_{m}$$, it follows that
$$\mathcal{X}^{2}=\bigcup_{m\geq1} \mathcal{X}_{m}.$$
Thus by the Baire category theorem, there exist $$m_{0}\geq1$$ and an open set $$S\in\mathcal{X}_{m_{0}}$$ such that $$M(y,x)\leq m_{0}$$ for all $$(x,y)\in\mathcal{S}$$. □

Now we have the following problem.

### Local Lipschitzity problem

Suppose T is a non-trivial rth-order Lipschitz mappings (that is, T is not of lower order on either $$T(\mathcal{X})$$ or $$\mathcal{X}$$). Can the constants M and $$m_{0}$$ appearing in Proposition  4.1 and Theorems 4.3 and 4.4 be exactly equal to 1?

The determination of whether or when this can be answered in the affirmative would demonstrate that there are subsets of elements in a complete metric space from which the Picard iterations are sharply convergent to the fixed point of a higher-order contraction mapping on the metric space in question.

## 5 Conclusion

In light of the fact that a higher-order Lipschitz mapping can be considered as a linear system of mapping in the form of the matrix inequality
$$\begin{pmatrix} d(Ty,Tx)\\ d(T^{2}y,T^{2}x)\\ \vdots\\ d(T^{r-1}y,T^{r-1}x)\\ d(T^{r}y,T^{r}x) \end{pmatrix} \le \begin{pmatrix} 0&1;&0;&\ldots&0;\\ 0&0;&1;&\ldots&0;\\ \vdots&\vdots&\vdots&\ddots&\vdots\\ 0&0;&0;&\ldots&1;\\ c_{0}&c_;{1}&c_;{2}&\ldots&c_;{r-1} \end{pmatrix} \begin{pmatrix} d(y,x)\\ d(Ty,Tx)\\ \vdots\\ d(T^{r-2}y,T^{r-2}x)\\ d(T^{r-1}y,T^{r-1}x) \end{pmatrix}$$
it is rather natural to consider mappings when the above matrix inequality can be satisfied but with non-negative $$r\times r$$ matrix rather than the companion matrix. We recall that the tools used in the local-Lipschitzity analysis of higher-order Lipschitz mappings (the Perron-Frobenius result in particular) are at our disposal to use when we consider a non-negative matrix instead of just a companion matrix. We therefore propose the following broader kind of mappings.

### Definition 5.1

Let $$(\mathcal{X},d)$$ be a (complete) a metric space. Then a countable collection of self-mappings $$\{T_{k}\}_{k\ge1}$$ on $$\mathcal{X}$$ is said to form an rth-order Lipschitz system of mappings if there exists an $$r\times r$$ non-negative matrix $$[c_{i,j}]_{i,j=1}^{r}$$ such that the following matrix inequality is satisfied:
$$\begin{pmatrix} d(T_{k+1}y,T_{k+1}x)\\ d(T_{k+2}y,T_{k+2}x)\\ \vdots\\ d(T_{k+r}y,T_{k+r}x) \end{pmatrix} \le \begin{pmatrix} c_{11}&c_;{12}&\ldots&c_;{1r}\\ c_{21}&c_;{22}&\ldots&c_;{2r}\\ \vdots&\vdots&\ddots&\vdots\\ c_{r,1}&c_;{r,2}&\ldots&c_;{rr} \end{pmatrix} \begin{pmatrix} d(T_{k}y,T_{k}x)\\ d(T_{k+1}y,T_{k+1}x)\\ \vdots\\ d(T_{k+r-1}y,T_{k+r-1}x) \end{pmatrix}.$$

When $$T_{k}:=T^{k}$$, then (essentially) we achieve an rth-order Lipschitz mapping. One can then similarly inquire as regards the fixed point theory of such Lipschitz system of mappings.

Footnotes
1

Schauder’s theorem is more general than this and relates to continuous self-mappings on compact convex subsets of Banach spaces.

## Declarations

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

## Authors’ Affiliations

(1)
Department of Mathematics, Kwame Nkrumah University of Science and Technology, PMB, Kumasi, Ghana

## References

Advertisement 