Skip to main content

The regularized CQ algorithm without a priori knowledge of operator norm for solving the split feasibility problem

Abstract

The split feasibility problem (SFP) is finding a point \(x\in C\) such that \(Ax\in Q\), where C and Q are nonempty closed convex subsets of Hilbert spaces \(H_{1}\) and \(H_{2}\), and \(A:H_{1}\rightarrow H_{2}\) is a bounded linear operator. Byrne’s CQ algorithm is an effective algorithm to solve the SFP, but it needs to compute \(\|A\|\), and sometimes \(\|A\| \) is difficult to work out. López introduced a choice of stepsize \(\lambda_{n}\), \(\lambda_{n}=\frac{\rho_{n}f(x_{n})}{\|\nabla f(x_{n})\| ^{2}}\), \(0<\rho_{n}<4\). However, he only obtained weak convergence theorems. In order to overcome the drawbacks, in this paper, we first provide a regularized CQ algorithm without computing \(\|A\|\) to find the minimum-norm solution of the SFP and then obtain a strong convergence theorem.

1 Introduction

Let \(H_{1}\) and \(H_{2}\) be real Hilbert spaces and let C and Q be nonempty closed convex subsets of \(H_{1}\) and \(H_{2}\), and let \(A:H_{1}\rightarrow H_{2}\) be a bounded linear operator. Let \(\mathbb {N}\) and \(\mathbb{R}\) denote the sets of positive integers and real numbers.

In 1994, Censor and Elfving [1] came up with the split feasibility problem (SFP) in finite-dimensional Hilbert spaces. In infinite-dimensional Hilbert spaces, it can be formulated as

$$ \textit{Find }x\in C\textit{ such that }Ax\in Q, $$
(1.1)

where C and Q are nonempty closed convex subsets of \(H_{1}\) and \(H_{2}\), and \(A:H_{1}\rightarrow H_{2}\) is a bounded linear operator. Suppose that SFP (1.1) is solvable, and let S denote its solution set. The SFP is widely applied to signal processing, image reconstruction and biomedical engineering [24].

So far, some authors have studied SFP (1.1) [517]. Others have also found a lot of algorithms to study the split equality fixed point problem and the minimization problem [1820]. Byrne’s CQ algorithm is an effective method to solve SFP (1.1). A sequence \(\{x_{n}\}\), generated by the formula

$$ x_{n+1}=P_{C}\bigl(x_{n}-\lambda_{n}A^{*}(I-P_{Q})Ax_{n} \bigr), \quad\forall n\geq0, $$
(1.2)

where the parameters \(\lambda_{n}\in(0,\frac{2}{\|A\|^{2}})\), \(P_{C}: H\rightarrow C\), and \(P_{Q}: H \rightarrow Q\), is a set of orthogonal projections.

As is well-known, Cencor and Elfving’s algorithm needs to compute \(A^{-1}\), and Byrne’s CQ algorithm needs to compute \(\|A\|\). However, they are difficult to calculate.

Consider the following convex minimization problem:

$$ \min_{x\in C}f(x), $$
(1.3)

where

$$\begin{aligned}& f(x)=\frac{1}{2}\big\| (I-P_{Q})Ax\big\| ^{2}, \end{aligned}$$
(1.4)
$$\begin{aligned}& \nabla f(x)=A^{*}(I-P_{Q})Ax, \end{aligned}$$
(1.5)

\(f(x)\) is differentiable and the gradient f is L-Lipschitz with \(L>0\).

The gradient-projection algorithm [21] is the most effective method to solve (1.3). A sequence \(\{x_{n}\}\) is generated by the recursive formula

$$ x_{n+1}=P_{C}(I-\lambda_{n} \nabla f)x_{n}, \quad\forall n\geq0, $$
(1.6)

where the parameter \(\lambda_{n}\in(0,\frac{2}{L})\). Then we know that Byrne’s CQ algorithm is a special case of the gradient-projection algorithm.

In Byrne’s CQ algorithm, \(\lambda_{n}\) depends on the operator norm \(\| A\|\). However, it is difficult to compute. In 2005, Yang [22] considered \(\lambda_{n}\) as follows:

$$ \lambda_{n}:=\frac{\rho_{n}}{\|\nabla f(x_{n})\|}, $$

where \(\rho_{n}>0\) and satisfies

$$ \sum_{n=0}^{\infty}\rho_{n} = \infty,\qquad \sum_{n=0}^{\infty}\rho_{n}^{2} < \infty. $$

In 2012, López [23] introduced \(\lambda_{n}\) as follows:

$$ \lambda_{n}:=\frac{\rho_{n}f(x_{n})}{\|\nabla f(x_{n})\|^{2}}, $$

where \(0<\rho_{n}<4\). However, López’s algorithm only has weak convergence.

In 2013, Yao [24] introduced a self-adaptive method for the SFP and obtained a strong convergence theorem. However, the algorithm is difficult to work out.

In general, there are two types of algorithms to solve SFPs. One is the algorithm which depends on the norm of the operator. The other is the algorithm without a priori knowledge of the operator norm. The first type of algorithm needs to calculate \(\|A\|\), but \(\|A\|\) is not easy to work out. The second type of algorithm also has a drawback. It always has weak convergence. If we want to obtain strong convergence, we have to use the composited iterative method, but then the algorithm is difficult to calculate. In order to overcome the drawbacks, we propose a new regularized CQ algorithm without a priori knowledge of the operator norm to solve the SFP and we obtain a strong convergence theorem.

Consider the following regularized minimization problem:

$$ \min_{x\in C}f_{\beta}(x):=f(x)+\frac{\beta}{2}\|x \|^{2}, $$
(1.7)

where the regularization parameter \(\beta>0\). A sequence \(\{x_{n}\}\) is generated by the formula

$$ x_{n+1}=P_{C}\bigl(I-\lambda_{n}(\nabla f+ \beta_{n}I)\bigr)x_{n}, \quad\forall n\geq0, $$
(1.8)

where \(\nabla f(x_{n})=A^{*}(I-P_{Q})Ax_{n}\), \(0<\beta_{n}<1\), and \(\lambda_{n}=\frac{\rho_{n}f(x_{n})}{\|\nabla f(x_{n})\|^{2}}\), \(0<\rho _{n}<4\). Then, under suitable conditions, the sequence \(\{x_{n}\}\) generated by (1.8) converges strongly to a point \(z\in S\), where \(z=P_{S}(0)\) is the minimum-norm solution of SFP (1.1).

2 Preliminaries

In this part, we introduce some lemmas and some properties that are used in the rest of the paper. Throughout this paper, let \(H_{1}\) and \(H_{2}\) be real Hilbert spaces, \(A:H_{1}\rightarrow H_{2}\) be a bounded linear operator and I be the identity operator on \(H_{1}\) or \(H_{2}\). If \(f:H\rightarrow\mathbb{R}\) is a differentiable functional, then the gradient of f is denoted by f. We use the sign ‘→’ to denote strong convergence and use the sign ‘’ to denote weak convergence.

Definition 2.1

See [25]

Let D be a nonempty subset of H, and let \(T:D\rightarrow H\). Then T is firmly nonexpansive if

$$\|Tx-Ty\|^{2}+\big\| (I-T)x-(I-T)y\big\| ^{2}\leq\|x-y \|^{2}, \quad\forall x, y\in D. $$

Lemma 2.2

See [26]

Let \(T:H\rightarrow H\) be an operator. Then the following are equivalent:

  1. (i)

    T is firmly nonexpansive,

  2. (ii)

    \(I-T\) is firmly nonexpansive,

  3. (iii)

    \(2T-I\) is nonexpansive,

  4. (iv)

    \(\|Tx-Ty\|^{2}\leq\langle x-y, Tx-Ty \rangle\), \(\forall x, y \in H\),

  5. (v)

    \(0\leq\langle Tx-Ty, (I-T)x-(I-T)y\rangle\).

Recall \(P_{C}: H\rightarrow C\) is an orthogonal projection, where C is a nonempty closed convex subset of H. Then to each point \(x\in H\), the unique point \(P_{C}x\in C\) satisfies the following property:

$$ \|x-P_{C}x\|=\inf_{y\in C}\|x-y\|=: d(x,C). $$

\(P_{C}\) also has the following characteristics.

Lemma 2.3

See [27]

For a given \(x\in H\),

  1. (i)

    \(z=P_{C}x \Longleftrightarrow\langle x-z,z-y\rangle\geq0\), \(\forall y\in C\),

  2. (ii)

    \(z=P_{C}x \Longleftrightarrow\|x-z\|^{2}\leq\|x-y\|^{2}-\|y-z\| ^{2}\), \(\forall y\in C\),

  3. (iii)

    \(\langle P_{C}x-P_{C}y, x-y\rangle\geq\|P_{C}x-P_{C}y\|^{2}\), \(\forall x,y\in H\).

Lemma 2.4

See [28]

Let f be given by (1.4). Then

  1. (i)

    f is convex and differential,

  2. (ii)

    \(\nabla f(x)=A^{*}(I-P_{Q})Ax\), \(\forall x\in H\),

  3. (iii)

    f is w-lsc on H,

  4. (iv)

    f is \(\|A\|^{2}\)-Lipschitz: \(\|\nabla f(x)-\nabla f(y)\|\leq\|A\|^{2}\|x-y\|\), \(\forall x,y\in H\).

Lemma 2.5

See [29]

Let \(\{a_{n}\}\) be a sequence of nonnegative real numbers such that

$$a_{n+1} \leq(1-\alpha_{n})a_{n} + \alpha_{n}\delta_{n}, \quad n \geq0, $$

where \(\{\alpha_{n}\}_{n=0}^{\infty}\) is a sequence in \((0,1)\) and \(\{ \delta_{n}\}_{n=0}^{\infty}\) is a sequence in \(\mathbb{R}\) such that

  1. (i)

    \(\sum_{n=0}^{\infty}\alpha_{n} = \infty\),

  2. (ii)

    \(\limsup_{n\rightarrow\infty}\delta_{n} \leq0\) or \(\sum_{n=0}^{\infty}\alpha_{n}|\delta_{n}| < \infty\).

Then \(\lim_{n\rightarrow\infty}a_{n} = 0\).

Lemma 2.6

See [30]

Let \(\{\gamma_{n}\}_{n\in\mathbb{N}}\) be a sequence of real numbers such that there exists a subsequence \(\{\gamma_{n_{i}}\}_{i\in \mathbb{N}}\) of \(\{\gamma_{n}\}_{n\in\mathbb{N}}\) such that \(\gamma _{n_{i}}<\gamma_{n_{i}+1}\) for all \(i\in\mathbb{N}\). Then there exists a nondecreasing sequence \(\{m_{k}\}_{k\in\mathbb{N}}\) of \(\mathbb{N}\) such that \(\lim_{k\rightarrow\infty}m_{k}=\infty\) and the following properties are satisfied by all (sufficiently large) numbers \(k\in \mathbb{N}\):

$$ \gamma_{m_{k}}\leq\gamma_{m_{k}+1},\qquad \gamma_{k}\leq \gamma_{m_{k}+1}. $$

In fact, \(m_{k}\) is the largest number n in the set \(\{1,\ldots,k\}\) such that the condition

$$ \gamma_{n}\leq\gamma_{n+1} $$

holds.

3 Main results

In this paper, we always assume that \(f:H\rightarrow\mathbb{R}\) is a real-valued convex function, where \(f(x)=\frac{1}{2}\|(I-P_{Q})Ax\|^{2}\), the gradient \(\nabla f(x)=A^{*}(I-P_{Q})Ax\), C and Q are nonempty closed convex subsets of real Hilbert spaces \(H_{1}\) and \(H_{2}\), and \(A:H_{1}\rightarrow H_{2}\) is a bounded linear operator.

Algorithm 3.1

Choose an initial guess \(x_{0}\in H\) arbitrarily. Assume that the nth iterate \(x_{n}\in C\) has been constructed and \(\nabla f(x_{n})\neq0\). Then we calculate the \((n+1)\)th iterate \(x_{n+1}\) via the formula

$$ x_{n+1}=P_{C}\bigl(x_{n}-\lambda_{n} \bigl(A^{*}(I-P_{Q})Ax_{n}+\beta_{n}x_{n} \bigr)\bigr), \quad\forall n\geq0, $$
(3.1)

where \(\lambda_{n}\) is chosen as follows:

$$\lambda_{n}=\frac{\rho_{n}f(x_{n})}{\|\nabla f(x_{n})\|^{2}} $$

with \(0<\rho_{n}<4\). If \(\nabla f(x_{n})= 0\), then \(x_{n+1}=x_{n}\) is a solution of SFP (1.1) and the iterative process stops. Otherwise, we set \(n:=n+1\) and go to (3.1) to evaluate the next iterate \(x_{n+2}\).

Theorem 3.1

Suppose that \(S\neq\emptyset\) and the parameters \(\{\beta_{n}\}\) and \(\{\rho_{n}\}\) satisfy the following conditions:

  1. (i)

    \(\{\beta_{n}\}\subset(0,1)\), \(\lim_{n\rightarrow\infty}\beta _{n}=0\), \(\sum_{n=1}^{\infty}\beta_{n} = \infty\),

  2. (ii)

    \(\varepsilon\leq\rho_{n}\leq4-\varepsilon\) for some \(\varepsilon>0\) small enough.

Then the sequence \(\{x_{n}\}\) generated by Algorithm 3.1 converges strongly to \(z\in S\), where \({z=P_{S}(0)}\).

Proof

Let \(x^{*}\in S\). Since minimization is an exactly fixed point of its projection mapping, we have \(x^{*}=P_{C}x^{*}\) and \(Ax^{*}=P_{Q}Ax^{*}\).

By (3.1) and the nonexpansivity of \(P_{C}\), we derive

$$ \begin{aligned}[b] \big\| x_{n+1}-x^{*}\big\| ^{2} ={}&\big\| P_{C}\bigl(x_{n}- \lambda_{n}\bigl(A^{*}(I-P_{Q})Ax_{n}+\beta _{n}x_{n}\bigr)\bigr)-P_{C}x^{*}\big\| ^{2} \\ \leq{}&\big\| (1-\lambda_{n}\beta_{n})x_{n}-\lambda _{n}A^{*}(I-P_{Q})Ax_{n}-x^{*}\big\| ^{2} \\ ={}&\bigg\| \lambda_{n}\beta_{n}\bigl(-x^{*}\bigr) +(1-\lambda_{n}\beta_{n}) \biggl(x_{n}- \frac{\lambda_{n}}{1-\lambda_{n}\beta _{n}}A^{*}(I-P_{Q})Ax_{n}-x^{*}\biggr) \bigg\| ^{2} \\ ={}&\lambda_{n}\beta_{n}\big\| x^{*}\big\| ^{2} +(1- \lambda_{n}\beta_{n})\bigg\| x_{n}-\frac{\lambda_{n}}{1-\lambda_{n}\beta _{n}}A^{*}(I-P_{Q})Ax_{n}-x^{*} \bigg\| ^{2} \\ & -\lambda_{n}\beta_{n}(1-\lambda_{n} \beta_{n})\bigg\| x_{n}-\frac{\lambda _{n}}{1-\lambda_{n}\beta_{n}}A^{*}(I-P_{Q})Ax_{n} \bigg\| ^{2} \\ \leq{}&\lambda_{n}\beta_{n}\big\| x^{*}\big\| ^{2} +(1-\lambda_{n}\beta_{n})\bigg\| x_{n}- \frac{\lambda_{n}}{1-\lambda _{n}\beta_{n}}A^{*}(I-P_{Q})Ax_{n}-x^{*}\bigg\| ^{2}. \end{aligned}$$
(3.2)

Since \(P_{Q}\) is firmly nonexpansive, from Lemma 2.2, we deduce that \(I-P_{Q}\) is also firmly nonexpansive. Hence, we have

$$ \begin{aligned}[b] \bigl\langle A^{*}(I-P_{Q})Ax_{n}, x_{n}-x^{*}\bigr\rangle &=\bigl\langle (I-P_{Q})Ax_{n}, Ax_{n}-Ax^{*} \bigr\rangle \\ &=\bigl\langle (I-P_{Q})Ax_{n}-(I-P_{Q})Ax^{*}, Ax_{n}-Ax^{*}\bigr\rangle \\ &\geq\big\| (I-P_{Q})Ax_{n}\big\| ^{2} \\ &=2f(x_{n}). \end{aligned}$$
(3.3)

Note that \(\nabla f(x_{n})=A^{*}(I-P_{Q})Ax_{n}\). From (3.3), we obtain

$$\begin{aligned} &\bigg\| x_{n}-\frac{\lambda_{n}}{1-\lambda_{n}\beta _{n}}A^{*}(I-P_{Q})Ax_{n}-x^{*} \bigg\| ^{2} \\ &\quad=\big\| x_{n}-x^{*}\big\| ^{2}+\frac{\lambda_{n}^{2}}{(1-\lambda_{n}\beta_{n})^{2}}\big\| A^{*}(I-P_{Q})Ax_{n}\big\| ^{2} -\frac{2\lambda_{n}}{1-\lambda_{n}\beta_{n}}\bigl\langle A^{*}(I-P_{Q})Ax_{n}, x_{n}-x^{*}\bigr\rangle \\ &\quad=\big\| x_{n}-x^{*}\big\| ^{2}+\frac{\lambda_{n}^{2}}{(1-\lambda_{n}\beta_{n})^{2}}\big\| \nabla f(x_{n})\big\| ^{2}-\frac{2\lambda_{n}}{1-\lambda_{n}\beta_{n}}\bigl\langle \nabla f(x_{n}), x_{n}-x^{*} \bigr\rangle \\ &\quad\leq\big\| x_{n}-x^{*}\big\| ^{2}+\frac{\lambda_{n}^{2}}{(1-\lambda_{n}\beta_{n})^{2}}\big\| \nabla f(x_{n})\big\| ^{2}-\frac{4\lambda_{n}}{1-\lambda_{n}\beta _{n}}f(x_{n}) \\ &\quad=\big\| x_{n}-x^{*}\big\| ^{2}+\frac{1}{(1-\lambda_{n}\beta_{n})^{2}}\cdot \frac{\rho _{n}^{2}f(x_{n})^{2}}{\|\nabla f(x_{n})\|^{4}}\cdot\big\| \nabla f(x_{n})\big\| ^{2} \\ &\qquad{} -\frac{4\rho_{n}f(x_{n})}{(1-\lambda_{n}\beta_{n})\|\nabla f(x_{n})\| ^{2}}\cdot f(x_{n}) \\ &\quad=\big\| x_{n}-x^{*}\big\| ^{2}+\frac{\rho_{n}^{2}f(x_{n})^{2}}{(1-\lambda_{n}\beta _{n})^{2}\|\nabla f(x_{n})\|^{2}}-\frac{4\rho_{n}f(x_{n})^{2}}{(1-\lambda _{n}\beta_{n})\|\nabla f(x_{n})\|^{2}} \\ &\quad=\big\| x_{n}-x^{*}\big\| ^{2} -\rho_{n}\biggl(4- \frac{\rho_{n}}{1-\lambda_{n}\beta_{n}}\biggr)\cdot\frac {f(x_{n})^{2}}{(1-\lambda_{n}\beta_{n})\|\nabla f(x_{n})\|^{2}}. \end{aligned}$$
(3.4)

By condition (ii), without loss of generality, we assume that \((4-\frac{\rho_{n}}{1-\lambda_{n}\beta_{n}})>0\) for all \(n\geq0 \). Thus from (3.2) and (3.4), we obtain

$$ \begin{aligned}[b] \big\| x_{n+1}-x^{*}\big\| ^{2} \leq{}&\lambda_{n} \beta_{n}\big\| x^{*}\big\| ^{2}+(1-\lambda_{n} \beta_{n}) \biggl(\big\| x_{n}-x^{*}\big\| ^{2} \\ &-\rho_{n}\biggl(4-\frac{\rho_{n}}{1-\lambda_{n}\beta_{n}}\biggr)\cdot\frac {f(x_{n})^{2}}{(1-\lambda_{n}\beta_{n})\|\nabla f(x_{n})\|^{2}} \biggr) \\ ={}&\lambda_{n}\beta_{n}\big\| x^{*}\big\| ^{2}+(1- \lambda_{n}\beta_{n})\big\| x_{n}-x^{*}\big\| ^{2} \\ & -\rho_{n}\biggl(4-\frac{\rho_{n}}{1-\lambda_{n}\beta_{n}}\biggr)\cdot\frac {f(x_{n})^{2}}{\|\nabla f(x_{n})\|^{2}} \\ \leq{}&\lambda_{n}\beta_{n}\big\| x^{*}\big\| ^{2}+(1- \lambda_{n}\beta_{n})\big\| x_{n}-x^{*}\big\| ^{2} \\ \leq{}&\max\bigl\{ \big\| x^{*}\big\| ^{2}, \big\| x_{n}-x^{*}\big\| ^{2} \bigr\} . \end{aligned}$$
(3.5)

Hence, \(\{x_{n}\}\) is bounded.

Let \(z=P_{S}0\). From (3.5), we deduce

$$ \begin{aligned}[b] 0 &\leq\rho_{n}\biggl(4-\frac{\rho_{n}}{1-\lambda_{n}\beta_{n}}\biggr)\frac {f(x_{n})^{2}}{\|\nabla f(x_{n})\|^{2}} \\ &\leq\lambda_{n}\beta_{n}\|z\|^{2}+(1- \lambda_{n}\beta_{n})\|x_{n}-z\| ^{2}- \|x_{n+1}-z\|^{2}. \end{aligned}$$
(3.6)

We consider the following two cases.

Case 1. One has \(\|x_{n+1}-z\|\leq\|x_{n}-z\|\) for every \(n\geq n_{0}\) large enough.

In this case, \(\lim_{n\rightarrow\infty}\|x_{n}-z\|\) exists as finite and hence

$$ \lim_{n\rightarrow\infty}\bigl(\|x_{n+1}-z\|-\|x_{n}-z\|\bigr)=0. $$
(3.7)

This, together with (3.6), implies that

$$ \rho_{n}\biggl(4-\frac{\rho_{n}}{1-\lambda_{n}\beta_{n}}\biggr)\frac{f(x_{n})^{2}}{\| \nabla f(x_{n})\|^{2}} \rightarrow0. $$

Since \(\liminf_{n\rightarrow\infty}\rho_{n}(4-\frac{\rho_{n}}{1-\lambda _{n}\beta_{n}})\geq\varepsilon_{0}\) (where \(\varepsilon_{0}>0\) is a constant), we get

$$\frac{f(x_{n})^{2}}{\|\nabla f(x_{n})\|^{2}}\rightarrow0. $$

Noting that \(\|\nabla f(x_{n})\|^{2}\) is bounded, we deduce immediately that

$$ \lim_{n\rightarrow\infty}f(x_{n})=0. $$
(3.8)

Next, we prove that

$$ \limsup_{n\rightarrow\infty}\langle-z, x_{n}-z\rangle\leq0. $$
(3.9)

Since \(\{x_{n}\}\) is bounded, there exists a subsequence \(\{x_{n_{i}}\} \) satisfying \(x_{n_{i}}\rightharpoonup\hat{z}\) and

$$ \limsup_{n\rightarrow\infty}\langle-z, x_{n}-z\rangle=\lim _{i\rightarrow\infty}\langle -z, x_{n_{i}}-z\rangle. $$

By the lower semicontinuity of f, we get

$$ 0\leq f(\hat{z})\leq\liminf_{i\rightarrow\infty}f(x_{n_{i}})=\lim _{n\rightarrow\infty}f(x_{n})=0. $$

So

$$ f(\hat{z})=\frac{1}{2}\big\| (I-P_{Q})A\hat{z}\big\| ^{2}=0. $$

That is, is a minimizer of f, and \(\hat{z}\in S\). Therefore

$$ \begin{aligned}[b] \limsup_{n\rightarrow\infty}\langle-z, x_{n}-z\rangle &=\lim _{i\rightarrow\infty}\langle -z, x_{n_{i}}-z\rangle \\ &=\langle-z, \hat{z}-z\rangle \\ &\leq 0.\end{aligned} $$
(3.10)

Then we have

$$\begin{aligned} \|x_{n+1}-z\|^{2} ={}&\big\| P_{C}\bigl(x_{n}- \lambda_{n}A^{*}(I-P_{Q})Ax_{n}-\lambda_{n} \beta _{n}x_{n}\bigr)-P_{C}z\big\| ^{2} \\ \leq{}&\big\| (1-\lambda_{n}\beta_{n})x_{n}- \lambda_{n}A^{*}(I-P_{Q})Ax_{n}-z\big\| ^{2} \\ ={}&\bigg\| \lambda_{n}\beta_{n}(-z) +(1-\lambda_{n}\beta_{n}) \biggl(x_{n}- \frac{\lambda_{n}}{1-\lambda_{n}\beta _{n}}A^{*}(I-P_{Q})Ax_{n}-z\biggr)\bigg\| ^{2} \\ ={}&(\lambda_{n}\beta_{n})^{2}\|z\|^{2} \\ &+(1-\lambda_{n}\beta_{n})^{2}\bigg\| x_{n}- \frac{\lambda_{n}}{1-\lambda _{n}\beta_{n}}A^{*}(I-P_{Q})Ax_{n}-z\bigg\| ^{2} \\ &+2(1-\lambda_{n}\beta_{n})\lambda_{n} \beta_{n}\biggl\langle x_{n}-\frac {\lambda_{n}}{1-\lambda_{n}\beta_{n}}A^{*}(I-P_{Q})Ax_{n}-z,-z \biggr\rangle \\ \leq{}&(1-\lambda_{n}\beta_{n})^{2} \|x_{n}-z\|^{2}+(\lambda_{n}\beta_{n})^{2} \| z\|^{2} \\ & +2(1-\lambda_{n}\beta_{n})\lambda_{n} \beta_{n}\langle x_{n}-z, -z\rangle+2\lambda_{n}^{2} \beta_{n}\bigl\langle \nabla f(x_{n}), z\bigr\rangle \\ \leq{}&(1-\lambda_{n}\beta_{n})\|x_{n}-z \|^{2} \\ &+\lambda_{n}\beta_{n}\bigl(\lambda _{n}\beta_{n}\|z\|^{2}+2(1-\lambda_{n}\beta_{n})\langle x_{n}-z, -z \rangle+2\lambda_{n}\big\| \nabla f(x_{n})\big\| \cdot\|z\|\bigr). \end{aligned}$$

Note that \(\|\nabla f(x_{n})\|^{2}\) is bounded, and that \(\lambda_{n}\| \nabla f(x_{n})\|=\frac{\rho_{n}f(x_{n})}{\|\nabla f(x_{n})\|^{2}}\cdot\| \nabla f(x_{n})\|\). Thus \(\lambda_{n}\|\nabla f(x_{n})\|\rightarrow0\) by (3.8). From Lemma 2.5, we deduce that

$$x_{n}\rightarrow z. $$

Case 2. There exists a subsequence \(\{\|x_{n_{j}}-z\|\}\) of \(\{\| x_{n}-z\|\}\) such that

$$\|x_{n_{j}}-z\|< \|x_{n_{j}+1}-z\| \quad\text{for all }j\geq1. $$

By Lemma 2.6, there exists a strictly nondecreasing sequence \(\{m_{k}\} \) of positive integers such that \(\lim_{k\rightarrow\infty}m_{k}=+\infty \) and the following properties are satisfied by all numbers \(k\in \mathbb{N}\):

$$ \|x_{m_{k}}-z\|\leq\|x_{m_{k}+1}-z\|,\qquad \|x_{k}-z\|\leq \|x_{m_{k}+1}-z\|. $$
(3.11)

We have

$$\begin{aligned} \|x_{n+1}-z\| ={}&\big\| P_{C}\bigl(x_{n}- \lambda_{n}A^{*}(I-P_{Q})Ax_{n}-\lambda_{n} \beta _{n}x_{n}\bigr)-P_{C}z\big\| \\ \leq{}&\big\| (1-\lambda_{n}\beta_{n})x_{n}- \lambda_{n}A^{*}(I-P_{Q})Ax_{n}-z\big\| \\ ={}&\bigg\| \lambda_{n}\beta_{n}(-z) +(1-\lambda_{n}\beta_{n}) \biggl(x_{n}- \frac{\lambda_{n}}{1-\lambda_{n}\beta _{n}}A^{*}(I-P_{Q})Ax_{n}-z\biggr)\bigg\| \\ \leq{}&\lambda_{n}\beta_{n}\big\| (-z)\big\| +(1-\lambda_{n} \beta_{n})\bigg\| x_{n}-\frac {\lambda_{n}}{1-\lambda_{n}\beta_{n}}A^{*}(I-P_{Q})Ax_{n}-z \bigg\| \\ \leq{}&\lambda_{n}\beta_{n}\|z\|+(1-\lambda_{n} \beta_{n})\|x_{n}-z\|. \end{aligned}$$

Consequently,

$$\begin{aligned} 0 &\leq\lim_{k\rightarrow\infty}\bigl(\|x_{m_{k}+1}-z\|-\|x_{m_{k}}-z \|\bigr) \\ &\leq\limsup_{n\rightarrow\infty}\bigl(\|x_{n+1}-z\|-\|x_{n}-z \|\bigr) \\ &\leq\limsup_{n\rightarrow\infty}\bigl(\lambda_{n} \beta_{n}\|z\|+(1-\lambda _{n}\beta_{n}) \|x_{n}-z\|-\|x_{n}-z\|\bigr) \\ &=\limsup_{n\rightarrow\infty}\lambda_{n}\beta_{n}\bigl(\|z \|-\|x_{n}-z\|\bigr) \\ &=0. \end{aligned}$$

Hence,

$$ \lim_{k\rightarrow\infty}\bigl(\|x_{m_{k}+1}-z\|-\|x_{m_{k}}-z\|\bigr)=0. $$
(3.12)

By a similar argument to that of Case 1, we prove that

$$\begin{aligned}& \limsup_{k\rightarrow\infty}\langle-z, x_{m_{k}}-z \rangle\leq0, \\& \|x_{m_{k}+1}-z\|^{2}\leq(1-\lambda_{m_{k}} \beta_{m_{k}})\|x_{m_{k}}-z\| ^{2}+\lambda_{m_{k}} \beta_{m_{k}}\sigma_{m_{k}}, \end{aligned}$$
(3.13)

where

$$\sigma_{m_{k}}=\lambda_{m_{k}}\beta_{m_{k}}\|z \|^{2}+2(1-\lambda _{m_{k}}\beta_{m_{k}})\langle x_{m_{k}}-z, -z\rangle+2\lambda_{m_{k}}\big\| \nabla f(x_{m_{k}})\big\| \cdot\|z\|. $$

In particular, from (3.13), we get

$$ \lambda_{m_{k}}\beta_{m_{k}}\|x_{m_{k}}-z\|^{2} \leq\|x_{m_{k}}-z\|^{2}-\| x_{m_{k}+1}-z\|^{2}+ \lambda_{m_{k}}\beta_{m_{k}}\sigma_{m_{k}}. $$
(3.14)

Since \(\|x_{m_{k}}-z\|\leq\|x_{m_{k}+1}-z\|\), we deduce that

$$ \|x_{m_{k}}-z\|^{2}-\|x_{m_{k}+1}-z\|^{2}\leq0. $$

Then, from (3.14), we have

$$ \lambda_{m_{k}}\beta_{m_{k}}\|x_{m_{k}}-z\|^{2} \leq\lambda_{m_{k}}\beta _{m_{k}}\sigma_{m_{k}}. $$

Then

$$ \limsup_{k\rightarrow\infty}\|x_{m_{k}}-z\|^{2}\leq\limsup _{k\rightarrow \infty}\sigma_{m_{k}}\leq0. $$
(3.15)

Then, from (3.12), we deduce that

$$ \limsup_{k\rightarrow\infty}\|x_{m_{k}+1}-z\|=0. $$
(3.16)

Thus, from (3.11) and (3.16), we conclude that

$$ \limsup_{k\rightarrow\infty}\|x_{k}-z\|\leq\limsup _{k\rightarrow\infty }\|x_{m_{k}+1}-z\|=0. $$

Therefore, \(x_{n}\rightarrow z\). This completes the proof. □

4 Conclusion

Recently, the SFP has been studied extensively by many authors. However, some algorithms need to compute \(\|A\|\), and this is not an easy thing to work out. Others do not need to compute \(\|A\|\), but the algorithms always have weak convergence. If we want to obtain strong convergence theorems, the algorithms are complex and difficult to calculate. We try to get over the drawbacks. In this article, we use the regularized CQ algorithm without computing \(\|A\|\) to find the minimum-norm solution of the SFP, where \(\lambda_{n}=\frac{\rho _{n}f(x_{n})}{\|\nabla f(x_{n})\|^{2}}\), \(0<\rho_{n}<4\). Then, under suitable conditions, the explicit strong convergence theorem is obtained.

References

  1. Censor, Y, Elfving, T: A multiprojection algorithm using Bregman projections in a product space. Numer. Algorithms 8, 221-239 (1994)

    Article  MathSciNet  MATH  Google Scholar 

  2. Censor, Y, Bortfeld, T, Martin, B, Trofimov, A: A unified approach for inversion problems in intensity-modulated radiation therapy. Phys. Med. Biol. 51, 2353-2365 (2003)

    Article  Google Scholar 

  3. Censor, Y, Elfving, T, Kopf, N, Bortfeld, T: The multiple-sets split feasibility problem and its applications for inverse problem. Inverse Probl. 21, 2071-2084 (2005)

    Article  MathSciNet  MATH  Google Scholar 

  4. López, G, Martín-Márquez, V, Xu, HK: Perturbation techniques for nonexpansive mappings with applications. Nonlinear Anal., Real World Appl. 10, 2369-2383 (2009)

    Article  MathSciNet  MATH  Google Scholar 

  5. López, G, Martín-Márquez, V, Xu, HK: Iterative algorithms for the multiple-sets split feasibility problem. In: Censor, Y, Jiang, M, Wang, G (eds.) Biomedical Mathematics: Promising Directions in Imaging, Therapy Planning and Inverse Problems, pp. 243-279. Medical Physics Publishing, Madison (2010)

    Google Scholar 

  6. Xu, HK: Iterative methods for the split feasibility problem in infinite-dimensional Hilbert space. Inverse Probl. 26, Article ID 105018 (2010)

    Article  MathSciNet  MATH  Google Scholar 

  7. Dang, Y, Gao, Y: The strong convergence of a KM-CQ-like algorithm for a split feasibility problem. Inverse Probl. 27, Article ID 015007 (2011)

    Article  MathSciNet  MATH  Google Scholar 

  8. Qu, B, Xiu, N: A note on the CQ algorithm for the split feasibility problem. Inverse Probl. 21, 1655-1665 (2005)

    Article  MathSciNet  MATH  Google Scholar 

  9. Wang, F, Xu, HK: Approximating curve and strong convergence of the CQ algorithm for the split feasibility problem. J. Inequal. Appl. 2010, Article ID 102085 (2010)

    MathSciNet  MATH  Google Scholar 

  10. Xu, HK: A variable Krasnosel’skii-Mann algorithm and the multiple-set split feasibility problem. Inverse Probl. 22, 2021-2034 (2006)

    Article  MATH  Google Scholar 

  11. Yang, Q: The relaxed CQ algorithm for solving the split feasibility problem. Inverse Probl. 20, 1261-1266 (2004)

    Article  MathSciNet  MATH  Google Scholar 

  12. Byrne, C: Iterative oblique projection onto convex sets and the split feasibility problem. Inverse Probl. 18, 441-453 (2002)

    Article  MathSciNet  MATH  Google Scholar 

  13. Byrne, C: A unified treatment of some iterative algorithms in signal processing and image reconstruction. Inverse Probl. 20, 103-120 (2004)

    Article  MathSciNet  MATH  Google Scholar 

  14. Yao, YH, Wu, JG, Liou, YC: Regularized methods for the split feasibility problem. Abstr. Appl. Anal. 2012, Article ID 140679 (2012)

    MathSciNet  MATH  Google Scholar 

  15. Yao, YH, Liou, YC, Yao, JC: Iterative algorithms for the split variational inequality and fixed point problems under nonlinear transformations. J. Nonlinear Sci. Appl. 10, 843-854 (2017)

    Article  MathSciNet  Google Scholar 

  16. Yao, YH, Agarwal, RP, Postolache, M, Liou, YC: Algorithms with strong convergence for the split common solution of the feasibility problem and fixed point problem. Fixed Point Theory Appl. 2014, Article ID 183 (2014)

    Article  MathSciNet  Google Scholar 

  17. Latif, A, Qin, X: A regularization algorithm for a splitting feasibility problem in Hilbert spaces. J. Nonlinear Sci. Appl. 10, 3856-3862 (2017)

    Article  Google Scholar 

  18. Chang, SS, Wang, L, Zhao, YH: On a class of split equality fixed point problems in Hilbert spaces. J. Nonlinear Var. Anal. 1, 201-212 (2017)

    Google Scholar 

  19. Tang, JF, Chang, SS, Dong, J: Split equality fixed point problem for two quasi-asymptotically pseudocontractive mappings. J. Nonlinear Funct. Anal. 2017, Article ID 26 (2017)

    Google Scholar 

  20. Shi, LY, Ansari, QH, Wen, CF, Yao, JC: Incremental gradient projection algorithm for constrained composite minimization problems. J. Nonlinear Var. Anal. 1, 253-264 (2017)

    Google Scholar 

  21. Xu, HK: Averaged mappings and the gradient-projection algorithm. J. Optim. Theory Appl. 150, 360-378 (2011)

    Article  MathSciNet  MATH  Google Scholar 

  22. Yang, Q: On variable-step relaxed projection algorithm for variational inequalities. J. Math. Anal. Appl. 302, 166-179 (2005)

    Article  MathSciNet  MATH  Google Scholar 

  23. López, G, Martín-Márquez, V, Wang, FH, Xu, HK: Solving the split feasibility problem without prior knowledge of matrix norms. Inverse Probl. (2012). doi:10.1088/0266-5611/28/8/085004

    MathSciNet  MATH  Google Scholar 

  24. Yao, YH, Postolache, M, Liou, YC: Strong convergence of a self-adaptive method for the split feasibility problem. Fixed Point Theory Appl. 2013, Article ID 201 (2013)

    Article  MathSciNet  MATH  Google Scholar 

  25. Bauschke, HH, Combettes, PL: Convex Analysis and Monotone Operator Theory in Hilbert Spaces. Springer, Berlin (2011)

    Book  MATH  Google Scholar 

  26. Goebel, K, Kirk, WA: Topics on Metric Fixed Point Theory. Cambridge University Press, Cambridge (1990)

    Book  MATH  Google Scholar 

  27. Takahashi, W: Nonlinear Functional Analysis. Yokohama Publishers, Yokohama (2000)

    MATH  Google Scholar 

  28. Aubin, JP: Optima and Equilibria: An Introduction to Nonlinear Analysis. Springer, Berlin (1993)

    Book  MATH  Google Scholar 

  29. Xu, HK: Viscosity approximation methods for nonexpansive mappings. J. Math. Anal. Appl. 298, 279-291 (2004)

    Article  MathSciNet  MATH  Google Scholar 

  30. Maingé, PE: Strong convergence of projected subgradient methods for non-smooth and non-strictly convex minimization. Set-Valued Anal. 17(7-8), 899-912 (2008)

    Article  MATH  Google Scholar 

Download references

Acknowledgements

The authors thank the referees for their helping comments, which notably improved the presentation of this paper. This work was supported by the Fundamental Research Funds for the Central Universities [grant number 3122017072]. MT was supported by the Foundation of Tianjin Key Laboratory for Advanced Signal Processing. H-FZ was supported in part by the Technology Innovation Funds of Civil Aviation University of China for Graduate (Y17-39).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ming Tian.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

All the authors read and approved the final manuscript.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Tian, M., Zhang, HF. The regularized CQ algorithm without a priori knowledge of operator norm for solving the split feasibility problem. J Inequal Appl 2017, 207 (2017). https://doi.org/10.1186/s13660-017-1480-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13660-017-1480-2

MSC

Keywords