Open Access

Strong convergence of a relaxed CQ algorithm for the split feasibility problem

Journal of Inequalities and Applications20132013:197

https://doi.org/10.1186/1029-242X-2013-197

Received: 15 November 2012

Accepted: 9 April 2013

Published: 22 April 2013

Abstract

The split feasibility problem (SFP) is finding a point in a given closed convex subset of a Hilbert space such that its image under a bounded linear operator belongs to a given closed convex subset of another Hilbert space. The most popular iterative method is Byrne’s CQ algorithm. López et al. proposed a relaxed CQ algorithm for solving SFP where the two closed convex sets are both level sets of convex functions. This algorithm can be implemented easily since it computes projections onto half-spaces and has no need to know a priori the norm of the bounded linear operator. However, their algorithm has only weak convergence in the setting of infinite-dimensional Hilbert spaces. In this paper, we introduce a new relaxed CQ algorithm such that the strong convergence is guaranteed. Our result extends and improves the corresponding results of López et al. and some others.

MSC:90C25, 90C30, 47J25.

Keywords

split feasibility problemrelaxed CQ algorithmHilbert spacestrong convergencebounded linear operator

1 Introduction

The split feasibility problem (SFP) was proposed by Censer and Elfving in [1]. It can mathematically be formulated as the problem of finding a point x satisfying the property:
x C , A x Q ,
(1.1)

where A is a given M × N real matrix (where A is the transpose of A), C and Q are nonempty, closed and convex subsets in R N and R M , respectively. This problem has received much attention [2] due to its applications in signal processing and image reconstruction, with particular progress in intensity-modulated radiation therapy [35], and many other applied fields.

We assume the SFP (1.1) is consistent, and let S be the solution set, i.e.,
S = { x C : A x Q } .
It is easily seen that S is closed convex. Many iterative methods can be used to solve the SFP (1.1); see [616]. Byrne [6, 17] was among others the first to propose the so-called CQ algorithm, which is defined as follows:
x n + 1 = P C ( x n τ n A ( I P Q ) A x n ) ,
(1.2)

where τ n ( 0 , 2 A 2 ) , and P C and P Q are the orthogonal projections onto the sets C and Q, respectively. Compared with Censer and Elfving’ algorithm [1], the Byrne’ algorithm is easily executed since it only deal with orthogonal projections with no need to compute matrix inverses.

The CQ algorithm (1.2) can be obtained from optimization. In fact, if we introduce the (convex) objective function
f ( x ) = 1 2 ( I P Q ) A x 2 ,
(1.3)
and analyze the minimization problem
min x C f ( x ) ,
(1.4)
then the CQ algorithm (1.2) comes immediately as a special case of the gradient projection algorithm (GPA)(For more details about the GPA, the reader is referred to [18]). Since the convex objective f ( x ) is differentiable, and has a Lipschitz gradient, which is given by
f ( x ) = A ( I P Q ) A x ,
(1.5)
the GPA for solving the minimization problem (1.4) generates a sequence ( x n ) recursively as
x n + 1 = P C ( x n τ n f ( x n ) ) ,
(1.6)

where the stepsize τ n is chosen in the interval ( 0 , 2 L ) , where L is the Lipschitz constant of f.

Observe that in algorithms (1.2) and (1.6) mentioned above, in order to implement the CQ algorithm, one has to compute the operator (matrix) norm A , which is in general not an easy work in practice. To overcome this difficulty, some authors proposed different adaptive choices of selecting the stepsize τ n (see [6, 14, 19]). For instance, very recently López et al. introduced a new way of selecting the stepsize [19] as follows:
τ n : = ρ n f ( x n ) f ( x n ) 2 , 0 < ρ n < 4 .
(1.7)

The computation of a projection onto a general closed convex subset is generally difficult. To overcome this difficulty, Fukushima [20] suggested a so-called relaxed projection method to calculate the projection onto a level set of a convex function by computing a sequence of projections onto half-spaces containing the original level set. In the setting of finite-dimensional Hilbert spaces, this idea was followed by Yang [13], who introduced the relaxed CQ algorithms for solving SFP (1.1) where the closed convex subsets C and Q are level sets of convex functions.

Recently, for the purpose of generality, the SFP (1.1) is studied in a more general setting. For instance, Xu [12] and López et al. [19] considered the SFP (1.1) in infinite-dimensional Hilbert spaces (i.e., the finite-dimensional Euclidean spaces R N and R M are replaced with general Hilbert spaces). Very recently, López et al. proposed a relaxed CQ algorithm with a new adaptive way of determining the stepsize sequence ( τ n ) for solving the SFP (1.1) where the closed convex subsets C and Q are level sets of convex functions. This algorithm can be implemented easily since it computes projections onto half-spaces and has no need to know a priori the norm of the bounded linear operator. However, their algorithm has only weak convergence in the setting of infinite-dimensional Hilbert spaces. In this paper, we introduce a new relaxed CQ algorithm such that the strong convergence is guaranteed in infinite-dimensional Hilbert spaces. Our result extends and improves the corresponding results of López et al. and some others.

The rest of this paper is organized as follows. Some useful lemmas are listed in Section 2. In Section 3, the strong convergence of the new relaxed CQ algorithm of this paper is proved.

2 Preliminaries

Throughout the rest of this paper, we denote by H or K a Hilbert space, A is a bounded linear operator from H to K, and by I the identity operator on H or K. If f : H R is a differentiable function, then we denote by f the gradient of the function f. We will also use the notations:

  • → denotes strong convergence.

  • denotes weak convergence.

  • ω w ( x n ) = { x { x n k } { x n }  such that  x n k x } denotes the weak ω-limit set of { x n } .

Recall that a mapping T : H H is said to be nonexpansive if
T x T y x y , x , y H .
T : H H is said to be firmly nonexpansive if, for x , y H ,
T x T y 2 x y 2 ( I T ) x ( I T ) y 2 .
Recall that a function f : H R is called convex if
f ( λ x + ( 1 λ ) y ) λ f ( x ) + ( 1 λ ) f ( y ) , λ ( 0 , 1 ) , x , y H .
A differentiable function f is convex if and only if there holds the inequality:
f ( z ) f ( x ) + f ( x ) , z x , z H .
Recall that an element g H is said to be a subgradient of f : H R at x if
f ( z ) f ( x ) + g , z x , z H .

This relation is called the subdifferentiable inequality.

A function f : H R is said to be subdifferentiable at x, if it has at least one subgradient at x. The set of subgradients of f at the point x is called the subdifferentiable of f at x, and it is denoted by f ( x ) . A function f is called subdifferentiable, if it is subdifferentiable at all x H . If a function f is differentiable and convex, then its gradient and subgradient coincide.

A function f : H R is said to be weakly lower semi-continuous (w-lsc) at x if x n x implies
f ( x ) lim inf n f ( x n ) .
We know that the orthogonal projection P C from H onto a nonempty closed convex subset C H is a typical example of a firmly nonexpansive mapping, which is defined by
P C x : = arg min y C x y 2 , x H .
(2.1)
It is well known that P C is characterized by the inequality (for x H )
x P C x , y P C x 0 , y C .
(2.2)

The following lemma is not hard to prove (see [17, 21]).

Lemma 2.1 Let f be given as in (1.3). Then
  1. (i)

    f is convex and differential.

     
  2. (ii)

    f ( x ) = A ( I P Q ) A x , x H .

     
  3. (iii)

    f is w-lsc on H.

     
  4. (iv)

    f is A 2 -Lipschitz: f ( x ) f ( y ) A 2 x y , x , y H .

     

The following are characterizations of firmly nanexpansive mappings (see [22]).

Lemma 2.2 Let T : H H be an operator. The following statements are equivalent.
  1. (i)

    T is firmly nonexpansive.

     
  2. (ii)

    T x T y 2 x y , T x T y , x , y H .

     
  3. (iii)

    I T is firmly nonexpansive.

     

Lemma 2.3 [23]

Assume ( α n ) is a sequence of nonnegative real numbers such that
α n + 1 ( 1 γ n ) α n + γ n σ n , n = 0 , 1 , 2 , ,
where ( γ n ) is a sequence in ( 0 , 1 ) , and ( σ n ) is a sequence in such that
  1. (i)

    n = 0 γ n = .

     
  2. (ii)

    lim sup n σ n 0 , or n = 0 | γ n σ n | < .

     

Then lim n α n = 0 .

3 Iterative Algorithm

In this section, we turn to consider a new relaxed CQ algorithm in the setting of infinite-dimensional Hilbert spaces for solving SFP (1.1) where the closed convex subsets C and Q are of the particular structure, i.e. level sets of convex functions given as follows:
C = { x H : c ( x ) 0 } and Q = { y K : q ( y ) 0 } ,
(3.1)

where c : H R and q : K R are convex functions. We assume that both c and q are subdifferentiable on H and K, respectively, and that ∂c and ∂q are bounded operators (i.e., bounded on bounded sets). By the way, we mention that every convex function defined on a finite-dimensional Hilbert space is subdifferentiable and its subdifferential operator is a bounded operator (see [24]).

Set
C n = { x H : c ( x n ) ξ n , x n x } ,
(3.2)
where ξ n c ( x n ) , and
Q n = { y K : q ( A x n ) ζ n , A x n y } ,
(3.3)

where ζ n q ( A x n ) .

Obviously, C n and Q n are half-spaces and it is easy to verify that C n C and Q n Q hold for every n 0 from the subdifferentiable inequality. In what follows, we define
f n ( x ) = 1 2 ( I P Q n ) A x 2 , n 0 ,
where Q n is given as in (3.3). We then have
f n ( x ) = A ( I P Q n ) A x .

Firstly, we recall the relaxed CQ algorithm of López et al. [19] for solving the SFP (1.1) where C and Q are given in (3.1) as follows.

Algorithm 3.1 Choose an arbitrary initial guess x 0 . Assume x n has been constructed. If f n ( x n ) = 0 , then stop; otherwise, continue and construct x n + 1 via the formula
x n + 1 = P C n ( x n τ n f n ( x n ) ) ,
where C n is given as (3.2), and
τ n = ρ n f n ( x n ) f n ( x n ) 2 , 0 < ρ n < 4 .
(3.4)

López et al. proved that under some certain conditions the sequence ( x n ) generated by Algorithm 3.1 converges weakly to a solution of the SFP (1.1). Since the projections onto half-spaces C n and Q n have closed forms and τ n is obtained adaptively via the formula (3.4) (no need to know a priori the norm of operator A), the above relaxed CQ algorithm 3.1 is implementable. But the weak convergence is its a weakness. To overcome this weakness, inspired by Algorithm 3.1, we will introduce a new relaxed CQ algorithm for solving the SFP (1.1) where C and Q are given in (3.1) so that the strong convergence is guaranteed.

It is well known that Halpern’s algorithm has a strong convergence for finding a fixed point of a nonexpansive mapping [25, 26]. Then we are in a position to give our algorithm. The algorithm given below is referred to as a Halpern-type algorithm [27].

Algorithm 3.2 Let u H , and start an initial guess x 0 H arbitrarily. Assume that the n th iterate x n has been constructed. If f n ( x n ) = 0 , then stop ( x n is a approximate solution of SFP (1.1)). Otherwise, continue and calculate the ( n + 1 ) th iterate x n + 1 via the formula:
x n + 1 = P C n [ α n u + ( 1 α n ) ( x n τ n f n ( x n ) ) ] ,
(3.5)

where the sequence ( α n ) ( 0 , 1 ) and ( τ n ) and ( ρ n ) are given as in (3.4).

The convergence result of Algorithm 3.2 is stated in the next theorem.

Theorem 3.3 Assume that ( α n ) and ( ρ n ) satisfy the assumptions:

(a1) lim n α n = 0 and n = 0 α n = .

(a2) inf n ρ n ( 4 ρ n ) > 0 .

Then the sequence ( x n ) generated by Algorithm  3.2 converges in norm to P S u .

Proof We may assume that the sequence ( x n ) is infinite, that is, Algorithm 3.2 does not terminate in a finite number of iterations. Thus f n ( x n ) 0 for all n 0 . Recall that S is the solution set of the SFP (1.1),
S = { x C : A x Q } .
In the consistent case of the SFP (1.1), S is nonempty, closed and convex. Thus, the metric projection P S is well-defined. We set z = P S u . Since z S C n and the projection operator P C n is nonexpansive, we obtain
x n + 1 z 2 = P C n [ α n u + ( 1 α n ) ( x n τ n f n ( x n ) ) ] z 2 α n ( u z ) + ( 1 α n ) ( x n τ n f n ( x n ) z ) 2 ( 1 α n ) x n τ n f n ( x n ) z 2 + 2 α n u z , x n + 1 z .
Note that I P Q n is firmly nonexpansive and f n ( z ) = 0 , it is deduced from Lemma 2.2 that
f n ( x n ) , x n z = ( I P Q n ) A x n , A x n A z ( I P Q n ) A x n 2 = 2 f n ( x n ) ,
which implies that
x n τ n f n ( x n ) z 2 = x n z 2 + τ n f n ( x n ) 2 2 τ n f n ( x n ) , x n z x n z 2 + τ n 2 f n ( x n ) 2 4 τ n f n ( x n ) x n z 2 ρ n ( 4 ρ n ) f n 2 ( x n ) f n ( x n ) 2 .
Thus, we have
x n + 1 z 2 ( 1 α n ) x n τ n f n ( x n ) z 2 + 2 α n u z , x n + 1 z ( 1 α n ) x n z 2 + 2 α n u z , x n + 1 z ( 1 α n ) ρ n ( 4 ρ n ) f n 2 ( x n ) f n ( x n ) 2 .
(3.6)
Now we prove ( x n ) is bounded. Indeed, we have from (3.6) that
x n + 1 z 2 ( 1 α n ) x n z 2 + 2 α n u z , x n + 1 z ( 1 α n ) x n z 2 + 1 4 α n x n + 1 z 2 + 4 α n u z 2 ,
and consequently
x n + 1 z 2 1 α n 1 1 4 α n x n z 2 + 3 4 α n 1 1 4 α n 16 3 u z 2 .
It turns out that
x n + 1 z max { x n z , 16 3 u z } ,
and inductively
x n z max { x 0 z , 16 3 u z } ,
and this means that ( x n ) is bounded. Since α n 0 , with no loss of generality, we may assume that there is σ > 0 so that ρ n ( 4 ρ n ) ( 1 α n ) σ for all n. Setting s n = x n z 2 , from the last inequality of (3.6), we get the following inequality:
s n + 1 s n + α n s n + σ f n 2 ( x n ) f n ( x n ) 2 2 α n u z , x n + 1 z .
(3.7)

Now, following an idea in [28], we prove s n 0 by distinguishing two cases.

Case 1: ( s n ) is eventually decreasing (i.e. there exists k 0 such that s n > s n + 1 holds for all n k ). In this case, ( s n ) must be convergent, and from (3.7) it follows that
σ f n 2 ( x n ) f n ( x n ) 2 M α n + ( s n s n + 1 ) ,
(3.8)
where M > 0 is a constant such that 2 x n + 1 z u z M for all n N . Using the condition (a1), we have from (3.8) that f n 2 ( x n ) f n ( x n ) 2 0 . Thus, to verify that f n ( x n ) 0 , it suffices to show that ( f ( x n ) ) is bounded. In fact, it follows from Lemma 2.1 that (noting that f n ( z ) = 0 due to z S )
f n ( x n ) = f n ( x n ) f n ( z ) A 2 x n z .

This implies that f n ( x n ) is bounded and it yields f n ( x n ) 0 , namely ( I P Q n ) A x n 0 .

Since ∂q is bounded on bounded sets, there exists a constant η > 0 such that ζ n η for all n 0 . From (3.3) and the trivial fact that P Q n ( A x n ) Q n , it follows that
q ( A x n ) ζ n , A x n P Q n ( A x n ) η ( I P Q n ) A x n 0 .
(3.9)
If x ω w ( x n ) , and ( x n k ) is a subsequence of ( x n ) such that x n k x , then the w-lsc of q and (3.9) imply that
q ( A x ) lim inf k q ( A x n k ) 0 .
It turns out that A x Q . Next, we turn to prove x C . For convenience, we set y n : = α n u + ( 1 α n ) ( x n τ n f n ( x n ) ) . In fact, since the P C n is firmly nonexpansive, it concludes that
P C n x n P C n z 2 x n z 2 ( I P C n ) x n 2 .
(3.10)
On the other hand, we have
x n + 1 z 2 = P C n y n P C n z 2 = P C n y n P C n x n + P C n x n P C n z 2 P C n x n P C n z 2 + 2 P C n y n P C n x n , x n + 1 z ,
(3.11)
and
P C n y n P C n x n y n x n = α n u + ( 1 α n ) ( x n τ n f n ( x n ) ) x n α n u x n + τ n f n ( x n ) .
(3.12)
Noting that ( x n ) is bounded, we have from (3.10)-(3.12) that
x n + 1 z 2 x n z 2 ( I P C n ) x n 2 + 2 ( α n u x n + τ n f n ( x n ) ) x n + 1 z x n z 2 ( I P C n ) x n 2 + ( α n + f n ( x n ) f n ( x n ) ) M ,
(3.13)
where M is some positive constant. Clearly, from (3.13), it turns out that
( I P C n ) x n 2 s n s n + 1 + ( α n + f n ( x n ) f n ( x n ) ) M .
(3.14)
Thus, we assert that ( I P C n ) x n 0 due to the fact that s n s n + 1 + ( α n + f n ( x n ) f n ( x n ) ) M 0 . Moreover, by the definition of C n , we obtain
c ( x n ) ξ n , x n P C n ( x n ) δ x n P C n x n 0 ( n ) ,
where δ is a constant such that ξ n δ for all n 0 . The w-lsc of c then implies that
c ( x ) lim inf k c ( x n k ) = 0 .
Consequently, x C , and hence ω w ( x n ) S . Furthermore, due to (2.2), we get
lim sup n u z , x n z = max w ω w ( x n ) u P S u , w P S u 0 .
(3.15)
Taking into account of (3.7), we have
s n + 1 ( 1 α n ) s n + 2 α n u z , x n + 1 z .
(3.16)

Applying Lemma 2.3 to (3.16), we obtain s n 0 .

Case 2: ( s n ) is not eventually decreasing, that is, we can find an integer n 0 such that s n 0 s n 0 + 1 . Now we define
V n : = { n 0 k n : s k s k + 1 } , n > n 0 .
It is easy to see that V n is nonempty and satisfies V n V n + 1 . Let
ψ ( n ) : = max V n , n > n 0 .
It is clear that ψ ( n ) as n (otherwise, ( s n ) is eventually decreasing). It is also clear that s ψ ( n ) s ψ ( n ) + 1 for all n > n 0 . Moreover,
s n s ψ ( n ) + 1 , n > n 0 .
(3.17)
In fact, if ψ ( n ) = n , then the inequity (3.17) is trivial; if ψ ( n ) < n , from the definition of ψ ( n ) , there exists some i N such that ψ ( n ) + i = n , we deduce that
s ψ ( n ) + 1 > s ψ ( n ) + 2 > > s ψ ( n ) + i = s n ,
and the inequity (3.17) holds again. Since s ψ ( n ) s ψ ( n ) + 1 for all n > n 0 , it follows from (3.8) that
σ f ψ ( n ) 2 ( x ψ ( n ) ) f ψ ( n ) ( x ψ ( n ) ) 2 M α ψ ( n ) 0 ,
so that f ψ ( n ) ( x ψ ( n ) ) 0 as n (noting that f ψ ( n ) ( x ψ ( n ) ) is bounded). By the same argument to the proof in case 1, we have ω w ( x ψ ( n ) ) S . On the other hand, noting s ψ ( n ) s ψ ( n ) + 1 again, we have from (3.12) and (3.14) that
x ψ ( n ) x ψ ( n ) + 1 x ψ ( n ) P C ψ ( n ) x ψ ( n ) + P C ψ ( n ) x ψ ( n ) P C ψ ( n ) y ψ ( n ) α ψ ( n ) + f ψ ( n ) ( x ψ ( n ) ) f ψ ( n ) ( x ψ ( n ) ) × ( α ψ ( n ) + f ψ ( n ) ( x ψ ( n ) ) f ψ ( n ) ( x ψ ( n ) ) + 1 ) M ,
where M is a positive constant. Letting n yields that
x ψ ( n ) x ψ ( n ) + 1 0 ,
(3.18)
from which one can deduce that
lim sup n u z , x ψ ( n ) + 1 z = lim sup n u z , x ψ ( n ) z = max w ω w ( x ψ ( n ) ) u P S u , w P S u 0 .
(3.19)
Since s ψ ( n ) s ψ ( n ) + 1 , it follows from (3.7) that
s ψ ( n ) 2 u z , x ψ ( n ) + 1 z , n > n 0 .
(3.20)
Combining (3.19) and (3.20) yields
lim sup n s ψ ( n ) 0 ,
(3.21)
and hence s ψ ( n ) 0 , which together with (3.18) implies that
s ψ ( n ) + 1 ( x ψ ( n ) z ) + ( x ψ ( n ) + 1 x ψ ( n ) ) s ψ ( n ) + x ψ ( n ) + 1 x ψ ( n ) 0 ,
(3.22)

which, together with (3.17), in turn implies that s n 0 , that is, x n z . □

Remark 3.4 Since u can be chosen in H arbitrarily, one can compute the minimum-norm solution of SFP (1.1) where C and Q are given in (3.1) by taking u = 0 in Algorithm 3.2 whether 0 C or 0 C .

Declarations

Acknowledgements

The authors wish to thank the referees for their helpful comments, which notably improved the presentation of this manuscript. This work was supported by the Fundamental Research Funds for the Central Universities (ZXH2012K001) and in part by the Foundation of Tianjin Key Lab for Advanced Signal Processing.

Authors’ Affiliations

(1)
College of Science, Civil Aviation University of China
(2)
Tianjin Key Laboratory for Advanced Signal Processing, Civil Aviation University of China

References

  1. Censor Y, Elfving T: A multiprojection algorithm using Bregman projection in product space. Numer. Algorithms 1994, 8(2–4):221–239.MathSciNetView ArticleGoogle Scholar
  2. López G, Martín-Márquez V, Xu HK: Iterative algorithms for the multipul-sets split feasiblity problem. In Biomedical Mathematics Promising Directions in Imaging, Therapy Planning and Inverse Problems. Edited by: Censor Y, Jiang M, Wang G. Medical Physics Publishing, Madison, WI; 2010:243–279.Google Scholar
  3. Censor Y, Bortfeld T, Martin B, Trofimov A: A unified approach for inversion problems in intensity-modulated radiation therapy. Phys. Med. Biol. 2003, 51: 2353–2365.View ArticleGoogle Scholar
  4. Censor Y, Elfving T, Kopf N, Bortfeld T: The multiple-sets split feasibility problem and its applications for inverse problem. Inverse Probl. 2005, 21: 2071–2084. 10.1088/0266-5611/21/6/017MathSciNetView ArticleGoogle Scholar
  5. López G, Martín-Márquez V, Xu HK: Perturbation techniques for nonexpansive mappings with applications. Nonlinear Anal., Real World Appl. 2009, 10: 2369–2383. 10.1016/j.nonrwa.2008.04.020MathSciNetView ArticleGoogle Scholar
  6. Byrne C: Iterative oblique projection onto convex sets and the split feasibity problem. Inverse Probl. 2002, 18: 441–453. 10.1088/0266-5611/18/2/310MathSciNetView ArticleGoogle Scholar
  7. Dang Y, Gao Y: The strong convergence of a KM-CQ-like algorithm for a split feasibility problem. Inverse Probl. 2011., 27: Article ID 015007Google Scholar
  8. Qu B, Xiu N: A note on the CQ algorithm for the split feasibility problem. Inverse Probl. 2005, 21: 1655–1665. 10.1088/0266-5611/21/5/009MathSciNetView ArticleGoogle Scholar
  9. Schöpfer F, Schuster T, Louis AK: An iterative regularization method for the solution of the split feasibility problem in Banach spaces. Inverse Probl. 2008., 24: Article ID 055008Google Scholar
  10. Wang F, Xu HK: Approximating curve and strong convergence of the CQ algorithm for the split feasibility problem. J. Inequal. Appl. 2010., 2010: Article ID 102085Google Scholar
  11. Xu HK: A variable Krasonosel’skii-Mann algorithm and the multiple-set split feasibility problem. Inverse Probl. 2006, 22: 2021–2034. 10.1088/0266-5611/22/6/007View ArticleGoogle Scholar
  12. Xu HK: Iterative methods for the split feasibility problem in infinite-dimensional Hilbert spaces. Inverse Probl. 2010., 26: Article ID 105018Google Scholar
  13. Yang Q: The relaxed CQ algorithm for solving the split feasibility problem. Inverse Probl. 2004, 20: 1261–1266. 10.1088/0266-5611/20/4/014View ArticleGoogle Scholar
  14. Yang Q: On variable-set relaxed projection algorithm for variational inequalities. J. Math. Anal. Appl. 2005, 302: 166–179. 10.1016/j.jmaa.2004.07.048MathSciNetView ArticleGoogle Scholar
  15. Zhao J, Yang Q: Generalized KM theorems and their applications. Inverse Probl. 2006, 22(3):833–844. 10.1088/0266-5611/22/3/006View ArticleGoogle Scholar
  16. Zhao J, Yang Q: Self-adaptive projection methods for the multiple-sets split feasibility problem. Inverse Probl. 2011., 27: Article ID 035009Google Scholar
  17. Byrne C: A unified treatment of some iterative algorithms in signal processing and image reconstruction. Inverse Probl. 2004, 20(1):103–120. 10.1088/0266-5611/20/1/006MathSciNetView ArticleGoogle Scholar
  18. Figueiredo MA, Nowak RD, Wright SJ: Gradient projection for space reconstruction: application to compressed sensing and other inverse problems. IEEE J. Sel. Top. Signal Process. 2007, 1: 586–598.View ArticleGoogle Scholar
  19. López G, Martín-Márquez V, Wang FH, Xu HK: Solving the split feasibility problem without prior knowledge of matrix norms. Inverse Probl. 2012. doi:10.1088/0266–5611/28/8/085004Google Scholar
  20. Fukushima M: A relaxed projection method for variational inequalities. Math. Program. 1986, 35: 58–70. 10.1007/BF01589441MathSciNetView ArticleGoogle Scholar
  21. Aubin JP: Optima and Equilibria: An Introduction to Nonlinear Analysis. Springer, Berlin; 1993.View ArticleGoogle Scholar
  22. Goebel K, Kirk WA: Topics on Metric Fixed Point Theory. Cambridge University Press, Cambridge; 1990.View ArticleGoogle Scholar
  23. Xu HK: Iterative algorithms for nonlinnear operators. J. Lond. Math. Soc. 2002, 66: 240–256. 10.1112/S0024610702003332View ArticleGoogle Scholar
  24. Bauschke HH, Borwein JM: On projection algorithms for solving convex feasibility problem. SIAM Rev. 1996, 38: 367–426. 10.1137/S0036144593251710MathSciNetView ArticleGoogle Scholar
  25. Suzuki T: A sufficient and necessary condition for Halpern-type strong convergence to fixed points of nonexpansive mappings. Proc. Am. Math. Soc. 2007, 135: 99–106.View ArticleGoogle Scholar
  26. Xu HK: Viscosity approximation methods for nonexpansive mappings. J. Math. Anal. Appl. 2004, 298: 279–291. 10.1016/j.jmaa.2004.04.059MathSciNetView ArticleGoogle Scholar
  27. Halpern B: Fixed points of nonexpanding maps. Bull. Am. Math. Soc. 1967, 73: 957–961. 10.1090/S0002-9904-1967-11864-0View ArticleGoogle Scholar
  28. Maingé PE: A hybrid extragradient-viscosity method for monotone operators and fixed point problems. SIAM J. Control Optim. 2008, 47: 1499–1515. 10.1137/060675319MathSciNetView ArticleGoogle Scholar

Copyright

© He and Zhao; licensee Springer 2013

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.