Skip to main content

Split feasibility and fixed-point problems for asymptotically quasi-nonexpansive mappings

Abstract

The purpose of this paper is to introduce and analyze a weakly convergent theorem by using the regularized method and the relaxed extragradient method for finding a common element of the solution set Γ of the split feasibility problem and Fix(T) of fixed points of asymptotically quasi-nonexpansive mappings T in the setting of infinite-dimensional Hilbert spaces. Consequently, we prove that the sequence generated by the proposed algorithm converges weakly to an element of Fix(T)Γ under mild assumptions.

MSC:47H09, 47J25, 65K10.

1 Introduction

In 1994, Censor and Elfving [1] first introduced the split feasibility problem (SFP) in finite-dimensional Hilbert spaces for modeling inverse problems which arise from phase retrievals and in medical image reconstruction [2]. It was found that the SFP can also be used to model intensity-modulated radiation therapy (IMRT) (see [36]). Very recently, Xu [7] considered the SFP in the framework of infinite-dimensional Hilbert spaces. In this setting, the SFP is formulated as the problem of finding a point x with the property

x CandA x Q,
(1.1)

where C and Q are the nonempty closed convex subsets of the infinite-dimensional real Hilbert spaces H 1 and H 2 , respectively. Let AB( H 1 , H 2 ), where B( H 1 , H 2 ) denotes the family of all bounded linear operators from H 1 to H 2 .

We use Γ to denote the solution set of the SFP, i.e.,

Γ={xC:AxQ}.

Assume that the SFP is consistent (i.e., (1.1) has a solution) so that Γ is closed, convex and nonempty. A special case of the SFP is the following convex constrained linear inverse problem:

find xCsuch that Ax=b,
(1.2)

which has extensively been investigated by using the Landweber iterative method [8]:

let  x 0  be arbitrary for  n = 0 , 1 , ,  let x n + 1 = x n + γ A T ( b A x n ) .

Comparatively, the SFP has received much less attention so far due to the complexity resulting from the set Q. Therefore, whether various versions of the projected Landweber iterative method [8] can be extended to solve the SFP remains an interesting open topic.

The original algorithm given in [1] involves the computation of the inverse A 1 (assuming the existence of the inverse of A):

x k + 1 = A 1 P Q ( P A ( C ) ( A x k ) ) ,k0,

where C,Q R n are closed convex sets, A is a full rank n×n matrix and A(C)={y R n |y=Ax,xC}, and thus has not become popular.

A more popular algorithm that solves the SFP seems to be the CQ algorithm of Byrne [2, 9] which is found to be a gradient-projection method (GPM) in convex minimization. It is also a special case of the proximal forward-backward splitting method [10]. The CQ algorithm only involves the computations of the projections P C and P Q onto the sets C and Q, respectively, and is therefore implementable in the case where P C and P Q have closed-form expressions (for example, C and Q are closed balls or half-spaces). It remains, however, a challenge on the CQ algorithm in the case where the projection P C and/or P Q fail to have closed-form expressions though theoretically we can prove the (weak) convergence of the algorithm.

Recently, Xu [7] gave a continuation of the study on the CQ algorithm and its convergence. He applied Mann’s algorithm to the SFP and proposed an averaged CQ algorithm, which was proved to be weakly convergent to a solution of the SFP. He derived a weak convergence result, which shows that for suitable choices of iterative parameters (including the regularization), the sequence of iterative solutions can converge weakly to an exact solution of the SFP. He also established the strong convergence result, which shows that the minimum-norm solution can be obtained. Later, Deepho and Kumam [11] extended the results of Xu [7] by introducing and studying the modified Halpern iterative scheme for solving the split feasibility problem (SFP) in the setting of infinite-dimensional Hilbert spaces.

Throughout this paper, we always assume that the SFP is consistent, that is, the solution set Γ of the SFP is nonempty. Let f: H 1 R be a continuous differentiable function. The minimization problem

min x C f(x):= 1 2 A x P Q A x 2
(1.3)

is ill-posed. Therefore (see [7]), consider the following Tikhonov regularized problem:

min x C f α (x):= 1 2 A x P Q A x 2 + 1 2 α x 2 ,
(1.4)

where α>0 is the regularization parameter.

We observe that the gradient

f α (x)=f(x)+αI= A (I P Q )A+αI
(1.5)

is (α+ A 2 )-Lipschitz continuous and α-strongly monotone.

Define the Picard iterates

x n + 1 α = P C ( I γ ( A ( I P Q ) A + α I ) ) x n α .
(1.6)

Xu [7] showed that if SFP (1.1) is consistent, then as n, x n α x α and consequently the strong lim α 0 x α exists and is the minimum-norm solution of the SFP. Note that (1.6) is double-step iteration. Xu [7] further suggested the following single step regularized method:

x n + 1 = P C (Iγ f α n ) x n = P C ( ( 1 α n γ n ) x n γ n A ( I P Q ) A x n ) .
(1.7)

He proved that the sequence { x n } generated by (1.7) converges in norm to the minimum-norm solution of the SFP provided the parameters { α n } and { γ n } satisfy the following conditions:

  1. (i)

    α n 0 and 0< γ n < α n A 2 + α n ;

  2. (ii)

    n α n γ n =;

  3. (iii)

    | γ n + 1 γ n | + γ n | α n + 1 α n | ( α n + 1 γ n + 1 ) 2 0.

Motivated by the idea of the relaxed extragradient method and Xu’s regularization, Ceng, Ansari and Yao [12] presented the following relaxed extragradient method with regularization for finding a common element of the solution set of the split feasibility problem and the set Fix(S) of fixed points of a nonexpansive mapping S:

{ x 0 = x C chosen arbitrarily , y n = ( 1 β n ) + β n P C ( x n λ f α n ( x n ) ) , x n + 1 = γ n x n + ( 1 γ n ) S P C ( y n λ f α n ( y n ) ) , n 0 .
(1.8)

They only obtained the weak convergence of iterative algorithm (1.8).

The purpose of this paper to study and analyze an relaxed extragradient method with regularization for finding a common element of the solution set Γ of the SFP and the set solutions of fixed points for asymptotically quasi-nonexpansive mappings and a Lipschitz continuous mapping in a real Hilbert space. We prove that the sequence generated by the proposed method converges weakly to an element x ˆ in Fix(T)Γ.

2 Preliminaries

We first recall some definitions, notations, and conclusions which will be needed in proving our main results. Let H be a real Hilbert space with the inner product , and and let C be a closed and convex subset of H. Let E be a Banach space. A mapping T:EE is said to be demi-closed at origin if for any sequences { x n }E with x n x and (IT) x n 0, x =T x . A Banach space E is said to have the Opial property if for any sequence { x n } with x n x ,

lim inf n x n x < lim inf n x n y,yE with y x .

Remark 2.1 It is well known that each Hilbert space possesses the Opial property.

Definition 2.2 Let H be a real Hilbert space, let C be a nonempty and closed convex subset. We denote by Fix(T) the set of fixed points of T, that is, Fix(T)={xC:x=Tx}. Then T is said to be

  1. (i)

    nonexpansive ifTxTyxy for all x,yC;

  2. (ii)

    quasi-nonexpansive ifTxpxp for all xC and pF(T);

  3. (iii)

    asymptotically nonexpansive if there exist a sequence k n 1 and lim n k n =1 such that

    T n x T n y k n xy

for all x,yC and n1;

  1. (iv)

    asymptotically quasi-nonexpansive if there exist a sequence k n 1 and lim n k n =1 such that

    T n x p k n xp

for all xC, pF(T) and n1;

  1. (v)

    uniformlyL-Lipschitzian if there exists a constant L>0 such that

    T n x T n y Lxy

for all x,yC and n1.

Remark 2.3 By the above definitions, it is clear that:

  1. (i)

    a nonexpansive mapping is an asymptotically quasi-nonexpansive mapping;

  2. (ii)

    a quasi-nonexpansive mapping is an asymptotically-quasi nonexpansive mapping;

  3. (iii)

    an asymptotically nonexpansive mapping is an asymptotically quasi-nonexpansive mapping.

Proposition 2.4 (see [9])

We have the following assertions.

  1. (i)

    Tis nonexpansive if and only if the complementITis 1 2 -ism.

  2. (ii)

    IfTisν-ism andγ>0, thenγTis ν γ -ism.

  3. (iii)

    Tis averaged if and only if the complementITisν-ism for someν> 1 2 .

Indeed, forα(0,1), Tisα-averaged if and only ifITis 1 2 α -ism.

Proposition 2.5 (see [9, 13])

We have the following assertions.

  1. (i)

    IfT=(1α)S+αVfor someα(0,1), Sis averaged andVis nonexpansive, thenTis averaged.

  2. (ii)

    Tis firmly nonexpansive if and only if the complementITis firmly nonexpansive.

  3. (iii)

    IfT=(1α)S+αVfor someα(0,1), Sis firmly nonexpansive andVis nonexpansive, thenTis averaged.

  4. (iv)

    The composite of finite many averaged mappings is averaged. That is, if each of the mappings { T i } i = 1 n is averaged, then so is the composite T 1 T 2 T N . In particular, if T 1 is α 1 -averaged and T 2 is α 2 -averaged, where α 1 , α 2 (0,1), then the composite T 1 T 2 isα-averaged, whereα= α 1 + α 2 α 1 α 2 .

  5. (v)

    If the mappings { T i } i = 1 n are averaged and have a common fixed point, then

    i = 1 n Fix( T i )=Fix( T 1 T N ).

Lemma 2.6 (see [14], demiclosedness principle)

LetCbe a nonempty closed and convex subset of a real Hilbert spaceHand letS:CCbe a nonexpansive mapping withFix(S). If the sequence{ x n }Cconverges weakly toxand the sequence{(IS) x n }converges strongly toy, then(IS)x=y; in particular, ify=0, thenxFix(S).

Lemma 2.7 (see [15])

Let the sequences { a n } and { u n } of real numbers satisfy

a n + 1 (1+ u n ) a n ,n1,

where a n 0, u n 0and n = 1 u n <. Then

  1. (1)

    lim n a n exists;

  2. (2)

    if lim inf n a n =0, then lim n a n =0.

The following lemma gives some characterizations and useful properties of the metric projection P C in a Hilbert space.

For every point xH, there exists a unique nearest point in C, denoted by P C x, such that

x P C xxy,yC,
(2.1)

where P C is called the metric projection ofH onto C. We know that P C is a nonexpansive mapping of H onto C.

Proposition 2.8For givenxHandzC:

  1. (i)

    z= P C xif and only ifxz,yz0for allyC.

  2. (ii)

    z= P C xif and only if x z 2 x y 2 y z 2 for allyC.

  3. (iii)

    For allyH, P C x P C y,xy P C x P C y 2 .

Lemma 2.9 (see [16])

LetHbe a real Hilbert space. Then the following equations hold:

  1. (i)

    x y 2 = x 2 y 2 2xy,yfor allx,yH;

  2. (ii)

    t x + ( 1 t ) y 2 =t x 2 +(1t) y 2 t(1t) x y 2 for allt[0,1]andx,yH.

Let K be a nonempty closed convex subset of a real Hilbert space H and let F:KH be a monotone mapping. The variational inequality problem (VIP) is to find xK such that

Fx,yx0,yK.

The solution set of the VIP is denoted by VIP(K,F). It is well known that

xVI(K,F)x= P K (xλFx),λ>0.

A set-valued mapping T:H 2 H is called monotone if for all x,yH, fTx and gTy imply

xy,fg0.

A monotone mapping T:H 2 H is called maximal if its graph G(T) is not properly contained in the graph of any other monotone mapping. It is known that a monotone mapping T is maximal if and only if for (x,f)H×H, xy,fg0 for every (y,g)G(T) implies fTx. Let F:KH be a monotone and k-Lipschitz continuous mapping and let N K v be the normal cone to K at vK, that is,

N K v= { w H : v u , w 0 , u K } .

Define

Tv={ F v + N K v if  v K , if  v K .

Then T is maximal monotone and 0Tv if and only if vVI(K,F); see [15] for more details.

We can use fixed point algorithms to solve the SFP on the basis of the following observation.

Let λ>0 and assume that x Γ. Then A x Q, which implies that (I P Q )A x =0, and thus λ A (I P Q )A x =0. Hence, we have the fixed point equation (Iλ A (I P Q )A) x = x . Requiring that x C, we consider the fixed point equation

P C (Iλf) x = P C ( I λ A ( I P Q ) A ) x = x .
(2.2)

It is proved in [[7], Proposition 3.2] that the solutions of fixed point equation (2.2) are exactly the solutions of the SFP; namely, for given x H 1 , x solves the SFP if and only if x solves fixed point equation (2.2).

Proposition 2.10 (see [12])

Given x H 1 , the following statements are equivalent.

  1. (i)

    x solves the SFP;

  2. (ii)

    x solves fixed point equation (2.2);

  3. (iii)

    x solves the variational inequality problem (VIP) of finding x Csuch that

    f ( x ) , x x 0,xC,
    (2.3)

wheref= A (I P Q )Aand A is the adjoint ofA.

Proof (i) (ii). See the proof in [[7], Proposition 3.2].

  1. (ii)

    (iii). Observe that

    P C ( I λ A ( I P Q ) A ) x = x ( I λ A ( I P Q ) A ) x x , x x 0 , x C λ A ( I P Q ) A x , x x 0 , x C f ( x ) , x x 0 , x C ,

where f= A (I P Q )A. □

Remark 2.11 It is clear from Proposition 2.10 that

Γ:=Fix ( P C ( I λ f ) ) =VI(C,f)

for any λ>0, where Fix( P C (Iλf)) and VI(C,f) denote the set of fixed points of P C (Iλf) and the solution set of VIP.

3 Main result

Theorem 3.1LetCbe a nonempty, closed, and convex subset of a real Hilbert spaceHand letT:CCbe a uniformlyL-Lipschitzian and asymptotically quasi-nonexpansive mappings withFix(T)Γand{ k n }[1,)for allnNsuch that n = 1 ( k n 1)<. Let{ x n }and{ y n }be the sequences inCgenerated by the following algorithm:

{ x 0 = x C chosen arbitrarily , y n = P C ( I λ n f α n ) x n , x n + 1 = β n x n + ( 1 β n ) T n y n , n 0 ,
(3.1)

where f α n =f+ α n I= A (I P Q )A+ α n I, and three sequences{ α n }, { λ n }, and{ β n }satisfy the conditions:

  1. (i)

    n = 1 α n <,

  2. (ii)

    { λ n }[a,b]for somea,b(0, 1 A 2 )and n = 1 | λ n + 1 λ n |<,

  3. (iii)

    { β n }[c,d]for somec,d(0,1).

Then the sequences{ x n }and{ y n }converge weakly to an element x ˆ Fix(T)Γ.

Proof We first show that P C (Iλ f α ) is ζ-averaged for each λ n (0, 2 α + A 2 ), where

ζ= 2 + λ ( α + A 2 ) 4 .

Indeed, it is easy to see that f= A (I P Q )A is 1 A 2 -ism, that is,

f ( x ) f ( y ) , x y 1 A 2 f ( x ) f ( y ) 2 .

Observe that

( α + A 2 ) f α ( x ) f α ( y ) , x y = ( α + A 2 ) [ α x y 2 + f ( x ) f ( y ) , x y ] = α 2 x y 2 + α f ( x ) f ( y ) , x y + α A 2 x y 2 + A 2 f ( x ) f ( y ) , x y α 2 x y 2 + 2 α f ( x ) f ( y ) , x y + f ( x ) f ( y ) 2 = α ( x y ) + f ( x ) f ( y ) 2 = f ( x ) f ( y ) 2 .

Hence, it follows that f α =αI+ A (I P Q )A is 1 α + A 2 -ism. Thus, λ f α is 1 λ ( α + A 2 ) -ism. By Proposition 2.4(iii) the composite (Iλ f α ) is λ ( α + A 2 ) 2 -averaged. Therefore, noting that P C is 1 2 -averaged and utilizing Proposition 2.5(iv), we know that for each λ(0, 2 α + A 2 ), P C (Iλ f α ) is ζ-averaged with

ζ= 1 2 + λ ( α + A 2 ) 2 1 2 λ ( α + A 2 ) 2 = 2 + λ ( α + A 2 ) 4 (0,1).

This shows that P C (Iλ f α ) is nonexpansive. Furthermore, for { λ n }[a,b] with a,b(0, 1 A 2 ), utilizing the fact that lim n 1 α n + A 2 = 1 A 2 , we may assume that

0<a λ n b< 1 A 2 ,n0.

Consequently, it follows that for each integer n0, P C (I λ n f α n ) is ζ n -averaged with

ζ n = 1 2 + λ n ( α n + A 2 ) 2 1 2 λ n ( α n + A 2 ) 2 = 2 + λ n ( α n + A 2 ) 4 (0,1).

This immediately implies that P C (I λ n f α n ) is nonexpansive for all n0.

We divide the remainder of the proof into several steps.

Step 1. We prove that { x n } is bounded. Indeed, we take a fixed pFix(T)Γ arbitrarily. Then we get P C (I λ n f)p=p for λ n (0, 2 A 2 ). Since P C and (I λ n f α n ) are nonexpansive mappings, then we have

y n p = P C ( I λ n f α n ) x n P C ( I λ n f ) p P C ( I λ n f α n ) x n P C ( I λ n f α n ) p + P C ( I λ n f α n ) p P C ( I λ n f ) p x n p + ( I λ n f α n ) p ( I λ n f ) p = x n p + λ n f p λ n f α n p = x n p + λ n f p f α n p = x n p + λ n f p f p α n p x n p + α n λ n p .
(3.2)

Observe that

x n + 1 p = β n x n + ( 1 β n ) T n y n p β n x n p + ( 1 β n ) T n y n p β n x n p + ( 1 β n ) k n y n p β n x n p + ( 1 β n ) k n ( x n p + λ n α n p ) = β n x n p + ( 1 β n ) k n x n p + ( 1 β n ) k n α n λ n p = ( 1 + ( k n 1 ) ( 1 β n ) ) x n p + ( 1 β n ) k n α n λ n p .
(3.3)

Since n = 1 ( k n 1)<, according to Lemma 2.7 and (i), (ii) and (3.3), we obtain that

lim n x n p exists for each pFix(T)Γ.
(3.4)

This implies that { x n } is bounded and { y n } is also bounded.

It follows that

T n x n p k n x n p.

Hence { T n x n p} is bounded.

Step 2. We prove that

lim n y n T y n =0.

In fact, it follows from (3.2) that

y n p 2 = ( x n p + α n λ n p ) 2 x n p 2 + 2 α n λ n p x n p + α n 2 λ n 2 p 2 = x n p 2 + α n ( 2 λ n p x n p + α n λ n 2 p 2 ) = x n p 2 + α n M ,

where M= sup n 0 {2 λ n p x n p+ α n λ n 2 p 2 }<.

It follows that

T n y n p 2 ( k n y n p ) 2 = k n 2 y n p 2 = k n 2 x n p 2 + α n k n 2 M .

Also, observe that

x n + 1 p 2 = β n x n + ( 1 β n ) T n y n p 2 β n x n p 2 + ( 1 β n ) T n y n p 2 β n ( 1 β n ) T n y n x n 2 β n x n p 2 + ( 1 β n ) ( k n 2 x n p 2 + α n k n 2 M ) β n ( 1 β n ) T n y n x n 2 = β n x n p 2 + ( 1 β n ) k n 2 x n p 2 + ( 1 β n ) k n 2 α n M β n ( 1 β n ) T n y n x n 2 = ( k n 2 β n ( k n 2 1 ) ) x n p 2 + ( 1 β n ) k n 2 α n M β n ( 1 β n ) T n y n x n 2 .

Hence, we have

β n ( 1 β n ) T n y n x n 2 ( k n 2 β n ( k n 2 1 ) ) x n p 2 x n + 1 p 2 + ( 1 β n ) k n 2 α n M .
(3.5)

By the conditions (i), (iii) and lim n k n =1, we can conclude that

lim n T n y n x n =0.
(3.6)

Consider that since y n = P C ( x n λ n f α n x n ) and by Proposition 2.8(ii), we have

y n p 2 x n λ n f α n ( x n ) p 2 x n λ n f α n ( x n ) y n 2 = x n p 2 x n y n 2 + 2 λ n f α n ( x n ) , p y n = x n p 2 x n y n 2 + 2 λ n ( f α n ( x n ) f α n ( p ) , p x n + f α n ( p ) , p x n + f α n ( x n ) , x n y n ) x n p 2 x n y n 2 + 2 λ n ( f α n ( p ) , p x n + f α n ( x n ) , x n y n ) = x n p 2 x n y n 2 + 2 λ n [ ( α n I + f ) p , p x n + f α n ( x n ) , x n y n ] = x n p 2 x n y n 2 + 2 λ n [ α n p , p x n + f α n ( x n ) , x n y n ] = x n p 2 x n y n 2 + 2 λ n α n p , p x n + 2 λ n f α n ( x n ) , x n y n = x n p 2 x n y n 2 + 2 λ n α n p , p x n 2 λ n f α n ( x n ) , y n x n = x n p 2 x n y n 2 + 2 λ n α n p , p x n 2 λ n f α n ( x n ) , y n p + p x n = x n p 2 x n y n 2 + 2 λ n α n p , p x n 2 λ n f α n ( x n ) , y n p 2 λ n f α n ( x n ) , p x n x n p 2 x n y n 2 + 2 λ n α n p p x n 2 λ n f α n ( x n ) y n p 2 λ n f α n ( x n ) p x n .
(3.7)

Consequently, utilizing Lemma 2.9(ii) and (3.7), we conclude that

x n + 1 p 2 = β n x n + ( 1 β n ) T n y n p 2 = β n x n + ( 1 β n ) T n y n ( β n + ( 1 β n ) ) p 2 = β n x n + ( 1 β n ) T n y n β n p ( 1 β n ) p 2 = β n ( x n p ) + ( 1 β n ) ( T n y n p ) 2 = β n x n p 2 + ( 1 β n ) T n y n p 2 β n ( 1 β n ) x n T n y n 2 β n x n p 2 + ( 1 β n ) k n 2 y n p 2 β n ( 1 β n ) x n T n y n 2 = β n x n p 2 + ( 1 β n ) k n 2 [ x n p 2 x n y n 2 + 2 λ n α n p p x n 2 λ n f α n ( x n ) y n p 2 λ n f α n ( x n ) p x n ] β n ( 1 β n ) x n T n y n 2 = ( β n + ( 1 β n ) k n 2 ) x n p 2 ( 1 β n ) k n 2 x n y n 2 + 2 ( 1 β n ) k n 2 λ n α n p p x n 2 ( 1 β n ) k n 2 λ n f α n ( x n ) y n p 2 ( 1 β n ) k n 2 λ n f α n ( x n ) p x n β n ( 1 β n ) x n T n y n 2 = ( k n 2 β n ( k n 2 1 ) ) x n p 2 ( 1 β n ) k n 2 x n y n 2 + 2 ( 1 β n ) k n 2 λ n α n p p x n 2 ( 1 β n ) k n 2 λ n f α n ( x n ) y n p 2 ( 1 β n ) k n 2 λ n f α n ( x n ) p x n β n ( 1 β n ) x n T n y n 2 .

It follows that we get

( 1 β n ) k n 2 x n y n 2 2 ( 1 β n ) k n 2 λ n α n p p x n + 2 ( 1 β n ) k n 2 λ n f α n ( x n ) ( y n p + p x n ) + β n ( 1 β n ) x n T n y n 2 ( k n 2 β n ( k n 2 1 ) ) x n p 2 x n + 1 p 2 .
(3.8)

So, taking n, since lim n 0 k n =1, (i)-(iii), (3.6) and (3.8), we can conclude that

lim n 0 y n x n =0.
(3.9)

Consider

x n + 1 x n = β n x n x n + ( 1 β n ) T n y n = ( 1 β n ) x n + ( 1 β n ) T n y n ( 1 β n ) T n y n x n .
(3.10)

From (3.6) we obtain

x n + 1 x n (1 β n ) T n y n x n 0(as n).
(3.11)

Observe that

T n y n y n = T n y n x n + x n y n T n y n x n + x n y n .

So, from (3.6) and (3.9), we get

lim n T n y n y n =0.
(3.12)

We compute that

y n + 1 y n = P C ( x n + 1 λ n + 1 f α n + 1 x n + 1 ) P C ( x n λ n f α n x n ) = P C ( I λ n + 1 f α n + 1 ) x n + 1 P C ( I λ n f α n ) x n P C ( I λ n + 1 f α n + 1 ) x n + 1 P C ( I λ n + 1 f α n + 1 ) x n + P C ( I λ n + 1 f α n + 1 ) x n P C ( I λ n f α n ) x n x n + 1 x n + ( I λ n + 1 f α n + 1 ) x n ( I λ n f α n ) x n = x n + 1 x n + x n λ n + 1 f α n + 1 x n ( x n λ n f α n x n ) = x n + 1 x n + λ n f α n x n λ n + 1 f α n + 1 x n = x n + 1 x n + λ n ( f + α n ) x n λ n + 1 ( f + α n + 1 ) x n = x n + 1 x n + λ n f x n + λ n α n x n ( λ n + 1 f x n + λ n + 1 α n + 1 x n ) = x n + 1 x n + ( λ n λ n + 1 ) f x n + λ n α n x n λ n + 1 α n + 1 x n = x n + 1 x n + ( λ n λ n + 1 ) f x n + λ n α n x n λ n α n + 1 x n + λ n α n + 1 x n λ n + 1 α n + 1 x n = x n + 1 x n + ( λ n λ n + 1 ) f x n + λ n ( α n α n + 1 ) x n + ( λ n λ n + 1 ) α n + 1 x n x n + 1 x n + | λ n λ n + 1 | f x n + λ n | α n α n + 1 | x n + α n + 1 | λ n λ n + 1 | x n .

From the conditions (i), (ii) and (3.11), we obtain that

y n + 1 y n 0(as n).
(3.13)

Since T is uniformly L-Lipschitzian continuous, then

y n T y n y n y n + 1 + y n + 1 T n + 1 y n + 1 + T n + 1 y n + 1 T n + 1 y n + T n + 1 y n T y n y n y n + 1 + y n + 1 T n + 1 y n + 1 + L y n y n + 1 + L T n y n y n .

Since lim n y n + 1 y n =0 and lim n y n T n y n =0, it follows that

lim n y n T y n =0.
(3.14)

Step 3. We show that x ˆ Fix(T)Γ.

Since f= A (I P Q )A is Lipschitz continuous, from (3.9), we have

lim n f ( x n ) f ( y n ) =0.

Since { x n } is bounded, there is a subsequence { x n i } of { x n } that converges weakly to some  x ˆ .

First, we show that x ˆ Γ. Since x n y n 0, it is known that y n i x ˆ .

Put

S w 1 ={ f w 1 + N C w 1 if  w 1 C , if  w 1 C ,

where N C w 1 ={z H 1 : w 1 u,z0,uC}. Then S is maximal monotone and 0S w 1 if and only if w 1 VI(C,f); (see [17]) for more details. Let ( w 1 ,z)G(S), we have

zS w 1 =f w 1 + N C w 1 ,

and hence

zf w 1 N C w 1 .

So, we have

w 1 u,zf w 1 0,uC.

On the other hand, from

y n = P C (I λ n f α n ) x n and w 1 C,

we have

x n λ n f α n x n y n , y n w 1 0,

and

w 1 y n , y n x n λ n + f α n x n 0.

Therefore, from zf w 1 N C w 1 and y n i C, it follows that

w 1 y n i , z w 1 y n i , f w 1 w 1 y n i , f w 1 w 1 y n i , y n i x n i λ n i + f α n i x n i = w 1 y n i , f w 1 w 1 y n i , y n i x n i λ n i + f x n i α n i w 1 y n i , x n i = w 1 y n i , f w 1 f y n i + w 1 y n i , f y n i f x n i w 1 y n i , y n i x n i λ n i α n i w 1 y n i , x n i w 1 y n i , f y n i f x n i w 1 y n i , y n i x n i λ n i α n i w 1 y n i , x n i .

Hence, we obtain

w 1 x ˆ ,z0as i.

Since S is maximal monotone, we have x ˆ S 1 0, and hence x ˆ VI(C,f). Thus, it is clear that x ˆ Γ.

Next, we show that x ˆ Fix(T). Indeed, since y n i x ˆ and y n i T y n i 0 by (3.14) and Lemma 2.6, we get x ˆ Fix(T). Therefore, we have x ˆ Fix(T)Γ.

Let { x n j } be another subsequence of { x n } such that { x n j } x ¯ . Then x ¯ Fix(T)Γ. Let us show that x ˆ = x ¯ . Assume that x ˆ x ¯ . From the Opial condition [18], we have

lim n x n x ˆ = lim n i inf x n i x ˆ < lim n i inf x n i x ¯ = lim n x n x ¯ = lim n j inf x n j x ¯ < lim n j inf x n j x ˆ = lim n x n x ˆ .

This is a contradiction. Thus, we have x ˆ = x ¯ . This implies

x n x ˆ Fix(T)Γ.

Further, from x n y n 0, it follows that y n x ˆ . This shows that both sequences { x n } and { y n } converge weakly to x ˆ Fix(T)Γ. This completes the proof. □

Utilizing Theorem 3.1, we have the following new results in the setting of real Hilbert spaces.

Take T n I(identitymappings) in Theorem 3.1. Therefore the conclusion follows.

Corollary 3.2LetCbe a nonempty, closed, and convex subset of a real Hilbert spaceH. Suppose thatΓ. Let{ x n }be a sequence inCgenerated by the following algorithm:

{ x 0 = x C chosen arbitrarily , x n + 1 = β n x n + ( 1 β n ) P C ( I λ n f α n ) x n , n 0 ,
(3.15)

where f α n =f+ α n I= A (I P Q )A+ α n I, and the sequences{ α n }, { λ n }, and{ β n }satisfy the conditions:

  1. (i)

    n = 1 α n <,

  2. (ii)

    { λ n }[a,b]for somea,b(0, 1 A 2 )and n = 1 | λ n + 1 λ n |<,

  3. (iii)

    { β n }[c,d]for somec,d(0,1).

Then the sequence{ x n }converges weakly to an element x ˆ Γ.

Take P C I(identitymappings) in Theorem 3.1. Therefore the conclusion follows.

Corollary 3.3LetCbe a nonempty, closed, and convex subset of a real Hilbert spaceHand letT:CCbe a uniformlyL-Lipschitzian and quasi-nonexpansive mapping withFix(T)and{ k n }[1,)for allnNsuch that n = 1 ( k n 1)<. Let{ x n }be the sequence inCgenerated by the following algorithm:

{ x 0 = x C chosen arbitrarily , x n + 1 = β n x n + ( 1 β n ) T n x n , n 0 ,
(3.16)

and let the sequence{ β n }satisfy the condition{ β n }[c,d]for somec,d(0,1). Then the sequence{ x n }converges weakly to an element x ˆ Fix(T).

Remark 3.4 Theorem 3.1 improves and extends [[7], Theorem 5.7] in the following aspects:

  1. (a)

    The iterative algorithm [[7], Theorem 5.7] is extended for developing our relaxed extragradient algorithm with regularization in Theorem 3.1.

  2. (b)

    The technique of proving weak convergence in Theorem 3.1 is different from that in [[7], Theorem 5.7] because of our technique to use asymptotically quasi-nonexpansive mappings and the property of maximal monotone mappings.

  3. (c)

    The problem of finding a common element of Fix(T)Γ for asymptotically quasi-nonexpansive mappings which is more general than that for nonexpansive mappings and the problem of finding a solution of the SFP in [[7], Theorem 5.7].

References

  1. Censor Y, Elving T: A multiprojection algorithm using Bregman projections in product space. Numer. Algorithms 1994, 8: 221–239. 10.1007/BF02142692

    Article  MathSciNet  MATH  Google Scholar 

  2. Byrne C: Iterative oblique projection onto convex subsets and the split feasibility problem. Inverse Probl. 2002, 18: 441–453. 10.1088/0266-5611/18/2/310

    Article  MathSciNet  MATH  Google Scholar 

  3. Censor Y, Bortfeld T, Martin B, Trofimov A: A unified approach for inversion problems in intensity-modulated radiation therapy. Phys. Med. Biol. 2006, 51: 2353–2365. 10.1088/0031-9155/51/10/001

    Article  Google Scholar 

  4. Censor Y, Elving T, Kopf N, Bortfeld T: The multiple-set split feasibility problem and its applications for inverse problem. Inverse Probl. 2005, 21: 2071–2084. 10.1088/0266-5611/21/6/017

    Article  MATH  Google Scholar 

  5. Censor Y, Motova A, Segal A: Perturbed projections and subgradient projections for the multiple-set split feasibility problem. J. Math. Anal. Appl. 2007, 327: 1244–1256. 10.1016/j.jmaa.2006.05.010

    Article  MathSciNet  MATH  Google Scholar 

  6. Censor Y, Segal A: Iterative projection methods in biomedical inverse problem. In Mathematical Methods in Biomedical Imaging and Intensity-Modulate Therapy, IMRT. Edited by: Censor Y, Jiang M, Louis AK. Edizioni della Normale, Pisa; 2008:65–96.

    Google Scholar 

  7. Xu HK: Iterative methods for the split feasibility problem in infinite-dimensional Hilbert spaces. Inverse Probl. 2010., 26(10): Article ID 105018

  8. Landweber L: An iterative formula for Fredholm integral equations of the first kind. Am. J. Math. 1951, 73: 615–625. 10.2307/2372313

    Article  MathSciNet  MATH  Google Scholar 

  9. Byrne C: A unified treatment of some iterative algorithms in signal processing and image reconstruction. Inverse Probl. 2004, 26: 103–120.

    Article  MathSciNet  MATH  Google Scholar 

  10. Combettes PL, Wajs V: Signal recovery by proximal forward-backward splitting multiscale model. Simulation 2005, 4: 1168–1200.

    MathSciNet  MATH  Google Scholar 

  11. Deepho J, Kumam P: A modified Halpern’s iterative scheme for solving split feasibility problems. Abstr. Appl. Anal. 2012., 2012: Article ID 876069

    Google Scholar 

  12. Ceng LC, Ansari QH, Yao JC: Relaxed extragradient method for finding minimum-norm solutions of the split feasibility problems. Nonlinear Anal. 2012, 75(4):2116–2125. 10.1016/j.na.2011.10.012

    Article  MathSciNet  MATH  Google Scholar 

  13. Combettes PL: Solving monotone inclusions via compositions of nonexpansive averaged operator. Optimization 2004, 53(5–6):475–504. 10.1080/02331930412331327157

    Article  MathSciNet  MATH  Google Scholar 

  14. Geobel K, Kirk WA Cambridge Studies in Advanced Mathematics 28. In Topics in Metric Fixed Point Theory. Cambridge University Press, Cambridge; 1990.

    Chapter  Google Scholar 

  15. Tan KK, Xu HK: Approximating fixed points of nonexpansive mappings by the Ishikawa iteration process. J. Math. Anal. Appl. 1993, 178: 301–308. 10.1006/jmaa.1993.1309

    Article  MathSciNet  MATH  Google Scholar 

  16. Marino G, Xu HK: Weak and strong convergence theorems for strict pseudo-contractions in Hilbert space. J. Math. Anal. Appl. 2007, 329: 336–346. 10.1016/j.jmaa.2006.06.055

    Article  MathSciNet  MATH  Google Scholar 

  17. Rockafellar RT: On the maximality of sums of nonlinear monotone operators. Trans. Am. Math. Soc. 1970, 149: 75–88. 10.1090/S0002-9947-1970-0282272-5

    Article  MathSciNet  MATH  Google Scholar 

  18. Opial Z: Weak convergence of the sequence of successive approximations for nonexpansive mappings. Bull. Am. Math. Soc. 1967, 73: 591–597. 10.1090/S0002-9904-1967-11761-0

    Article  MathSciNet  MATH  Google Scholar 

Download references

Acknowledgements

The authors thank the referee for comments and suggestions on this manuscript. The first author was supported by the Thailand Research Fund through the Royal Golden Jubilee Ph.D. Program (Grant No. PHD/0033/2554) and the King Mongkut’s University of Technology Thonburi (KMUTT). Moreover, this study was supported by the Higher Education Research Promotion and National Research University Project of Thailand, Office of the Higher Education Commission (Under Grant No. NRU56000508).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Poom Kumam.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

All authors contributed equally and significantly in writing this paper. All authors read and approved the final manuscript.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Deepho, J., Kumam, P. Split feasibility and fixed-point problems for asymptotically quasi-nonexpansive mappings. J Inequal Appl 2013, 322 (2013). https://doi.org/10.1186/1029-242X-2013-322

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1029-242X-2013-322

Keywords