Skip to main content

Conversion of algorithms by releasing projection for minimization problems

Abstract

The projection methods for solving the minimization problems have been extensively considered in many practical problems, for example, the least-square problem. However, the computational difficulty of the projection might seriously affect the efficiency of the method. The purpose of this paper is to construct two algorithms by releasing projection for solving the minimization problem with the feasibility sets such as the set of fixed points of nonexpansive mappings and the solution set of the equilibrium problem.

MSC:47J05, 47J25, 47H09, 65J15.

1 Introduction

In the present paper, our main purpose is to solve the following minimization problem of finding x such that

x = min x Fix ( S ) EPA x,
(1.1)

where Fix(S) is the set of fixed points of nonexpansive mapping S and EPA is the solution set of the following equilibrium problem:

Find zCsuch thatF(z,y)+Az,yz0,yC,
(1.2)

where C is a nonempty closed convex subset of a real Hilbert space H, F:C×CR is a bifunction and A:CH is an α-inverse-strongly monotone mapping. The reasons why we focus on the above minimization problem (1.1) are mainly in two respects.

Reason 1 This problem is motivated by the following least-square problem:

{ B x = b , x Ω ,
(1.3)

where Ω is a nonempty closed convex subset of a real Hilbert space H, B is a bounded linear operator from H to another real Hilbert space H 1 , B is the adjoint of B and b is a given point in H 1 . The least-squares solution to (1.3) is the least-norm minimizer of the minimization problem

min x Ω Bxb.
(1.4)

For some related works, please see Reich and Xu [1], Sabharwal and Potter [2], Xu [3] and Yao et al. [4].

Reason 2 The problem (1.2) is very general in the sense that it includes optimization problems, variational inequalities, minimax problems and the Nash equilibrium problem in noncooperative games as special cases. At the same time, fixed point algorithms for non-expansive mappings have received vast investigations due to their extensive applications in a variety of applied areas of the inverse problem, partial differential equations, image recovery and signal processing.

Based on the above facts, it is an interesting topic to construct algorithms for solving the above problems. Now we next briefly review some historic approaches which relate to the problems (1.2) and (1.4).

For solving the equilibrium problem, Combettes and Hirstoaga [5] introduced an iterative algorithm of finding the best approximation to the initial data and proved a strong convergence theorem. Moudafi [6] introduced an iterative algorithm and proved a weak convergence theorem. In 2007, Takahashi and Takahashi [7] introduced the following new scheme for finding a common element of the set of solutions of the equilibrium problem and the set of fixed point points of a nonexpansive mapping:

{ F ( u n , y ) + A x n , y u n + 1 λ n y u n , u n x n 0 , y C , x n + 1 = β n x n + ( 1 β n ) S [ α n u + ( 1 α n ) u n ] , n N .

Subsequently, algorithms constructed for solving the equilibrium problems and fixed point problems have been further developed by some authors. For some works related to the equilibrium problem, fixed point problems and the variational inequality problem, please see Blum and Oettli [8], Chang et al. [9], Chantarangsi et al. [10], Cianciaruso et al. [11], Colao et al. [12, 13], Fang et al. [14], Jung [15], Mainge [16], Mainge and Moudafi [17], Moudafi and Théra [18], Nadezhkina and Takahashi [19], Noor et al. [20], Peng et al. [21], Peng and Yao [22], Plubtieng and Punpaeng [23], Takahashi and Takahashi [24], Yao et al. [25], Yao and Liou [26] and the references therein.

We observe that the solution set of (1.3) has a unique element with a minimum norm and finding the least-squares solution of the constrained linear inverse problem is equivalent to finding the minimum-norm fixed point of the nonexpansive mapping x P C (xλ B (Bxb)). Hence, a natural idea is that we can use projection to construct algorithms for finding the minimum-norm solution. By using this idea, Yao and Liou [26] constructed two algorithms for solving the minimization problem (1.1):

x t =μ P C [ ( 1 t ) S x t ] +(1μ) T r ( x t rA x t ),t(0,1),
(1.5)

and

x n + 1 = μ n P C [ α n f ( x n ) + ( 1 α n ) S x n ] +(1 μ n ) T r ( x n rA x n ),n0.
(1.6)

Remark 1.1 It is well known that projection methods are used extensively in a variety of methods in optimization theory. Apart from theoretical interest, the main advantage of projection methods, which makes them successful in real-word applications, is computational. The field of projection methods is vast; see, e.g., Bauschke and Borwein [27], Combettes [28], Combettes and Pesquet [29]. However, it is clear that if the set C is simple enough, so that the projection onto it is easily executed, then this method is particularly useful; but if C is a general closed and convex set, then a minimal distance problem has to be solved in order to obtain the next iterative. This might seriously affect the efficiency of the method. Hence, it is a very interesting work of solving (1.1) without involving projection.

Motivated and inspired by the results in the literature, in this paper we suggest two algorithms:

{ F ( u t , y ) + 1 r y u t , u t ( t f + ( 1 t ) I r A ) x t 0 , y C , x t = μ S x t + ( 1 μ ) u t , t ( 0 , 1 r 2 α ) ,

and

{ F ( u n , y ) + 1 r y u n , u n ( α n f + ( 1 α n ) I r A ) x n 0 , y C , x n + 1 = μ S x n + ( 1 μ ) u n , n 0 .

It is shown that under some mild conditions, the net { x t } and the sequences { x n } converge strongly to x ˜ which is the unique solution of the VI:

x ˜ Fix(S)EFA, ( I f ) x ˜ , x x ˜ 0,xFix(S)EFA.

In particular, if we take f=0, then the net { x t } and the sequences { x n } converge in norm to a solution of the minimization problem (1.1). It should be pointed out that our suggested algorithms solve the above minimization problem (1.1) without involving the metric projection.

2 Preliminaries

Let C be a nonempty closed convex subset of a real Hilbert space H. Recall that a mapping A:CH is called α-inverse-strongly monotone if there exists a positive real number α such that AxAy,xyα A x A y 2 , x,yC. It is clear that any α-inverse-strongly monotone mapping is monotone and 1 α -Lipschitz continuous. A mapping S:CC is said to be nonexpansive if SxSyxy, x,yC. Denote the set of fixed points of S by Fix(S).

Let F:C×CR be a bifunction. Throughout this paper, we assume that a bifunction F:C×CR satisfies the following conditions:

  1. (H1)

    F(x,x)=0 for all xC;

  2. (H2)

    F is monotone, i.e., F(x,y)+F(y,x)0 for all x,yC;

  3. (H3)

    for each x,y,zC, lim t 0 F(tz+(1t)x,y)F(x,y);

  4. (H4)

    for each xC, yF(x,y) is convex and lower semicontinuous.

The metric (or nearest point) projection from H onto C is the mapping P C :HC which assigns to each point xC the unique point P C xC satisfying the property

x P C x= inf y C xy=:d(x,C).

It is well known that P C is a nonexpansive mapping and satisfies

xy, P C x P C y P C x P C y 2 ,x,yH.

We need the following lemmas for proving our main results.

Lemma 2.1 ([5])

Let C be a nonempty closed convex subset of a real Hilbert space H. Let F:C×CR be a bifunction which satisfies conditions (H1)-(H4). Let r>0 and xC. Then there exists zC such that

F(z,y)+ 1 r yz,zx0,yC.

Further, if T r (x)={zC:F(z,y)+ 1 r yz,zx0,yC}, then the following hold:

  1. (i)

    T r is single-valued and T r is firmly nonexpansive, i.e., for any x,yH, T r x T r y 2 T r x T r y,xy;

  2. (ii)

    EP is closed and convex and EP=Fix( T r ).

Lemma 2.2 ([30])

Let C be a nonempty closed convex subset of a real Hilbert space H. Let the mapping A:CH be α-inverse strongly monotone and r>0 be a constant. Then we have

( I r A ) x ( I r A ) y 2 x y 2 +r(r2α) A x A y 2 ,x,yC.

In particular, if 0r2α, then IrA is nonexpansive.

Lemma 2.3 ([31])

Let C be a closed convex subset of a real Hilbert space H, and S:CC be a nonexpansive mapping. Then the mapping IS is demiclosed. That is, if { x n } is a sequence in C such that x n x weakly and (IS) x n y strongly, then (IS) x =y.

Lemma 2.4 ([32])

Assume that { a n } is a sequence of nonnegative real numbers such that

a n + 1 (1 γ n ) a n + δ n γ n ,

where { γ n } is a sequence in (0,1) and { δ n } is a sequence such that

  1. (1)

    n = 1 γ n =;

  2. (2)

    lim sup n δ n 0 or n = 1 | δ n γ n |<.

Then lim n a n =0.

3 Main results

In this section, we convert algorithms (1.5) and (1.6) by releasing projection P C and construct two algorithms for finding the minimum norm element x of Γ:=EPAFix(S).

Let S:CC be a nonexpansive mapping and A:CH be an α-inverse strongly monotone mapping. Let F:C×CR be a bifunction which satisfies conditions (H1)-(H4). Let r and μ be two constants such that r(0,2α) and μ(0,1). In order to find a solution of the minimization problem (1.1), we construct the following implicit algorithm

{ F ( u t , y ) + 1 r y u t , u t ( ( 1 t ) I r A ) x t 0 , y C , x t = μ S x t + ( 1 μ ) u t , t ( 0 , 1 r 2 α ) .
(3.1)

We will show that the net { x t } defined by (3.1) converges to a solution of the minimization problem (1.1). As matter of fact, in this paper, we study the following general algorithm: Taking a ρ-contraction f:CH, for each t(0,1 r 2 α ), let { x t } be the net defined by

{ F ( u t , y ) + 1 r y u t , u t ( t f + ( 1 t ) I r A ) x t 0 , y C , x t = μ S x t + ( 1 μ ) u t , t ( 0 , 1 r 2 α ) .
(3.2)

It is clear that if f=0, then (3.2) reduces to (3.1). Next, we show that (3.2) is well defined. From Lemma 2.1, we know that u t = T r [tf( x t )+(1t) x t rA x t ]. We define a mapping W t :=μS+(1μ) T r [tf+(1t)IrA]. From Lemma 2.2, for 0<t<1 r 2 α , the mapping I r 1 t A is nonexpansive. Also, note that the mappings S and T r are nonexpansive, then we have

W t x W t y = μ ( S x S y ) + ( 1 μ ) ( T r [ t f ( x ) + ( 1 t ) x r A x ] T r [ t f ( y ) + ( 1 t ) y r A y ] ) μ S x S y + ( 1 μ ) T r [ t f ( x ) + ( 1 t ) ( x r 1 t A x ) ] T r [ t f ( y ) + ( 1 t ) ( y r 1 t A y ) ] μ x y + ( 1 μ ) t f ( x ) f ( y ) + ( 1 μ ) ( 1 t ) ( x r 1 t A x ) ( y r 1 t A y ) μ x y + ( 1 μ ) t ρ x y + ( 1 μ ) ( 1 t ) x y = [ 1 ( 1 μ ) ( 1 ρ ) t ] x y .

This indicates that W t is a contraction. Using the Banach contraction principle, there exists a unique fixed point x t of W t in C. Hence, (3.2) is well defined.

In the sequel, we assume:

  1. (1)

    C is a nonempty closed convex subset of a real Hilbert space H;

  2. (2)

    S:CC is a nonexpansive mapping, A:CH is an α-inverse strongly monotone mapping and f:CH is a ρ-contraction;

  3. (3)

    F:C×CR is a bifunction which satisfies conditions (H1)-(H4);

  4. (4)

    Γ.

In order to prove our first main result, we need the following propositions.

Proposition 3.1 The net { x t } generated by the implicit method (3.2) is bounded.

Proof Take zΓ. It is clear that Sz=z= T r (zrAz)= T r [tz+(1t)(z r 1 t Az)] for all t(0,1 r 2 α ). Since T r and I r 1 t A are nonexpansive, we have

u t z = T r [ t f ( x t ) + ( 1 t ) ( x t r 1 t A x t ) ] T r [ t z + ( 1 t ) ( z r 1 t A z ) ] = t ( f ( x t ) z ) + [ ( 1 t ) ( x t r 1 t A x t ) ( z r 1 t A z ) ] t f ( x t ) z + ( 1 t ) ( x t r 1 t A x t ) ( z r 1 t A z ) t f ( x t ) f ( z ) + t f ( z ) z + ( 1 t ) x t z t ρ x t z + ( 1 t ) x t z + t f ( z ) z = [ 1 ( 1 ρ ) t ] x t z + t f ( z ) z .
(3.3)

It follows from (3.2) that

x t z = μ ( S x t z ) + ( 1 μ ) ( u t z ) μ S x t z + ( 1 μ ) u t z μ x t z + ( 1 μ ) u t z .

Hence,

x t z u t z [ 1 ( 1 ρ ) t ] x t z+t f ( z ) z ,
(3.4)

that is,

x t z f ( z ) z 1 ρ .

So, { x t } is bounded. Hence { u t }, {S x t }, {A x t } and {f( x t )} are also bounded. This completes the proof. □

Proposition 3.2 The net { x t } generated by the implicit method (3.2) is relatively norm compact as t0.

Proof From (3.3) and Lemma 2.2, we have

u t z 2 t f ( x t ) z 2 + ( 1 t ) ( x t r 1 t A x t ) ( z r 1 t A z ) 2 t f ( x t ) z 2 + ( 1 t ) [ x t z 2 + r 1 t ( r 1 t 2 α ) A x t A z 2 ] ( 1 t ) x t z 2 + r ( r 1 t 2 α ) A x t A z 2 + t f ( x t ) z 2 .
(3.5)

From (3.4) and (3.5), we have

x t z 2 u t z 2 ( 1 t ) x t z 2 + r ( r 1 t 2 α ) A x t A z 2 + t f ( x t ) z 2 .

Thus,

r ( 2 α r 1 t ) A x t A z 2 t ( f ( x t ) z 2 x t z 2 ) 0.

Since lim inf t 0 + r(2α r 1 t )>0, we derive

lim t 0 + A x t Az=0.
(3.6)

From Lemma 2.1 and Lemma 2.2, we obtain

u t z 2 = T r ( t f ( x t ) + ( 1 t ) x t r A x t ) T r ( z r A z ) 2 t f ( x t ) + ( 1 t ) x t r A x t ( z r A z ) , u t z = 1 2 ( t f ( x t ) + ( 1 t ) x t r A x t ( z r A z ) 2 + u t z 2 t f ( x t ) + ( 1 t ) x t r ( A x t A z ) u t 2 ) .
(3.7)

It follows that

u t z 2 t f ( x t ) + ( 1 t ) x t r A x t ( z r A z ) 2 t f ( x t ) + ( 1 t ) x t r ( A x t A z ) u t 2 .

By the nonexpansivity of I r 1 t A, we have

t f ( x t ) + ( 1 t ) x t r A x t ( z r A z ) 2 = ( 1 t ) ( ( x t r 1 t A x t ) ( z r 1 t A z ) ) + t ( f ( x t ) z ) 2 ( 1 t ) ( x t r 1 t A x t ) ( z r 1 t A z ) 2 + t f ( x t ) z 2 ( 1 t ) x t z 2 + t f ( x t ) z 2 .

Thus

x t z 2 u t z 2 ( 1 t ) x t z 2 + t f ( x t ) z 2 t f ( x t ) + ( 1 t ) x t r ( A x t A z ) u t 2 .

Hence

t ( f ( x t ) x t ) r ( A x t A z ) ( u t x t ) 2 t ( f ( x t ) z 2 x t z 2 ) 0.

Since A x t Az0 (by (3.6)), we deduce

lim t 0 + x t u t =0.

So

lim t 0 + x t S x t = lim t 0 + (1μ) x t u t =0.
(3.8)

Next we show that { x t } is relatively norm compact as t0+. Let { t n }(0,1) be a sequence such that t n 0 as n. Put x n := x t n and u n := u t n . From (3.8), we get

x n S x n 0.
(3.9)

By (3.7), we deduce

u t z 2 t f ( x t ) f ( z ) , u t z + t f ( z ) z , u t z + ( 1 t ) x t r 1 t A x t ( z r 1 t A z ) , u t z [ 1 ( 1 ρ ) t ] x t z u t z + t f ( z ) z , u t z 1 ( 1 ρ ) t 2 x t z 2 + 1 2 u t z 2 + t f ( z ) z , u t z ,

that is,

u t z 2 [ 1 ( 1 ρ ) t ] x t z 2 +2t f ( z ) z , u t z .

Hence,

x t z 2 u t z 2 [ 1 ( 1 ρ ) t ] x t z 2 +2t f ( z ) z , u t z .

It follows that

x t z 2 2 1 ρ f ( z ) z , u t z .

In particular,

x n z 2 2 1 ρ f ( z ) z , u n z .
(3.10)

Since { x n } is bounded, without loss of generality, we may assume that { x n } converges weakly to a point x C. Also S x n x and u n x . Noticing (3.9) we can use Lemma 2.3 to get x Fix(S).

Now we show x EPA. Since u n = T λ ( t n f( x n )+(1 t n ) x n rA x n ) for any yC, we have

F( u n ,y)+A x n ,y u n + 1 r y u n , u n ( t n f ( x n ) + ( 1 t n ) x n ) 0.

From (H2), we have

A x n ,y u n + 1 r y u n , u n ( t n f ( x n ) + ( 1 t n ) x n ) F(y, u n ).
(3.11)

Put z t =ty+(1t) x for all t(0,1 λ 2 α ) and yC. Then we have z t C. So, from (3.11), we have

z t u n , A z t z t u n , A z t z t u n , A x n 1 r z t u n , u n ( t n f ( x n ) + ( 1 t n ) x n ) + F ( z t , u n ) = z t u n , A z t A u n + z t u n , A u n A x n 1 r z t u n , u n x n t n ( f ( x n ) x n ) + F ( z t , u n ) .

Since A is Lipschitz continuous and u n x n 0, we have A u n A x n 0. Further, from the monotonicity of A, we have z t u n ,A z t A u n 0. So, from (H4), we have

z t x , A z t F ( z t , x ) as n.
(3.12)

From (H1), (H4) and (3.12), we also have

0 = F ( z t , z t ) t F ( z t , y ) + ( 1 t ) F ( z t , x ) t F ( z t , y ) + ( 1 t ) z t x , A z t = t F ( z t , y ) + ( 1 t ) t y x , A z t

and hence

0F( z t ,y)+(1t) y x , A z t .

Letting t0, we have, for each yC,

0F ( x , y ) + y x , A x .

This implies x EPA. Therefore we can substitute x for z in (3.10) to get

x n x 2 2 1 ρ f ( x ) x , u n x ,zFix(S)EPA.

Consequently, the weak convergence of { x n } (and { u n }) to x actually implies that x n x . This has proved the relative norm-compactness of the net { x t } as t0+. This completes the proof. □

Now we show our first main result.

Theorem 3.3 The net { x t } generated by the implicit method (3.2) converges in norm, as t0+, to the unique solution x of the following variational inequality:

x Γ, ( I f ) x , x x 0,xΓ.
(3.13)

In particular, if we take f=0, then the net { x t } converges in norm, as t0+, to a solution of the minimization problem (1.1).

Proof Now we return to (3.10) and take the limit as n to get

x z 2 2 1 ρ z f ( z ) , z x ,zΓ.
(3.14)

In particular, x solves the following variational inequality

x Γ, ( I f ) z , z x 0,zΓ

or the equivalent dual variational inequality

x Γ, ( I f ) x , z x 0,zΓ.

Therefore, x =( P Γ f) x . That is, x is the unique fixed point in Γ of the contraction P Γ f. Clearly, this is sufficient to conclude that the entire net { x t } converges in norm to x as t0.

Finally, if we take f=0, then (3.14) is reduced to

x z 2 z , z x ,zΓ.

Equivalently,

x 2 x , z ,zΓ.

This clearly implies that

x z,zΓ.

Therefore, x is a solution of the minimization problem (1.1). This completes the proof. □

Next, we introduce an explicit algorithm for finding a solution of the minimization problem (1.1).

Algorithm 3.4 For given x 0 C arbitrarily, let the sequence { x n } be generated iteratively by

{ F ( u n , y ) + 1 r y u n , u n ( α n f + ( 1 α n ) I r A ) x n 0 , y C , x n + 1 = μ S x n + ( 1 μ ) u n , n 0 ,
(3.15)

where { α n } is a real number sequence in [0,1].

Next, we give our second main result.

Theorem 3.5 Assume that the sequence { α n } satisfies the conditions: lim n α n =0, n = 0 α n = and lim n α n + 1 α n =1. Then the sequence { x n } generated by (3.15) converges strongly to x ˜ which is the unique solution of the variational inequality (3.13). In particular, if f=0, then the sequence { x n } converges strongly to a solution of the minimization problem (1.1).

Proof Pick zΓ. From Lemma 2.2, we know that u n = T r [ α n f( x n )+(1 α n ) x n rA x n ]. Set z n = α n f( x n )+(1 α n ) x n rA x n for all n. From (3.15), we get

u n z = T r z n T r ( z r A z ) z n ( z r A z ) = ( α n f ( x n ) + ( 1 α n ) ( x n r A x n 1 α n ) ) ( α n z + ( 1 α n ) ( z r A z 1 α n ) ) = ( 1 α n ) ( ( x n r A x n 1 α n ) ( z r A z 1 α n ) ) + α n ( f ( x n ) z ) ( 1 α n ) x n z + α n f ( x n ) f ( z ) + α n f ( z ) z [ 1 ( 1 ρ ) α n ] x n z + α n f ( z ) z .
(3.16)

Hence,

x n + 1 z μ S x n z + ( 1 μ ) u n z μ x n z + ( 1 μ ) [ 1 ( 1 ρ ) α n ] x n z + ( 1 μ ) α n f ( z ) z = [ 1 ( 1 μ ) ( 1 ρ ) α n ] x n z + ( 1 μ ) α n f ( z ) z .

By induction, we have

x n + 1 zmax { x 0 z , f ( z ) z 1 ρ } .

Therefore, { x n } is bounded. Hence, {A x n }, { u n }, {S x n } are also bounded.

From (3.16), we obtain

u n z 2 (1 α n ) ( x n r A x n 1 α n ) ( z r A z 1 α n ) 2 + α n f ( x n ) z 2 .

Since A is α-inverse strongly monotone, we know from Lemma 2.3 that

( x n r A x n 1 α n ) ( z r A z 1 α n ) 2 x n z 2 + r ( r 2 ( 1 α n ) α ) ( 1 α n ) 2 A x n A z 2 .

It follows that

u n z 2 (1 α n ) x n z 2 + r ( r 2 ( 1 α n ) α ) ( 1 α n ) A x n A z 2 + α n f ( z ) z 2 .
(3.17)

Note that

u n + 1 u n T r z n + 1 T r z n z n + 1 z n .
(3.18)

From Lemma 2.3, we know that IλA is nonexpansive for all λ(0,2α). Thus, we have I λ n + 1 1 α n + 1 A is nonexpansive for all n due to the fact that r 1 α n + 1 (0,2α). Then we get

z n + 1 z n = α n + 1 f ( x n + 1 ) + ( 1 α n + 1 ) x n + 1 r A x n + 1 ( α n f ( x n ) + ( 1 α n ) x n r A x n ) ( 1 α n + 1 ) ( x n + 1 r 1 α n + 1 A x n + 1 ) ( 1 α n ) ( x n r 1 α n A x n ) + α n + 1 f ( x n + 1 ) f ( x n ) + | α n + 1 α n | f ( x n ) ( 1 α n + 1 ) ( I r 1 α n + 1 A ) x n + 1 ( I r 1 α n + 1 A ) x n + ( 1 α n + 1 ) ( x n r 1 α n + 1 A x n ) ( 1 α n ) ( x n r 1 α n A x n ) + α n + 1 ρ x n + 1 x n + | α n + 1 α n | f ( x n ) [ 1 ( 1 ρ ) α n + 1 ] x n + 1 x n + | α n + 1 α n | ( f ( x n ) + x n ) .
(3.19)

From (3.15), (3.18) and (3.19), we obtain

x n + 2 x n + 1 μ S x n + 1 S x n + ( 1 μ ) u n + 1 u n μ x n + 1 x n + ( 1 μ ) u n + 1 u n μ x n + 1 x n + ( 1 μ ) [ 1 ( 1 ρ ) α n + 1 ] x n + 1 x n + ( 1 μ ) | α n + 1 α n | ( f ( x n ) + x n ) = [ 1 ( 1 μ ) ( 1 ρ ) α n + 1 ] x n + 1 x n + ( 1 μ ) ( 1 ρ ) α n + 1 | α n + 1 α n | ( 1 ρ ) α n + 1 ( f ( x n ) + x n ) .

By Lemma 2.4, we get

lim n x n + 1 x n =0.

From (3.15) and (3.17), we have

x n + 1 z 2 μ S x n z 2 + ( 1 μ ) u n z 2 μ x n z 2 + ( 1 μ ) ( 1 α n ) x n z 2 + ( 1 μ ) α n f ( z ) z 2 + ( 1 μ ) r ( r 2 ( 1 α n ) α ) ( 1 α n ) A x n A z 2 .

Then we obtain

( 1 μ ) r ( 2 ( 1 α n ) α r ) ( 1 α n ) A x n A z 2 x n z 2 x n + 1 z 2 + ( 1 μ ) α n f ( z ) z 2 ( x n z x n + 1 z ) x n + 1 x n + ( 1 μ ) α n f ( z ) z 2 .

Since lim n α n =0, lim n x n + 1 x n =0 and lim inf n (1μ) r ( 2 ( 1 α n ) α r ) ( 1 α n ) >0, we have

lim n A x n Az=0.
(3.20)

Next, we show x n u n 0. By using the firm nonexpansivity of T λ n , we have

u n z 2 = T r z n T r ( z r A z ) 2 z n ( z r A z ) , u n z = 1 2 ( z n ( z r A z ) 2 + u n z 2 z n ( z r A z ) ( u n z ) 2 ) = 1 2 ( z n ( z r A z ) 2 + u n z 2 α n ( f ( x n ) x n ) + ( x n u n ) r ( A x n A z ) 2 ) .

From (3.16) and (3.17), we have

z n ( z r A z ) 2 (1 α n ) x n z 2 + α n f ( x n ) z 2 .

Thus,

u n z 2 1 2 ( ( 1 α n ) x n z 2 + α n f ( x n ) z 2 + u n z 2 α n ( f ( x n ) x n ) + ( x n u n ) r ( A x n A z ) 2 ) .

That is,

u n z 2 ( 1 α n ) x n z 2 + α n f ( x n ) z 2 α n ( f ( x n ) x n ) + ( x n u n ) r ( A x n A z ) 2 = ( 1 α n ) x n z 2 + α n f ( x n ) z 2 x n u n 2 + 2 r x n u n , A x n A z 2 α n f ( x n ) x n , x n u n α n ( f ( x n ) x n ) r ( A x n A z ) 2 ( 1 α n ) x n z 2 + α n f ( x n ) z 2 x n u n 2 + 2 r x n u n A x n A z + 2 α n f ( x n ) x n x n u n .

It follows that

x n + 1 z 2 μ x n z 2 + ( 1 μ ) ( 1 α n ) x n z 2 + ( 1 μ ) α n f ( x n ) z 2 ( 1 μ ) x n u n 2 + 2 r x n u n A x n A z + 2 α n f ( x n ) x n x n u n = [ 1 ( 1 μ ) α n ] x n z 2 + ( 1 μ ) α n f ( x n ) z 2 ( 1 μ ) x n u n 2 + 2 r x n u n A x n A z + 2 α n f ( x n ) x n x n u n .

Hence,

( 1 μ ) x n u n 2 x n z 2 x n + 1 z 2 + ( 1 μ ) α n f ( x n ) z 2 + 2 r x n u n A x n A z + 2 α n f ( x n ) x n x n u n ( x n z + x n + 1 z ) x n + 1 x n + ( 1 μ ) α n f ( x n ) z 2 + 2 r x n u n A x n A z + 2 α n f ( x n ) x n x n u n .

Since x n + 1 x n 0, α n 0 and A x n Az0, we deduce

lim n x n u n =0.
(3.21)

This together with x n + 1 x n 0 implies that

lim n S x n x n =0.
(3.22)

Put x ˜ = lim t 0 + x t , where { x t } is the net defined by (3.2). We will finally show that x n x ˜ .

Set v n = x n λ n 1 α n (A x n A x ˜ ) for all n. Take z= x ˜ in (3.17) to get A x n A x ˜ 0. First, we prove lim sup n x ˜ f( x ˜ ), x n x ˜ 0. We take a subsequence { v n i } of { v n } such that

lim sup n x ˜ f ( x ˜ ) , x n x ˜ = lim i x ˜ f ( x ˜ ) , x n i x ˜ .

It is clear that { x n i } is bounded due to the boundedness of { x n }. Then there exists a subsequence { x n i j } of { x n i } which converges weakly to some point wC. Hence, { x n i j } also converges weakly to w. From (3.22), we have

lim j x n i j S x n i j =0.
(3.23)

By the demi-closedness principle of the nonexpansive mapping (see Lemma 2.3) and (3.23), we deduce wFix(S). Furthermore, by a similar argument as that of Theorem 3.3, we can show that w is also in EPA. Hence, we have wFix(S)EPA. This implies that

lim sup n x ˜ f ( x ˜ ) , x n x ˜ = lim j x ˜ f ( x ˜ ) , x n i j x ˜ = x ˜ f ( x ˜ ) , w x ˜ 0 .

From (3.15), we have

x n + 1 x ˜ 2 μ x n x ˜ 2 + ( 1 μ ) u n x ˜ 2 μ x n x ˜ 2 + ( 1 μ ) T r z n T r ( x ˜ r A x ˜ ) 2 μ x n x ˜ 2 + ( 1 μ ) z n ( x ˜ r A x ˜ ) 2 = μ x n x ˜ 2 + ( 1 μ ) α n f ( x n ) + ( 1 α n ) x n r A x n ( x ˜ r A x ˜ ) 2 = ( 1 μ ) α n ( f ( x n ) x ˜ ) + ( 1 α n ) ( ( x n r 1 α n A x n ) ( x ˜ r 1 α n A x ˜ ) ) 2 + μ x n x ˜ 2 = ( 1 μ ) ( ( 1 α n ) 2 ( x n r 1 α n A x n ) ( x ˜ r 1 α n A x ˜ ) 2 + 2 α n ( 1 α n ) f ( x n ) x ˜ , ( x n r 1 α n A x n ) ( x ˜ r 1 α n A x ˜ ) + α n 2 f ( x n ) x ˜ 2 ) + μ x n x ˜ 2 μ x n x ˜ 2 + ( 1 μ ) ( ( 1 α n ) 2 x n x ˜ 2 + 2 α n ( 1 α n ) f ( x n ) f ( x ˜ ) , x n x ˜ + 2 α n ( 1 α n ) f ( x ˜ ) x ˜ , x n x ˜ 2 r α n f ( x n ) x ˜ , A x n A x ˜ + α n 2 f ( x n ) x ˜ 2 ) μ x n x ˜ 2 + ( 1 μ ) ( ( 1 α n ) 2 x n x ˜ 2 + 2 α n ( 1 α n ) ρ x n x ˜ 2 + 2 α n ( 1 α n ) f ( x ˜ ) x ˜ , x n x ˜ + 2 r α n f ( x n ) x ˜ A x n A x ˜ + α n 2 f ( x n ) x ˜ 2 ) [ 1 2 ( 1 μ ) ( 1 ρ ) α n ] x n x ˜ 2 + ( 1 μ ) α n 2 ( x n x ˜ 2 + f ( x n ) x ˜ 2 ) + 2 ( 1 μ ) α n ( 1 α n ) f ( x ˜ ) x ˜ , x n x ˜ + 2 r ( 1 μ ) α n f ( x n ) x ˜ A x n A x ˜ = [ 1 2 ( 1 μ ) ( 1 ρ ) α n ] x n x ˜ 2 + 2 ( 1 μ ) ( 1 ρ ) α n { α n 1 ρ ( x n x ˜ 2 + f ( x n ) x ˜ 2 ) + 1 α n 1 ρ f ( x ˜ ) x ˜ , x n x ˜ + r 1 ρ f ( x n ) x ˜ A x n A x ˜ } .

It is clear that n 2(1μ)(1ρ) α n = and

lim sup n { α n 1 ρ ( x n x ˜ 2 + f ( x n ) x ˜ 2 ) + 1 α n 1 ρ f ( x ˜ ) x ˜ , x n x ˜ + r 1 ρ f ( x n ) x ˜ A x n A x ˜ } 0 .

We can therefore apply Lemma 2.4 to conclude that x n x ˜ .

Finally, if we take f=0, by a similar argument as that in Theorem 3.3, we deduce immediately that x ˜ is a minimum norm element in Γ. This completes the proof. □

References

  1. Reich S, Xu HK: An iterative approach to a constrained least squares problem. Abstr. Appl. Anal. 2003, 8: 503–512.

    Article  MathSciNet  Google Scholar 

  2. Sabharwal A, Potter LC: Convexly constrained linear inverse problems: iterative least-squares and regularization. IEEE Trans. Signal Process. 1998, 46: 2345–2352. 10.1109/78.709518

    Article  MATH  MathSciNet  Google Scholar 

  3. Xu HK: An iterative approach to quadratic optimization. J. Optim. Theory Appl. 2003, 116: 659–678. 10.1023/A:1023073621589

    Article  MATH  MathSciNet  Google Scholar 

  4. Yao Y, Liou YC, Yao JC: Convergence theorem for equilibrium problems and fixed point problems of infinite family of nonexpansive mappings. Fixed Point Theory Appl. 2007., 2007: Article ID 64363

    Google Scholar 

  5. Combettes PL, Hirstoaga A: Equilibrium programming in Hilbert spaces. J. Nonlinear Convex Anal. 2005, 6: 117–136.

    MATH  MathSciNet  Google Scholar 

  6. Moudafi A: Weak convergence theorems for nonexpansive mappings and equilibrium problems. J. Nonlinear Convex Anal. 2008, 9: 37–43.

    MATH  MathSciNet  Google Scholar 

  7. Takahashi S, Takahashi W: Viscosity approximation methods for equilibrium problems and fixed point problems in Hilbert spaces. J. Math. Anal. Appl. 2007, 331: 506–515. 10.1016/j.jmaa.2006.08.036

    Article  MATH  MathSciNet  Google Scholar 

  8. Blum E, Oettli W: From optimization and variational inequalities to equilibrium problems. Math. Stud. 1994, 63: 123–145.

    MATH  MathSciNet  Google Scholar 

  9. Chang SS, Lee HWJ, Chan CK: A new method for solving equilibrium problem fixed point problem and variational inequality problem with application to optimization. Nonlinear Anal. 2009, 70: 3307–3319. 10.1016/j.na.2008.04.035

    Article  MATH  MathSciNet  Google Scholar 

  10. Chantarangsi W, Jaiboon C, Kumam P: A viscosity hybrid steepest descent method for generalized mixed equilibrium problems and variational inequalities for relaxed cocoercive mapping in Hilbert spaces. Abstr. Appl. Anal. 2010., 2010: Article ID 390972

    Google Scholar 

  11. Cianciaruso F, Marino G, Muglia L, Yao Y: A hybrid projection algorithm for finding solutions of mixed equilibrium problem and variational inequality problem. Fixed Point Theory Appl. 2010., 2010: Article ID 383740

    Google Scholar 

  12. Colao V, Acedo GL, Marino G: An implicit method for finding common solutions of variational inequalities and systems of equilibrium problems and fixed points of infinite family of nonexpansive mappings. Nonlinear Anal. 2009, 71: 2708–2715. 10.1016/j.na.2009.01.115

    Article  MATH  MathSciNet  Google Scholar 

  13. Colao V, Marino G, Xu HK: An iterative method for finding common solutions of equilibrium and fixed point problems. J. Math. Anal. Appl. 2008, 344: 340–352. 10.1016/j.jmaa.2008.02.041

    Article  MATH  MathSciNet  Google Scholar 

  14. Fang YP, Huang NJ, Yao JC: Well-posedness by perturbations of mixed variational inequalities in Banach spaces. Eur. J. Oper. Res. 2010, 201: 682–692. 10.1016/j.ejor.2009.04.001

    Article  MATH  MathSciNet  Google Scholar 

  15. Jung JS: Strong convergence of composite iterative methods for equilibrium problems and fixed point problems. Appl. Math. Comput. 2009, 213: 498–505. 10.1016/j.amc.2009.03.048

    Article  MATH  MathSciNet  Google Scholar 

  16. Mainge PE: Projected subgradient techniques and viscosity methods for optimization with variational inequality constraints. Eur. J. Oper. Res. 2010, 205: 501–506. 10.1016/j.ejor.2010.01.042

    Article  MATH  MathSciNet  Google Scholar 

  17. Mainge PE, Moudafi A: Coupling viscosity methods with the extragradient algorithm for solving equilibrium problems. J. Nonlinear Convex Anal. 2008, 9: 283–294.

    MATH  MathSciNet  Google Scholar 

  18. Moudafi A, Théra M Lecture Notes in Economics and Mathematical Systems 477. In Proximal and Dynamical Approaches to Equilibrium Problems. Springer, Berlin; 1999:187–201.

    Google Scholar 

  19. Nadezhkina N, Takahashi W: Weak convergence theorem by an extragradient method for nonexpansive mappings and monotone mappings. J. Optim. Theory Appl. 2006, 128: 191–201. 10.1007/s10957-005-7564-z

    Article  MATH  MathSciNet  Google Scholar 

  20. Noor MA, Yao Y, Chen R, Liou YC: An iterative method for fixed point problems and variational inequality problems. Math. Commun. 2007, 12: 121–132.

    MATH  MathSciNet  Google Scholar 

  21. Peng JW, Wu SY, Yao JC: A new iterative method for finding common solutions of a system of equilibrium problems, fixed-point problems, and variational inequalities. Abstr. Appl. Anal. 2010., 2010: Article ID 428293

    Google Scholar 

  22. Peng JW, Yao JC: A new hybrid-extragradient method for generalized mixed equilibrium problems and fixed point problems and variational inequality problems. Taiwan. J. Math. 2008, 12: 1401–1433.

    MATH  MathSciNet  Google Scholar 

  23. Plubtieng S, Punpaeng R: A new iterative method for equilibrium problems and fixed point problems of nonexpansive mappings and monotone mappings. Appl. Math. Comput. 2008, 197: 548–558. 10.1016/j.amc.2007.07.075

    Article  MATH  MathSciNet  Google Scholar 

  24. Takahashi S, Takahashi W: Strong convergence theorem for a generalized equilibrium problem and a nonexpansive mapping in a Hilbert space. Nonlinear Anal. 2008, 69: 1025–1033. 10.1016/j.na.2008.02.042

    Article  MATH  MathSciNet  Google Scholar 

  25. Yao Y, Cho YJ, Liou YC: Algorithms of common solutions for variational inclusions, mixed equilibrium problems and fixed point problems. Eur. J. Oper. Res. 2011, 212: 242–250. 10.1016/j.ejor.2011.01.042

    Article  MATH  MathSciNet  Google Scholar 

  26. Yao Y, Liou YC: Composite algorithms for minimization over the solutions of equilibrium problems and fixed point problems. Abstr. Appl. Anal. 2010., 2010: Article ID 763506

    Google Scholar 

  27. Bauschke HH, Borwein JM: On projection algorithms for solving convex feasibility problems. SIAM Rev. 1996, 38: 367–426. 10.1137/S0036144593251710

    Article  MATH  MathSciNet  Google Scholar 

  28. Combettes PL: Strong convergence of block-iterative outer approximation methods for convex optimization. SIAM J. Control Optim. 2000, 38: 538–565. 10.1137/S036301299732626X

    Article  MATH  MathSciNet  Google Scholar 

  29. Combettes PL, Pesquet JC: Proximal thresholding algorithm for minimization over orthonormal bases. SIAM J. Optim. 2007, 18: 1351–1376.

    Article  MATH  MathSciNet  Google Scholar 

  30. Takahashi W, Toyoda M: Weak convergence theorems for nonexpansive mappings and monotone mappings. J. Optim. Theory Appl. 2003, 118: 417–428. 10.1023/A:1025407607560

    Article  MATH  MathSciNet  Google Scholar 

  31. Geobel K, Kirk WA Cambridge Studies in Advanced Mathematics 28. In Topics in Metric Fixed Point Theory. Cambridge University Press, Cambridge; 1990.

    Chapter  Google Scholar 

  32. Xu HK: Iterative algorithms for nonlinear operators. J. Lond. Math. Soc. 2002, 66: 240–256. 10.1112/S0024610702003332

    Article  MATH  Google Scholar 

Download references

Acknowledgements

The first author was supported in part by NSFC 11071279 and NSFC 71161001-G0105. The second author was supported in part by NSC 101-2628-E-230-001-MY3.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shin Min Kang.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

All authors read and approved the final manuscript.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License ( https://creativecommons.org/licenses/by/2.0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Yao, Y., Liou, YC. & Kang, S.M. Conversion of algorithms by releasing projection for minimization problems. Fixed Point Theory Appl 2013, 114 (2013). https://doi.org/10.1186/1687-1812-2013-114

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1687-1812-2013-114

Keywords