- Research Article
- Open access
- Published:

# A General Iterative Approach to Variational Inequality Problems and Optimization Problems

*Fixed Point Theory and Applications*
**volumeÂ 2011**, ArticleÂ number:Â 284363 (2011)

## Abstract

We introduce a new general iterative scheme for finding a common element of the set of solutions of variational inequality problem for an inverse-strongly monotone mapping and the set of fixed points of a nonexpansive mapping in a Hilbert space and then establish strong convergence of the sequence generated by the proposed iterative scheme to a common element of the above two sets under suitable control conditions, which is a solution of a certain optimization problem. Applications of the main result are also given.

## 1. Introduction

Let be a real Hilbert space with inner product and induced norm . Let be a nonempty closed convex subset of and be self-mapping on . We denote by the set of fixed points of and by the metric projection of onto .

Let be a nonlinear mapping of into . The variational inequality problem is to find a such that

We denote the set of solutions of the variational inequality problem (1.1) by . The variational inequality problem has been extensively studied in the literature; see [1â€“5] and the references therein.

Recently, in order to study the problem (1.1) coupled with the fixed point problem, many authors have introduced some iterative schemes for finding a common element of the set of the solutions of the problem (1.1) and the set of fixed points of nonexpansive mappings; see [6â€“9] and the references therein. In particular, in 2005, Iiduka and Takahashi [8] introduced an iterative scheme for finding a common point of the set of fixed points of a nonexapansive mapping and the set of solutions of the problem (1.1) for an inverse-strong monotone mapping : and

where and . They proved that the sequence generated by (1.2) strongly converges strongly to . In 2010, Jung [10] provided the following new composite iterative scheme for the fixed point problem and the problem (1.1): and

where is a contraction with constant ,,, and . He proved that the sequence generated by (1.3) strongly converges strongly to a point in , which is the unique solution of a certain variational inequality.

On the other hand, the following optimization problem has been studied extensively by many authors:

where , are infinitely many closed convex subsets of such that , , is a real number, is a strongly positive bounded linear operator on (i.e., there is a constant such that , for all ), and is a potential function for (i.e., for all ). For this kind of optimization problems, see, for example, Deutsch and Yamada [11], Jung [10], and Xu [12, 13] when and for a given point in .

In 2007, related to a certain optimization problem, Marino and Xu [14] introduced the following general iterative scheme for the fixed point problem of a nonexpansive mapping:

where and . They proved that the sequence generated by (1.5) converges strongly to the unique solution of the variational inequality

which is the optimality condition for the minimization problem

where is a potential function for . The result improved the corresponding results of Moudafi [15] and Xu [16].

In this paper, motivated by the above-mentioned results, we introduce a new general composite iterative scheme for finding a common point of the set of solutions of the variational inequality problem (1.1) for an inverse-strongly monotone mapping and the set of fixed points of a nonexapansive mapping and then prove that the sequence generated by the proposed iterative scheme converges strongly to a common point of the above two sets, which is a solution of a certain optimization problem. Applications of the main result are also discussed. Our results improve and complement the corresponding results of Chen et al. [6], Iiduka and Takahashi [8], Jung [10], and others.

## 2. Preliminaries and Lemmas

Let be a real Hilbert space and let be a nonempty closed convex subset of . We write to indicate that the sequence converges weakly to . implies that converges strongly to .

First we recall that a mapping is a *contraction* on if there exists a constant such thatâ€‰â€‰, . A mapping is called *nonexpansive* if . We denote by the set of fixed points of .

For every point , there exists a unique nearest point in , denoted by , such that

for all . is called the *metric projection* of onto . It is well known that is nonexpansive and satisfies

for every . Moreover, is characterized by the properties:

In the context of the variational inequality problem for a nonlinear mapping , this implies that

It is also well known that satisfies the *Opial condition*, that is, for any sequence with , the inequality

holds for every with .

A mapping of into is called *inverse-strongly monotone* if there exists a positive real number such that

for all ; see [4, 7, 17]. For such a case, is called -inverse-strongly monotone. We know that if , where is a nonexpansive mapping of into itself and is the identity mapping of , then is -inverse-strongly monotone and . A mapping of into is called *strongly monotone* if there exists a positive real number such that

for all . In such a case, we say is -strongly monotone. If is -strongly monotone and -*Lipschitz continuous*, that is, for all , then is -inverse-strongly monotone. If is an -inverse-strongly monotone mapping of into , then it is obvious that is -Lipschitz continuous. We also have that for all and ,

So, if , then is a nonexpansive mapping of into . The following result for the existence of solutions of the variational inequality problem for inverse strongly-monotone mappings was given in Takahashi and Toyoda [9].

Proposition 2.1.

Letâ€‰â€‰ be a bounded closed convex subset of a real Hilbert space and let be an -inverse-strongly monotone mapping of into . Then, is nonempty.

A set-valued mapping is called *monotone* if for all , , and imply . A monotone mapping is *maximal* if the graph of is not properly contained in the graph of any other monotone mapping. It is known that a monotone mapping is maximal if and only if for , for every implies . Let be an inverse-strongly monotone mapping of into and let be the *normal cone* to at , that is, , and define

Then is maximal monotone and if and only if ; see [18, 19].

We need the following lemmas for the proof of our main results.

Lemma 2.2.

In a real Hilbert space , there holds the following inequality:

for all .

Lemma 2.3 (Xu [12]).

Let be a sequence of nonnegative real numbers satisfying

where and satisfy the following conditions:

(i) and or, equivalently, ;

(ii) or ;

(iii).

Then .

Lemma 2.4 (Marino and Xu [14]).

Assume that is a strongly positive linear bounded operator on a Hilbert space with constant and . Then .

The following lemma can be found in [20, 21] (see also Lemma 2.2 in [22]).

Lemma 2.5.

Let be a nonempty closed convex subset of a real Hilbert space , and let be a proper lower semicontinunous differentiable convex function. If is a solution to the minimization problem

then

In particular, if solves the optimization problem

then

where is a potential function for .

## 3. Main Results

In this section, we present a new general composite iterative scheme for inverse-strongly monotone mappings and a nonexpansive mapping.

Theorem 3.1.

Let be a nonempty closed convex subset of a real Hilbert space such that . Let be an -inverse-strongly monotone mapping of into and a nonexpansive mapping of into itself such that . Let and let be a strongly positive bounded linear operator on with constant and a contraction of into itself with constant . Assume that and . Let be a sequence generated by

where , , and . Let , , and satisfy the following conditions:

(i); ;

(ii) for all and for some ;

(iii) for some , with ;

(iv), , .

Then converges strongly to , which is a solution of the optimization problem

where is a potential function for .

Proof.

We note that from the control condition (i), we may assume, without loss of generality, that . Recall that if is bounded linear self-adjoint operator on , then

Observe that

which is to say that is positive. It follows that

Now we divide the proof into several steps.

Step 1.

We show that is bounded. To this end, let and for every . Let . Since is nonexpansive and from (2.4), we have

Similarly, we have

Now, set . Let . Then, from (IS) and (3.4), we obtain

From (3.5) and (3.6), it follows that

By induction, it follows from (3.7) that

Therefore, is bounded. So , , , , , , and are bounded. Moreover, since and , and are also bounded. And by the condition (i), we have

Step 2.

We show that and . Indeed, since and are nonexpansive and , we have

Similarly, we get

Simple calculations show that

So, we obtain

Also observe that

By (3.11), (3.13), and (3.14), we have

where , , and . From the conditions (i) and (iv), it is easy to see that

Applying Lemma 2.3 to (3.15), we obtain

Moreover, by (3.10) and (3.13), we also have

Step 3.

We show that and . Indeed,

which implies that

Obviously, by (3.9) and Step 2, we have as . This implies that

By (3.9) and (3.21), we also have

Step 4.

We show that and . To this end, let . Since and , we have

So we obtain

Since from the condition (i) and from Step 3, we have . Moreover, from (2.4) we obtain

and so

Thus

Then, we have

Since , and , we get . Also by (3.21)

Step 5.

We show that . In fact, since

from (3.9) and (3.29), we have .

Step 6.

We show that

where is a solution of the optimization problem (OP1). First we prove that

Since is bounded, we can choose a subsequence of such that

Without loss of generality, we may assume that converges weakly to .

Now we will show that . First we show that . Assume that . Since and , by the Opial condition and Step 5, we obtain

which is a contradiction. Thus we have .

Next, let us show that . Let

Then is maximal monotone. Let . Since and , we have

On the other hand, from , we have and hence

Therefore, we have

Since in Step 4 and is -inverse-strongly monotone, we have as . Since is maximal monotone, we have and hence .

Therefore, . Now from Lemma 2.5 and Step 5, we obtain

By (3.9) and (3.39), we conclude that

Step 7.

We show that and , where is a solution of the optimization problem (OP1). Indeed from (IS) and Lemma 2.2, we have

that is,

where , , and

From (i), in Steps 3, and 6, it is easily seen that , , and . Hence, by Lemma 2.3, we conclude as . This completes the proof.

As a direct consequence of Theorem 3.1, we have the following results.

Corollary 3.2.

Let , and be as in Theorem 3.1. Let be a sequence generated by

where and . Let and satisfy the conditions (i), (ii), and (iv) in Theorem 3.1. Then converges strongly to , which is a solution of the optimization problem

where is a potential function for .

Corollary 3.3.

Let , , , , , , , , , and be as in Theorem 3.1. Let be a sequence generated by

where , , and . Let , and satisfy the conditions (i), (ii), (iii), and (iv) in Theorem 3.1. Then converges strongly to , which is a solution of the optimization problem

where is a potential function for .

Remark 3.4.

## 4. Applications

In this section, as in [6, 8, 10], we prove two theorems by using Theorem 3.1. First of all, we recall the following definition.

A mapping is called *strictly pseudocontractive* if there exists with such that

for every . If , then is nonexpansive. Put , where is a strictly pseudo-contractive mapping with constant . Then is -inverse-strongly monotone; see [2]. Actually, we have, for all ,

On the other hand, since is a real Hilbert space, we have

Hence we have

Using Theorem 3.1, we found a strong convergence theorem for finding a common fixed point of a nonexpansive mapping and a strictly pseudo-contractive mapping.

Theorem 4.1.

Let , , , , , , , , , and be as in Theorem 3.1. Let be an -strictly pseudo-contractive mapping of into itself such that . Let be a sequence generated by

where , , and . Let , , and satisfy the conditions (i), (ii), (iii), and (iv) in Theorem 3.1. Then converges strongly to , which is a solution of the optimization problem

where is a potential function for .

Proof.

Put . Then is -inverse-strongly monotone. We have and . Thus, the desired result follows from Theorem 3.1.

Using Theorem 3.1, we also obtain the following result.

Theorem 4.2.

Let be a real Hilbert space. Let be an -inverse-strongly monotone mapping of into and a nonexpansive mapping of into itself such that . Let ,â€‰â€‰and let be a strongly positive bounded linear operator on with constant and a contraction with constant . Assume that and . Let be a sequence generated by

where , , and . Let , , and satisfy the conditions (i), (ii), (iii), and (iv) in Theorem 3.1. Then converges strongly to , which is a solution of the optimization problem

where is a potential function for .

Proof.

We have . So, putting , by Theorem 3.1, we obtain the desired result.

Remark 4.3.

## References

Browder FE:

**Nonlinear monotone operators and convex sets in Banach spaces.***Bulletin of the American Mathematical Society*1965,**71:**780â€“785. 10.1090/S0002-9904-1965-11391-XBruck, RE Jr.:

**On the weak convergence of an ergodic iteration for the solution of variational inequalities for monotone operators in Hilbert space.***Journal of Mathematical Analysis and Applications*1977,**61**(1):159â€“164. 10.1016/0022-247X(77)90152-4Lions J-L, Stampacchia G:

**Variational inequalities.***Communications on Pure and Applied Mathematics*1967,**20:**493â€“519. 10.1002/cpa.3160200302Liu F, Nashed MZ:

**Regularization of nonlinear ill-posed variational inequalities and convergence rates.***Set-Valued Analysis*1998,**6**(4):313â€“344. 10.1023/A:1008643727926Yamada I:

**The hybrid steepest descent method for the variational inequality problem over the intersection of fixed point sets of nonexpansive mappings.**In*Inherently Parallel Algorithms in Feasibility and Optimization and Their Applications (Haifa, 2000), Studies in Computational Mathematics*.*Volume 8*. North-Holland, Amsterdam, The Netherlands; 2001:473â€“504.Chen J, Zhang L, Fan T:

**Viscosity approximation methods for nonexpansive mappings and monotone mappings.***Journal of Mathematical Analysis and Applications*2007,**334**(2):1450â€“1461. 10.1016/j.jmaa.2006.12.088Iiduka H, Takahashi W, Toyoda M:

**Approximation of solutions of variational inequalities for monotone mappings.***Panamerican Mathematical Journal*2004,**14**(2):49â€“61.Iiduka H, Takahashi W:

**Strong convergence theorems for nonexpansive mappings and inverse-strongly monotone mappings.***Nonlinear Analysis: Theory, Methods & Applications*2005,**61**(3):341â€“350. 10.1016/j.na.2003.07.023Takahashi W, Toyoda M:

**Weak convergence theorems for nonexpansive mappings and monotone mappings.***Journal of Optimization Theory and Applications*2003,**118**(2):417â€“428. 10.1023/A:1025407607560Jung JS:

**A new iteration method for nonexpansive mappings and monotone mappings in Hilbert spaces.***Journal of Inequalities and Applications*2010,**2010:**-16.Deutsch F, Yamada I:

**Minimizing certain convex functions over the intersection of the fixed point sets of nonexpansive mappings.***Numerical Functional Analysis and Optimization*1998,**19**(1â€“2):33â€“56.Xu H-K:

**Iterative algorithms for nonlinear operators.***Journal of the London Mathematical Society*2002,**66**(1):240â€“256. 10.1112/S0024610702003332Xu H-K:

**An iterative approach to quadratic optimization.***Journal of Optimization Theory and Applications*2003,**116**(3):659â€“678. 10.1023/A:1023073621589Marino G, Xu H-K:

**A general iterative method for nonexpansive mappings in Hilbert spaces.***Journal of Mathematical Analysis and Applications*2006,**318**(1):43â€“52. 10.1016/j.jmaa.2005.05.028Moudafi A:

**Viscosity approximation methods for fixed-points problems.***Journal of Mathematical Analysis and Applications*2000,**241**(1):46â€“55. 10.1006/jmaa.1999.6615Xu H-K:

**Viscosity approximation methods for nonexpansive mappings.***Journal of Mathematical Analysis and Applications*2004,**298**(1):279â€“291. 10.1016/j.jmaa.2004.04.059Browder FE, Petryshyn WV:

**Construction of fixed points of nonlinear mappings in Hilbert space.***Journal of Mathematical Analysis and Applications*1967,**20:**197â€“228. 10.1016/0022-247X(67)90085-6Rockafellar RT:

**On the maximality of sums of nonlinear monotone operators.***Transactions of the American Mathematical Society*1970,**149:**75â€“88. 10.1090/S0002-9947-1970-0282272-5Rockafellar RT:

**Monotone operators and the proximal point algorithm.***SIAM Journal on Control and Optimization*1976,**14**(5):877â€“898. 10.1137/0314056Oden JT:

*Qualitative Methods on Nonlinear Mechanics*. Prentice-Hall, Englewood Cliffs, NJ, USA; 1986.Yao Y, Noor MA, Zainab S, Liou Y-C:

**Mixed equilibrium problems and optimization problems.***Journal of Mathematical Analysis and Applications*2009,**354**(1):319â€“329. 10.1016/j.jmaa.2008.12.055Jung JS:

**Iterative algorithms with some control conditions for quadratic optimizations.***Panamerican Mathematical Journal*2006,**16**(4):13â€“25.Jung JS, Cho YJ, Agarwal RP:

**Iterative schemes with some control conditions for a family of finite nonexpansive mappings in Banach spaces.***Fixed Point Theory and Applications*2005,**2005**(2):125â€“135.

## Acknowledgment

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science and Technology (2010-0017007).

## Author information

### Authors and Affiliations

### Corresponding author

## Rights and permissions

**Open Access** This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

## About this article

### Cite this article

Jung, J. A General Iterative Approach to Variational Inequality Problems and Optimization Problems.
*Fixed Point Theory Appl* **2011**, 284363 (2011). https://doi.org/10.1155/2011/284363

Received:

Accepted:

Published:

DOI: https://doi.org/10.1155/2011/284363