- Research
- Open access
- Published:
A projection method for bilevel variational inequalities
Journal of Inequalities and Applications volume 2014, Article number: 205 (2014)
Abstract
A fixed point iteration algorithm is introduced to solve bilevel monotone variational inequalities. The algorithm uses simple projection sequences. Strong convergence of the iteration sequences generated by the algorithm to the solution is guaranteed under some assumptions in a real Hilbert space.
MSC:65K10, 90C25.
1 Introduction
Let C be a nonempty closed convex subset of a real Hilbert space ℋ with the inner product and the norm . We denote weak convergence and strong convergence by notations ⇀ and →, respectively. The bilevel variational inequalities, shortly (BVI), are formulated as follows:
where , denotes the set of all solutions of the variational inequalities:
and . We denote the solution set of problem (BVI) by Ω.
Bilevel variational inequalities are special classes of quasivariational inequalities (see [1–4]) and of equilibrium with equilibrium constraints considered in [5]. However, they cover some classes of mathematical programs with equilibrium constraints (see [6]), bilevel minimization problems (see [7]), variational inequalities (see [8–13]), minimum-norm problems of the solution set of variational inequalities (see [14, 15]), bilevel convex programming models (see [16]) and bilevel linear programming in [17].
Suppose that . It is well known in convex programming that if f is convex and differentiable on , then is a solution to
if and only if is the solution to the bilevel variational inequalities , where ∇f is the gradient of f. Then the bilevel variational inequalities (BVI) are written in a form of mathematical programs with equilibrium constraints as follows:
If f, g are two convex and differentiable functions, then problem (BVI) (where and ) becomes the following bilevel minimization problem (see [7]):
In a special case for all , problem (BVI) becomes the minimum-norm problems of the solution set of variational inequalities as follows:
where is the projection of 0 onto . A typical example is the least-squares solution to the constrained linear inverse problem in [18]. For solving this problem under the assumption that the subset is nonempty closed convex, is α-inverse strongly monotone and , Yao et al. in [14] introduced the following extended extragradient method:
They showed that under certain conditions over parameters, the sequence converges strongly to .
Recently, Anh et al. in [19] introduced an extragradient algorithm for solving problem (BVI) in the Euclidean space . Roughly speaking the algorithm consists of two loops. At each iteration k of the outer loop, they applied the extragradient method for the lower variational inequality problem. Then, starting from the obtained iterate in the outer loop, they computed an -solution of problem . The convergence of the algorithm crucially depends on the starting points and the parameters chosen in advance. More precisely, they presented the following scheme
Under assumptions that F is strongly monotone and Lipschitz continuous, G is pseudomonotone and Lipschitz continuous on C, the sequences of parameters were chosen appropriately. They showed that two iterative sequences and converged to the same point which is a solution of problem (BVI). However, at each iteration of the outer loop, the scheme requires computing an approximation solution to a variational inequality problem.
There exist some other solution methods for bilevel variational inequalities when the cost operator has some monotonicity (see [16, 19–21]). In all of these methods, solving auxiliary variational inequalities is required. In order to avoid this requirement, we combine the projected gradient method in [10] for solving variational inequalities and the fixed point property that is a solution to problem if and only if it is a fixed point of the mapping , where . Then, the strong convergence of proposed sequences is considered in a real Hilbert space.
In this paper, we are interested in finding a solution to bilevel variational inequalities (BVI), where the operators F and G satisfy the following usual conditions:
(A1) G is η-inverse strongly monotone on ℋ and F is β-strongly monotone on C.
(A2) F is L-Lipschitz continuous on C.
(A3) The solution set Ω of problem (BVI) is nonempty.
The purpose of this paper is to propose an algorithm for directly solving bilevel pseudomonotone variational inequalities by using the projected gradient method and fixed point techniques.
The rest of this paper is divided into two sections. In Section 2, we recall some properties for monotonicity, the metric projection onto a closed convex set and introduce in detail a new algorithm for solving problem (BVI). The third section is devoted to the convergence analysis for the algorithm.
2 Preliminaries
We list some well-known definitions and the projection under the Euclidean norm which will be used in our analysis.
Definition 2.1 Let C be a nonempty closed convex subset in â„‹. We denote the projection on C by , i.e.,
The operator is said to be
-
(i)
γ-strongly monotone on C if for each ,
-
(ii)
η-inverse strongly monotone on C if for each ,
-
(iii)
Lipschitz continuous with constant (shortly L-Lipschitz continuous) on C if for each ,
If and , then φ is called nonexpansive on C.
We know that the projection has the following well-known basic properties.
Property 2.2
-
(a)
, .
-
(b)
, .
-
(c)
, .
To prove the main theorem of this paper, we need the following lemma.
Lemma 2.3 (see [21])
Let be β-strongly monotone and L-Lipschitz continuous, and . Then the mapping for all satisfies the inequality
where .
Lemma 2.4 (see [22])
Let â„‹ be a real Hilbert space, C be a nonempty closed and convex subset of â„‹ and be a nonexpansive mapping. Then (I is the identity operator on â„‹) is demiclosed at , i.e., for any sequence in C such that and , we have .
Lemma 2.5 (see [19])
Let be a sequence of nonnegative real numbers such that
where and is a sequence in â„› such that
-
(a)
,
-
(b)
or .
Then .
Now we are in a position to describe an algorithm for problem (BVI). The proposed algorithm can be considered as a combination of the projected gradient and fixed point methods. Roughly speaking the algorithm consists of two steps. First, we use the well-known projected gradient method for solving the variational inequalities (), where and . The method generates a sequence converging strongly to the unique solution of problem under assumptions that G is L-Lipschitz continuous and α-strongly monotone on C with the step-size . Next, we use the Banach contraction-mapping fixed-point principle for finding the unique fixed point of the contraction-mapping , where F is β-strongly monotone and L-Lipschitz continuous, I is the identity mapping, and . The algorithm is presented in detail as follows.
Algorithm 2.6 (Projection algorithm for solving (BVI))
Step 0. Choose , , a positive sequence , λ, μ such that
Step 1. Compute
Update , and go to Step 1.
Note that in the case for all , Algorithm 2.6 becomes the projected gradient algorithm as follows:
3 Convergence results
In this section, we state and prove our main results.
Theorem 3.1 Let C be a nonempty closed convex subset of a real Hilbert space â„‹. Let two mappings and satisfy assumptions (A1)-(A3). Then the sequences and in Algorithm 2.6 converge strongly to the same point .
Proof For conditions (2.1), we consider the mapping defined by
Using Property 2.2(a), G is η-inverse strongly monotone and conditions (2.1), for each , we have
Combining this and Lemma 2.3, we get
where . Thus, is a contraction on â„‹. By the Banach contraction principle, there is the unique fixed point such that . For each , set
Due to this and Property 2.2(a), we get that the mapping is contractive on â„‹. Then there exists the unique point such that . Set . It follows from (3.2) that
Thus, , and hence . Therefore, there exists a subsequence of the sequence such that . Combining this and the assumption , we get
It follows from (3.1) that the mapping is nonexpansive on â„‹. Using Lemma 2.4, (3.3) and , we obtain , which implies . Now we will prove that .
Set , and , where I is the identity mapping. Since and , we have
and
Then
By the Schwarz inequality, we have
and
Combining (3.4), (3.5) and (3.6), we get
Then we have
Let , and hence the sequence converges strongly to . This implies that the sequence also converges strongly to .
Otherwise, by using (3.2), we have
Moreover, by Lemma 2.3, we have
This implies that
and hence
So, we have
Let
Then
Since is bounded, for all , we have
Applying Lemma 2.5, . Combining this and the fact that the sequence converges strongly to , the sequence also converges strongly to the unique solution to problem (BVI). □
Now we consider the special case for all . It is easy to see that F is Lipschitz continuous with constant and strongly monotone with constant on â„‹. Problem (BVI) becomes the minimum-norm problems of the solution set of the variational inequalities.
Corollary 3.2 Let C be a nonempty closed convex subset of a real Hilbert space ℋ. Let be η-inverse strongly monotone. The iteration sequence is defined by
The parameters satisfy the following:
Then the sequences and converge strongly to the same point .
References
Anh PN, Muu LD, Hien NV, Strodiot JJ: Using the Banach contraction principle to implement the proximal point method for multivalued monotone variational inequalities. J. Optim. Theory Appl. 2005, 124: 285–306. 10.1007/s10957-004-0926-0
Baiocchi C, Capelo A: Variational and Quasivariational Inequalities: Applications to Free Boundary Problems. Wiley, New York; 1984.
Facchinei F, Pang JS: Finite-Dimensional Variational Inequalities and Complementary Problems. Springer, New York; 2003.
Xu MH, Li M, Yang CC: Neural networks for a class of bi-level variational inequalities. J. Glob. Optim. 2009, 44: 535–552. 10.1007/s10898-008-9355-1
Moudafi A: Proximal methods for a class of bilevel monotone equilibrium problems. J. Glob. Optim. 2010, 47: 287–292. 10.1007/s10898-009-9476-1
Luo ZQ, Pang JS, Ralph D: Mathematical Programs with Equilibrium Constraints. Cambridge University Press, Cambridge; 1996.
Solodov M: An explicit descent method for bilevel convex optimization. J. Convex Anal. 2007, 14: 227–237.
Anh PN: An interior-quadratic proximal method for solving monotone generalized variational inequalities. East-West J. Math. 2008, 10: 81–100.
Anh PN: An interior proximal method for solving pseudomonotone non-Lipschitzian multivalued variational inequalities. Nonlinear Anal. Forum 2009, 14: 27–42.
Anh PN, Muu LD, Strodiot JJ: Generalized projection method for non-Lipschitz multivalued monotone variational inequalities. Acta Math. Vietnam. 2009, 34: 67–79.
Daniele P, Giannessi F, Maugeri A: Equilibrium Problems and Variational Models. Kluwer Academic, Dordrecht; 2003.
Giannessi F, Maugeri A, Pardalos PM: Equilibrium Problems: Nonsmooth Optimization and Variational Inequality Models. Kluwer Academic, Dordrecht; 2004.
Konnov IV: Combined Relaxation Methods for Variational Inequalities. Springer, Berlin; 2000.
Yao, Y, Marino, G, Muglia, L: A modified Korpelevich’s method convergent to the minimum-norm solution of a variational inequality. Optimization, 1–11, iFirst (2012)
Zegeye H, Shahzad N, Yao Y: Minimum-norm solution of variational inequality and fixed point problem in Banach spaces. Optimization 2013. 10.1080/02331934.2013.764522
Trujillo-Cortez R, Zlobec S: Bilevel convex programming models. Optimization 2009, 58: 1009–1028. 10.1080/02331930701763330
Glackin J, Ecker JG, Kupferschmid M: Solving bilevel linear programs using multiple objective linear programming. J. Optim. Theory Appl. 2009, 140: 197–212. 10.1007/s10957-008-9467-2
Sabharwal A, Potter LC: Convexly constrained linear inverse problems: iterative least-squares and regularization. IEEE Trans. Signal Process. 1998, 46: 2345–2352. 10.1109/78.709518
Anh PN, Kim JK, Muu LD: An extragradient method for solving bilevel variational inequalities. J. Glob. Optim. 2012, 52: 627–639. 10.1007/s10898-012-9870-y
Kalashnikov VV, Kalashnikova NI: Solving two-level variational inequality. J. Glob. Optim. 1996, 8: 289–294. 10.1007/BF00121270
Iiduka H: Strong convergence for an iterative method for the triple-hierarchical constrained optimization problem. Nonlinear Anal. 2009, 71: e1292-e1297. 10.1016/j.na.2009.01.133
Suzuki T: Strong convergence of Krasnoselskii and Mann’s type sequences for one parameter nonexpansive semigroups without Bochner integrals. J. Math. Anal. Appl. 2005, 305: 227–239. 10.1016/j.jmaa.2004.11.017
Acknowledgements
The authors are very grateful to the anonymous referees for their really helpful and constructive comments that helped us very much in improving the paper.
Author information
Authors and Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
The main idea of this paper was proposed by TTHA. The revision is made by LBL and TVA. All authors read and approved the final manuscript.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Anh, T.T., Long, L.B. & Anh, T.V. A projection method for bilevel variational inequalities. J Inequal Appl 2014, 205 (2014). https://doi.org/10.1186/1029-242X-2014-205
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/1029-242X-2014-205