- Research
- Open Access
- Published:

# Modified Newton-type methods for the NCP by using a class of one-parametric NCP-functions

*Journal of Inequalities and Applications*
**volume 2012**, Article number: 286 (2012)

## Abstract

In this paper, we propose a new Newton-type method for solving the nonlinear complementarity problem (NCP) based on a class of one-parametric NCP-functions, where an approximate Newton direction can be obtained by solving a modified Newton equation in each iteration. The method is shown to be globally convergent without any additional assumption. To investigate the fast convergence of this class of methods, we propose a modified version of the proposed method and show the method is globally and locally superlinearly convergent. The preliminary numerical results show the effectiveness of the modified method.

## 1 Introduction

Consider the nonlinear complementarity problem NCP(F)

where F:{\mathcal{R}}^{n}\to {\mathcal{R}}^{n} is a continuously differentiable function. We assume that *F* is a {P}_{0}-function throughout this paper. It is well known that NCP(F) can be reformulated as a system of nonsmooth equations, where the so-called NCP-function plays an important role in this class of methods.

**Definition 1** A function \varphi :{\mathcal{R}}^{2}\to \mathcal{R} is called an NCP-function if it satisfies

Over the past two decades, a variety of NCP-functions have been studied (see, for example, [1–7]). Among them, a popular NCP-function is the well-known Fischer-Burmeister NCP-function [3] defined as

In this paper, we use a family of NCP-functions based on the FB function, which was introduced by Kanzow and Kleinmichel [6],

where *λ* is a fixed parameter such that \lambda \in (0,4). In the case of \lambda =2, the NCP-function {\varphi}_{\lambda} obviously reduces to the Fischer-Burmeister function.

By using {\varphi}_{\lambda} defined by (1), the NCP is equivalent to a system of nonsmooth equations

Let {\theta}_{\lambda}(x)=\frac{1}{2}{\parallel {\mathrm{\Phi}}_{\lambda}(x)\parallel}^{2}. Then solving NCP(F) is equivalent to solving the unconstrained minimization {min}_{x\in {R}^{n}}{\theta}_{\lambda}(x) with the optimal value 0.

Kanzow and Kleinmichel [6] studied the properties of {\mathrm{\Phi}}_{\lambda} and {\theta}_{\lambda} and proposed the corresponding semismooth Newton method. Their method first attempted to use the Newton direction, but if the Newton equation is unsolvable or the Newton direction is not a direction of sufficient decrease for {\theta}_{\lambda}, then it switches to the steepest descent direction. In this paper, we propose a Newton-type method for the {P}_{0}-NCP(F), where, in each iteration, we need to construct an approximation of \partial {\mathrm{\Phi}}_{\lambda}(x) (the Clarke subdifferential of {\mathrm{\Phi}}_{\lambda} at *x*, which is defined in the next section), which is nonsingular, and hence the direction-finding problem can be solved only by solving a system of perturbed Newton equations. We show that the proposed method is globally convergent without any additional assumption. The proposed method is similar to the one discussed by Yamashita and Fukushima [8], where the NCP-function {\varphi}_{\mathrm{FB}} was used. Since {\varphi}_{\mathrm{FB}} is a special case of {\varphi}_{\lambda}, the proposed method can be used more widely. However, it is hard for us to discuss the locally fast convergence of the proposed method. In order to investigate the locally fast convergence of this class of methods, we revise the proposed method. We show that the modified method is globally and locally superlinearly convergent. The preliminary numerical results show the effectiveness of the modified method.

## 2 Preliminaries

In this section, we recall some basic concepts and known results.

**Definition 2** F:{\mathcal{R}}^{n}\to {\mathcal{R}}^{n} is called a {P}_{0}-function if

**Definition 3** A matrix M\in {\mathcal{R}}^{n\times n} is a {P}_{0}-matrix if each of its principal minors is nonnegative.

It is known that the Jacobian of every continuously differentiable {P}_{0}-function is a {P}_{0}-matrix. The following theorem will play an important role in our analysis. Notice that, for a vector *a*, {D}_{a} denotes the diagonal matrix with the *i* th diagonal element being {a}_{i}.

**Theorem 4** (see [8])

*Let* *M* *be a* {P}_{0}-*matrix*, {D}_{a} *and* {D}_{b} *be negative definite diagonal matrices*. *Then* {D}_{a}+{D}_{b}M *is nonsingular*.

Let \mathrm{\Phi}:{\mathcal{R}}^{n}\to {\mathcal{R}}^{n} be locally Lipschitz continuous; by Rademacher’s theorem, Φ is differentiable almost everywhere.

**Definition 5** Let {D}_{\mathrm{\Phi}} denote the set \{x\in {\mathcal{R}}^{n}|\mathrm{\Phi}\text{is differentiable at}x\}, then the *B*-subdifferential of Φ at *x* is defined as

The Clarke subdifferential of Φ at *x* is defined as

where *co* denotes the convex hull of a set.

By the definition of {\mathrm{\Phi}}_{\lambda}, we know that {\mathrm{\Phi}}_{\lambda} is not differentiable at *x* if {x}_{i}=0={F}_{i}(x) for some *i*. However, since {\mathrm{\Phi}}_{\lambda} is locally Lipschitz continuous [[6], Lemma 2.1], {\partial}_{B}{\mathrm{\Phi}}_{\lambda}(x) is nonempty at every x\in {\mathcal{R}}^{n}. But how to specify the set {\partial}_{B}{\mathrm{\Phi}}_{\lambda}(x) exactly at *x* where \mathrm{\nabla}{\mathrm{\Phi}}_{\lambda}(x) does not exist?

To solve this problem, we construct two mappings \tilde{\mathcal{H}} and \stackrel{\u02c6}{\mathcal{H}} which approximate {\partial}_{B}{\mathrm{\Phi}}_{\lambda}. For a set *X*, we denote the power set of *X* by P(X).

Define the mapping \tilde{\mathcal{H}}:{\mathcal{R}}^{n}\to P({\mathcal{R}}^{n\times n}) as

where \tilde{\mathrm{\Omega}}:{\mathcal{R}}^{n}\to P({\mathcal{R}}^{2n}) is given by

with

Here, {C}_{\lambda} denotes the constant 2-\frac{\lambda (4-\lambda )}{8}, and

In the following, we define \stackrel{\u02c6}{\mathcal{H}} similarly to \tilde{\mathcal{H}}, which is a subset of \tilde{\mathcal{H}}.

The mapping \stackrel{\u02c6}{\mathcal{H}}:{\mathcal{R}}^{n}\to P({\mathcal{R}}^{n\times n}) is defined by

where \stackrel{\u02c6}{\mathrm{\Omega}}:{\mathcal{R}}^{n}\to P({\mathcal{R}}^{2n}) is defined by

Here Z(x)=\{z\in {\mathcal{R}}^{n}|{z}_{i}\ne 0,\text{if}i\in \beta \}, and *β* denotes the set \{i|{x}_{i}=0={F}_{i}(x)\}. The components of a vector g(x,z) are given by

and the components of a vector h(x,z) are given by

**Remark** From (2), we find that, for every x\in {\mathcal{R}}^{n}, (\tilde{a},\tilde{b})\in \tilde{\mathrm{\Omega}}(x) satisfies -\sqrt{{C}_{\lambda}}-1\le {\tilde{a}}_{i},{\tilde{b}}_{i}\le 0 (see [[6], Proposition 2.6]), and {\tilde{a}}_{i}, {\tilde{b}}_{i} do not vanish simultaneously. It is the same with elements in \stackrel{\u02c6}{\mathcal{H}}.

The mappings \tilde{\mathcal{H}} and \stackrel{\u02c6}{\mathcal{H}} have the following property which will play an important role in our analysis.

**Theorem 6** *For an arbitrary* x\in {\mathcal{R}}^{n}, *we have* \stackrel{\u02c6}{\mathcal{H}}(x)\subseteq {\partial}_{B}{\mathrm{\Phi}}_{\lambda}(x)\subseteq \tilde{\mathcal{H}}(x).

*Proof* {\partial}_{B}{\mathrm{\Phi}}_{\lambda}(x)\subseteq \tilde{\mathcal{H}}(x) was shown in [[6], Proposition 2.5]. Hence, we prove \stackrel{\u02c6}{\mathcal{H}}(x)\subseteq {\partial}_{B}{\mathrm{\Phi}}_{\lambda}(x) in the following.

For an arbitrary \stackrel{\u02c6}{H}\in \stackrel{\u02c6}{\mathcal{H}}(x), we shall build a sequence of points \{{y}^{k}\} where {\mathrm{\Phi}}_{\lambda} is differentiable at every {y}^{k} and such that \mathrm{\nabla}{\mathrm{\Phi}}_{\lambda}{({y}^{k})}^{T} tends to \stackrel{\u02c6}{H}; then the theorem will be obtained by the definition of *B*-subdifferential.

Let {y}^{k}=x+{\epsilon}^{k}z, where z\in Z(x) and \{{\epsilon}^{k}\} is a sequence of positive numbers converging to 0. If i\notin \beta, either {x}_{i}\ne 0 or {F}_{i}(x)\ne 0, and {z}_{i}\ne 0 for all i\in \beta.

We can see, by continuity, that if {\epsilon}^{k} is small enough, then for each *i*, either {y}_{i}^{k}\ne 0 or {F}_{i}({y}^{k})\ne 0, so {\mathrm{\Phi}}_{\lambda} is differentiable at {y}^{k}. If i\notin \beta, by continuity, the *i* th row of \mathrm{\nabla}{\mathrm{\Phi}}_{\lambda}{({y}^{k})}^{T} tends to the *i* th row of \stackrel{\u02c6}{H}. So, we only concern the case of i\in \beta.

From [[6], Proposition 2.5], we know that the *i* th row of \mathrm{\nabla}{\mathrm{\Phi}}_{\lambda}{({y}^{k})}^{T} is

where

By Taylor-expansion, we have, for each i\in \beta,

Substituting (4) into (3) and passing to limit, we have, by the continuity of ∇*F*, that the rows of \mathrm{\nabla}{\mathrm{\Phi}}_{\lambda}{({y}^{k})}^{T} tend to the corresponding rows of \stackrel{\u02c6}{H} when i\in \beta. Hence, \mathrm{\nabla}{\mathrm{\Phi}}_{\lambda}{({y}^{k})}^{T} tends to \stackrel{\u02c6}{H}. □

In this paper, we present two algorithms. The first one, which is presented in Section 3, uses matrices obtained by perturbing \tilde{H}\in \tilde{\mathcal{H}}. We will establish its global convergence. While in Section 4, we present another algorithm based on \stackrel{\u02c6}{H}\in \stackrel{\u02c6}{\mathcal{H}}, which is a restricted version of the first one. The second algorithm can be superlinearly convergent.

## 3 Algorithm and global convergence

Considering the Newton-type method, the direction-finding problem is solved by \tilde{H}d=-{\mathrm{\Phi}}_{\lambda}({x}^{k}), where \tilde{H}\in \tilde{\mathcal{H}}({x}^{k}). However, \tilde{H} is not necessarily nonsingular. In this section, we will perturb \tilde{H} to \tilde{G}, which is nonsingular. Then a search direction can be obtained by solving \tilde{G}d=-{\mathrm{\Phi}}_{\lambda}({x}^{k}). Now, let us construct \tilde{G} as follows.

First, mapping {\mathrm{\Lambda}}_{i}:{\mathcal{R}}^{n+2}\to P({\mathcal{R}}^{2}), i=1,2,\dots ,n are defined by

where \epsilon \in (0,1-\sqrt{{C}_{\lambda}/2}), and \sigma :{\mathcal{R}}^{+}\to {\mathcal{R}}^{+} is a nondecreasing continuous function such that \sigma (0)=0 and \sigma (t)>0 for all t>0.

Because \epsilon \in (0,1-\sqrt{\frac{{C}_{\lambda}}{2}}), it is obvious that for (a,b)\in \tilde{\mathrm{\Omega}}(x), the case of -\epsilon <{a}_{i} and -\epsilon <{b}_{i} will not happen.

In the following, we construct \tilde{G} as

where \tilde{p} and \tilde{q} are vectors such that

with (\tilde{a},\tilde{b})\in \tilde{\mathrm{\Omega}}(x), and ({\overline{a}}_{i},{\overline{b}}_{i})\in {\mathrm{\Lambda}}_{i}(x,{\tilde{a}}_{i},{\tilde{b}}_{i}), i=1,2,\dots ,n.

If {\theta}_{\lambda}(x)>0, the definition of {\mathrm{\Lambda}}_{i} and (5) imply that both {D}_{\tilde{p}}, {D}_{\tilde{p}} are negative definite matrices. Furthermore, we define \tilde{\mathcal{G}}:{\mathcal{R}}^{n}\to P({\mathcal{R}}^{n\times n}) as follows:

It is obvious that \tilde{G}={D}_{\tilde{p}}+{D}_{\tilde{q}}{F}^{\prime}(x) and \tilde{H}={D}_{\tilde{a}}+{D}_{\tilde{b}}{F}^{\prime}(x) are closely related. \tilde{G} is nonsingular under proper conditions.

**Theorem 7** *If* *x* *is not a solution of* NCP(F), *i*.*e*., {\theta}_{\lambda}(x)>0, *then every* \tilde{G}\in \tilde{\mathcal{G}}(x) *is nonsingular*.

*Proof* For every \tilde{G}\in \tilde{\mathcal{G}}(x), if {\theta}_{\lambda}(x)>0, then it follows from the definition of \tilde{G} that {D}_{\tilde{p}} and {D}_{\tilde{q}} are negative definite matrices.

Since *F* is a {P}_{0}-function, the Jacobian of *F* is a {P}_{0}-matrix. So, {F}^{\prime}({x}^{k}) is a {P}_{0}-matrix. Hence, by Theorem 4, \tilde{G} is nonsingular. □

By the mapping \tilde{\mathcal{G}}, we define \tilde{\mathrm{\Delta}}:{\mathcal{R}}^{n}\to P({\mathcal{R}}^{n}) as

It is easy to see that \tilde{\mathrm{\Delta}}(x) is nonempty for every *x* such that {\theta}_{\lambda}(x)>0. Now we give the first algorithm.

**Algorithm 1** Step 1. Initialization: choose \lambda \in (0,4), {x}^{0}\in {\mathcal{R}}^{n}, \rho \in (0,0.5), \beta \in (0,1), and set k:=0.

Step 2. Termination criterion: if {\theta}_{\lambda}(x)=0, stop. Otherwise, go to Step 3.

Step 3. Search direction calculation: find a vector {d}^{k}\in \tilde{\mathrm{\Delta}}({x}^{k}).

Step 4. Line search: let *m* be the smallest nonnegative integer such that

Step 5. Update: set {x}^{k+1}:={x}^{k}+{t}_{k}{d}^{k}, where {t}_{k}={\beta}^{m}, k:=k+1, and go to Step 2.

It is obvious that if {\theta}_{\lambda}({x}^{k})=0, then {x}^{k} is a solution of NCP(F). Next, we will prove the global convergence of Algorithm 1. First, we show that every d\in \tilde{\mathrm{\Delta}}(x) is a descent direction of {\theta}_{\lambda} at *x*.

**Lemma 8** (see [[8], Lemma 3.2])

*If* *x* *is not a solution of* NCP(F), *i*.*e*., {\theta}_{\lambda}(x)>0, *then every* d\in \tilde{\mathrm{\Delta}}(x) *satisfies the descent condition for* {\theta}_{\lambda}, *i*.*e*., \mathrm{\nabla}{\theta}_{\lambda}{(x)}^{T}d<0.

**Theorem 9** *Every accumulation point of a sequence* \{{x}^{k}\} *generated by Algorithm * 1 *is a solution of* NCP(F).

*Proof* Owing to Step 4, \{{\theta}_{\lambda}({x}^{k})\} is decreasing monotonically and nonnegative. It must converge to some {\theta}_{\lambda}^{\ast}\ge 0. We assume {\theta}_{\lambda}^{\ast}>0. Let {x}^{\ast} be an accumulation point of \{{x}^{k}\} and {\{{x}^{k}\}}_{k\in \mathcal{K}} be a subsequence converging to {x}^{\ast}.

\tilde{\mathrm{\Delta}} is uniformly compact near {x}^{\ast} and closed at {x}^{\ast} (see [[8], Lemma 3.4]), we assume, without loss of generality, that lim\underset{}{{}_{k\to \mathrm{\infty}}k\in \mathcal{K}}{d}^{k}={d}^{\ast}\in \tilde{\mathrm{\Delta}}({x}^{\ast}). From Lemma 8, we will get the contradiction if we can prove \mathrm{\nabla}{\theta}_{\lambda}{({x}^{\ast})}^{T}{d}^{\ast}=0. This can be obtained by considering the following two cases:

• Suppose that inf\{{t}_{k}\}\ge t>0. Then we have

It is obvious that \mathrm{\nabla}{\theta}_{\lambda}{({x}^{\ast})}^{T}{d}^{\ast}=0 is satisfied.

• Suppose that inf\{{t}_{k}\}=0. In this case, we assume lim\underset{k\to \mathrm{\infty}}{k\in \mathcal{K}}{t}_{k}=0 without loss of generality. By line search, we have

taking the limit yields \mathrm{\nabla}{\theta}_{\lambda}{({x}^{\ast})}^{T}{d}^{\ast}\ge \rho \mathrm{\nabla}{\theta}_{\lambda}{({x}^{\ast})}^{T}{d}^{\ast}. Since \rho \in (0,0.5), we have \mathrm{\nabla}{\theta}_{\lambda}{({x}^{\ast})}^{T}{d}^{\ast}\ge 0. Hence, \mathrm{\nabla}{\theta}_{\lambda}{({x}^{\ast})}^{T}{d}^{\ast}=0.

We get the contradiction. The proof is complete. □

## 4 Modified algorithm and fast convergence

In the above section, we established global convergence of Algorithm 1. It determines a search direction based on \tilde{\mathcal{H}} which contains the generalized Jacobian {\partial}_{B}{\mathrm{\Phi}}_{\lambda}(x). However, it is hard for us to show the superlinear convergence of Algorithm 1. In the following, we should modify the search direction properly to accelerate the convergence of algorithm. By the definition of \stackrel{\u02c6}{\mathcal{H}}, we know that \stackrel{\u02c6}{H}\in \stackrel{\u02c6}{\mathcal{H}} is not necessarily nonsingular. Can we perturb \stackrel{\u02c6}{H} similar to \tilde{H}? Next, we give a positive answer to this question.

Define \stackrel{\u02c6}{G} as

where \stackrel{\u02c6}{p} and \stackrel{\u02c6}{q} are vectors such that

with (\stackrel{\u02c6}{a},\stackrel{\u02c6}{b})\in \stackrel{\u02c6}{\mathrm{\Omega}}(x), and ({\overline{a}}_{i},{\overline{b}}_{i})\in {\mathrm{\Lambda}}_{i}(x,{\stackrel{\u02c6}{a}}_{i},{\stackrel{\u02c6}{b}}_{i}), i=1,2,\dots ,n.

If {\theta}_{\lambda}(x)>0, the definition of {\mathrm{\Lambda}}_{i} and (6) imply that both {D}_{\stackrel{\u02c6}{p}}, {D}_{\stackrel{\u02c6}{p}} are negative definite matrices.

Mapping \stackrel{\u02c6}{\mathcal{G}}:{\mathcal{R}}^{n}\to P({\mathcal{R}}^{n\times n}) is defined by

From Theorem 6, \stackrel{\u02c6}{\mathcal{H}}\subseteq \tilde{\mathcal{H}}. It is obvious that \stackrel{\u02c6}{\mathcal{G}}\subseteq \tilde{\mathcal{G}}. And from Theorem 7, every \stackrel{\u02c6}{G}\in \stackrel{\u02c6}{\mathcal{G}}(x) is nonsingular if {\theta}_{\lambda}(x)>0.

Define \stackrel{\u02c6}{\mathrm{\Delta}}:{\mathcal{R}}^{n}\to P({\mathcal{R}}^{n}) as

For any *x*, since \stackrel{\u02c6}{\mathcal{G}}(x)\subseteq \tilde{\mathcal{G}}(x), we obtain \stackrel{\u02c6}{\mathrm{\Delta}}(x)\subseteq \tilde{\mathrm{\Delta}}(x). Hence, \stackrel{\u02c6}{\mathrm{\Delta}}(x) is nonempty for every *x* such that {\theta}_{\lambda}(x)>0. Next, we will give the second algorithm. The search direction is chosen from \stackrel{\u02c6}{\mathrm{\Delta}}(x). The only difference from Algorithm 1 is the search direction.

**Algorithm 2** Step 1. Initialization: choose \lambda \in (0,4), {x}^{0}\in {\mathcal{R}}^{n}, \rho \in (0,0.5), \beta \in (0,1), and set k:=0.

Step 2. Termination criterion: if {\theta}_{\lambda}(x)=0, stop. Otherwise, go to Step 3.

Step 3. Search direction calculation: find a vector {d}^{k}\in \stackrel{\u02c6}{\mathrm{\Delta}}({x}^{k}).

Step 4. Line search: let *m* be the smallest nonnegative integer such that

Step 5. Update: set {x}^{k+1}:={x}^{k}+{t}_{k}{d}^{k}, where {t}_{k}={\beta}^{m}, k:=k+1, and go to Step 2.

Since \stackrel{\u02c6}{\mathrm{\Delta}}({x}^{k})\subseteq \tilde{\mathrm{\Delta}}({x}^{k}) at each {x}^{k} as mentioned above, global convergence of Algorithm 2 is directly obtained from Theorem 9. We state the following theorem without proof.

**Theorem 10** *Every accumulation point of a sequence* \{{x}^{k}\} *generated by Algorithm * 2 *is a solution of* NCP(F).

In the following, we focus our attention on the superlinear convergence rate of Algorithm 2. To begin with, we assume that the sequence \{{x}^{k}\} generated by Algorithm 2 has a unique limit point {x}^{\ast}.

**Lemma 11** *We have* \parallel {x}^{k}+{d}^{k}-{x}^{\ast}\parallel =o(\parallel {x}^{k}-{x}^{\ast}\parallel ).

*Proof* For each *k*, we have

where {\stackrel{\u02c6}{H}}_{k}\in \stackrel{\u02c6}{\mathcal{H}}({x}^{k}) is the matrix corresponding to {\stackrel{\u02c6}{G}}_{k}. Since {\mathrm{\Phi}}_{\lambda} is semismooth [[6], Lemma 2.2] and \stackrel{\u02c6}{\mathcal{H}}({x}^{k})\in {\partial}_{B}{\mathrm{\Phi}}_{\lambda}({x}^{k}) for each *k*, by Theorem 6, we have

(see the proof of [[9], Theorem 3.1]). Moreover, by the definition of {\mathrm{\Lambda}}_{i} and (6), we have

Consequently, it follows that

Since \sigma ({\theta}_{\lambda}({x}^{k}))\to 0 and \{\parallel {\stackrel{\u02c6}{G}}_{k}^{-1}\parallel \} is bounded (see the proof of [[8], Lemma 4.2]), we obtain the desired result. □

Now, we prove the superlinear convergence of Algorithm 2.

**Theorem 12** *Algorithm * 2 *has a superlinear rate of convergence*.

*Proof* We have {x}^{k+1}={x}^{k}+{d}^{k} for all *k* sufficiently large (see the proof of [[8], Lemma 4.6]). It then follows from Lemma 11 that

The proof is complete. □

## 5 Numerical results

In this section, we do some preliminary numerical experiments to test Algorithm 2 and compare its performance with that of the algorithms proposed in Chen and Pan [2] and Sun and Zeng [10].

First, we set \beta =0.5, \rho ={10}^{-4}, \epsilon =0.05 and \sigma (t)=0.1min\{1,t\}.

For z\in Z(x), we define

The stopping criterion for Algorithm 2 is {\theta}_{\lambda}({x}^{k})\le {10}^{-8}. The programs are coded in MATLAB and run on a personal computer with a 2.1 GHZ CPU processor.

The meaning of the columns in the tables are stated as follows:

iter: the total iteration number, resi: the value of {\theta}_{\lambda}(x).

**Problem 1** Let F(x)=Ax+q, where

The corresponding complementarity problem has the unique solution. Table 1 lists the test results for Problem 1 with different *n*, *λ* and initial points a={(-5,\dots ,-5)}^{T}, b={(0,\dots ,0)}^{T}, c={(1,\dots ,1)}^{T}, d={(8,\dots ,8)}^{T}.

From Table 1, we see that the test results for \lambda \in (0,1) are better than for other cases. Especially, the good numerical results are obtained when *λ* closes to 0. Then we compare the test results with Chen and Pan [2], where we set p=2, \epsilon =1.0\mathrm{e}-08, \sigma =1.0\mathrm{e}-10, \beta =0.2 for convenience. Table 2 lists the test results for [2].

Tables 1 and 2 indicate that Algorithm 2 performed much better than Chen and Pan [2] did on Problem 1.

**Problem 2** Free boundary problems can also be solved by the method we presented. The following problem arises from the discretization of a free boundary problem (see [10]). Let \mathrm{\Omega}=(0,1)\times (0,1) and a function *g* satisfy g(x,0)=x(1-x), g(x,y)=0 on x=0,1 or y=1.

Consider the following problem: find *u* such that

where f(u,x,y) is a continuously differentiable {P}_{0}-function. We discretize the problem by the five-point difference scheme with mesh-step *h*. Then we get the following complementarity problem: find x\in {\mathcal{R}}^{n} such that

Set initial point as {x}_{0}={(0,\dots ,0)}^{T}. Table 3 lists the test results with different functions *f*, *λ*, and *h*.

From Table 3, we have the following observations.

• Our test results become better when *λ* decreases. It is obvious that when \lambda =2 the result is not good enough. That is to say, Algorithm 2 with the NCP-function {\varphi}_{\lambda} (0<\lambda <2) is better than the one discussed in [8] where the Fischer-Burmeister function was used.

• Whether the function f(u,x,y) is linear or nonlinear, the test results are good. The results are especially better when *λ* closes to 0.

We compare the test results with Sun and Zeng [10] where we set \beta =0.5, c={0.5}^{4}. Table 4 lists the test results for [10] with different functions *f* and *h*.

Tables 3 and 4 indicate that Algorithm 2 performed as well as Sun and Zeng [10] did on Problem 2.

**Problem 3** We implemented Algorithm 2 for some test problems with all available starting points in MCPLIB [11]. The results are reported in Table 5 with seconds for unit of time.

The above examples indicate that the results are better when *λ* closes to 0. A reasonable interpretation for this is that the values of {g}_{i}(x,z) and {h}_{i}(x,z) become smaller when *λ* increases and hence causes some difficulty for Algorithm 2. This also implies that the performance of Algorithm 2 will become worse when *p* increases. When \lambda \to 0, the NCP obviously reduces to min\{x,F(x)\}=0. But it is a nonsmooth equation so we cannot use this method.

## 6 Concluding remarks

In this paper, we have studied a class of one-parametric NCP-functions {\varphi}_{\lambda}(\cdot ,\cdot ) which include the well-known Fischer-Burmeister function as a special case and proposed modified Newton-type algorithms for solving {P}_{0} complementarity problems.

Numerical results for the test problems have shown that this method is promising when \lambda \in (0,4). Moreover, our numerical results indicated that the performance of the modified Newton-type method becomes better when *λ* decreases, which is a new and important numerical result. We believe that Algorithm 2 can effectively solve more practical problems if they can be reformulated as an NCP(F). We leave this as a future research topic.

## References

Chen J-S:

**On some NCP-functions based on the generalized Fischer-Burmeister function.***Asia-Pac. J. Oper. Res.*2007, 24: 401–420. 10.1142/S0217595907001292Chen J-S, Pan S:

**A family of NCP functions and a descent method for the nonlinear complementarity problem.***Comput. Optim. Appl.*2008, 40: 389–404. 10.1007/s10589-007-9086-0Fischer A:

**A special Newton-type optimization methods.***Optimization*1992, 24: 269–284. 10.1080/02331939208843795Hu SL, Huang ZH, Chen J-S:

**Properties of a family of generalized NCP-functions and a derivative free algorithm for complementarity problems.***J. Comput. Appl. Math.*2009, 230(1):69–82. 10.1016/j.cam.2008.10.056Hu SL, Huang ZH, Lu N:

**Smoothness of a class of merit functions for the second-order cone complementarity problem.***Pac. J. Optim.*2010, 6(3):551–571.Kanzow C, Kleinmichel H:

**A new class of semismooth Newton-type methods for nonlinear complementarity problems.***Comput. Optim. Appl.*1998, 11: 227–251. 10.1023/A:1026424918464Lu LY, Huang ZH, Hu SL:

**Properties of a family of merit functions and a derivative-free method for the NCP.***Appl. Math. J. Chin. Univ. Ser. A*2010, 25(4):379–390. 10.1007/s11766-010-2179-zYamashita N, Fukushima M:

**Modified Newton methods for solving a semismooth reformulation of monotone complementarity problems.***Math. Program.*1997, 76: 469–491.Qi L:

**Convergence analysis of some algorithms for solving nonsmooth equations.***Math. Oper. Res.*1993, 18: 227–244. 10.1287/moor.18.1.227Sun Z, Zeng JP:

**A monotone semismooth Newton type method for a class of complementarity problem.***J. Comput. Appl. Math.*2011, 235: 1261–1274. 10.1016/j.cam.2010.08.012Billups SC, Dirkse SP, Soares MC:

**A comparison of algorithms for large scale mixed complementarity problems.***Comput. Optim. Appl.*1997, 7: 3–25. 10.1023/A:1008632215341

## Acknowledgements

This work was supported by the NSFC (50975200).

## Author information

### Affiliations

### Corresponding author

## Additional information

### Competing interests

The authors declare that they have no competing interests.

### Authors’ contributions

WX participated in the design of the algorithm. ZD performed the numerical experiment and statistical analysis. All authors read and approved the final manuscript.

## Rights and permissions

**Open Access** This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

## About this article

### Cite this article

Xie, W., Deng, Z. Modified Newton-type methods for the NCP by using a class of one-parametric NCP-functions.
*J Inequal Appl* **2012, **286 (2012). https://doi.org/10.1186/1029-242X-2012-286

Received:

Accepted:

Published:

DOI: https://doi.org/10.1186/1029-242X-2012-286

### Keywords

- nonlinear complementarity problem
- NCP-function
- generalized Newton method
- global convergence
- superlinear convergence