Modified HS conjugate gradient method for solving generalized absolute value equations

Abstract

We investigate a kind of generalized equations involving absolute values of variables as $$|A|x-|B||x|=b$$, where $$A \in R^{n\times n}$$ is a symmetric matrix, $$B \in R^{n\times n}$$ is a diagonal matrix, and $$b\in R^{n}$$. A sufficient condition for unique solvability of the proposed generalized absolute value equations is also given. By utilizing an equivalence relation to the unconstrained optimization problem, we propose a modified HS conjugate gradient method to solve the transformed unconstrained optimization problem. Only under mild conditions, the global convergence of the given method is also established. Finally, the numerical results show the efficiency of the proposed method.

Introduction

The absolute value equation of the type

$$Ax+B \vert x \vert =b$$
(1.1)

was investigated in [14, 22, 25,26,27,28]. If $$\det (B)\neq 0$$, then (1.1) can be reduced to the form

$$Ax- \vert x \vert =b.$$
(1.2)

The absolute value Eq. (1.2) has also been intensively studied, e.g., see [9, 12, 13, 15,16,17, 19, 30, 33, 34]. In this paper, we propose a new generalized absolute value equation (GAVE) problem of the form

$$\vert A \vert x- \vert B \vert \vert x \vert =b,$$
(1.3)

where $$A=(a_{ij})\in R^{n\times n}$$ is a symmetric matrix, $$B=(b_{ij})\in R^{n\times n}$$ is a diagonal matrix, the absolute values of matrices are defined as $$|A|=(|a_{ij}|)$$, $$|B|=(|b_{ij}|)$$, $$i, j=1,2,\ldots,n$$, $$b\in R^{n}$$ and $$|x|=(|x_{1}|,|x_{2}|,\ldots,|x_{n}|)^{T}$$. As we all know, the study of absolute value equations comes from linear complementarity problem. The general linear complementarity problem [5], which subsumes many mathematical programming problems, bimatrix games, and equilibrium programming problems, can be formulated as the absolute value equations of the forms such as (1.1)–(1.3). Mangasarian [14] showed that (1.1) is NP-hard. Prolopyev [22] stated the relations of (1.1) with linear complementarity problem and mixed integer programming problem. Rohn et al. [30] gave the sufficient conditions for unique solvability of AVE (1.2) and an iterative method to solve it. Mangasarian et al. [17] gave the existence and nonexistence results of (1.2) and proved the equivalence relations between (1.2) and the generalized linear complementarity problem. Hu et al. [9] proved that (1.2) can be equivalently reformulated as a standard linear complementarity problem without any assumption. In [16] and [15], Mangasarian proposed a concave minimization optimization method and a generalized Newton method, respectively. Zhang et al. [34] presented a generalized Newton method. Noor et al. [19] gave an iterative algorithm for solving (1.2). Yong [33] proposed a smoothing Newton algorithm to solve (1.2). Saheya et al. [31] focused on numerical comparisons based on four smoothing functions for (1.2). Bello Cruz et al. [2] showed the global Q-linear convergence of the inexact semi-smooth Newton method for solving (1.2). Ke et al. [10] studied a SOR-like iteration method for solving system of (1.2). Abdallah et al. [1] reformulated (1.2) as a sequence of concave minimization problems and gave a smoothing method to solve it. Cacceta et al. [4] proposed a smoothing Newton method with global and quadratic convergence for solving (1.2). Rohn [29] proved a theorem of alternatives for equation $$|Ax|-|B||x|=b$$ and gave some sufficient conditions for solvability of the equation. The current research on the methods for solving (1.1) and (1.2) is based mostly on nonlinear optimization techniques. Little attention, however, has been paid so far to the nonlinear conjugate gradient method with smaller storage capacity and faster convergence speed for solving GAVE (1.3). In this paper, we propose a modified HS conjugate gradient method to compute the solution of GAVE (1.3).

This paper is organized as follows. In Sect. 2, we provide a sufficient condition for the solution of GAVE (1.3). In Sect. 3, a modified HS conjugate gradient method for solving GAVE (1.3) is given. Under mild conditions, we prove the global convergence theorem of the given method. In Sect. 4, we present numerical results of the relevant numerical experiments to show the effectiveness and the efficiency of the proposed method.

Throughout the paper, lowercase $$x, y, \ldots$$ denote vectors, $$\beta ,\varepsilon ,\ldots$$ denote parameters, uppercase letters $$A, B, \ldots$$ denote matrices.

General absolute value equation and unconstrained optimization problem

We will start by showing that (1.3) is equivalent to an unconstrained optimization problem with an objective function that is continuously differentiable. Firstly, we introduce the relevant definition from [24, 25].

Definition 2.1

Given two matrices $$E, F \in R^{n \times n}$$ and $$F \geq 0$$, the set of matrices

$$\varSigma =\bigl\{ \widetilde{A} | \vert E \vert -F \leq \widetilde{A}\leq \vert E \vert +F \bigr\}$$

is called an interval matrix. Σ is called regular if Ã is nonsingular.

Theorem 2.1

Suppose that $$\widetilde{\varSigma }=\{\tilde{A}||A|-F \leq \widetilde{A}\leq |A|+F \}$$ is regular and $$|B| \leq F$$, then (1.3) has a unique solution.

Proof

By $$x^{+} = \max \{x,0\} = \frac{|x|+x}{2}$$, $$x^{-} =\min \{x,0\} = \frac{|x|-x}{2}$$, we get

$$\vert x \vert = x^{+} + x^{-},\qquad x=x^{+} - x^{-}.$$

Then (1.3) can be rewritten as

$$x^{+}=\bigl( \vert A \vert - \vert B \vert \bigr)^{-1}\bigl( \vert A \vert + \vert B \vert \bigr)x^{-}+\bigl( \vert A \vert - \vert B \vert \bigr)^{-1}b.$$
(2.1)

From $$|B| \leq F$$, we know that $$|A|-|B|, |A|+|B| \in \widetilde{\varSigma }$$ and $$(|A|-|B|)^{-1}$$ exists. Similar to Theorem 1 in [25], by [18, 23], we know that (2.1) has a unique solution. Hence (1.3) has a unique solution. We finish the proof. □

In the remaining part of this section, we transform (1.3) to an unconstraint optimization problem. Denote

$$f(x)=\bigl\langle { \vert A \vert x,x}\bigr\rangle -\bigl\langle { \vert B \vert \vert x \vert ,x}\bigr\rangle -2 \langle {b,x}\rangle ,$$
(2.2)

where A, B are defined similarly as (1.3). $$\langle {\cdot } \rangle$$ denotes the inner product of vectors, namely $$\forall x, y \in R ^{n}$$

$$\langle {x,y} \rangle =\sum_{i=1}^{n} x_{i}y_{i}.$$

Now, we give the related notation and lemmas.

Definition 2.2

Suppose that matrix $$A \in R^{n \times n}$$ is symmetric, then A is a positive definite matrix if and only if $$\langle x,Ax \rangle >0$$ is set up for arbitrarily nonzero vector $$x \in R^{n}$$.

In the remainder of this paper, we consider the matrices A and B such that $$|A|-|B|D$$ is positive definite for any arbitrary matrix D. If A is symmetric and B, D are both diagonal matrices, then $$|A|-|B|D$$ is symmetric. The diagonal matrix D is defined as $$D=\partial |x|=\operatorname{diag}(\operatorname{sign}(x))$$, where $$\operatorname{diag}(x)$$ denote a vector with components equal to 1, 0, −1 depending on whether the corresponding component of x is positive, zero, or negative.

Theorem 2.2

If matrix $$|A|-|B|D$$ is a positive definite matrix, then x is a solution of (1.3) x is a minimum of the function $$f(x)$$, where $$f(x)$$ is defined as (2.2).

Proof

Case I. For arbitrary $$\alpha \in R$$ and $$\nu \in R^{n}$$, by Taylor’s series, we get

$$f(x+\alpha \nu )=f(x)+\alpha \bigl\langle \nabla f(x),\nu \bigr\rangle + \frac{ \alpha ^{2}}{2} \bigl\langle \nabla ^{2} f(x)\nu ,\nu \bigr\rangle ,$$

where $$\nabla f(x)=2(|A|x-|B||x|-b)$$, $$\nabla ^{2}f(x)=2(|A|-|B|D)$$. Let $$C=|A|-|B|D$$, then C is a positive definite matrix and

$$f(x+\alpha \nu )=f(x)+2\alpha \bigl\langle \vert A \vert x- \vert B \vert \vert x \vert -b,\nu \bigr\rangle + \alpha ^{2} \langle C \nu ,\nu \rangle .$$

Let $$g:R^{n} \rightarrow R$$ be a function about α, we get

$$g(\alpha )=f(x)+2\alpha \bigl\langle \vert A \vert x- \vert B \vert \vert x \vert -b,\nu \bigr\rangle +\alpha ^{2} \langle C \nu ,\nu \rangle ,$$

then g has the minimum point with $$\langle C\nu , \nu \rangle >0$$,

$$\alpha =-\frac{ \langle \vert A \vert x- \vert B \vert \vert x \vert -b,\nu \rangle }{\langle C \nu , \nu \rangle },$$

and

$$g(\alpha )=f(x)-\frac{ \langle \vert A \vert x- \vert B \vert \vert x \vert -b,\nu \rangle ^{2}}{\langle C \nu ,\nu \rangle }.$$

So, we have

$$f(x+\alpha \nu ) \leq f(x),\quad \forall \nu \neq 0.$$

The above strict inequality is impossible. Then we have

$$\bigl\langle \vert A \vert x- \vert B \vert \vert x \vert -b,\nu \bigr\rangle =0.$$

And it follows that

$$f(x)=f(x+\alpha \nu ).$$

If $$x^{*}$$ satisfies $$|A|x^{*}-|B||x^{*}|=b$$, then $$\langle |A|x^{*}-|B||x ^{*}|-b,\nu \rangle =0$$ for arbitrary ν and $$f(x)$$ cannot be made any smaller than $$f(x^{*})$$. Then $$x^{*}$$ minimizes f.

Case II. Suppose that $$x^{*}$$ is the minimum point of $$f(x)$$, then $$\forall \nu \in R^{n}$$, $$\alpha \in R$$, it follows that

$$f\bigl(x^{*}+\alpha \nu \bigr) \geq f\bigl(x^{*}\bigr).$$

So,

$$\bigl\langle \vert A \vert x^{*}- \vert B \vert \bigl\vert x^{*} \bigr\vert -b,\nu \bigr\rangle =0.$$

Then the above equation implies that

$$\vert A \vert x^{*}- \vert B \vert \bigl\vert x^{*} \bigr\vert -b=0,$$

that is,

$$\vert A \vert x^{*}- \vert B \vert \bigl\vert x^{*} \bigr\vert =b.$$

This shows that $$x^{*}$$ is a solution of (1.3). Hence, this completes the proof. □

Therefore, GAVE (1.3) can be transformed into the following unconstrained optimization problem:

$$\min_{x \in R} f(x),$$

where f is defined by formula (2.2). It is well known that nonlinear conjugate gradient methods such as Hestenes–Stiefel (HS) method [8], Fletcher–Reeves (FR) method [7], Polak–Ribiere–Polyak (PRP) method [20, 21], Dai–Yuan (DY) method [6], and other methods [3, 11, 32, 35, 36] are very efficient for large-scale smooth optimization problems due to their simplicity and low storage. Moreover, we notice that some modified HS conjugate gradient methods are more efficient to solve the unconstrained optimization problem than classical methods, see [11, 32]. In the next section, we give a modified HS conjugate gradient method for (1.3). To develop an efficient optimization method for (1.3), we also use the Armijo-type line search globalization technique [36].

In this section, we firstly propose the modified HS conjugate gradient method based on [11] with Armijo-type line search based on [36]. Then we present the global convergence of the given method under mild conditions.

Algorithm 3.1

Step 0. Choose initial point $$x_{0}\in R^{n}$$ and constants $$\delta _{1}, \delta _{2}, \rho \in (0,1)$$, $$\varepsilon >0$$. Let $$k:=0$$.

Step 1. Denote $$g_{k}=\nabla f(x_{k})$$. If $$\|g_{k}\| \leq \varepsilon$$, stop. Otherwise, compute $$d_{k}$$ by

$$d_{k}= \textstyle\begin{cases} -g_{k},& \mbox{if }k=0, \\ - g_{k}+ \beta ^{\operatorname{NHS}}_{k} d_{k-1}-\beta ^{\operatorname{NHS}}_{k}\frac{g^{T}_{k}d_{k-1}}{ \Vert g_{k} \Vert ^{2}}g_{k}, & \mbox{if } k>0, \end{cases}$$
(3.1)

where

$$\beta _{k}^{\operatorname{NHS}}=\frac{g_{k}^{T}(g_{k}-g_{k-1})}{ z_{k} }, \quad z_{k}=\max \bigl\{ t \Vert d_{k-1} \Vert ,d_{k-1}^{T}(g_{k}-g_{k-1}) \bigr\} .$$

Step 2. Determine $$\alpha _{k}$$ by the Armijo-type line search, that is, $$\alpha _{k}=\max \{ \rho ^{j},j=0,1,2,\ldots\}$$ satisfying

$$f(x_{k}+\alpha _{k}d_{k})\leq f(x_{k})+\delta _{1} \alpha _{k} g_{k}^{T}d _{k}-\delta _{2} \alpha _{k}^{2} \Vert d_{k} \Vert ^{2}.$$
(3.2)

Step 3. Set $$x_{k+1}=x_{k}+\alpha _{k}d_{k}$$, $$k:=k+1$$. Go to Step 1.

To get the convergence of Algorithm 3.1, we only need the following mild assumption.

Assumption 3.1

The level set $$\varOmega =\{x|f(x)\leq f(x _{0})\}$$ is bounded.

Lemma 3.1

Let Assumption 3.1 hold, $$g(x)=2(|A|x-|B||x|-b)$$, then $$g(x)$$ satisfies the Lipschitz condition, that is,

$$\bigl\Vert g(x)-g(y) \bigr\Vert \leq L \Vert x-y \Vert ,$$

where $$\forall x,y\in N$$, N is a neighborhood of Ω and $$L>0$$ is a constant.

Proof

By $$g(x)=2(|A|x-|B||x|-b)$$, we get

\begin{aligned} \bigl\Vert g(x)-g(y) \bigr\Vert &=2 \bigl\Vert \vert A \vert x- \vert B \vert \vert x \vert -b- \vert A \vert y+ \vert B \vert \vert y \vert +b \bigr\Vert \\ & =2 \bigl\Vert \vert A \vert x- \vert B \vert \vert x \vert - \vert A \vert y+ \vert B \vert \vert y \vert \bigr\Vert \\ & =2 \bigl\Vert \vert A \vert (x-y)- \vert B \vert \bigl( \vert x \vert - \vert y \vert \bigr) \bigr\Vert \\ & \leq 2 \bigl\Vert \vert A \vert (x-y) \bigr\Vert +2 \bigl\Vert \vert B \vert \bigl( \vert x \vert - \vert y \vert \bigr) \bigr\Vert \\ &\leq 2\bigl(\vert \!\Vert A\vert \!\Vert + \vert \!\Vert B\vert \!\Vert \bigr) \bigl( \Vert x-y \Vert \bigr)=L \Vert x-y \Vert . \end{aligned}

Denote $$L=2(|\!|\!|A|\!|\!|+|\!|\!|B|\!|\!|)$$, we get this lemma. □

Remark 3.1

On account of the descent property of $$\{f(x_{k})\}$$, the sequence $$\{x_{k}\}$$ generated by Algorithm 3.1 is contained in Ω. Besides, it follows from Assumption 3.1 that there exists a constant $$\eta >0$$ such that

$$\bigl\Vert g(x) \bigr\Vert \leq \eta ,\quad \forall x \in \varOmega .$$

Lemma 3.2

([11])

Let $$\{d_{k}\}$$ be computed by Algorithm 3.1, then

$$g_{k}^{T}d_{k}=- \Vert g_{k} \Vert ^{2}$$
(3.3)

holds for arbitrary $$k>0$$.

From Assumption 3.1, Lemma 3.1, and Lemma 3.2, we can get the following lemma.

Lemma 3.3

([11])

Suppose that Assumption 3.1 holds. Let $$\{d_{k}\}$$ and $$\{x_{k}\}$$ be generated by Algorithm 3.1, then there exists a positive constant c such that

$$\Vert d_{k} \Vert \leq c,\quad \forall k>0.$$
(3.4)

Based on the above assumptions and lemmas, we now give the global convergence theorem of Algorithm 3.1.

Theorem 3.1

Suppose that Assumption 3.1 holds. If $$\{x_{k}\}$$ is generated by Algorithm 3.1, then

$$\liminf_{k\rightarrow \infty } \Vert g_{k} \Vert =0.$$
(3.5)

Proof

Now, assume that this theorem is not true, namely (3.5) does not hold, then there exists a positive constant $$\tau >0$$ such that

$$\Vert g_{k} \Vert \geq \tau ,\quad \forall k\geq 0.$$
(3.6)

From Assumption 3.1 and (3.2), it follows

$$\sum_{k\geq 0}\bigl(-\delta _{1} \alpha _{k} g_{k}^{T}d_{k}+\delta _{2} \alpha _{k}^{2} \Vert d_{k} \Vert ^{2}\bigr)< \infty ,$$

this with (3.3) indicates

$$\sum_{k\geq 0} \alpha _{k}^{2} \Vert d_{k} \Vert ^{2}< \infty ,$$

and

$$-\sum_{k\geq 0}\alpha _{k} g_{k}^{T}d_{k}= \sum _{k\geq 0} \alpha _{k} \Vert g_{k} \Vert ^{2} < \infty ,$$

then we obtain

$$\lim_{k\rightarrow \infty }\alpha _{k} \Vert g_{k} \Vert ^{2}=0.$$
(3.7)

If $$\liminf_{k \to \infty } {\alpha }_{k} > 0$$, then we have $$\liminf_{k\rightarrow \infty } \| g_{k}\|=0$$ by (3.7), which contradicts (3.6).

If $$\liminf_{k\rightarrow \infty } \alpha _{k}=0$$, then there exists a set $$K\in N$$ such that

$$\lim_{k\in K, k\rightarrow \infty }\alpha _{k}=0.$$
(3.8)

The Armijo-type line search rule suggests that $$\rho ^{-1}\alpha _{k}$$ does not satisfy line search condition (3.2) for k sufficiently enough, namely

\begin{aligned} f\bigl(x_{k}+\rho ^{-1} \alpha _{k}d_{k} \bigr)-f(x_{k}) >& \delta _{1}\rho ^{-1} \alpha _{k} g_{k}^{T}d_{k}-\delta _{2} \rho ^{-2}\alpha _{k}^{2} \Vert d_{k} \Vert ^{2} \\ =& -\delta _{1}\rho ^{-1} \alpha _{k} \Vert g_{k} \Vert ^{2}-\delta _{2} \rho ^{-2}\alpha _{k}^{2} \Vert d_{k} \Vert ^{2}. \end{aligned}
(3.9)

By the mean value theorem and Lemma 3.1, there exists $$\xi _{k} \in (0,1)$$ such that

\begin{aligned} f\bigl(x_{k}+\rho ^{-1} \alpha _{k}d_{k} \bigr)-f(x_{k}) &=\rho ^{-1} \alpha _{k} g\bigl(x _{k}+\xi _{k}\rho ^{-1} \alpha _{k}d_{k}\bigr)^{T}d_{k} \\ &=\rho ^{-1} \alpha _{k} g_{k}^{T}d_{k}+ \rho ^{-1} \alpha _{k} \bigl[g\bigl(x_{k}+ \xi _{k}\rho ^{-1} \alpha _{k}d_{k} \bigr)-g(x_{k})\bigr]^{T}d_{k} \\ &\leq \rho ^{-1} \alpha _{k} g_{k}^{T}d_{k} + L\rho ^{-2}\alpha _{k}^{2} \Vert d_{k} \Vert ^{2} \\ &=-\rho ^{-1} \alpha _{k} \Vert g_{k} \Vert ^{2}+L\rho ^{-2}\alpha _{k}^{2} \Vert d _{k} \Vert ^{2}. \end{aligned}

This together with Lemma 3.3 and (3.9) implies

$$\Vert g_{k} \Vert ^{2} < \frac{L+\delta _{2}}{(1-\delta _{1})\rho }c^{2} \alpha _{k}.$$

Then we obtain $$\liminf_{k\in K, k\rightarrow \infty } \|g_{k}\|=0$$ from (3.8), which also contradicts (3.6). The proof is completed. □

Remark 3.2

In Step 2 of Algorithm 3.1, we adopt the Armijo-type line search [36]. The following line searches are also well defined in Algorithm 3.1 since the search directions are descent. The Wolfe line search [6]

$$f(x_{k}+\alpha _{k}d_{k})\leq f(x_{k})+\rho \alpha _{k} g_{k}^{T}d_{k}$$

and

$$g_{k+1}^{T}d_{k}\geq \sigma g_{k}^{T}d_{k},\quad \rho \in \biggl(0,\frac{1}{2}\biggr), \sigma \in (\rho ,1),$$

and the standard Armijo line search [35]

$$f(x_{k}+\alpha _{k}d_{k})\leq f(x_{k})+\rho _{1} \alpha _{k} g_{k}^{T}d _{k},\quad \rho _{1} \in (0,1).$$
(3.10)

Numerical experiments

In this section, we present numerical results to show the efficiency of the modified HS conjugate gradient method (Algorithm 3.1). The numerical testing was carried out on a Lenovo PC with the use of Matlab. The following tables and figures list the numerical results for the given GAVE problems, where we set $$\varepsilon =10^{-6}$$, $$\rho =0.6$$, $$\rho _{1}=0.4$$, $$\delta _{1}=0.4$$, $$\delta _{2}=0.4$$, $$t=2$$.

Example 4.1

Consider GAVE (1.1), where

$$A= \begin{pmatrix} 7&2&2 \\ 2&7&2 \\ 2&2&7 \end{pmatrix}, \qquad B= \begin{pmatrix} 3&0&0 \\ 0&-3&0 \\ 0&0&-3 \end{pmatrix},\qquad b= \begin{pmatrix} 8 \\ 8 \\ 8 \end{pmatrix}.$$

The exact solution of Example 4.1 is $$(1,1,1)^{T}$$. The initial points in Algorithm 3.1 are taken randomly five times. The detailed numerical results are showed in Table 1 and Fig. 1. $$x^{*}$$ denotes the numerical solution, k denotes the number of iterations, and Val denotes $$\||A|x_{k}-|B||x_{k}|-b\|_{\infty }$$. From Table 1 and Fig. 1, we can see that Algorithm 3.1 is promising.

Example 4.2

Consider GAVE (1.1), where

\begin{aligned} &A= \begin{pmatrix} 6&-3&-3&-3&-3&-3 \\ -3&6&-3&-3&-3&-3 \\ -3&-3&6&-3&-3&-3 \\ -3&-3&-3&6&-3&-3 \\ -3&-3&-3&-3&6&-3 \\ -3&-3&-3&-3&-3&6 \end{pmatrix},\\ & B= \begin{pmatrix} 2&0&0&0&0&0 \\ 0&-1&0&0&0&0 \\ 0&0&2&0&0&0 \\ 0&0&0&-1&0&0 \\ 0&0&0&0&2&0 \\ 0&0&0&0&0&-1 \end{pmatrix},\qquad b= \begin{pmatrix} 19 \\ 20 \\ 19 \\ 20 \\ 19 \\ 20 \end{pmatrix}. \end{aligned}

The exact solution of this example is $$(1,1,1,1,1,1)^{T}$$. Compute this example by Algorithm 3.1 with random initial points uniformly distributed in $$(0,1)$$. The results of the numerical experiments are showed in Table 2, where $$x^{*}$$ denotes the numerical solution, k denotes the number of iterations, and Val denotes $$\||A|x_{k}-|B||x _{k}|-b\|_{\infty }$$. From Table 2 and Fig. 2, we can see that Algorithm 3.1 is also efficient to get the solution of this kind of GAVE.

Example 4.3

Consider GAVE (1.1), where $$A \in R^{n\times n}$$ whose diagonal elements are 2n and other elements are 1, $$B \in R^{n\times n}$$ whose diagonal elements are n and other elements are 0, and $$b=(2n-1)e$$. The exact solution of this example is $$(1,1,\ldots,1)^{T}$$. We use random initial points uniformly distributed in $$(0,1)$$ to compute this example by Algorithm 3.1 with Armijo line search (3.10) and Algorithm 3.1 stops at iteration $$x_{k}$$ if $$\||A|x_{k}-|B||x_{k}|-b\|<10^{-3}$$. The results of the numerical experiments are showed in Table 3, where n denotes the dimension of the vector, $$x^{*}$$ denotes the numerical solution, k denotes the number of iterations, and Val denotes $$\||A|x_{k}-|B||x _{k}|-b\|_{\infty }$$. Figure 3 represents the number of iterations with $$n=300$$. From Table 3 and Fig. 3, we can see that Algorithm 3.1 can also efficiently get the solution of this kind of GAVE.

Conclusions

Absolute value equation problem has been widely used in mathematical programming and other related areas of science and engineering. However, little attention has been paid to solving general absolute value equation problems by the nonlinear conjugate gradient method. In this paper, we provide a sufficient condition for the unique solution of general absolute value equation of the form as (1.3) and propose a modified HS conjugate gradient method to solve it. The global convergence of the nonlinear conjugate gradient method is proved under only one mild assumption. This method is also very easy to implement and is also very promising.

References

1. Abdallah, L., Haddou, M., Migot, T.: Solving absolute value equation using complementarity and smoothing functions. J. Comput. Appl. Math. 327, 196–207 (2018)

2. Bello Cruz, J.Y., Ferreira, O.P., Prudente, L.F.: On the global convergence of the inexact semi-smooth Newton method for absolute value equation. Comput. Optim. Appl. 65(1), 93–108 (2016)

3. Birgin, E.G., Martinez, J.M.: A spectral conjugate gradient method for unconstrained optimization. Appl. Math. Optim. 43, 117–128 (2001)

4. Caccetta, L., Qu, B., Zhou, G.L.: A globally and quadratically convergent method for absolute value equations. Comput. Optim. Appl. 48, 45–58 (2011)

5. Cottle, R.W., Pang, J.S., Stone, R.E.: The Linear Complementarity Problem. Academic Press, New York (1992)

6. Dai, Y.H., Yuan, Y.: A nonlinear conjugate gradient method with a strong global convergence property. SIAM J. Optim. 10, 177–182 (1999)

7. Fletcher, R., Reeves, C.M.: Function minimization by conjugate gradients. Comput. J. 7, 149–154 (1964)

8. Hestenes, M.R., Stiefel, E.L.: Methods of conjugate gradients for solving linear systems. J. Res. Natl. Bur. Stand. 49, 409–436 (1952)

9. Hu, S.L., Huang, Z.H.: A note on absolute value equations. Optim. Lett. 4(3), 417–424 (2010)

10. Ke, Y.F., Ma, C.F.: SOR-like iteration method for solving absolute value equations. Appl. Math. Comput. 311, 195–202 (2017)

11. Li, M., Qu, A.P.: Some sufficient descent conjugate gradient methods and their global convergence. Comput. Appl. Math. 33, 333–347 (2014)

12. Liu, C.H., Liu, H.W., Zhu, J.G.: A new semi-smooth Newton method for absolute value equations. Chin. J. Eng. Math. 30(1), 101–111 (2013)

13. Magasarian, O.L.: Knapsack feasibility as an absolute value equation solvable by successive linear programming. Optim. Lett. 3(2), 161–170 (2009)

14. Mangasarian, O.L.: Absolute value programming. Comput. Optim. Appl. 36(1), 43–53 (2007)

15. Mangasarian, O.L.: Absolute value equation solution via concave minimization. Optim. Lett. 1(1), 3–8 (2007)

16. Mangasarian, O.L.: A generalized Newton method for absolute values. Optim. Lett. 3(1), 101–108 (2009)

17. Mangasarian, O.L., Meyer, R.R.: Absolute value equations. Linear Algebra Appl. 419(2), 359–367 (2006)

18. Murty, K.G.: On the number of solutions to the complementarity problem and spanning properties of complementary cones. Linear Algebra Appl. 5, 65–108 (1972)

19. Noor, M.A., Iqbal, J., Noor, K.I., Al-Said, E.: On an iterative method for solving absolute value equations. Optim. Lett. 6(5), 1027–1033 (2012)

20. Polak, E., Ribière, G.: Note sur la convergence de methodes de directions conjuguees. Rev. Fr. Autom. Inform. Rech. Opér. 16, 35–43 (1969)

21. Polyak, B.T.: The conjugate gradient method in extremal problems. USSR Comput. Math. Math. Phys. 9(4), 94–112 (1969)

22. Prokopyev, O.: On equivalent reformulations for absolute value equations. Comput. Optim. Appl. 44(3), 363–372 (2009)

23. Rohn, J.: Systems of linear interval equations. Linear Algebra Appl. 126, 39–78 (1989)

24. Rohn, J.: Interval matrices: singularity and real eigenvalues. SIAM J. Matrix Anal. Appl. 14, 82–91 (1993)

25. Rohn, J.: A theorem of the alternatives for the equation $$Ax+B|x|=b$$. Linear Multilinear Algebra 52(6), 421–426 (2004)

26. Rohn, J.: On unique solvability of the absolute value equation. Optim. Lett. 3(4), 603–606 (2009)

27. Rohn, J.: An algorithm for solving the absolute value equation. Electron. J. Linear Algebra 18(1), 589–599 (2009)

28. Rohn, J.: An algorithm for computing all solutions of an absolute value equation. Optim. Lett. 6(5), 851–856 (2012)

29. Rohn, J.: A theorem of alternatives for equation $$|Ax|-|B||x|=b$$. Optim. Lett. 6(3), 585–591 (2012)

30. Rohn, J., Hooshyarbarkhsh, V., Farhadsefat, R.: An iterative method for solving absolute value equations and sufficient conditions for unique solvability. Optim. Lett. 8(1), 35–44 (2014)

31. Saheya, B., Yu, C.H., Chen, J.S.: Numerical comparisons based on four smoothing functions for absolute value equations. J. Appl. Math. Comput. 56, 131–149 (2018)

32. Sun, Q.Y., Liu, Q.: Global convergence of modified HS conjugate gradient method. J. Appl. Math. Comput. 22, 289–297 (2006)

33. Yong, L.Q.: A smoothing Newton method for absolute value equation. Int. J. Control. Autom. Syst. 9(2), 119–132 (2016)

34. Zhang, C., Wei, Q.J.: Global and finite convergence of a generalized Newton method for absolute value equations. J. Optim. Theory Appl. 143(2), 391–403 (2009)

35. Zhang, L., Zhou, W.J.: On the global convergence of the Hager–Zhang conjugate gradient method with Armijo line search. Acta Math. Sci. 28, 840–845 (2008)

36. Zhang, L., Zhou, W.J., Li, D.H.: Global convergence of a modified Fletcher–Reeves conjugate gradient method with Armijo-type line search. Numer. Math. 104(4), 561–572 (2006)

Not applicable.

Funding

This work was supported by the Shandong Provincial Nature Science Foundation, China (No. ZR2016AM29) and also by the National Natural Science Foundation of China (No. 11671220).

Author information

Authors

Contributions

All authors contributed equally. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Shouqiang Du.

Ethics declarations

Competing interests

The authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and Permissions

Li, Y., Du, S. Modified HS conjugate gradient method for solving generalized absolute value equations. J Inequal Appl 2019, 68 (2019). https://doi.org/10.1186/s13660-019-2018-6

• Accepted:

• Published:

• DOI: https://doi.org/10.1186/s13660-019-2018-6

• 65K05
• 90C30

Keywords

• Generalized absolute value equations
• Unconstrained optimization
• Modified HS conjugate gradient method
• Global convergence