- Research
- Open access
- Published:
Optimal lower and upper bounds for the geometric convex combination of the error function
Journal of Inequalities and Applications volume 2015, Article number: 382 (2015)
Abstract
For \(x\in R\), the error function \(\operatorname{erf}(x)\) is defined as
In this paper, we answer the question: what are the greatest value p and the least value q, such that the double inequality \(\operatorname {erf}(M_{p}(x,y;\lambda))\leq G(\operatorname{erf}(x),\operatorname {erf}(y);\lambda)\leq\operatorname{erf}(M_{q}(x,y;\lambda))\) holds for all \(x,y\geq1\) (or \(0< x,y<1\)) and \(\lambda\in(0,1)\)? Here, \(M_{r}(x,y;\lambda)=(\lambda x^{r}+(1-\lambda)y^{r})^{1/r}\) (\(r\neq0\)), \(M_{0}(x,y;\lambda)=x^{\lambda}y^{1-\lambda}\) and \(G(x,y;\lambda )=x^{\lambda}y^{1-\lambda}\) are the weighted power and the weighted geometric mean, respectively.
1 Introduction
For \(x\in R\), the error function \(\operatorname{erf}(x)\) is defined as
The most important properties of this function are collected, for example, in [1, 2]. In the recent past, the error function has been a topic of recurring interest, and a great number of results on this subject have been reported in the literature [3–16]. It might be surprising that the error function has application in the field of heat conduction besides probability [17, 18].
In 1933, Aumann [19] introduced a generalized notion of convexity, the so-called MN-convexity, when M and N are mean values. A function \(f:[0,\infty)\to[0,\infty)\) is MN-convex if \(f(M(x,y))\leq N(f(x),f(y))\) for \(x,y\in[0,\infty)\). The usual convexity is the special case when M and N both are arithmetic means. Furthermore, the applications of MN-convexity reveal a new world of beautiful inequalities which involve a broad range of functions from the elementary ones, such as sine and cosine function, to the special ones, such as the Γ function, the Gaussian hypergeometric function, and the Bessel function. For the details as regards MN-convexity and its applications the reader is referred to [20–25].
Let \(\lambda\in(0,1)\), we define \(A(x,y;\lambda)=\lambda x+(1-\lambda)y\), \(G(x,y;\lambda)=x^{\lambda}y^{1-\lambda}\), \(H(x,y;\lambda)=\frac{xy}{\lambda y+(1-\lambda)x}\) and \(M_{r}(x,y;\lambda)=(\lambda x^{r}+(1-\lambda)y^{r})^{1/r}\) (\(r\neq0\)), \(M_{0}(x,y;\lambda)=x^{\lambda}y^{1-\lambda}\). These are commonly known as weighted arithmetic mean, weighted geometric mean, weighted harmonic mean, and weighted power mean of two positive numbers x and y, respectively. Then it is well known that the inequalities
hold for all \(\lambda\in(0,1)\) and \(x,y>0\) with \(x\neq y\).
By elementary computations, one has
and
In [26], Alzer proved that \(c_{1}(\lambda)=\frac{\lambda+(1-\lambda )\operatorname{erf}(1)}{\operatorname{erf}(1/(1-\lambda))}\) and \(c_{2}(\lambda)=1\) are the best possible factors such that the double inequality
holds for all \(x, y \in[1,+\infty)\) and \(\lambda\in(0,1/2)\).
Inspired by (1.2), it is natural to ask: does the inequality \(\operatorname{erf}(M(x,y))\leq N(\operatorname{erf}(x),\operatorname {erf}(y))\) hold for other means M, N, such as geometric, harmonic or power means?
In [27, 28], the authors found the greatest values \(\alpha_{1}\), \(\alpha_{2}\) and the least values \(\beta_{1}\), \(\beta_{2}\), such that the double inequalities
and
hold for all \(x,y\geq1\) (or \(0< x,y<1\)) and \(\lambda\in(0,1)\).
In the following we answer the question: what are the greatest value p and the least value q, such that the double inequality
holds for all \(x,y\geq1\) (or \(0< x,y<1\)) and \(\lambda\in(0,1)\)?
2 Lemmas
In this section we present two lemmas, which will be used in the proof of our main results.
Lemma 2.1
Let \(r\neq0\), \(r_{0}=-1-\frac{2}{e\sqrt{\pi}\operatorname{erf}(1)}=-1.4926\ldots \) , and \(u(x)=\log\operatorname{erf}(x^{1/r})\). Then the following statements are true:
-
(1)
if \(r< r_{0}\), then \(u(x)\) is strictly convex on \([1,+\infty)\);
-
(2)
if \(r_{0}\leq r<0\), then \(u(x)\) is strictly concave on \((0,1]\);
-
(3)
if \(r>0\), then \(u(x)\) is strictly concave on \((0,+\infty)\).
Proof
Simple computations lead to
and
where
Then
and
We divide the proof into two cases.
Case 1. If \(r<0\), then (2.6), (2.5), and (2.3) lead to
and
Inequality (2.7) implies that \(g_{1}(x)\) is strictly decreasing on \([0,+\infty)\).
It follows from the monotonicity of \(g_{1}(x)\) and (2.8) that there exists \(x_{1}\in(0,+\infty)\), such that \(g(x)\) is strictly increasing on \([0,x_{1}]\) and strictly decreasing on \([x_{1},+\infty)\).
From the piecewise monotonicity of \(g(x)\) and (2.9) we clearly see that there exists \(x_{2}\in(0,+\infty)\), such that \(g(x)<0\) for \(x\in(0,x_{2})\) and \(g(x)>0\) for \(x\in(x_{2},+\infty)\).
Case 1.1. If \(r< r_{0}\), then from (2.10) we know that \(g(1)>0\). This leads to \(g(x)>0\) for \(x\in[1,+\infty)\). Therefore (2.2) leads to the conclusion that \(u(x)\) is strictly convex on \([1,+\infty)\).
Case 1.2. If \(r_{0}\leq r<0\), then (2.10) implies that \(g(1)\leq0\). This leads to \(g(x)\leq0\) for \(x\in(0,1]\). Therefore (2.2) leads to the conclusion that \(u(x)\) is strictly concave on \((0,1]\).
Case 2. If \(r>0\), then (2.5) and (2.3) imply that
and
for \(x\in(0,+\infty)\).
It follows from (2.11), (2.4), and (2.12) that \(g(x)<0\). Therefore (2.2) leads to the conclusion that \(u(x)\) is strictly concave on \((0,+\infty)\). □
Lemma 2.2
The function \(h(x)=2x^{2}+\frac{xe^{-x^{2}}}{\int_{0}^{x}e^{-t^{2}}\,dt}\) is strictly increasing on \((0,+\infty)\).
Proof
Simple computations lead to
where
and
for \(x\in(0,+\infty)\).
Hence, \(h(x)\) is strictly increasing on \((0,+\infty)\), as follows from (2.15), (2.14), and (2.13). □
3 Main results
Theorem 3.1
Let \(\lambda\in(0,1)\) and \(r_{0}=-1-\frac {2}{e\sqrt{\pi}\operatorname{erf}(1)}=-1.4926\ldots\) . Then the double inequality
holds for all \(x,y\geq1\) if and only if \(p=-\infty\) and \(q\geq r_{0}\).
Proof
First of all, we prove that inequality (3.1) holds if \(p=-\infty\) and \(q\geq r_{0}\). It follows from (1.1) that the first inequality in (3.1) is true if \(p=-\infty\). Since the weighted power mean \(M_{t}(x,y;\lambda)\) is strictly increasing with respect to t on R, thus we only need to prove that the second inequality in (3.1) is true if \(r_{0}\leq q<0\).
If \(r_{0}\leq q<0\), \(u(z)=\log\operatorname{erf}(z^{1/q})\), then Lemma 2.1(2) leads to
for \(\lambda\in(0,1)\) and \(s,t\in(0,1]\).
Let \(s=x^{q}\), \(t=y^{q}\), and \(x,y\geq1\). Then (3.2) leads to the second inequality in (3.1).
Second, we prove that the second inequality in (3.1) implies \(q\geq r_{0}\).
Let \(x\geq1\) and \(y\geq1\). Then the second inequality in (3.1) leads to
It follows from (3.3) that
and
Therefore,
follows from (3.3) and (3.4) together with Lemma 2.2.
Finally, we prove that the first inequality in (3.1) implies \(p=-\infty\). We distinguish two cases.
Case I. \(p\geq0\). Then for any fixed \(y\in[1,+\infty)\) we have
and
which contradicts the first inequality in (3.1).
Case II. \(-\infty< p<0\). Let \(x\geq1\), \(\alpha=\lambda ^{1/p}\) and \(y\to +\infty\). Then the first inequality in (3.1) leads to
It follows from (3.5) that
and
Note that \(\alpha>1\), then
It follows from (3.7) and (3.8) that there exists a sufficiently large \(\eta_{1}\in[1,+\infty)\), such that \(E'(x)>0\) for \(x\in(\eta_{1},+\infty)\). Hence \(E(x)\) is strictly increasing on \([\eta_{1},+\infty)\).
From the monotonicity of \(E(x)\) on \([\eta_{1},+\infty)\) and (3.6) we conclude that there exists \(\eta_{2}\in[1,+\infty)\), such that \(E(x)<0\) for \(x\in(\eta_{2},+\infty)\), this contradicts (3.5). □
Theorem 3.2
Let \(\lambda\in(0,1)\), then the double inequality
holds for all \(0< x,y<1\) if and only if \(\mu\leq r_{0}\) and \(\nu\geq 0\).
Proof
First of all, we prove that (3.9) holds if \(\mu\leq r_{0}\) and \(\nu\geq0\).
If \(\mu\leq r_{0}\), \(u(z)=\log\operatorname{erf}(z^{1/\mu})\), then Lemma 2.1(1) leads to
for \(\lambda\in(0,1)\), \(s,t>1\).
Let \(s=x^{\mu}\), \(t=y^{\mu}\), and \(0< x,y<1\). Then (3.10) leads to the first inequality in (3.9).
If \(\nu\geq0\), \(u(z)=\log\operatorname{erf}(z^{1/\nu})\), then Lemma 2.1(3) leads to
for \(\lambda\in(0,1)\), \(0< s,t<1\).
Therefore, the second inequality in (3.9) follows from \(s=x^{\nu}\), \(t=y^{\nu}\), and \(0< x,y<1\) together with (3.11).
Second, we prove that the second inequality in (3.9) implies \(\nu\geq0\).
Let \(0< x,y<1\). Then the second inequality in (3.9) leads to
It follows from (3.12) that
and
Hence, from (3.12) and (3.13) together with Lemma 2.2 we know that
Finally, we prove that the first inequality in (3.9) implies \(\mu\leq r_{0}\).
Let \(y\to1\). Then the first inequality in (3.9) leads to
for \(0< x<1\).
It follows from (3.14) that
and
Let
Then
and
If \(\mu>r_{0}\), then from (3.19) we clearly see that there exists a small \(\delta_{1}>0\), such that \(L_{1}'(x)<0\) for \(x\in(1-\delta_{1},1)\). Therefore, \(L_{1}(x)\) is strictly decreasing on \([1-\delta_{1},1]\).
The monotonicity of \(L_{1}(x)\) on \([1-\delta_{1},1]\) and (3.18) imply that there exists \(\delta_{2}>0\), such that \(L_{1}(x)>0\) for \(x\in(1-\delta_{2},1)\).
Hence, (3.16) and (3.17) lead to \(L(x)\) being strictly increasing on \([1-\delta_{2},1]\). It follows from the monotonicity of \(L(x)\) and (3.15) that there exists \(\delta_{3}>0\), such that \(L(x)<0\) for \(x\in(1-\delta_{3},1)\), this contradicts (3.14). □
References
Abramowitz, M, Stegun, I (eds.): Handbook of Mathematical Functions with Formulas, Graphs, and Mathematical Tables. Dover, New York (1965)
Oldham, K, Myland, J, Spanier, J: An Atlas of Functions: With Equator, the Atlas Function Calculator, 2nd edn. Springer, New York (2009)
Clendenin, W: Rational approximations for the error function and for similar functions. Commun. ACM 4, 354-355 (1961)
Hart, RG: A close approximation related to the error function. Math. Comput. 20, 600-602 (1966)
Cody, WJ: Rational Chebyshev approximations for the error function. Math. Comput. 23, 631-637 (1969)
Matta, F, Reichel, A: Uniform computation of the error function and other related functions. Math. Comput. 25, 339-344 (1971)
Bajić, B: On the computation of the inverse of the error function by means of the power expansion. Bull. Math. Soc. Sci. Math. Roum. 17(65), 115-121 (1973)
Blair, JM, Edwards, CA, Johnson, JH: Rational Chebyshev approximations for the inverse of the error function. Math. Comput. 30(136), loose microfiche suppl., 7-68 (1976)
Bhaduri, RK, Jennings, BK: Note on the error function. Am. J. Phys. 44(6), 590-592 (1976)
Zimmerman, IH: Extending Menzel’s closed-form approximation for the error function. Am. J. Phys. 44(6), 592-593 (1976)
Elbert, Á, Laforgia, A: An inequality for the product of two integrals relating to the incomplete gamma function. J. Inequal. Appl. 5, 39-51 (2000)
Gawronski, W, Müller, J, Reinhard, M: Reduced cancellation in the evaluation of entire functions and applications to the error function. SIAM J. Numer. Anal. 45(6), 2564-2576 (2007)
Baricz, Á: Mills’ ratio: monotonicity patterns and functional inequalities. J. Math. Anal. Appl. 340, 1362-1370 (2008)
Alzer, H: Functional inequalities for the error function. II. Aequ. Math. 78(1-2), 113-121 (2009)
Dominici, D: Some properties of the inverse error function. Contemp. Math. 457, 191-203 (2008)
Temme, NM: Error functions, Dawson’s and Fresnel integrals. In: NIST Handbook of Mathematical Functions, pp. 159-171. U.S. Dept. Commerce, Washington (2010)
Kharin, SN: A generalization of the error function and its application in heat conduction problems. In: Differential Equations and Their Applications, vol. 176, pp. 51-59 (1981) (in Russian)
Chaudhry, MA, Qadir, A, Zubair, SM: Generalized error functions with applications to probability and heat conduction. Int. J. Appl. Math. 9(3), 259-278 (2002)
Aumann, G: Konvexe Funktionen und die Induktion bei Ungleichungen zwischen Mittelwerten. Münchner Sitzungsber. 109, 403-415 (1933)
Anderson, GD, Vamanamurthy, MK, Vuorinen, M: Generalized convexity and inequalities. J. Math. Anal. Appl. 335(2), 1294-1308 (2007)
Gronau, D: Selected topics on functional equations. In: Functional Analysis, IV (Dubrovnik, 1993). Various Publ. Ser. (Aarhus), vol. 43, pp. 63-84. Aarhus Univ., Aarhus (1994)
Gronau, D, Matkowski, J: Geometrical convexity and generalization of the Bohr-Mollerup theorem on the gamma function. Math. Pannon. 4, 153-160 (1993)
Gronau, D, Matkowski, J: Geometrically convex solutions of certain difference equations and generalized Bohr-Mollerup type theorems. Results Math. 26, 290-297 (1994)
Matkowski, J: \(\mathbf{L}^{p}\)-Like paranorms. In: Selected Topics in Functional Equations and Iteration Theory. Proceedings of the Austrian-Polish Seminar (Graz, 1991). Grazer Math. Ber., vol. 316, pp. 103-138. Karl-Franzens-Univ. Graz, Graz (1992)
Niculescu, CP: Convexity according to the geometric mean. Math. Inequal. Appl. 3, 155-167 (2000)
Alzer, H: Error function inequalities. Adv. Comput. Math. 33(3), 349-379 (2010)
Xia, W, Chu, Y: Optimal inequalities for the convex combination of error function. J. Math. Inequal. 9(1), 85-99 (2015)
Chu, Y, Li, Y, Xia, W, Zhang, X: Best possible inequalities for the harmonic mean of error function. J. Inequal. Appl. 2014, 525 (2014)
Acknowledgements
This research was supported by the Natural Science Foundation of China under Grants 61174076, 61374086, 11371125, and 11401191, and the Natural Science Foundation of Zhejiang Province under Grant LY13A010004. The authors wish to thank the anonymous referees for their careful reading of the manuscript and their fruitful comments and suggestions.
Author information
Authors and Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
All authors contributed equally to the writing of this paper. All authors read and approved the final manuscript.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Li, YM., Xia, WF., Chu, YM. et al. Optimal lower and upper bounds for the geometric convex combination of the error function. J Inequal Appl 2015, 382 (2015). https://doi.org/10.1186/s13660-015-0906-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13660-015-0906-y