Skip to main content

Wavelet estimations for densities and their derivatives with Fourier oscillating noises

Abstract

By developing the classical kernel method, Delaigle and Meister provide a nice estimation for a density function with some Fourier-oscillating noises over a Sobolev ball W 2 s (L) and over L 2 risk (Delaigle and Meister in Stat. Sin. 21:1065-1092, 2011). The current paper extends their theorem to Besov ball B r , q s (L) and L p risk with p,q,r[1,] by using wavelet methods. We firstly show a linear wavelet estimation for densities in B r , q s (L) over L p risk, motivated by the work of Delaigle and Meister. Our result reduces to their theorem, when p=q=r=2. Because the linear wavelet estimator is not adaptive, a nonlinear wavelet estimator is then provided. It turns out that the convergence rate is better than the linear one for rp. In addition, our conclusions contain estimations for density derivatives as well.

1 Introduction and preliminary

One of the fundamental deconvolution problems is to estimate a density function f X of a random variable X, when the available data W 1 , W 2 ,, W n are independent and identically distributed (i.i.d.) with

W j = X j + δ j (j=1,2,,n).

We assume that all X j and δ j are independent and the density function f δ of the noise δ is known.

Let the Fourier transform f f t of fL(R) be defined by f f t (t)= R f(x) e i t x dx in this paper. When f δ f t satisfies

| f δ f t ( t ) | c ( 1 + | t | ) α
(1)

with c>0 and α>0, there exist lots of optimal estimations for f X [16]. However, many noise densities f δ have zeros in the Fourier transform domain, i.e., the inequality (1) does not hold. For example, Sun et al. described an experiment where data on the velocity of halo stars in the Milky Way are collected, and where the measurement errors are assumed to be uniformly distributed [7]. The classical kernel method provides a slower convergence rate in that case [810]. Delaigle and Meister [11] developed a new method for a density f δ with

| f δ f t ( t ) | c | sin ( π t λ ) | v ( 1 + | t | ) α .
(2)

Here c,λ,α>0 and vN (non-negative integer set). Such noises are called Fourier-oscillating. Clearly, (2) allows f δ f t having zeros for v1. When v=0, (2) reduces to (1) (non-zero case).

Delaigle and Meister defined a kernel estimator f ˆ n for a density f X in a Sobolev space and prove that with EX denoting the expectation of a random variable X,

sup f X W 2 s ( L ) E a b | f ˆ n ( x ) f X ( x ) | 2 dx=O ( n 2 s 2 s + 2 α + 1 )
(3)

under the assumption (2) (Theorem 4.1 in [11]). Here, W 2 s (L) stands for the Sobolev ball with the radius L. This above convergence rate attains the same one as in the non-zero case [14, 6]. In particular, it does not depend on the parameter v.

It seems that many papers deal with L 2 estimations. However, L p estimations (1p+) are important [5, 12]. On the other hand, Besov spaces contain many classical spaces (e.g., L 2 Sobolev spaces and Hölder spaces) as special examples. The current paper extends (3) from W 2 s (L) to the Besov ball B r , q s (L), and from L 2 to L p risk estimations. In addition, our results contain estimations for d th derivatives f X ( d ) of f X . The next section provides a linear wavelet estimation for f X ( d ) over a Besov ball B r , q s (L) and over L p risk (p,q,r1). It turns out that our estimation reduces to (3), when d=0, p=r=q=2. Moreover, we show a nonlinear wavelet estimation which improves the linear one for rp in the last part.

1.1 Wavelet basis

The fundamental method to construct a wavelet basis comes from the concept of multiresolution analysis (MRA [13]). It is defined as a sequence of closed subspaces { V j } of the square integrable function space L 2 (R) satisfying the following properties:

  1. (i)

    V j V j + 1 , jZ (the integer set);

  2. (ii)

    f(x) V j if and only if f(2x) V j + 1 for each jZ;

  3. (iii)

    j Z V j ¯ = L 2 (R) (the space j Z V j is dense in L 2 (R));

  4. (iv)

    There exists φ(x) L 2 (R) (scaling function) such that { φ ( x k ) } k Z forms an orthonormal basis of V 0 = span ¯ { φ ( x k ) } k Z .

With the standard notation h j , k (x):= 2 j 2 h( 2 j xk) in wavelet analysis, we can find a corresponding wavelet function

ψ(x)= k Z ( 1 ) k + 1 h ¯ 1 k φ 1 , k (x)with  h k =φ, φ 1 , k

such that, for a fixed jZ, { ψ j , k } k Z constitutes an orthonormal basis of the orthogonal complement W j of V j in V j + 1 [13]. Then each f L 2 (R) has an expansion in L 2 (R) sense,

f= k Z α j , k φ j , k + l = j k Z β l , k ψ l , k ,

where α j , k =f, φ j , k , β l , k =f, ψ l , k .

A family of important examples are Daubechies wavelets D 2 N (x), which are compactly supported in time domain [14]. They can be smooth enough with increasing supports as N gets large, although D 2 N do not have analytic formulas except for N=1.

As usual, let P j and Q j be the orthogonal projections from L 2 (R) to V j and W j , respectively,

P j f= k Z α j , k φ j , k , Q j f= k Z β j , k ψ j , k =( P j + 1 P j )f.

The following simple lemma is fundamental in our discussions. We use f p to denote L p (R) norm for f L p (R), and λ p do l p (Z) norm for λ l p (Z), where

l p (Z):= { { λ = { λ k } , k Z | λ k | p < + } , 1 p < ; { λ = { λ k } , sup k Z | λ k | < + } , p = .

By using Proposition 8.3 in [15], we have the following conclusion.

Lemma 1.1 Let h be a Daubechies scaling function or the corresponding wavelet. Then there exists c 2 c 1 >0 such that, for λ={ λ k } l p (Z) and 1p,

c 1 2 j ( 1 2 1 p ) λ p k Z λ k h j , k p c 2 2 j ( 1 2 1 p ) λ p .

1.2 Besov spaces

One of the advantages of wavelet bases is that they can characterize Besov spaces. To introduce those spaces (see [15]), we need the Sobolev spaces with integer order W p n (R):={f L p (R), f ( n ) L p (R)} and f W p n := f p + f ( n ) p . Then L p (R) can be considered as W p 0 (R).

For 1p,q, s=n+α with nN and α(0,1], the Besov spaces are defined by

B p , q s (R):= { f W p n ( R ) , { 2 j α ω p 2 ( f ( n ) , 2 j ) } j Z l q ( Z ) } ,

with the associated norm f B p , q s := f W p n + { 2 j α ω p 2 ( f ( n ) , 2 j ) } j Z l q , where

ω p 2 (f,t):= sup | h | t f ( + 2 h ) 2 f ( + h ) + f ( ) p

stands for the smoothness modulus of f. It should be pointed out that B 2 , 2 s (R)= W 2 s (R).

According to Theorem 9.6 in [15], the following result holds.

Lemma 1.2 Let φ= D 2 N be a Daubechies scaling function with large N and ψ be the corresponding wavelet. If f L p (R), 1p,q, s>0, α 0 , k =f, φ 0 , k , and β j , k =f, ψ j , k , then the following assertions are equivalent:

  1. (i)

    f B p , q s (R);

  2. (ii)

    α 0 , l p + { 2 j ( s + 1 2 1 p ) β j , l p } j 0 l q <+;

  3. (iii)

    { 2 j s P j f f p } j 0 l q , where P j is the projection operator to V j .

In each case,

f B p , q s α 0 , l p + { 2 j ( s + 1 2 1 p ) β j , l p } j 0 l q P 0 f p + { 2 j s P j f f p } j 0 l q .

Here and throughout, AB denotes ACB for some constant C>0; AB means BA; we use AB standing for both AB and BA.

Note that l p 1 is continuously embedded into l p 2 for p 1 p 2 . Then the above lemma implies that

B p 1 , q s 1 (R) B p 2 , q s 2 (R)for  p 1 p 2  and  s 1 1 p 1 = s 2 1 p 2 .

2 Linear wavelet estimation

We shall provide a linear wavelet estimation for a compactly supported density function f X and its derivatives f X ( d ) under Fourier-oscillating noises in this section, motivated by the work of Delaigle and Meister. It turns out that our result generalizes their theorem.

As in [11], we define

p(x)= m = 0 v ( v m ) ( 1 ) v m f X ( x 2 π m λ ) .
(4)

Then pL(R) and p f t (t)= ( e i 2 π t λ 1 ) v f X f t (t). Delaigle and Meister found that

f X (x)= m = 0 J η m p ( x 2 π m λ ) ,
(5)

where J and η m depend only on v and the support length of f X .

Let φ= D 2 N be the Daubechies scaling function with N large enough. Since both f X and φ have compact supports, the set K j :={kZ: f X , φ j , k 0} is finite and the cardinality | K j | 2 j . Then with α j , k = f X ( d ) , φ j , k ,

P j f X ( d ) = k K j α j , k φ j , k .

It is easy to see α j , k = ( 1 ) d f X , ( φ j , k ) ( d ) . This with (5) and the Plancherel formula leads to

α j , k = ( 1 ) d 2 π m = 0 J η m e i 2 π m t λ p f t ( t ) , [ ( φ j , k ) ( d ) ] f t ( t ) .
(6)

Note that p f t (t)= ( e 2 π i t λ 1 ) v f W f t ( t ) f δ f t ( t ) and [ ( φ j , k ) ( d ) ] f t (t)= 2 j 2 + d j e i k 2 j t [ φ ( d ) ] f t ( 2 j t). Then the identity (6) reduces to

α j , k = ( 1 ) d 2 π 2 j 2 + d j ξ(t) [ φ ( d ) ] f t ( 2 j t ) ¯ f δ f t ( t ) e i k 2 j t f W f t (t)dt
(7)

where ξ(t)= m = 0 J η m e i 2 π m t λ ( e i 2 π t λ 1 ) v . Since the empirical estimator for f W f t is 1 n l = 1 n e i W l t , it is natural to define a linear wavelet estimator

f ˆ n , d l i n (x)= k K j α ˆ j , k φ j , k (x),
(8)

with

α ˆ j , k = 1 n l = 1 n ( 1 ) d 2 π 2 j 2 + d j ξ(t) [ φ ( d ) ] f t ( 2 j t ) ¯ f δ f t ( t ) e i k 2 j t e i W l t dt.
(9)

When v=d=0, p(x)= f X (x) and η m = δ m , 0 . Then our estimator f ˆ n , d l i n reduces to the classical linear estimator for the case f δ f t having no zeros (see e.g., [25]).

Let ψ be the Daubechies wavelet function corresponding to the scaling function D 2 N and β j , k = f ( d ) , ψ j , k . Similar to (9), we define

β ˆ j , k = 1 n l = 1 n ( 1 ) d 2 π 2 j 2 + d j ξ(t) [ ψ ( d ) ] f t ( 2 j t ) ¯ f δ f t ( t ) e i k 2 j t e i W l t dt.
(10)

Then it can easily be seen that E α ˆ j , k = α j , k , E β ˆ j , k = β j , k and E f ˆ n , d l i n = P j f X ( d ) .

For a<b, L>0, we consider the subset of B p , q s (R),

B p , q s (a,b,L):= { f B p , q s ( R ) : f B p , q s L , f ( x ) 0 , R f ( x ) d x = 1 , supp f [ a , b ] }

in this paper.

Lemma 2.1 Let φ= D 2 N (N large enough), ψ be the corresponding wavelet and f δ f t satisfy (2). If f X B r , q s + d (a,b,L), r,q[1,+], s> 1 r and 2 j n, then, for p[1,+),

E | α ˆ j , k α j , k | p n p 2 2 j p ( α + d ) ,E | β ˆ j , k β j , k | p n p 2 2 j p ( α + d ) .

Proof One shows only the first inequality; the second one is similar. Define

Z l , k := ( 1 ) d 2 π 2 j 2 + d j ξ(t) [ φ ( d ) ] f t ( 2 j t ) ¯ f δ f t ( t ) e i k 2 j t e i W l t dt.

Then α ˆ j , k = 1 n l = 1 n Z l , k and

α ˆ j , k α j , k = 1 n l = 1 n ( Z l , k E Z l , k ):= 1 n l = 1 n Y l , k .
(11)

Clearly, E Y l , k =0. One estimates | Y l , k | and E | Y l , k | 2 in order to use the Rosenthal inequality: By the assumption (2), | Z l , k | 2 j 2 + d j | e i 2 π t λ 1 | v | [ φ ( d ) ] f t ( 2 j t ) | | f δ f t ( t ) | dt 2 j 2 + d j ( 1 + | t | ) α | [ φ ( d ) ] f t ( 2 j t)|dt = 2 j 2 + d j ( 1 + | 2 j t | ) α | [ φ ( d ) ] f t (t)|dt 2 j ( α + d + 1 2 ) ( 1 + | t | ) α | [ φ ( d ) ] f t (t)|dt. Because φ= D 2 N , the last integration is finite for large N. Hence,

| Y l , k |2| Z l , k | 2 j ( α + d + 1 2 ) .
(12)

Since f X B r , q s + d (R) B , q s + d 1 r (R) (s> 1 r ), f X <+ and f W = f X f δ f X f δ 1 <+. This with the Parseval identity shows

E Z l , k 2 = | ( 1 ) d 2 π 2 j 2 + d j ξ ( t ) [ φ ( d ) ] f t ( 2 j t ) ¯ f δ f t ( t ) e i k 2 j t e i t y d t | 2 f W ( y ) d y f W | 2 j 2 + d j ξ ( t ) [ φ ( d ) ] f t ( 2 j t ) ¯ f δ f t ( t ) e i k 2 j t | 2 d t .

Furthermore, one obtains E Z l , k 2 | 2 j 2 + d j ( 1 + | t | ) α [ φ ( d ) ] f t ( 2 j t ) | 2 dt 2 2 j ( α + d ) thanks to (2). Hence,

E Y l , k 2 =E | Z l , k E Z l , k | 2 E Z l , k 2 2 2 j ( α + d ) .
(13)

According to (11) and the Rosenthal inequality,

E | α ˆ j , k α j , k | p { n p ( l = 1 n E | Y l , k | 2 ) p 2 , p [ 1 , 2 ) , n p l = 1 n E | Y l , k | p + n p ( l = 1 n E | Y l , k | 2 ) p 2 , p [ 2 , + ) .

Combining this with (13), one obtains E | α ˆ j , k α j , k | p n p 2 2 j p ( α + d ) for 1p<2, which is the desired conclusion. When p2,

E | Y l , k | p Y l , k p 2 E | Y l , k | 2 2 j ( α + d + 1 2 ) ( p 2 ) 2 2 j ( α + d ) = 2 j ( p 2 1 + α p + d p )

due to (12) and (13). Moreover, E | α ˆ j , k α j , k | p n 1 p 2 j ( p 2 1 + α p + d p ) + n p 2 2 j p ( α + d ) . Since 2 j n,

n 1 p 2 j ( p 2 1 + α p + d p ) = n p 2 ( 2 j n ) p 2 1 2 j p ( α + d ) n p 2 2 j p ( α + d ) .

Finally, E | α ˆ j , k α j , k | p n p 2 2 j p ( α + d ) . This completes the proof of the first part of Lemma 2.1. □

Now, we are in a position to state our first theorem.

Theorem 2.1 Let f ˆ n , d l i n be defined by (8)-(9), r,q[1,+], p[1,+) and s> 1 r . Then with s :=s ( 1 r 1 p ) + and x + =max{0,x},

sup f X B r , q s + d ( a , b , L ) E f ˆ n , d l i n f X ( d ) p p n s p 2 s + 2 α + 2 d + 1 .

Proof It is easy to see that f ˆ n , d l i n f X ( d ) p f ˆ n , d l i n f X ( d ) r for rp, because L r ([a,b]) is continuously embedded into L p ([a,b]). Moreover, E f ˆ n , d l i n f X ( d ) p p E f ˆ n , d l i n f X ( d ) r p ( E f ˆ n , d l i n f X ( d ) r r ) p r thanks to Jensen’s inequality.

When rp, s 1 p =s 1 r and B r , q s + d (a,b,L) B p , q s + d (a,b,L). Then

sup f X B r , q s + d ( a , b , L ) E f ˆ n , d l i n f X ( d ) p p sup f X B p , q s + d ( a , b , L ) E f ˆ n , d l i n f X ( d ) p p .

Therefore, it suffices to prove the theorem, for r=p,

sup f X B p , q s + d ( a , b , L ) E f ˆ n , d l i n f X ( d ) p p n s p 2 s + 2 α + 2 d + 1 .
(14)

If f X B p , q s + d (R), then f X ( d ) B p , q s (R) and

P j f X ( d ) f X ( d ) p p 2 j s p
(15)

due to Lemma 1.2. On the other hand, f ˆ n , d l i n P j f X ( d ) = k K j ( α ˆ j , k α j , k ) φ j , k and

E f ˆ n , d l i n P j f X ( d ) p p 2 j ( p 2 1 ) k K j E | α ˆ j , k α j , k | p 2 j p 2 sup k K j E | α ˆ j , k α j , k | p

because of Lemma 1.1 and | K j | 2 j . This with Lemma 2.1 leads to E f ˆ n , d l i n P j f X ( d ) p p ( 2 j n ) p 2 2 j p ( α + d ) . Combining this with (15), one obtains

E f ˆ n , d l i n f X ( d ) p p E f ˆ n , d l i n P j f X ( d ) p p + P j f X ( d ) f X ( d ) p p ( 2 j n ) p 2 2 j p ( α + d ) + 2 j s p .

Take 2 j n 1 2 s + 2 α + 2 d + 1 . Then the inequality (14) follows, and the proof of Theorem 2.1 is finished. □

Remark 2.1 If p=q=r=2 and d=0, then s =s, B r , q s + d (a,b,L)= W 2 s (a,b,L), Theorem 2.1 reduces to Theorem 4.1 in [11].

Remark 2.2 From the choice 2 j n 1 2 s + 2 α + 2 d + 1 in the proof of Theorem 2.1, we find that our estimator is not adaptive, because it depends on the parameter s of B r , q s (R). In order to avoid that shortcoming, we study a nonlinear estimation in the next part.

3 Nonlinear estimation

This section is devoted to an adaptive nonlinear estimation, which also improves the convergence rate of the linear one in some cases. The idea of proof comes from [12]. Choose r 0 >s,

2 j 0 n 1 2 r 0 + 2 α + 2 d + 1 and 2 j 1 n 1 2 α + 2 d + 1 .
(16)

Let α ˆ j , k and β ˆ j , k be defined by (9) and (10), respectively, and

β ˜ j , k = { β ˆ j , k , | β ˆ j , k | > T 2 j ( α + d ) j / n , 0 , otherwise ,

where the constant T will be determined in the proof of Theorem 3.1. Then we define a nonlinear wavelet estimator

f ˆ n , d n o n (x):= k K j 0 α ˆ j 0 , k φ j 0 , k (x)+ j = j 0 j 1 k I j β ˜ j , k ψ j , k (x),

where K j 0 :={kZ: f X , φ j 0 , k 0}, and I j :={kZ: f X , ψ j , k 0}. Clearly, the cardinality | I j | 2 j since both f X and ψ have compact supports.

Lemma 3.1 If j 2 j n, then there exists c 0 >0 such that, for each T T 0 >0,

P { | β ˆ j , k β j , k | > T 2 2 j ( α + d ) j / n } 2 c 0 T j .

Proof By the definitions of β j , k and β ˆ j , k , β ˆ j , k β j , k = 1 n l = 1 n ( Z l , k E Z l , k ):= 1 n l = 1 n Y l , k , where

Z l , k := ( 1 ) d 2 π 2 j 2 + d j m = 0 J η m e i t 2 π m λ ( e 2 π i t λ 1 ) v [ ψ ( d ) ] f t ( 2 j t ) ¯ f δ f t ( t ) e i 2 j k t e i W l t dt.

Then E Y l , k =0 and with λ= T 2 2 j ( α + d ) j n ,

P { | β ˆ j , k β j , k | > T 2 2 j ( α + d ) j n } 2exp ( n λ 2 2 ( E Y l , k 2 + Y l , k λ / 3 ) )
(17)

thanks to the classical Bernstein inequality in [15]. On the other hand, E Y l , k 2 + Y l , k λ/3 2 2 j ( α + d ) + 2 ( α + d + 1 2 ) j T 6 2 j ( α + d ) j n CT 2 2 j ( α + d ) because of (12), (13), and j 2 j n. Hence, n λ 2 2 ( E Y l , k 2 + Y l , k λ / 3 ) n T 2 4 j n 2 2 j ( α + d ) 2 C T 2 2 j ( α + d ) = T 8 C j, and (17) reduces to

P { | β ˆ j , k β j , k | > T 2 2 j ( α + d ) j / n } 2 e T 8 C j = 2 c 0 T j

with c 0 = 1 8 C log 2 e. This completes the proof of Lemma 2.1. □

Theorem 3.1 Under the assumptions of Theorem  2.1, there exist θ>0 and T 0 >0 such that, for T T 0 ,

sup f X B r , q s + d ( a , b , L ) E f ˆ n , d n o n f X ( d ) p p { ( ln n ) θ n s p 2 s + 2 α + 2 d + 1 , r ( ( 2 α + 2 d + 1 ) p 2 s + 2 α + 2 d + 1 , p ] ; ( ln n ) θ n s p 2 ( s 1 / r ) + 2 α + 2 d + 1 , r [ 1 , ( 2 α + 2 d + 1 ) p 2 s + 2 α + 2 d + 1 ] .
(18)

Proof Similar to [12], one defines

μ:=min { s 2 s + 2 α + 2 d + 1 , s 2 ( s 1 / r ) + 2 α + 2 d + 1 } ,

and

ω:=sr+ ( α + d + 1 2 ) (pr).
(19)

It is easy to check that ω<0 holds if and only if r> ( 2 α + 2 d + 1 ) p 2 s + 2 α + 2 d + 1 , and μ= s 2 s + 2 α + 2 d + 1 as well as ω0 if and only if r ( 2 α + 2 d + 1 ) p 2 s + 2 α + 2 d + 1 , and μ= s 2 ( s 1 / r ) + 2 α + 2 d + 1 . Then the conclusion of Theorem 3.1 can be rewritten as

sup f X B r , q s + d ( a , b , L ) E f ˆ n , d n o n f X ( d ) p p ( ln n ) θ n μ p .
(20)

Choose j 0 (s,r,q) and j 1 (s,r,q) such that

2 j 0 ( s , r , q ) n 1 2 μ 2 α + 2 d + 1 and 2 j 1 ( s , r , q ) n μ s .

Then it can easily be shown by (16) that 2 j 0 2 j 0 ( s , r , q ) 2 j 1 ( s , r , q ) 2 j 1 . Clearly,

f ˆ n , d n o n f X ( d ) p p P j 0 ( f ˆ n , d n o n f X ( d ) ) p p + D j 0 , j 1 ( f ˆ n , d n o n f X ( d ) ) p p + P j 1 f X ( d ) f X ( d ) p p ,
(21)

where

D j 0 , j 1 f= j = j 0 j 1 k Z β j , k ψ j , k .

By the assumption rp, s :=s( 1 r 1 p )=s 1 r + 1 p and B r , q s + d (a,b,L) is continuously embedded into B p , q s + d (a,b,L). Since f X B p , q s + d (R), f X ( d ) B p , q s (R) and P j 1 f X ( d ) f X ( d ) p p 2 j 1 s p thanks to Lemma 1.2. This with 2 j 1 ( s , r , q ) 2 j 1 and the definition of μ leads to

P j 1 f X ( d ) f X ( d ) p p 2 j 1 ( s , r , p ) s p n μ p .
(22)

Note that P j 0 ( f ˆ n , d n o n f X ( d ) )= k K j 0 ( α ˆ j 0 , k α j 0 , k ) φ j 0 , k . Then E P j 0 ( f ˆ n , d n o n f X ( d ) ) p p 2 j 0 ( p 2 1 ) k K j 0 E | α ˆ j 0 , k α j 0 , k | p 2 j 0 p 2 n p 2 2 j 0 p ( α + d ) due to Lemma 1.1, | K j 0 | 2 j 0 and Lemma 2.1. By 2 j 0 2 j 0 ( s , r , q ) and the choice 2 j 0 ( s , r , q ) n 1 2 μ 2 α + 2 d + 1 ,

E P j 0 ( f ˆ n , d n o n f X ( d ) ) p p ( 2 j 0 ( s , r , q ) n ) p / 2 2 j 0 ( s , r , q ) p ( α + d ) n μ p .
(23)

According to (21)-(23), it is sufficient to prove E D j 0 , j 1 ( f ˆ n , d n o n f X ( d ) ) p p ( ln n ) θ n μ p : Define

B ˆ j : = { k : | β ˆ j , k | > T 2 j ( α + d ) j / n } , S ˆ j = B ˆ j c ; B j : = { k : | β j , k | > T 2 2 j ( α + d ) j / n } , S j = B j c ; B j : = { k : | β j , k | > 2 T 2 j ( α + d ) j / n } , S j = B j c .

Then D j 0 , j 1 ( f ˆ n , d n o n f X ( d ) ) = j = j 0 j 1 k I j ( β ˆ j , k β j , k )[I{k B ˆ j S j } + I{k B ˆ j B j }] ψ j , k j = j 0 j 1 k I j β j , k [I{k S ˆ j S j } + I{k S ˆ j B j }] ψ j , k = e b s + e s b + e b b + e s s , where

e b s : = j = j 0 j 1 k I j ( β ˆ j , k β j , k ) I { k B ˆ j S j } ψ j , k , e s b : = j = j 0 j 1 k I j β j , k I { k S ˆ j B j } ψ j , k , e b b : = j = j 0 j 1 k I j ( β ˆ j , k β j , k ) I { k B ˆ j B j } ψ j , k , e s s : = j = j 0 j 1 k I j β j , k I { k S ˆ j S j } ψ j , k .

In order to conclude Theorem 3.1, one needs only to show that

E e b s p p +E e s b p p +E e b b p p +E e s s p p ( ln n ) θ n μ p .
(24)

By (16), j 1 j 0 lnn and j = j 0 j 1 g j p p ( j 1 j 0 + 1 ) p 1 j = j 0 j 1 g j p p ( ln n ) p 1 j = j 0 j 1 g j p p . This with Lemma 1.1 shows that, for f ˆ (x)= j = j 0 j 1 k I j f ˆ j , k ψ j , k (x), there exists θ>0 such that

E f ˆ p p ( ln n ) θ sup j 0 j j 1 2 j ( p 2 1 ) k I j E | f ˆ j , k | p .
(25)

To estimate E e b s p p , one takes f ˆ j , k :=( β ˆ j , k β j , k )I{k B ˆ j S j }. Then, for each k B ˆ j S j , | β ˆ j , k β j , k || β ˆ j , k || β j , k |> T 2 2 j ( α + d ) j n and B ˆ j S j D ˆ j :={k:| β ˆ j , k β j , k |> T 2 2 j ( α + d ) j n }. This with the Hölder inequality shows

E | f ˆ j , k | p E | β ˆ j , k β j , k | p I{k D ˆ j } ( E | β ˆ j , k β j , k | 2 p ) 1 2 ( E I { k D ˆ j } ) 1 2 .

Clearly, EI{k D ˆ j }=P{| β ˆ j , k β j , k |> T 2 2 j ( α + d ) j n }. Furthermore, using | I j | 2 j , Lemma 2.1, and Lemma 3.1, one obtains k I j E | f ˆ j , k | p 2 j n p 2 2 j p ( α + d ) 2 c 0 T 2 j . Moreover, by (25),

E e b s p p ( ln n ) θ sup j 0 j j 1 n p 2 2 j ( p 2 + α p + d p c 0 T 2 ) .

Choose T p + 2 α p + 2 d p c 0 . Then p 2 +αp+dp c 0 T 2 0, and sup j 0 j j 1 n p 2 2 j ( p 2 + α p + d p c 0 T 2 ) n p 2 2 j 0 ( p 2 + α p + d p c 0 T 2 ) n p 2 2 j 0 p 2 2 j 0 p ( α + d ) . Similar to (23), one has

E e b s p p ( ln n ) θ n μ p .
(26)

In the proof of (26), one needs to choose T 0 c 0 1 (p+2αp+2dp).

Now, one considers E e s b p p : For k S ˆ j B j , | β ˆ j , k β j , k || β j , k || β ˆ j , k |>T 2 j ( α + d ) j n and S ˆ j B j D ˆ j :={k:| β ˆ j , k β j , k |> T 2 2 j ( α + d ) j n }. By Lemma 3.1, EI{k S ˆ j B j }EI{k D ˆ j } 2 c 0 T j . Since f X ( d ) B r , q s (R) B p , q s (R), β j , p := f X ( d ) , ψ j , p f X ( d ) B p , q s 2 j ( s + 1 / 2 1 / p ) and

k I j | β j , k | p EI { k S j ˆ B j } β j , p p 2 c 0 T j f X ( d ) B p , q s p 2 j ( s p + p 2 1 + c 0 T ) .

Moreover, it follows from the definition of e s b and (25) that

E e s b p p ( ln n ) θ sup j 0 j j 1 2 j ( s p + c 0 T ) ( ln n ) θ 2 j 0 ( s p + c 0 T ) .

By (16), one can choose T j 1 j 0 c 0 j 0 s p (independent of n) so that j 1 s p j 0 ( s p+ c 0 T). Hence, this above inequality reduces to E e s b p p ( ln n ) θ 2 j 1 s p . Similar arguments to (22) lead to

E e s b p p ( ln n ) θ n μ p .
(27)

In this above proof, one needs to choose T 0 ( c 0 j 0 ) 1 ( j 1 j 0 ) s p.

For E e b b p p ( ln n ) θ n μ p , one uses | I j | 2 j , (25), and Lemma 2.1 to find

I:=E j = j 0 j 0 ( s , r , q ) k I j ( β ˆ j , k β j , k ) I { k B ˆ j B j } ψ j , k p p ( ln n ) θ sup j 0 j j 0 ( s , r , q ) n p 2 2 j p ( α + d + 1 2 ) .

Recall that j 0 j 0 (s,r,q). Then I ( ln n ) θ ( n 1 2 j 0 ( s , r , q ) ) p 2 2 j 0 ( s , r , q ) p ( α + d ) , which reduces to I ( ln n ) θ n μ p by similar arguments of (23). It remains to show

II:=E j = j 0 ( s , r , q ) j 1 k I j ( β ˆ j , k β j , k ) I { k B j ˆ B j } ψ j , k p p ( ln n ) θ n μ p .
(28)

By Lemma 2.1 and the definition of B j , k I j E | ( β ˆ j , k β j , k ) I { k B j ˆ B j } | p k I j E | β ˆ j , k β j , k | p I{k B j } n p 2 2 j p ( α + d ) k B j | 2 β j , k T 1 2 j ( α + d ) n j 1 | r . According to Lemma 1.2, β j , r r f X ( d ) B r , q s r 2 j ( s r + r / 2 1 ) . Hence,

k I j E | ( β ˆ j , k β j , k ) I { k B j ˆ B j } | p n p r 2 ( ln n ) θ 2 j ( s r + r 2 1 α p d p + α r + d r ) .

Combining this above inequality with (25), one obtains

II ( ln n ) θ n p r 2 sup j 0 ( s , r , q ) j j 1 2 j ω := A n ,

where ω:=sr+(α+d+ 1 2 )(pr)=sr r 2 + p 2 +αp+dpαrdr as defined in (19).

When ω0, A n ( ln n ) θ n p r 2 2 j 0 ( s , r , q ) ω . By the choice 2 j 0 ( s , r , q ) n 1 2 μ 2 α + 2 d + 1 ,

A n ( ln n ) θ n p r 2 n 1 2 μ 2 α + 2 d + 1 [ s r + ( α + d + 1 / 2 ) ( p r ) ] = ( ln n ) θ n μ p n r ( μ 1 2 μ 2 α + 2 d + 1 s ) .

Since μ= s 2 s + 2 α + 2 d + 1 for ω0, μ 1 2 μ 2 α + 2 d + 1 s=0 and A n ( ln n ) θ n μ p . Then one obtains the desired inequality (28).

When ω>0, r< ( 2 α + 2 d + 1 ) p 2 s + 2 α + 2 d + 1 and μ= s 2 ( s 1 / r ) + 2 α + 2 d + 1 . Take

p 1 = 2 α p + 2 d p + p 2 2 ( s 1 r ) + 2 α + 2 d + 1 .

Then p p 1 2 =μp and r< p 1 in that case. With f ˆ j , k =( β ˆ j , k β j , k )I{k B j ˆ B j }, one knows from Lemma 2.1 and the definition of B j that

k I j E | f ˆ j , k | p n p 2 2 j p ( α + d ) k B j | 2 β j , k T 2 j ( α + d ) n j | p 1 ( ln n ) θ n p p 1 2 2 j ( p p 1 ) ( α + d ) β j , p 1 p 1 .

Since r< p 1 , β j , p 1 β j , r . This with (25) leads to

E e b b p p ( ln n ) θ n p p 1 2 sup j 0 j j 1 [ 2 j ( p 2 + 2 α p + 2 d p 2 p 1 ) 2 j ( α + d ) β j , r ] p 1 .

Note that 2 j p 2 + 2 α p + 2 d p 2 p 1 2 j ( α + d ) = 2 j ( s 1 r + 1 2 ) due to the definition of p 1 , as well as f X ( d ) B r , q s (a,b,L) implies 2 j ( s 1 r + 1 2 ) β j , r f X ( d ) B r , q s . Then

E e b b p p ( ln n ) θ n p p 1 2 = ( ln n ) θ n μ p .
(29)

Finally, one estimates E e s s p p : Define f ˆ j , k := β j , k I{k S ˆ j S j }. Then

k I j | f ˆ j , k | p k S j | β j , k | p r | β j , k | r ( 2 T 2 j ( α + d ) j n ) p r 2 j ( s + 1 2 1 r ) r f X ( d ) B r , q s r

due to rp and the definition of S j . Using (25) and ω:=sr+(α+d+ 1 2 )(pr) in (19), one obtains

E e s s p p =E j = j 0 j 1 k I j f ˆ j , k ψ j , k p p = { ( ln n ) θ 2 j 0 ω n p r 2 , ω 0 , ( ln n ) θ 2 j 1 ω n p r 2 , ω > 0 .
(30)

When ω0, E j = j 0 ( s , r , q ) j 1 k I j β j , k I { k S ˆ j S j } ψ j , k p p ( ln n ) θ n p r 2 2 j 0 ( s , r , q ) ω . Recall that ω0 holds if and only if μ= s 2 s + 2 α + 2 d + 1 and 2 j 0 ( s , r , q ) n 1 2 μ 2 α + 2 d + 1 . Then it can be checked that 2 j 0 ( s , r , q ) ω n p r 2 = n μ p . Hence,

E j = j 0 ( s , r , q ) j 1 k I j β j , k I { k S ˆ j S j } ψ j , k p p ( ln n ) θ n μ p .
(31)

On the other hand, Lemma 1.1 tells that

E j = j 0 j 0 ( s , r , q ) k S ˆ j S j β j , k ψ j , k p p j = j 0 j 0 ( s , r , q ) 2 j ( p 2 1 ) k S ˆ j S j E | β j , k | p .

Since k S j , | β j , k |2T 2 j ( α + d ) j n and k S ˆ j S j | β j , k | p k I j 2 j p ( α + d ) ( ln n ) θ n p 2 ( ln n ) θ n p 2 2 j 2 j p ( α + d ) . Moreover,

E j = j 0 j 0 ( s , r , q ) k I j β j , k I { k S ˆ j S j } ψ j , k p p ( ln n ) θ n p 2 sup j 0 j j 0 ( s , r , q ) 2 j ( p 2 1 ) 2 j 2 j p ( α + d ) ( ln n ) θ n p 2 2 j 0 ( s , r , q ) ( α + d + 1 2 ) p ( ln n ) θ n μ p ,
(32)

where the last inequality comes from the choice 2 j 0 ( s , r , q ) n 1 2 μ 2 α + 2 d + 1 . Combining this with (31), one has, for ω0,

E e s s p p ( ln n ) θ n μ p .
(33)

It remains to show E e s s p p n μ p for ω>0. By Lemma 1.1,

k I j β j , k I { k S ˆ j S j } ψ j , k p 2 j ( 1 2 1 p ) ( k I j | β j , k I { k S ˆ j S j } | p ) 1 / p 2 j ( 1 2 1 p ) β j , l p .

This with Lemma 1.2 shows that

[ j = j 1 ( s , r , q ) j 1 ( 2 j s k I j β j , k I { k S ˆ j S j } ψ j , k p ) q ] 1 / q [ j = j 1 ( s , r , q ) j 1 ( 2 j ( s + 1 2 1 p ) β j , l p ) q ] 1 / q f X ( d ) B p , q s .
(34)

When q=1,

j = j 1 ( s , r , q ) j 1 k I j β j , k I { k S ˆ j S j } ψ j , k p j = j 1 ( s , r , q ) j 1 2 j s k I j β j , k I { k S ˆ j S j } ψ j , k p 2 j 1 ( s , r , q ) s 2 j 1 ( s , p , q ) s .

When q=+,

j = j 1 ( s , r , q ) j 1 k I j β j , k I { k S ˆ j S j } ψ j , k p j = j 1 ( s , r , q ) j 1 2 j s f X ( d ) B p , q s ( ln n ) θ 2 j 1 ( s , p , q ) s .

For 1<q<+, by the Hölder inequality,

j = j 1 ( s , r , q ) j 1 k I j β j , k I { k S ˆ j S j } ψ j , k p [ j = j 1 ( s , r , q ) j 1 2 j s q ] 1 q [ j = j 1 ( s , r , q ) j 1 ( 2 j s k I j β j , k I { k S ˆ j S j } ψ j , k p ) q ] 1 q 2 j 1 ( s , r , q ) s .

Using (34) and the choice 2 j 1 ( s , r , q ) n μ s , one obtains

E j = j 1 ( s , r , q ) j 1 k I j β j , k I { k S ˆ j S j } ψ j , k p p ( ln n ) θ n μ p .

On the other hand,

E j = j 0 j 1 ( s , r , q ) k I j β j , k I { k S ˆ j S j } ψ j , k p p ( ln n ) θ 2 j 1 ( s , r , q ) ω n p r 2

thanks to (30). According to the choice of 2 j 1 ( s , r , q ) n μ s and

μ= s 1 / r + 1 / p 2 ( s 1 / r ) + 2 α + 2 d + 1 = s 2 ( s 1 / r ) + 2 α + 2 d + 1

for ω>0, one finds 2 j 1 ( s , r , q ) ω n p r 2 n μ p by direct computations. Hence, E e s s p p n μ p in each case, which is (33). Now, (24) follows from (26), (27), (29), and (33). The proof is done. □

Remark 3.1 We find easily from Theorem 2.1 and Theorem 3.1 that the nonlinear wavelet estimator converges faster than the linear one for rp. Moreover, the nonlinear estimator is adaptive, while the linear one is not.

Remark 3.2 This paper studies wavelet estimations of a density and its derivatives with Fourier-oscillating noises. The remaining problems include the optimality of the above estimations, numerical experiments as well as the corresponding regression problems. We shall investigate those problems in the future.

References

  1. Walter GG: Density estimation in the presence of noise. Stat. Probab. Lett. 1999, 41: 237–246. 10.1016/S0167-7152(98)00160-6

    Article  MATH  Google Scholar 

  2. Pensky M, Vidakovic B: Adaptive wavelet estimator for nonparametric density deconvolution. Ann. Stat. 1999, 27: 2033–2053. 10.1214/aos/1017939249

    Article  MathSciNet  MATH  Google Scholar 

  3. Fan J: On the optimal rates of convergence for nonparametric deconvolution problems. Ann. Stat. 1991, 19: 1257–1272. 10.1214/aos/1176348248

    Article  MATH  Google Scholar 

  4. Fan J, Koo J: Wavelet deconvolution. IEEE Trans. Inf. Theory 2002, 48: 734–747. 10.1109/18.986021

    Article  MathSciNet  MATH  Google Scholar 

  5. Lounici K, Nickl R: Global uniform risk bounds for wavelet deconvolution estimators. Ann. Stat. 2011, 39: 201–231. 10.1214/10-AOS836

    Article  MathSciNet  MATH  Google Scholar 

  6. Li R, Liu Y: Wavelet optimal estimations for a density with some additive noises. Appl. Comput. Harmon. Anal. 2014, 36: 416–433. 10.1016/j.acha.2013.07.002

    Article  MathSciNet  MATH  Google Scholar 

  7. Sun, J, Morrison, H, Harding, P, Woodroofe, M: Density and mixture estimation from data with measurement errors. Technical report (2002)

  8. Devroye L: Consistent deconvolution in density estimation. Can. J. Stat. 1989, 17: 235–239. 10.2307/3314852

    Article  MathSciNet  MATH  Google Scholar 

  9. Hall P, Meister A: A ridge-parameter approach to deconvolution. Ann. Stat. 2007, 35: 1535–1558. 10.1214/009053607000000028

    Article  MathSciNet  MATH  Google Scholar 

  10. Meister A: Deconvolution from Fourier-oscillating error densities under decay and smoothness restrictions. Inverse Probl. 2008, 24: 1–14.

    Article  MathSciNet  Google Scholar 

  11. Delaigle A, Meister A: Nonparametric function estimation under Fourier-oscillating noise. Stat. Sin. 2011, 21: 1065–1092. 10.5705/ss.2009.082

    Article  MathSciNet  MATH  Google Scholar 

  12. Donoho DL, Johnstone IM, Kerkyacharian G, Picard D: Density estimation by wavelet thresholding. Ann. Stat. 1996, 24: 508–539.

    Article  MathSciNet  MATH  Google Scholar 

  13. Hernández E, Weiss G: A First Course on Wavelets. CRC Press, Boca Raton; 1996.

    Book  MATH  Google Scholar 

  14. Daubechies I: Ten Lectures on Wavelets. SIAM, Philadelphia; 1992.

    Book  MATH  Google Scholar 

  15. Härdle W, Kerkyacharian G, Picard D, Tsybakov A Lecture Notes in Statist. In Wavelets, Approximation, and Statistical Applications. Springer, New York; 1998.

    Chapter  Google Scholar 

Download references

Acknowledgements

This paper is supported by the National Natural Science Foundation of China (No. 11271038).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Youming Liu.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

All authors contributed equally and significantly in writing this article. All authors read and approved the final manuscript.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0), which permits use, duplication, adaptation, distribution, and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Guo, H., Liu, Y. Wavelet estimations for densities and their derivatives with Fourier oscillating noises. J Inequal Appl 2014, 236 (2014). https://doi.org/10.1186/1029-242X-2014-236

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1029-242X-2014-236

Keywords