Wavelet deconvolution in a periodic setting
Abstract
Summary. Deconvolution problems are naturally represented in the Fourier domain, whereas thresholding in wavelet bases is known to have broad adaptivity properties. We study a method which combines both fast Fourier and fast wavelet transforms and can recover a blurred function observed in white noise with O{n log (n)2} steps. In the periodic setting, the method applies to most deconvolution problems, including certain ‘boxcar’ kernels, which are important as a model of motion blur, but having poor Fourier characteristics. Asymptotic theory informs the choice of tuning parameters and yields adaptivity properties for the method over a wide class of measures of error and classes of function. The method is tested on simulated light detection and ranging data suggested by underwater remote sensing. Both visual and numerical results show an improvement over competing approaches. Finally, the theory behind our estimation paradigm gives a complete characterization of the ‘maxiset’ of the method: the set of functions where the method attains a near optimal rate of convergence for a variety of Lp loss functions.
1. Deconvolution in white noise
(1)
(2)Our goal is to recover the unknown function f from the noisy blurred observations (1). The blurring function g in convolution (2) is assumed to be known. Further, we assume that the function f is periodic on the unit interval T and that g has a certain degree of smoothness. There is an extensive statistical literature on deconvolution problems; in particular wavelet methods have received considerable attention over the last decade. References that are particularly relevant to the present work include Donoho (1995), Abramovich and Silverman (1998), Pensky and Vidakovic (1999), Fan and Koo (2002), Kalifa and Mallat (2003) and Neelamani et al. (2003): these works in turn contain further references to previous literature.
An important application setting that is modelled by expression (1) is that of motion blur in signals or images; see for example Bertero and Boccacci (1998). Here g is taken as a ‘boxcar’g(x)=(2a)−1 𝕀[−a,a](x), of half‐width a. Owing to oscillations in the Fourier coefficients of g, this situation escapes the assumptions of much wavelet literature, but recent work of Neelamani et al. (2004) studied it explicitly with their ForWaRD algorithm.
Our aim in this paper is to study a wavelet deconvolution algorithm which can be applied to many deconvolution problems including certain cases of boxcar blur. We are particularly interested in obtaining adaptivity properties relative to a variety of error measures and function classes. Our theoretical investigation is conducted by using model (1), but examples and software are provided for data sampled at a discrete set of n equally spaced points.
(3)
, provided that the boxcar width is ‘badly approximated’ by rational numbers. This notion is recalled in Section 2.2; it includes boxcars of width that is given by quadratic irrational numbers such as √5; see also remark 8 following proposition 2.


(a) Log‐spectrum of a Γ(1, 0.0065) probability distribution function (smooth blur) and (b) log‐spectrum of a boxcar function g(x)=(1/2a) 𝕀[−a,a](x) with a=1/√353 (boxcar blur)
For both theoretical and practical convenience, we use band‐limited wavelet basis functions, and in particular the (periodized) Meyer wavelet basis for which fast algorithms are available; Kolaczyk (1994) and Donoho and Raimondo (2004). Our method can thus perform deconvolution in O{n pt log (n)2} steps. The WaveD software package that was used to prepare most figures and tables in this paper is available at http://www.usyd.edu.au:8000/u/marcr/. It is intended for use with Wavelab; see Buckheit et al. (1995).
We begin in Section 2 by describing an application of statistical deconvolution to remote sensing. There follow short reviews of continued fractions, periodized Meyer wavelets, Besov spaces and wavelet shrinkage. Section 3 describes our method specifically, and its relationship with the wavelet–vaguelette approach of Donoho (1995) and Abramovich and Silverman (1998). Section 4 is concerned with numerical performance and competing approaches. In the implementation of our method the choice of tuning parameters is informed by asymptotic minimax theory: this is discussed in Section 5. Proofs are summarized in Appendices A and B.
2. Motivations and preliminaries
2.1. Illustration from remote sensing
Deconvolution is a common problem in many areas of signal and image processing; see for example Jain (1989). Here we shall focus on light detection and ranging (LIDAR) remote sensing as in Je Park et al. (1997) and Harsdorf and Reuter (2000). LIDAR uses a laser device which emits pulses, reflections of which are gathered by a telescope that is aligned with the laser. The return signal is used to determine the distance and position of the reflecting material. Accordingly, the distance resolution is limited by the time resolution of the LIDAR instrument. If the system response function of the LIDAR detector is longer than the time resolution interval the measured LIDAR signal is blurred and the effective accuracy of the LIDAR decreases. This loss of precision can be corrected by deconvolution. In practice, measured LIDAR signals are corrupted by additional noise which renders direct deconvolution impossible. Borrowed from Harsdorf and Reuter (2000), we have depicted an ideal LIDAR signal in Fig. 1; this will be our target function f for numerical illustrations throughout this paper. The system response function of the LIDAR detector (denoted g(t) in expression (1)) is calibrated a posteriori once the LIDAR instrument has been built. We follow Harsdorf and Reuter (2000) and use system response functions that have a strong low pass characteristic. In the WaveD software that was used to plot most figures in this paper, the system response function parameters can be changed by the user to accommodate different calibration settings. In Fig. 2, we illustrate a smooth blur and a boxcar blur scenario in the Fourier domain. These two examples of system response function shapes illustrate the possibilities that are offered by our assumption C, under which near optimal rates are achievable. Finally, Fig. 3 shows artificial LIDAR data for a combination of different noise levels and system response functions.

Ideal LIDAR signal as in Harsdorf and Reuter (2000), corresponding to data for underwater LIDAR

Simulated LIDAR signals (1) with ti = i/n, n = 2048 , corresponding to the system response func‐ tions of Fig. 2: (a) smooth blur with low (standard deviation sd = 0.05) noise level; (b) smooth blur with medium (sd = 0.5) noise level; (c) smooth blur with high (sd = 1) noise level; (d) boxcar blur with low (sd = 0.05) noise level; (e) boxcar blur with medium (sd = 0.5) noise level; (f) boxcar blur with high (sd = 1) noise level
2.2. Boxcar blur and the continued fractions algorithm
(4)The convolution problem (1) that is associated with the boxcar, later referred to as boxcar blur, has the problem that, for rationala=p/q, the coefficients gk vanish for any integer k multiple of q. Hence, even without noise some frequencies are lost and f cannot be fully recovered. The problem is less severe for irrational numbers, and particularly for those which are ‘badly approximable’ (BA) by rational numbers. We briefly review the key tool in constructing such numbers.
2.2.1. The continued fractions algorithm
(5)
(6)
(7)
(8)
(9)
(10)- (a)
The denominators qn grow at least geometrically:
(11) - (b)
For all n0,

Hence, the size of the elements in the continued fractions algorithm (5) determines the quality of best rational approximation to a. It is customary to define families of irrational numbers a according to the size of their elements as follows.

Definition 2. A rational number a is called BA of order n if a is the convergent of order k of a BA number (a=pk/qk) and if qk−1n<qk.
The set of all BAs contains quadratic irrational numbers (e.g. √5). For the boxcar blur, we prove (proposition 2 in Section 4) that condition C holds with
for any scale a chosen in the set of BAs. In the finite sample implementation (of size n) of model (1), our method will remain numerically stable for any scale a that is chosen in the set of BA rational numbers of order at least n (see remark 8 below proposition 2) and satisfying a uniform bound (in a) on sup n{an(a)}. We refer to Johnstone and Raimondo (2002) for a discussion of cases outside the BA numbers.
2.3. Periodized Meyer wavelet transforms
(12)
(13)
(14)
(15)Here and in the rest of the paper (Φ,Ψ) will denote the periodized Meyer scaling and wavelet functions (Fig. 4). Thus, for any periodic function f an expansion that is similar to equation (13) holds with periodized basis functions (Φ,Ψ) and bivariate index κ restricted to the set I={(j,k):j0 and k=0,1,…,2j−1}. We use this basis for the following reasons.

Periodized Meyer scaling and wavelet function: (a) Φ3,4; (b) Ψ4,5
- (a)
The Meyer wavelet is band limited. In particular, we have
, where
denotes the Fourier transform of ψ.
- (b)
An efficient algorithm, due to Kolaczyk (1994), is available to compute the periodized Meyer wavelet transforms. It requires only O{n pt log (n)2} steps to derive an empirical version of the coefficients (14) from a sample of size n of f.
Band‐limited wavelets have been used in the deconvolution setting by Walter and Shen (1999), Shen and Walter (2002), Pensky and Vidakovic (1999) and Fan and Koo (2002). General information on band‐limited wavelet bases may be found, for example, in Mallat (1999), Hernàndez and Weiss (1996) and Walter (1994).
2.4. A wide class of target functions
and r1. For this, define for every measurable function f
and similarly
for positive integer N. Let


(16)The Besov spaces have proved to be an interesting scale for studying the properties of statistical procedures. The index s indicates the degree of smoothness of the function. Owing to the differential averaging effects of the integration parameters π and r, the Besov spaces capture a variety of smoothness features in a function including spatially inhomogeneous behaviour; see Donoho et al. (1995).
2.5. Wavelet shrinkage
(17)
and
are estimated wavelet coefficients and I0 and I1 are sets of indices. I0 ={(j0,k):k=0,1,…,2j0−1} corresponds to a coarse resolution level j0 and I1={(j,k):k=0,1,…,2j−1,j0jj1} indexes details up to a fine resolution level j1. The procedure (17) is non‐linear since only large coefficients
are kept; here λj is a threshold parameter. The choices of parameters j0, j1 and λj as well as estimators
and
depend on the problem at hand. For deconvolution problem (1) this will be discussed in the next section.
3. Wavelet deconvolution in the Fourier domain
3.1. Inverse estimation paradigm
. Letting h=f * g we have
(18)
(19)
the Fourier coefficients of Ψ, i.e.
. Combining equation (18) with Plancherel's identity we obtain
(20)
, we can recover wavelet coefficients
(21)
are known Fourier coefficients but the hls are not directly observable; in equation (19) we take yl as an (unbiased) estimator of hl and let
(22)
of α is defined in a similar fashion with Φ in place of Ψ.
—it is easily seen that this set does not depend on k. Indeed, from the compact support of the Meyer wavelet, we have

3.2. The wavelet deconvolution method
(23)
(24)
(25)If it is necessary to compute an estimate of the noise standard deviation σ, we adapt the method of Donoho et al. (1995) that was developed for direct data. If yJ,k = 〈Yn,ΨJ,k〉 denote the finest scale wavelet coefficients of the observed data, then
, where mad is the median absolute deviation.
We summarize the main steps of our wavelet deconvolution method, and we illustrate it in Fig. 5. Here and in the rest of the paper we refer to this as the WaveD method.

WaveD method applied to the LIDAR signals of Fig. 3 (smooth blur) (the threshold values are summarized in Table 1; levels 4–7 are shown as 100% thresholding occurs at level 8 or below): (a) low noise, estimated wavelet coefficients (22); (b) low noise, estimated wavelet coefficients (22) after shrinkage (24) with ηS = √2; (c) low noise, estimated LIDAR signal; (d) medium noise, estimated wavelet coefficients (22); (e) medium noise, estimated wavelet coefficients (22) after shrinkage (24) with ηS = √2; (f) medium noise, estimated LIDAR signal; (g) high noise, estimated wavelet coefficients (22); (h) high noise, estimated wavelet coefficients (22) after shrinkage (24) with ηS = √2; (i) high noise, estimated LIDAR signal
- (a)
Compute Fourier coefficients yl and gl and recover wavelet coefficients (22) by using Kolaczyck's algorithm (which requires only O{n pt log (n)2} operations); see Figs 5(a), 5(d) and 5(g).
- (b)
Compute, if needed, an estimate of the noise standard deviation
as described above. Find thresholds λj:=λn,j as shown in equation (24) and illustrated in Table 1 for the boxcar.
- (c)
Apply hard thresholding
; see Figs 5(b), 5(e) and 5(h). Finally, invert the wavelet transform to obtain an estimate of f; see Figs 5(c), 5(f) and 5(i). (In the WaveD software the (default) maximum resolution level j1 is determined from the data as follows: j1 is set to be the level preceeding j(100%) where j(100%) is the smallest level where 100% of thresholding occurs; see Table 1.)
- (d)
Cycle‐spin the WaveD estimator in the fashion of Coifman and Donoho (1995) (optional). This improves visual and numerical performance and was used in Fig. 7 and Table 2 later. We refer to Donoho and Raimondo (2004) for an efficient algorithm which cycle‐spins the WaveD estimator over all circulant shifts.
defined by expression (24)†
| Noise level | η | Thresholds for the following levels of j: | |||
|---|---|---|---|---|---|
| j = 5 | j = 6 | j = 7 | j = 8 | ||
Low ( )
|
√2 | 0.0036 (31.35%) | 0.0066 (62.5%) | 0.0128 (95.31%) | 0.0242 (100%) |
Medium ( )
|
√2 | 0.0375 (68.75%) | 0.0688 (98.44%) | 0.1328 (100%) | Maximum level: j1=6 |
High ( )
|
√2 | 0.0729 (90.63%) | 0.1338 (100%) | Maximum level: j1=5 | |
Low ( )
|
√6 | 0.0063 (31.25%) | 0.0115 (81.25%) | 0.0222 (99.22%) | 0.0418 (100%) |
Medium ( )
|
√6 | 0.0649 (87.5%) | 0.1192 (100%) | Maximum level: j1=5 | |
High ( )
|
√6 | 0.1263 (96.88%) | 0.2318 (100%) | Maximum level: j1=5 | |
- †Indicated in parentheses are the corresponding fractions of shrunken coefficients. The first three rows correspond to the smaller choice ηS=√2. The last three rows correspond to the conservative choice ηL=√{2(2ν+1)} with ν=1. (In the WaveD software the (default) maximum resolution level (25) is set to be the level preceeding j(100%) where j(100%) is the smallest level where 100% of thresholding occurs.)

(a) WaveD (low noise scenario); (b) WaveD (medium noise scenario); (c) WaveD (high noise scenario); (d) ForWaRD (low noise scenario); (e) ForWaRD (medium noise scenario); (f) ForWaRD (high noise scenario); (g) FoRD (low noise scenario); (h) FoRD (medium noise scenario); (i) FoRD (high noise scenario)

| Method | Blur | Means for the following levels of noise: | |||
|---|---|---|---|---|---|
| σ low=0.05 | σ med=0.5 | σ high=1 | σ lim=1.25 | ||
| WaveD | Smooth | 0.0024 | 0.0180 | 0.0388 | 0.0519 |
| ForWaRD | Smooth | 0.0027 | 0.0208 | 0.0642 | 0.0950 |
| FoRD | Smooth | 0.0084 | 0.0906 | 0.3201 | 0.3352 |
| WaveD | Boxcar | 0.0223 | 0.0753 | 0.0831 | 0.0900 |
| ForWaRD‡ | Boxcar | 0.0110 | 0.0573 | 0.0906 | 0.1030 |
| FoRD | Boxcar | 0.0237 | 0.0950 | 0.3470 | 0.3610 |
- †The results are means of 1000 independent simulations of model (1) with n=2048 as in Fig. 3. For each scenario, numbers in bold indicate the method which has the smallest MSE.
3.3. Connection with the wavelet–vaguelette decomposition

. In terms of Fourier coefficients, and setting sjl=gl/κj,
(26)
. If we observe that

are instead evaluated in the Fourier domain using the original wavelets Ψ and filter g.
The key additional property that is needed to establish a WVD is that the systems (𝒰) and (𝒱) each form Riesz bases—this property allows lower bounds and hence minimax rates of convergence to be established over the Besov classes that are considered in this paper. The lower bound arguments are given in detail in Donoho (1995) for L2‐loss and, as noted there, the methods can be extended to more general loss measures.
For the dilation homogeneous operators on ℝ that were principally studied in Donoho (1995) and Abramovich and Silverman (1998), the vaguelette systems are multiples of translates and dilates of a single mother vaguelette 𝒰0,0 or 𝒱0,0, and the Riesz basis property can be established as in Donoho (1995). See also Lee and Lucier (2001).
and condition C holds and if in addition we have, for constants C0 and C that are independent of j,
(27)
Candidate vaguelettes corresponding to the boxcar convolution of Fig. 2: (a) 𝒱4,5; (b) 𝒰4,5; (c) 𝒱5,6; (d) 𝒰5,6
These conditions hold if gl ≡ C|l|−ν or more generally if c0|l|−ν |gl| c1|l|−ν and Δrgl |l|−ν−r for r = 1 and r = 2. For such convolution operators, we obtain lower bounds to the rates of convergence over Besov spaces via the methods of Donoho (1995).
However, the sufficient conditions (27) are not satisfied by the boxcar kernel, owing to the oscillations in gl which inflate Δgl. We do not yet know whether the systems (𝒰) and (𝒱) form Riesz bases in this case, and so yield vaguelettes; see Fig. 6. Thus the only lower bounds that are currently available for this kernel are those which were established for Fourier hyperrectangles and ellipsoids in Johnstone and Raimondo (2004).
3.4. Remarks on tuning
(28)
is an estimated scale from the data and
(29)For deconvolution problems (1), our main result (proposition 1 in Section 4.1) states that, for any constant η2√{8π(p∨2)} that is sufficiently large, the choices (24) and (25) are near optimal for a wide variety of target function (Appendix B.1) and Lp loss functions.
It may be seen that the finest scale j1 that is suggested by equation (25) is considerably smaller than that given by equation (29) in the direct case. The size of the thresholds in equation (24) may in principle be larger than in equation (28), but in practice smaller thresholds than suggested by the proof may be desirable.
(30)We have tested our method with two choices of the parameter η: ηL=√{2(2ν+1)} as in Abramovich and Silverman (1998) and a smaller value ηS=√2 that is similar to that of direct estimation (28). The values of corresponding thresholds (for smooth blur) are given in Table 1. In our simulation study (Section 4), we used the WaveD estimator with ηS as it led to slightly better results than the conservative choice ηL.
4. Numerical performances of the wavelet deconvolution method
We compare several approaches to deconvolution which differ in the degree to which Fourier and wavelet filtering are balanced. On the one hand, we have Wiener‐filter‐like methods which have no wavelet component but use only Fourier inversion together with a regularization parameter (below referred to as the FoRD method). On the other hand we have wavelet decomposition approaches (like the WaveD method) where we perform Fourier inversion with no regularization but use wavelet smoothing to remove noise. Between those two approaches lies the ForWaRD method which combines Fourier regularization with wavelet smoothing. We give only a brief description of the ForWaRD and FoRD methods, referring to Neelamani et al. (2004) and references therein for further details.
4.1. Fourier regularized deconvolution
(31)
(32)Then we take
as an estimator of f by using the Fourier series with coefficients
. Here α is a regularization parameter which balances the suppression of noise with signal distortion. Small values of α give an unbiased but noisy estimate whereas large values of α suppress the noise but also distort the signal. We use the terminology FoRD of Neelamani et al. (2004) since our comparisons below use the regularization parameter choice in code that was graciously provided by R. Neelamani.
4.2. Wavelet‐regularized deconvolution
First apply filtering (31), deriving an estimator
of f. In the second step, we further smooth
by using data‐driven level‐dependent wavelet thresholding (17). Here also α plays the role of a regularization parameter which balances the level of noise and signal distortion. For α=0 the ForWaRD estimator is similar to a compactly supported WVD estimator of f whereas for α>0 the ForWaRD estimator is a hybrid of the FoRD and WVD methods. Although it is difficult to derive an optimal choice of the regularization parameter α (see Neelamani et al. (2004)), a data‐driven algorithm to compute the ForWaRD estimator is available at http://www.dsp.rice.edu/software/ward.shtml.
We compared the WaveD method with the FoRD and ForWaRD methods in a simulation study using the LIDAR target that is depicted in Fig. 1. Performance was tested with different blurring types and noise levels as illustrated in Fig. 3. For each combination of noise level and blurring type we computed the Monte Carlo approximation to the
. Our results are illustrated in Fig. 7 for smooth blur and summarized in Table 2.
4.3. Analysis of the results
Regularized Fourier filtering tends to distort the original signal and the superiority of the WaveD method becomes more apparent as the noise level increases. For smooth blur WaveD outperformed both ForWaRD and FoRD in all cases with larger margins in high noise scenarios. For all the methods we observe smaller margins and poorer performances in the boxcar blur scenario (which confirms a larger DIP). For boxcar blur WaveD outperformed ForWaRD for larger noise levels whereas ForWaRD outperformed WaveD for smaller noise levels. Both WaveD and ForWaRD outperformed FoRD, whose performance is limited because of its linear nature.
Remark 1. Cycle spinning the WaveD estimator mitigates its lack of translation invariance and leads to a better visual appearance and smaller MSE. We refer to Donoho and Raimondo (2004) for an efficient algorithm which cycle‐spins the WaveD estimator over all circulant shifts (used in the computations for Table 2 and Fig. 7).
Remark 2. Boundary corrections to deal with non‐periodic signals and an extension to the deconvolution of two‐dimensional data are currently under investigation by the authors.
5. Asymptotic theory
5.1. Near optimality for a wide range of smoothness classes
with π1, s1/π and

and maximum level (25) is such that
(33)
(34)
(35)Remark 3. There is an ‘elbow effect’ or ‘phase transition’ in the rates of convergence, switching from condition (34) to condition (35) as the assumed smoothness decreases. The existence of this effect is familiar from the direct observation case (e.g. Donoho et al. (1995) and references cited therein, where conditions (34) and (35) are respectively referred to as the ‘dense’ and ‘sparse’ cases). The additional presence of the parameter ν makes the sparse case (35) relevant even for quadratic loss, p=2.
Remark 4. For p = 2 and smooth convolutions, the rate α and dense case condition (34) are consistent with results of Donoho (1995), Abramovich and Silverman (1998) and Fan and Koo (2002). Pensky and Vidakovic (1999) obtained similar rates (in the density model) with the additional restriction π=2 (Hilbert–Sobolev function classes) so that constraint (34) does not appear.
Remark 5. For p=2 Kalifa and Mallat (2003) proposed a related procedure which can be applied for hyperbolic convolution where the convolution kernel depends on the sample size. Such convolutions do not satisfy condition C; hence their results are not directly comparable with ours.
Remark 6. For p = 2 and severely ill‐posed convolutions such as boxcar blur, our results agree with the degree of ill‐posedness
that was derived in Johnstone and Raimondo (2002).
Remark 7. For p≠2 our result seems to be new in the deconvolution context.
The following proposition gives some examples of blurring type where near optimal results of proposition 1 are achievable by our estimator.
- (a)
For ordinary smooth convolution, |gl|∼C|l|−ν, where gl denotes the Fourier coefficients of g and ν>0, assumption C is satisfied.
- (b)
For boxcar blur, g(x) = (1/2a)𝕀[−a,a](x), where a is a BA number, assumption C is satisfied with
.
Remark 8. Combining the results of propositions 1 and 2 for boxcar blur, we see that rates (34) and (35) hold with
provided that the boxcar width is a BA irrational number. In the finite sample implementation of model (1) on a computer, Fourier coefficients gl are computed up to l = n or l = −n, or more precisely for blocks Cj that are wholly contained in [−n,n]. Hence condition C needs only to be satisfied for any j>0 where 2j+rn. An examination of the proof of proposition 2, part (b), shows that the latter condition holds for those BA rational numbers that are of order greater than n discussed in Section 2.2.
Remark 9. For almost all irrational numbers (i.e. except for a set of Lebesgue measure 0), the boxcar blur is also known to have a degree of ill‐posedness of
, ignoring logarithmic terms (Johnstone and Raimondo, 2002). Whether the WaveD estimator can be tuned to achieve rates that are similar to expression (3) for boxcar blur in the almost all case remains open.
Remark 10. In the direct estimation setting, alternatives to co‐ordinatewise non‐linear thresholding also have broad adaptivity properties (e.g. Efromovich (1999)); whether such results extend to the deconvolution is an issue for further work.
5.2. The maxiset approach


(36)
(37)
(38)
(39)
(40)Remark 11. Through condition (40) and in the light of Appendix B.1 the theorem gives the maxiset of the method, i.e. the set of functions where the method attains a given rate of convergence. This way of measuring the performances of statistical procedures has been particularly successful in the nonparametric framework. It has often the advantage of giving less arbitrary and pessimistic comparisons of procedures than the minimax approach.
are included in the maxiset defined in expression (40) for q chosen such that
provides the rate that is given in inequality (33). In particular, for appropriate choices of j1 and Λn (see expression (42)), we have that
(41)Acknowledgements
We are grateful to all the referees, who provided us with helpful suggestions that have improved the original version significantly. This project began while Gérard Kerkyacharian and Dominique Picard visited the University of Sydney, partly funded by the University of Sydney. IMJ was supported in part by National Science Foundation grant DMS 00‐72661 and National Institutes of Health grant ROI EB001988‐08.
Appendices
Appendix A: Proofs
A.1. Outline of the proof of proposition 1
(42)In this setting, and under assumption C, we shall prove the following claims.
- (a)
Inequalities (37) and (38) hold with η2√{8π(p∨2)} (claim 1).
- (b)
The basis (σjΨjk) satisfies condition (63) (see Appendix B.2) as soon as there is a constant C such that, for any finite subset Λ of ℕ,
(43)(claim 2). Note that for p=2 condition (63) holds without any condition on σj.
(44) - (c)
Conditions (36), (43) and (44) are satisfied (claim 3).
Hence, under the assumptions of proposition 1, theorem 1 applies to the wavelet‐based estimator (17) which combined with remarks following theorem 1 gives the rate (33). To complete the proof we shall now prove the claims.
A.1.1. Proof of claim 1
. Taking the expectation in equation (19),
(45)
(46)
and that
(47)
since, for the Meyer wavelet,
. Using definition (23) and recalling that for the Meyer wavelet |Cj|=4π·2j:
(48)
s are Gaussian,
(49)
(50)
A.1.2. Proof of claim 2
The proof of claim 2 is a direct application of theorem 2 (see Appendix B).
A.1.3. Proof of claim 3

A.1.4. Proof of proposition 2

(52)
(53)
. Result (53) follows from condition (9) and the following lemma (see Johnstone and Raimondo (2002)). We refer to Section 2.2 for the notion of a BA number.
(54)Our task, then, is to show that Σl ∈ Cj║la║−2≅22j.


. Since qm+τ−1Kτqm−1, we combine over τ to obtain

Noting that qm−12j, we recover the upper bound.
For the lower bound, a little care is needed to construct intervals [N+1,N+q]⊂Cj on which to apply the lower bound of condition (54) in lemma 1. Define qm as before. Set N=2j and consider the following three cases.
- (a)
For qm>2j+1, set q=qm−1. Since qm−12j, we have N+q=2j+qm−12j+1<qm=q′ and so [N+1,N+q]⊂Cj. In addition,

- (b)
For qm2j+1 and qm+12qm+qm−1 (where the second condition corresponds to am+12 in expression (5)), set q=qm. We have N+q=2j+qm3×2j so [N+1,N+q]⊂Cj and N+q<2qm<qm+1=q′.
- (c)
Finally, suppose that qm2j+1 and qm+1=qm+qm−1. Now set q=qm+1, so that N+q=2j+qm+12j+qm+qm−12j+2j+1+2j4×2j. In addition N+q<qm+qm+1qm+2=q′, and
.

Appendix B
B.1. Embedding of Besov spaces
(56)
First, we observe that condition (56) will be satisfied when f belongs to
. Hence, we only need to prove that
is included in
. For this we shall use two types of Besov embeddings, setting appropriate conditions on s, π, r and q.
- (a)
In the periodic setting, we have
(57) - (b)
In the general case, we have the standard ‘Sobolev embeddings’
(58)
To prove condition (56), we are interested in taking ρ=p. For the case pπ, s>0 implies that only the dense case (34) can occur; hence we need to prove that
. This is always true for s>0 since 1−q/p=2s/(1+2ν+2s).
For the case p>π, we must prove that, in the dense case (34),
. This is equivalent to 2sσ′+(1+2ν)(1/p−1/π)0. But in this case the left‐hand side is greater than (1+2ν)(p/π−1)(s−1/π)≥0. In the sparse case (35), we must check that σ′≥(2ν+1)s′/(1+2ν+2s′), but this is equivalent to (2ν+1)p/{(2ν+1)p−2+2pσ′}1 or s1/π.
into lq,∞(μ). First let us mention that we shall simplify the problem by considering the embedding into

(59)
It remains to study the more intricate cases where we do not have π=r=q.
- (a)
Let q be defined by the relation (59); if 0<rq and
then
(60)
- (b)
Let q be defined by
if 0<rq and
(61)then
(62)

For
this not a restriction, since we are considering 1<p<∞. Moreover, in this case the first member of inequality (62) is always true if we deal with 1π, as 


:

We shall now use embeddings (57) and (58), taking ρ=q.
- (a)
If
and rq we have qπ; hence, using condition (57),
(Let us observe in addition that p>q⇔s>0.)
- (b)
If
and rq we shall use condition (58) to find an embedding with a different order of smoothness. This explains our definition of q. Solving
and using condition (58) gives
We now must check that p>q, but this is equivalent to
B.2. Temlyakov inequalities

(63)Obviously the left‐hand side is always true for p2 with c=1, whereas the right‐hand side is always true for p2 with C=1. In this section, we shall prove the following result.

(64)Proof. We start by proving the theorem for the Haar basis. We introduce the weighted Haar basis (2j/2σjh) where as usual h(x)=hj,k(x)=h(2jx−k) and h(x)=1[0,1](2x)−1[0,1](2x−1).
(65)


(66)

Then the sequence (en(x))n ∈ ℕ satisfies the Temlyakov property also.
Theorem 2 follows from lemma 2 since, for f = 1[0,1],f*(x) ≅ c(1 ∧ 1/|x|), and obviously, for all x ∈ ℝ,|f(x)|f*(x). Combining this with the assumption of theorem 2 we have that |ψ(x)|C h*(x) and |h(x)|C ψ*(x), which is obvious.
To complete the proof we derive the lemma.





B.3. Vaguelette properties
We use (u,κ ∈ I) to denote a generic system of candidate vaguelettes on the circle T. With
, then
stands for
in the case of the (𝒱) system and
in the case of the (𝒰) system.
Adapting the definition of Meyer and Coifman (1997), chapter 8, page 56, we say that {u,κ ∈ I} is a system of periodic vaguelettes on T if there are exponents 0<β<α and a constant C such that
- (a)
|u(t)|2j/2C(1+|2jt−k|)−1−α, t ∈ T,
- (b)
∫Tu(t) dt=0 and
- (c)
|u(s)−u(t)|2j(1/2+β)C|s−t|β,s,t ∈ T.
(In what follows, α=1 and 0<β<1.)

From the remarks around Donoho (1995), theorem 2, this is sufficient for the Riesz basis property.
. For the Hölder condition (c),
(67)
, we have

(68)
(69)
, we have t2∼|1− exp (−2πit)|2 and, setting Δfl=fl+1−fl and Δ2fl=Δ(Δfl),

(70)
If |t|2−j, we simply bound
and retracing the argument from condition (67) to condition (68) with β=0 we obtain condition (69).
References
Citing Literature
Number of times cited according to CrossRef: 64
- V. F. Kravchenko, V. I. Pustovoit, A. V. Yurin, Signal Reconstruction Using New Biorthogonal Frequency-Modified Kravchenko Wavelets When Representing the Measurement Process as a Convolution Model, Journal of Communications Technology and Electronics, 10.1134/S106422692004004X, 65, 4, (423-448), (2020).
- Rida Benhaddou, Anisotropic functional deconvolution for the irregular design: A minimax study, Communications in Statistics - Theory and Methods, 10.1080/03610926.2020.1818783, (1-13), (2020).
- undefined Kogut, undefined Bakuła, Improvement of Full Waveform Airborne Laser Bathymetry Data Processing based on Waves of Neighborhood Points, Remote Sensing, 10.3390/rs11101255, 11, 10, (1255), (2019).
- Rida Benhaddou, Qing Liu, Anisotropic functional deconvolution with long-memory noise: the case of a multi-parameter fractional Wiener sheet, Journal of Nonparametric Statistics, 10.1080/10485252.2019.1604953, (1-29), (2019).
- Rida Benhaddou, Qing Liu, Minimax adaptive wavelet estimator for the anisotropic functional deconvolution model with unknown kernel, Communications in Statistics - Theory and Methods, 10.1080/03610926.2019.1617880, (1-20), (2019).
- Rida Benhaddou, Blind deconvolution model in periodic setting with fractional Gaussian noise, Communications in Statistics - Theory and Methods, 10.1080/03610926.2017.1417431, 48, 3, (438-449), (2018).
- Rida Benhaddou, Minimax lower bounds for the simultaneous wavelet deconvolution with fractional Gaussian noise and unknown kernels, Statistics & Probability Letters, 10.1016/j.spl.2018.05.002, 140, (91-95), (2018).
- Mathias Trabs, Bayesian inverse problems with unknown operators, Inverse Problems, 10.1088/1361-6420/aac3aa, 34, 8, (085001), (2018).
- Kai Ding, Qingquan Li, Jiasong Zhu, Chisheng Wang, Minglei Guan, Zhipeng Chen, Chao Yang, Yang Cui, Jianghai Liao, An Improved Quadrilateral Fitting Algorithm for the Water Column Contribution in Airborne Bathymetric Lidar Waveforms, Sensors, 10.3390/s18020552, 18, 2, (552), (2018).
- Rida Benhaddou, On minimax convergence rates under -risk for the anisotropic functional deconvolution model , Statistics & Probability Letters, 10.1016/j.spl.2017.07.008, 130, (120-125), (2017).
- Rui Li, YouMing Liu, Supersmooth density estimations over L p risk by wavelets, Science China Mathematics, 10.1007/s11425-016-0294-3, 60, 10, (1901-1922), (2017).
- Fabienne Comte, Charles‐A. Cuenod, Marianna Pensky, Yves Rozenholc, Laplace deconvolution on the basis of time domain data and its application to dynamic contrast‐enhanced imaging, Journal of the Royal Statistical Society: Series B (Statistical Methodology), 10.1111/rssb.12159, 79, 1, (69-94), (2016).
- Rida Benhaddou, Deconvolution model with fractional Gaussian noise: A minimax study, Statistics & Probability Letters, 10.1016/j.spl.2016.05.022, 117, (201-208), (2016).
- J. Andrés Christen, Bruno Sansó, Mario Santana-Cibrian, Jorge X. Velasco-Hernández, Bayesian deconvolution of oil well test data using Gaussian processes, Journal of Applied Statistics, 10.1080/02664763.2015.1077374, 43, 4, (721-737), (2015).
- Rafal Kulik, Theofanis Sapatinas, Justin Rory Wishart, Multichannel deconvolution with long range dependence: Upper bounds on the -risk , Applied and Computational Harmonic Analysis, 10.1016/j.acha.2014.04.004, 38, 3, (357-384), (2015).
- Chisheng Wang, Qingquan Li, Yanxiong Liu, Guofeng Wu, Peng Liu, Xiaoli Ding, A comparison of waveform processing algorithms for single-wavelength LiDAR bathymetry, ISPRS Journal of Photogrammetry and Remote Sensing, 10.1016/j.isprsjprs.2014.11.005, 101, (22-35), (2015).
- Mihee Lee, Ling Wang, Haipeng Shen, Peter Hall, Guang Guo, J.S. Marron, Least squares sieve estimation of mixture distributions with boundary effects, Journal of the Korean Statistical Society, 10.1016/j.jkss.2014.07.003, 44, 2, (187-201), (2015).
- Sergios Agapiou, Johnathan M. Bardsley, Omiros Papaspiliopoulos, Andrew M. Stuart, Analysis of the Gibbs Sampler for Hierarchical Inverse Problems, SIAM/ASA Journal on Uncertainty Quantification, 10.1137/130944229, 2, 1, (511-544), (2014).
- Rui Li, Youming Liu, Wavelet optimal estimations for a density with some additive noises, Applied and Computational Harmonic Analysis, 10.1016/j.acha.2013.07.002, 36, 3, (416-433), (2014).
- Rida Benhaddou, Rafal Kulik, Marianna Pensky, Theofanis Sapatinas, Multichannel deconvolution with long-range dependence: A minimax study, Journal of Statistical Planning and Inference, 10.1016/j.jspi.2013.12.008, 148, (1-19), (2014).
- Anestis Antoniadis, Marianna Pensky, Theofanis Sapatinas, Nonparametric regression estimation based on spatially inhomogeneous data: minimax global convergence rates and adaptivity, ESAIM: Probability and Statistics, 10.1051/ps/2012024, 18, (1-41), (2013).
- Justin Rory Wishart, Wavelet deconvolution in a periodic setting with long-range dependent errors, Journal of Statistical Planning and Inference, 10.1016/j.jspi.2012.12.001, 143, 5, (867-881), (2013).
- H Yang, Z B Zhang, Image deblurring based on ForIcM: Fourier shrinkage and incomplete measurements, The Imaging Science Journal, 10.1179/1743131X12Y.0000000001, 60, 6, (344-351), (2013).
- Nicolai Bissantz, Hajo Holzmann, Asymptotics for spectral regularization estimators in statistical inverse problems, Computational Statistics, 10.1007/s00180-012-0309-1, 28, 2, (435-453), (2012).
- Mohammad Abbaszadeh, Christophe Chesneau, Hassan Doosti, Nonparametric estimation of density under bias and multiplicative censoring via wavelet methods, Statistics & Probability Letters, 10.1016/j.spl.2012.01.016, 82, 5, (932-941), (2012).
- Karin Knešaurek, Fourier-wavelet restoration in PET/CT brain studies, Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment, 10.1016/j.nima.2012.06.032, 689, (29-34), (2012).
- Maik Schwarz, Sébastien Van Bellegem, Jean-Pierre Florens, Nonparametric Frontier Estimation from Noisy Data, Exploring Research Frontiers in Contemporary Statistics and Econometrics, 10.1007/978-3-7908-2349-3, (45-64), (2012).
- Christophe Chesneau, Wavelet density estimators for the deconvolution of a component from a mixture, Sankhya A, 10.1007/s13171-011-0017-x, 73, 2, (245-266), (2012).
- Christopher E. Parrish, Inseong Jeong, Robert D. Nowak, R. Brent Smith, Empirical Comparison of Full-Waveform Lidar Algorithms, Photogrammetric Engineering & Remote Sensing, 10.14358/PERS.77.8.825, 77, 8, (825-838), (2011).
- Christophe Chesneau, Adaptive Wavelet Estimator for a Function and its Derivatives in an Indirect Convolution Model, Journal of Statistical Theory and Practice, 10.1080/15598608.2011.10412030, 5, 2, (303-326), (2011).
- Laurent Cavalier, Inverse Problems in Statistics, Inverse Problems and High-Dimensional Estimation, 10.1007/978-3-642-19989-9_1, (3-96), (2011).
- Stefanie Biedermann, Nicolai Bissantz, Holger Dette, Edmund Jones, Optimal designs for indirect regression, Inverse Problems, 10.1088/0266-5611/27/10/105003, 27, 10, (105003), (2011).
- ASHISH KHARE, UMA SHANKER TIWARY, MOONGU JEON, DAUBECHIES COMPLEX WAVELET TRANSFORM BASED MULTILEVEL SHRINKAGE FOR DEBLURRING OF MEDICAL IMAGES IN PRESENCE OF NOISE, International Journal of Wavelets, Multiresolution and Information Processing, 10.1142/S0219691309003100, 07, 05, (587-604), (2011).
- Marine Carrasco, Jean-Pierre Florens, A SPECTRAL METHOD FOR DECONVOLVING A DENSITY, Econometric Theory, 10.1017/S026646661000040X, 27, 3, (546-581), (2010).
- Ashish Khare, Manish Khare, Yongyeon Jeong, Hongkook Kim, Moongu Jeon, Despeckling of medical ultrasound images using Daubechies complex wavelet transform, Signal Processing, 10.1016/j.sigpro.2009.07.008, 90, 2, (428-439), (2010).
- Nasreddine Hajlaoui, Caroline Chaux, Guillaume Perrin, Frédéric Falzon, Amel Benazza-Benyahia, Satellite image restoration in the context of a spatially varying point spread function, Journal of the Optical Society of America A, 10.1364/JOSAA.27.001473, 27, 6, (1473), (2010).
- Laurent Cavalier, Marc Raimondo, Multiscale density estimation with errors in variables, Journal of the Korean Statistical Society, 10.1016/j.jkss.2009.09.001, 39, 4, (417-429), (2010).
- C. Chesneau, J. Fadili, J.-L. Starck, Stein block thresholding for image denoising, Applied and Computational Harmonic Analysis, 10.1016/j.acha.2009.07.003, 28, 1, (67-88), (2010).
- Clément Marteau, Risk hull method for spectral regularization in linear statistical inverse problems, ESAIM: Probability and Statistics, 10.1051/ps/2009011, 14, (409-434), (2010).
- Christophe Chesneau, On adaptive wavelet estimation of a quadratic functional from a deconvolution problem, Annals of the Institute of Statistical Mathematics, 10.1007/s10463-009-0232-6, 63, 2, (405-429), (2009).
- C. Chesneau, M. J. Fadili, J. L. Starck, undefined, 2009 16th IEEE International Conference on Image Processing (ICIP), 10.1109/ICIP.2009.5413572, (1329-1332), (2009).
- C. Marteau, On the stability of the risk hull method for projection estimators, Journal of Statistical Planning and Inference, 10.1016/j.jspi.2008.09.010, 139, 6, (1821-1835), (2009).
- J.-C. Pesquet, A. Benazza-Benyahia, C. Chaux, A SURE Approach for Digital Signal/Image Deconvolution Problems, IEEE Transactions on Signal Processing, 10.1109/TSP.2009.2026077, 57, 12, (4616-4632), (2009).
- Leming Qu, undefined, 2009 Fourth International Conference on Frontier of Computer Science and Technology, 10.1109/FCST.2009.18, (241-247), (2009).
- V.M. Patel, G.R. Easley, D.M. Healy, Shearlet-Based Deconvolution, IEEE Transactions on Image Processing, 10.1109/TIP.2009.2029594, 18, 12, (2673-2685), (2009).
- Glenn R. Easley, Dennis M. Healy,, Carlos A. Berenstein, Image Deconvolution Using a General Ridgelet and Curvelet Domain, SIAM Journal on Imaging Sciences, 10.1137/080720796, 2, 1, (253-283), (2009).
- Athanasia Petsa, Theofanis Sapatinas, Minimax convergence rates under the -risk in the functional deconvolution model, Statistics & Probability Letters, 10.1016/j.spl.2009.03.028, 79, 13, (1568-1576), (2009).
- E. A. Lebedeva, V. Yu. Protasov, Meyer wavelets with least uncertainty constant, Mathematical Notes, 10.1134/S0001434608110096, 84, 5-6, (680-687), (2009).
- Nicolai Bissantz, Gerda Claeskens, Hajo Holzmann, Axel Munk, Testing for lack of fit in inverse regression—with applications to biophotonic imaging, Journal of the Royal Statistical Society: Series B (Statistical Methodology), 10.1111/j.1467-9868.2008.00670.x, 71, 1, (25-48), (2008).
- Maarten Jansen, Guy P. Nason, B. W. Silverman, Multiscale methods for data on graphs and irregular multidimensional situations, Journal of the Royal Statistical Society: Series B (Statistical Methodology), 10.1111/j.1467-9868.2008.00672.x, 71, 1, (97-125), (2008).
- R. Stepanov, T. G. Arshakian, R. Beck, P. Frick, M. Krause, Magnetic field structures of galaxies derived from analysis of Faraday rotation measures, and perspectives for the SKA, Astronomy & Astrophysics, 10.1051/0004-6361:20078678, 480, 1, (45-59), (2008).
- Mikhail Ermakov, Minimax and bayes estimation in deconvolution problem, ESAIM: Probability and Statistics, 10.1051/ps:2007038, 12, (327-344), (2008).
- Alle Meije Wink, Hans Hoogduin, Jos BTM Roerdink, Data-driven haemodynamic response function extraction using Fourier-wavelet regularised deconvolution, BMC Medical Imaging, 10.1186/1471-2342-8-7, 8, 1, (2008).
- A. Benazza-Benyahia, J.-C. Pesquet, undefined, 2008 3rd International Symposium on Communications, Control and Signal Processing, 10.1109/ISCCSP.2008.4537471, (1536-1541), (2008).
- Елена Александровна Лебедева, Elena Alexandrovna Lebedeva, Владимир Юрьевич Протасов, Vladimir Yur'evich Protasov, Всплески Мейера с наименьшей константой неопределенностиMeyer Wavelets with Least Uncertainty Constant, Математические заметкиMatematicheskie Zametki, 10.4213/mzm4002, 84, 5, (732-740), (2008).
- L Cavalier, Nonparametric statistical inverse problems, Inverse Problems, 10.1088/0266-5611/24/3/034004, 24, 3, (034004), (2008).
- Nicolai Bissantz, Hajo Holzmann, Statistical inference for inverse problems, Inverse Problems, 10.1088/0266-5611/24/3/034009, 24, 3, (034009), (2008).
- Alexander Meister, Deconvolution from Fourier-oscillating error densities under decay and smoothness restrictions, Inverse Problems, 10.1088/0266-5611/24/1/015003, 24, 1, (015003), (2007).
- Laurent Cavalier, Marc Raimondo, Wavelet Deconvolution With Noisy Eigenvalues, IEEE Transactions on Signal Processing, 10.1109/TSP.2007.893754, 55, 6, (2414-2424), (2007).
- A. Meister, Deconvolving compactly supported densities, Mathematical Methods of Statistics, 10.3103/S106653070701005X, 16, 1, (63-76), (2007).
- Christophe Chesneau, Regression with random design: A minimax study, Statistics & Probability Letters, 10.1016/j.spl.2006.05.010, 77, 1, (40-53), (2007).
- N. Bissantz, T. Hohage, A. Munk, F. Ruymgaart, Convergence Rates of General Regularization Methods for Statistical Inverse Problems and Applications, SIAM Journal on Numerical Analysis, 10.1137/060651884, 45, 6, (2610-2636), (2007).
- Nicolai Bissantz, Gerda Claeskens, Hajo Holzmann, Axel Munk, Testing for Lack of Fit in Inverse Regression - With Applications to Photonic Imaging, SSRN Electronic Journal, 10.2139/ssrn.1093257, (2007).
- undefined Leming Qu, P.S. Routh, undefined Kyungduk Ko, Wavelet deconvolution in a periodic setting using cross-validation, IEEE Signal Processing Letters, 10.1109/LSP.2005.863657, 13, 4, (232-235), (2006).



)
)
)
)
)
)

. Now by 




; from formula
, we obtain


, we have
for r=0,1,2. Some calculation shows that



