SlideShare uma empresa Scribd logo
1 de 7
Baixar para ler offline
1146                                                                                IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 6, JUNE 2001




                                        Robust Wavelet Denoising
                                               Sylvain Sardy, Paul Tseng, and Andrew Bruce


   Abstract—For extracting a signal from noisy data, waveshrink                  cients for          to have a good mean squared error
and basis pursuit are powerful tools both from an empirical and
asymptotic point of view. They are especially efficient at estimating
spatially inhomogeneous signals when the noise is Gaussian. Their                                  MSE                E
performance is altered when the noise has a long tail distribution,
for instance, when outliers are present.                                         where the expectation is taken over .
   We propose a robust wavelet-based estimator using a robust loss                  Waveshrink uses orthonormal wavelets, which has two im-
function. This entails solving a nontrivial optimization problem
and appropriately choosing the smoothing and robustness param-                   portant consequences: First, the least squares estimate is simply
eters. We illustrate the advantage of the robust wavelet denoising                           , where is the matrix of discretized , and de-
procedure on simulated and real data.                                            notes the transpose of ; second,          is an unbiased estimate
   Index Terms—Basis pursuit, block coordinate relaxation, inte-                 of , and its covariance matrix is       so that the estimated least
rior point, robustness, wavelet, waveshrink.                                     squares coefficients are independent if the noise is Gaussian. For
                                                                                 a smaller mean squared error at the cost of introducing some bias,
                                                                                 Donoho and Johnstone [1] apply the hard or the soft function
                           I. INTRODUCTION

S      UPPOSE we observe a signal
       ated from
                                                                      gener-
                                                                                                              sign                               (3)

                                                                          (1)    where               is the identity function on , and where
                                                                                 is for              and zero otherwise. For Gaussian noise, the
where the equally spaced sampling locations          are points                  shrinkage can be applied to           component-wise because its
on the line for one dimensional (1-D) signals or on a grid for                   components are independent.
images. For now, we assume that the s are identically and                           The hard and soft estimates are, interestingly, the closed-form
independently distributed Gaussian random variables with                         solution to two optimization problems that are, in general, dif-
mean zero and variance one. Our goal is to denoise the signal                    ficult to solve unless is orthonormal.
 , i.e., to find a good estimate     of the underlying signal                        • Best Subset:               is “the best subset of size ” with
                          . The hat on top of a letter is the                             as the          th smallest element of         in the sense
notation used throughout this paper to indicate the estimate of                         that it minimizes the residual sum of squares among all
the corresponding parameter. Waveshrink [1] and basis pursuit                           sets with nonzero wavelet coefficients.
[2] are two nonparametric expansion based estimators. They                           • -Penalized Least Squares:                  is the closed-form
assume that can be well represented by a linear combination                             solution to the following optimization problem:
of wavelet basis functions , namely
                                                                                                                                                 (4)
                                                                          (2)
                                                                                        This property leads to the relaxation algorithm of Sec-
                                                                                        tion II-A and to the definition of basis pursuit.
where                             are the wavelet coefficients.                     When is no longer orthonormal but overcomplete, the least
Waveshrink is defined for orthonormal wavelets only (i.e.,                       squares estimate no longer has independent components, and the
        ), whereas basis pursuit can also use an “overcomplete”                  shrinkage idea cannot be applied as such. Basis pursuit general-
basis (i.e.,            ). The advantage of an overcomplete                      izes to that situation using the optimization problem formulation
wavelet dictionary is discussed by Chen et al. [2]. The goal of                  (4), whose solution is not trivial when is not orthonormal.
Waveshrink and basis pursuit is to estimate the wavelet coeffi-                     The selection of the smoothing parameter is important.
                                                                                 Several ways of selecting have been proposed for Gaussian
  Manuscript received February 1, 2000; revised February 22, 2001. This work     noise. They are based on a minimax argument (see Donoho
was supported in part by an SBIR Phase I contract with the Naval Air Warfare     and Johnstone [1] for real-valued noise and Sardy [3] for com-
Center at China Lake, CA. The associate editor coordinating the review of this   plex-valued noise) or on minimizing the Stein unbiased risk es-
paper and approving it for publication was Prof. Paulo S. R. Diniz.
  S. Sardy is with the Department of Mathematics, Swiss Federal Institute of     timate (SURE) (see Donoho and Johnstone [4]). Nason [5] se-
Technology, Lausanne, Switzerland.                                               lects the smoothing parameter by cross validation.
  P. Tseng is with the Department of Mathematics, University of Washington,         The predictive performance of waveshrink and basis pursuit
Seattle, WA 98195 USA.
  A. G. Bruce is with the MathSoft, Inc., Seattle, WA 98109 USA.                 deteriorates when the noise is not Gaussian. It is because of the
  Publisher Item Identifier S 1053-587X(01)03879-X.                                 loss function in (4). It arises naturally as the log-likelihood
                                                             1053–587X/01$10.00 © 2001 IEEE
SARDY et al.: ROBUST WAVELET DENOISING                                                                                                      1147



                                                                              where                         . We use the Huber loss function
                                                                              [10], which is a hybrid between for small residuals and for
                                                                              large residuals, namely

                                                                                                                                            (6)

                                                                              where           is some cutpoint. Using an even loss function,
                                                                              we implicitly assume that the noise is symmetric around zero.
                                                                              Note that when            and          , it becomes the and
                                                                              loss functions, respectively, both of which have the advantage
                                                                              of being convex.
                                                                                 Our proposal of a robust wavelet denoising procedure raises
                                                                              two issues. On the one hand, we must solve the nontrivial opti-
                                                                              mization problem defined in (5) for a given pair          ; on the
                                                                              other hand, we must select the smoothing parameter and the
                                                                              cutpoint . In Section II, we propose two algorithms to solve
                                                                              the robust wavelet denoising optimization problem (5): a block
                                                                              coordinate relaxation algorithm and an interior point algorithm.
                                                                              In Section III, we discuss the problem of selecting the cutpoint
Fig. 1. Nonrobust estimation of heavisine. (Top) True signal. (Bottom left)
Noisy signal. (Bottom right) Nonrobust estimate.                                 and the smoothing parameter . In Section IV, we give the
                                                                              result of a simulation to compare the efficiency of the two algo-
                                                                              rithms and to compare the denoising performance of the robust
when the noise is Gaussian, but it is no longer appropriate when
                                                                              versus the nonrobust estimators. In Section V, we illustrate ro-
the departure from Gaussianity is too strong. In such a case, the
                                                                              bust basis pursuit on two real data sets. We conclude the paper
quadratic loss function pulls the estimated function toward the
                                                                              in Section VI.
outliers. We illustrate this phenomenon on a 1-D signal in Fig. 1;
the true and noisy signal (90% standard Gaussian noise and 10%
Gaussian noise with a standard deviation of 4) are plotted on the                          II. TWO OPTIMIZATION ALGORITHMS
left side; on the right side, basis pursuit gives poor estimation near        A. Block Coordinate Relaxation (BCR) Algorithm
the outliers. The aim of this paper is to develop a robust wavelet-             The BCR algorithm relies on the following observation. The
based estimator that is less affected by a long-tailed noise.                 Huber loss function (6) may be rewritten as
   Some work has already been done in this direction. Bruce
et al. [6] preprocess the estimation of the wavelet coefficients                                                                            (7)
by a “fast and robust smooth/cleaner” at each multiresolution
level to downweight the effect of the outliers in the estimation              This nontrivial fact can be inferred from results on infimal con-
of the wavelet coefficients. Kovac and Silverman [7] prepro-                  volution as discussed in Rockafellar [11, ch. 16]. Interestingly,
cess the original signal to remove “bad” observations by means                (5) becomes
of a rough statistical test involving a running median smoother
with a window of, say, 5; their procedure has the drawback of                                                                               (8)
losing information by throwing out “bad” observations. For the
situation of a known symmetric long tail noise, Averkamp and                  The reformulated problem (8) has a separable structure to which
Houdré [8] derive minimax rules to select the smoothing pa-                   the BCR algorithm of Sardy et al. [12] can be applied. It solves
rameter. Krim and Schick [9] derive a robust estimator of the                 exactly a succession of subproblems using the soft shrinkage
wavelet coefficients based on minimax description length; their
                                                                              function (3).
assumption of independent noise in the wavelet domain is not
realistic, however.
   In this paper, we propose a different approach that has the                BCR Algorithm:
advantages of having a simple definition, of assuming a re-                   1) Choose an initial guess for             , e.g.,      ;
alistic independent contamination in the measurements , and                   2) Partition           into two matrices:      of orthonormal
of being able to deal with overcomplete wavelets as well. Its                    columns; of the remaining           columns. Define      and
challenges are in finding an efficient algorithm (see Section II)                  as the corresponding coefficients in ;
and in choosing appropriately two tuning parameters (see Sec-                 3) Define the residual vector               . Find the improved
tion III). Since the nonrobust behavior is due to the loss func-                     by solving the subproblem
tion, we simply replace it by a robust loss function and define
the coefficient estimate        of the robust wavelet denoising pro-                            arg
cedure as the solution to
                                                                                 using soft shrinkage;
                                                                       (5)    4) If convergence criterion is not met, go to step 1;
1148                                                                     IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 6, JUNE 2001



   The BCR algorithm assumes that the matrix (here             ) is   where is the log-barrier penalty that is chosen identically for
the union of a finite number of orthonormal matrices        . This    all the penalty terms. Letting      be the th canonical basis
assumption is verified for many wavelet dictionaries including        vector, the first-order optimality condition for the log-barrier
nondecimated wavelets, wavelet packets, local cosine packets,         subproblem is
chirplets [13], and brushlets [14]. Sardy et al. [12] propose two
strategies for choosing      in step 2 of the BCR algorithm and
prove convergence for real and complex-valued signals.

B. Interior Point (IP) Algorithm
   The interior point algorithm has the advantage of not re-
quiring to be the union of orthonormal blocks. It does not            Letting
apply to complex-valued signals, however, and is computation-
ally less efficient than the BCR algorithm (see Section IV-A).
   1) Transformation to Quadratic Programming: First, we
rewrite the optimization problem (5) as

                                  subject to                   (9)

                                                                      and letting
By attaching Lagrange multipliers to the linear constraints, this,
in turn, can be written as
                                                                                                      and

                                                                      the first-order optimality condition is a set of nonlinear equa-
The dual to this problem, which is obtained by exchanging the         tions
order of “min” and “max,” is



                                                                                                                                     (12)
                                                              (10)    where        diag     and      diag     with         and       .
                                                                      The variables , , and are called, respectively, the primal, the
where      is the th column of .                                      dual, and the dual slack variables. This IP problem could alter-
   Since the objective function in (9) is convex, defined every-      natively have been derived from (8) and its dual; for instance,
where, and the constraint is linear, it is known from convex du-                corresponds to , and corresponds to          in (7).
ality theory (see, e.g., Rockaffelar [11, th. 28.2 and 28.4]) that      In an interior point approach, one typically takes a single
the duality gap between the primal (9) and the dual (10) prob-        Newton step to solve the nonlinear system (12) inexactly and
lems is zero. Using (6) and some algebra, the dual problem (10)       then decreases . More precisely, given             ,         , ,
is                                                                    and        , one computes the Newton direction                 ,
                                                                      which is obtained by solving the following system of linear
                           with                               (11)    equations:

This is a quadratic programming problem. Notice that in the
case of        , the dual problem (11) is a linear programming
problem, implying that the primal problem can be transformed                                                                         (13)
into a linear programming problem. For brevity, we omit the
derivation (see Sardy [15]).                                          and then, one updates the variables according to
   2) Interior Point Algorithm: We solve the quadratic pro-
gramming problem using a primal-dual log-barrier interior                                                                            (14)
point algorithm inspired by Chen et al. [2]. The log-barrier                                                                         (15)
subproblem corresponding to (11) is                                                                                                  (16)

                                                                      where               ,       , and         are chosen to maintain
                                                                                    and           . A popular choice is

                                                                                                                                     (17)
                                                                                                                                     (18)
SARDY et al.: ROBUST WAVELET DENOISING                                                                                                   1149



and empirically, the choice of             has worked well. The               In one situation, the pair          can be tuned “by eye.” For
parameter may be updated in many ways. For example, Chen                   instance, in the first application of Section V, the noise is non-
et al. [2] suggested                                   . Typically,        Gaussian, the signal to recover is known to be an aircraft. and
only a small number of interior-point iterations is required to            the signal-to-noise ratio (SNR) is inherent to the infrared sensor
obtain a solution of desired accuracy.                                     used. In that situation, the smoothing parameter and the cut-
   3) Conjugate Gradient Solver for the Newton Step: Most of               point can be tuned on a training set of images and then used
the computational effort is spent in computing the Newton direc-           on future images.
tion at each iteration. From (13), we have that    is the solution            In the other situation, the underlying signal is not known,
of                                                                         and neither is the SNR; therefore, an automatic selection
                                                                           of the pair           is needed. Our radar application of Sec-
                                                                   (19)    tion V is an example of this situation. Several procedures
                                                                           have been developed to select the smoothing parameter
where                   is a diagonal matrix. The dual slack and           for non-Gaussian noise. Nason [5] observes, on a simulation
primal Newton directions are then obtained by                              using i.i.d., Student noise, that the -based cross validation
and                               . We adopt the algorithm of Chen         gives a better prediction than the minimax or SURE rules
et al. [2] and use the conjugate gradient method to solve the              derived for Gaussian noise. With smoothing splines, Xiang and
dense           system (19). Because multiplication by and                 Wahba [17] develop a generalized cross validation criterion
are typically fast (on the order of              or               oper-    for a differentiable smoothness penalty and for a known noise
ations), the conjugate gradient method is attractive. In practice,         distribution in the exponential family. The knowledge of a
however, the number of conjugate gradient iterations required              specific noise distribution is also required by Averkamp and
to solve (19) accurately can become very large as                    ap-   Houdré [8], who develop a minimax rule for specific long tail
proaches a solution, thus degrading the performance of the IP              noise distributions; their selection can only do so much to cope
algorithm.                                                                 with the problem of using the unnatural loss function. Crouse
   4) Finding an Initial Point: The IP algorithm requires an ini-          et al. [18] propose wavelet-Markov models for the dependence
tial point               satisfying             and             , which    between scales, and they employ an EM algorithm to estimate
ideally would not be too far from the solution. Let the ridge re-          their parameters by maximum likelihood. EM is typically
gression estimate                   (obtained by replacing            in   slow and gets trapped into local maxima, however [this cannot
(4) by        ) be an initial guess for the coefficients. Let              happen with our convex cost function (5)].
             and                         . With               and             In this paper, we propose a pragmatic approach that does not
       , let                     and                        . Then, the    require the specific knowledge of the noise distribution. First,
primal variables                                          are positive.    an estimate of scale is required. We use the median absolute
In addition, let             sign                       , and let          deviation of the high-frequency wavelet coefficients of Donoho
                   . Then, the dual variables                              and Johnstone [4]. This estimate of scale is robust to features of
satisfy                          ,                      , and the dual     the underlying signal and to outliers especially if, as suggested
slack variables                     are positive.                          by Kovac and Silverman [7], the Haar wavelet (which support
   5) Convergence: Although there have been many conver-                   has length two) is used. For the cutpoint in (6), we follow
gence studies of IP algorithms, the algorithms that work well in           Huber [10] and choose              . The default in software pack-
practice, including the one described above, often have no guar-           ages is often              based on the following statistical con-
antee of convergence. Specifically, convergence requires the ex-           sideration: Suppose you observe i.i.d. data with mean . Then,
istence of positive constants , , such that                                the asymptotic relative efficiency of the Huber estimate of
     and                       , for                , at all iterations.   is asymptotically 95% efficient with respect to the sample av-
We can enforce convergence by updating in a more conserva-                 erage when the noise is Gaussian (G). We can read this value
tive manner, but this would slow down its convergence in prac-             on the continuous curve (G) of Fig. 2. The relative efficiency is
tice. (See, e.g., Kojima et al. [16] for discussions of these issues       also plotted as a function of for the distributions used in the
in linear programming problems.)                                           simulation. Fig. 2 gives a guideline for the selection of in our
   A stopping rule for the IP algorithm is when all of the fol-            nonparametric regression problem: We see that a value of be-
lowing conditions are satisfied for a small               :                tween one and three gives overall a good efficiency. Based on
                                                                           the simulation of Section IV, we recommend using a cutpoint
        Primal feasibility:                                                of at least         . Finally, for the smoothing parameter , we
          Dual feasibility:                                                use the minimax         developed by Donoho and Johnstone [1]
                                                                           since the residuals within         do not depart dramatically from
              Duality gap:                                         (20)
                                                                           Gaussian residuals.


                   III. SELECTION OF        AND
                                                                                                  IV. SIMULATIONS

   The selection of the smoothing parameter and the cutpoint               A. Computational Efficiency
  is a difficult problem. Two different situations can be distin-            Empirically, Sardy et al. [12] found the BCR algorithm to
guished.                                                                   be more efficient than the IP algorithm at solving basis pursuit.
1150                                                                                  IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 6, JUNE 2001




Fig. 2. Asymptotic relative efficiency as a function of c of the Huber estimate
with respect to the sample average of an i.i.d. sample generated from the
distributions used in the simulation: Gaussian (G), Gaussian mixture (C), and
Student t with three degrees of freedom (T).

                                                                                   Fig. 4. Robust estimation of heavisine. (Top) True signal. (Bottom left) Noisy
                                                                                   signal. (Bottom right) Robust estimate to compare with Fig. 1.




Fig. 3. Decrease in the robust basis pursuit objective function as a function of
the CPU time (in seconds) for the BCR algorithm and the IP algorithm.

                                                                                            Fig. 5. Four signals used in the Monte Carlo experiment.
We observe the same advantage for solving its robust version
(5). The reason is that the matrix         now used only has to                    serves the two discontinuities and downweights the influence of
be augmented by the       columns of the identity matrix. Fig. 3                   the outliers.
illustrates, on four contaminated signals, the superiority of the                     We perform a Monte Carlo experiment to evaluate the relative
BCR algorithm that is up to 10 times faster than the IP algorithm                  performance of the nonrobust and robust wavelet-based estima-
in achieving the desired precision [           in (20)].                           tors with three noise scenarios:
                                                                                     • (G) standard Gaussian noise;
B. Predictive Performance                                                            • (C) a mixture of 90% Gaussian(0, 1) and 10% Gaussian(0,
  To illustrate graphically on a 1-D signal the advantage of                           16) at random locations;
using a robust procedure, Fig. 4 shows the robust estimation of                      • (T) Student noise with three degrees of freedom.
heavisine for the same contaminated data as in Fig. 1. The ro-                     We use the four test functions plotted in Fig. 5 and defined in
bust procedure gives a better reconstruction; in particular, it pre-               Donoho and Johnstone [1, Tab. 1]: blocks, bumps, heavisine, and
SARDY et al.: ROBUST WAVELET DENOISING                                                                                                         1151



                                                             TABLE I
                                                         1
 RELATIVE PREDICTIVE PERFORMANCE OF BASIS PURSUIT (c = ) AND ROBUST BASIS PURSUIT (c = 2:0; c = 1:345). ESTIMATED MEAN SQUARED ERROR
                         2
                       ( 100) WITH A STANDARD ERROR OF ROUGHLY 3%. (G: GAUSSIAN; C: CONTAMINATED; T: STUDENT t




Doppler. We normalize them such that their “standard devia-
tion” is equal to 7


                                 where                          (21)

We choose two sample sizes of                     and               ,
and the “s8” wavelet packet dictionary with all but four levels.
The minimax smoothing parameters             are          2.232 and
           2.594. Following the discussion of Section III, the
smoothing parameter is set to                   and the cutpoint of
the Huber loss function to           with              and          .
   For each combination of noise (G, C, T) of underlying func-
tion (blocks, bumps, heavisine, Doppler), of sample size (
      ,            ), and of procedure (nonrobust, robust), we es-
timate the MSE by averaging (40)(1024)/ model errors (i.e.,
40 for              and 10 for             to get the same number
of points is generated for the two sample sizes). Table I reports
the estimated MSEs of the competing estimators for the 24 sce-
narios.
   In light of Table I, a cutpoint of at least          is advisable
for robust basis pursuit; the standard value of              derived
in the parametric context from asymptotic considerations is not                                                                   2
                                                                        Fig. 6. Top: (Left) Noisy infrared sensor data and (Right) 3 3 robust median
large enough. With a cutpoint of        , the gain in efficiency can    filter denoised image. Bottom: (Left) Nonrobust and (Right) robust wavelet
be dramatic for non-Gaussian noise using robust basis pursuit.          based denoised images.
Its counterperformance on the bumps signal is due to the na-
ture of the signal whose features are difficult to distinguish with     the 2-D nondecimated wavelet transform with the “s8” wavelet.
noise in the upper tail when the sampling is light (               );   In this application, we know the underlying image we want to
with an heavier sampling (                  ), robust basis pursuit     recover; therefore, we tried several values of and (which was
again beats the nonrobust estimator for non-Gaussian noise.             feasible in a finite time thanks to the BCR algorithm) and chose
                                                                                           for the best visually appealing reconstruction
                                                                        of the aircraft. Using this parameter setting                 , the
                        V. APPLICATIONS
                                                                        robust procedure can be used to clean future images.
   The first data is an image taken by a long-wavelength infrared          The second data are radar glint observations and consist of
sensor. Just visible above the center of the image is an A-37                        angles of a target in degrees. The signal contains a
trainer aircraft flying above the Sierra Nevada at some distance        number of glint spikes, causing the apparent signal to behave
from the sensor platform. The 128 by 128 original image plotted         erratically. From physical considerations, a good model for the
in the top left of Fig. 6 clearly shows some “outlier pixels.” A        true signal is a low-frequency oscillation about . The esti-
standard median filter (with a        window) gets rid of the bad       mated standard deviation is                . To get a nice “noise-
pixels but does not preserve the aircraft well (top right). The two     free” visual display, we choose the universal threshold
bottom plots of Fig. 6 show the denoised image using (left) basis                             , and for the robust version, we choose
pursuit and (right) robust basis pursuit. The robust denoising                         . Fig. 7 shows the (top) original signal and (left:
procedure has the definite advantage of cleaning the image of           nonrobust; right: robust) the denoised estimates at the bottom.
the bad pixels while preserving the outline of the airplane. To         The robust estimate is a low-frequency oscillation, as expected,
clean the image with the two wavelet-based techniques, we used          whereas the nonrobust estimate remains jagged. Note that the
1152                                                                               IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 6, JUNE 2001



                                                                                  [2] S. S. Chen, D. L. Donoho, and M. A. Saunders, “Atomic decomposition
                                                                                      by basis pursuit,” SIAM J. Sci. Comput., vol. 20, no. 1, pp. 33–61, 1999.
                                                                                  [3] S. Sardy, “Minimax threshold for denoising complex signals with
                                                                                      waveshrink,” IEEE Trans. Signal Processing, vol. 48, pp. 1023–1028,
                                                                                      Apr. 2000.
                                                                                  [4] D. L. Donoho and I. M. Johnstone, “Adapting to unknown smoothness
                                                                                      via wavelet shrinkage,” J. Amer. Statist. Assoc., vol. 90, pp. 1200–1224,
                                                                                      Dec. 1995.
                                                                                  [5] G. P. Nason, “Wavelet function estimation using cross-validation,” in
                                                                                      Wavelets and Statistics, A. Antoniadis and G. Oppenheim, Eds. New
                                                                                      York: Springer-Verlag, 1995, pp. 261–280.
                                                                                  [6] A. G. Bruce, I. M. Donoho, and R. D. Martin, “Denoising and robust
                                                                                      nonlinear wavelet analysis,” Wavelet Applicat., vol. 2242, Apr. 1994.
                                                                                  [7] A. Kovac and B. W. Silverman, “Extending the scope of wavelet regres-
                                                                                      sion methods by coefficient-dependent thresholding,” J. Amer. Statist.
                                                                                      Assoc., vol. 95, pp. 172–183, 2000.
                                                                                  [8] R. Averkamp and C. Houdré, “Wavelet thresholding for non (neces-
                                                                                      sarily) gaussian noise: A preliminary report,” Georgia Inst. Technol.,
                                                                                      Atlanta, Tech. Rep., 1996.
                                                                                  [9] H. Krim and I. C. Schick, “Minimax description length for signal de-
                                                                                      noising and optimized representation,” IEEE Trans. Inform. Theory, vol.
                                                                                      45, pp. 898–908, May 1999.
                                                                                 [10] P. J. Huber, Robust Statistics. New York: Wiley, 1981.
                                                                                 [11] R. Rockafellar, Convex Analysis. Princeton, NJ: Princeton Univ. Press,
                                                                                      1970.
                                                                                 [12] S. Sardy, A. Bruce, and P. Tseng, “Block coordinate relaxation methods
                                                                                      for nonparametric wavelet denoising,” J. Comput. Graph. Statist., vol.
                                                                                      9, no. 2, pp. 361–379, 2000.
                                                                                 [13] S. Mann and S. Haykin, “Adaptive "chirplet" transform: And adaptive
                                                                                      generalization of the wavelet transform,” Opt. Eng., vol. 31, no. 6, pp.
                                                                                      1243–1256, June 1992.
Fig. 7. Top: (Left) Noisy radar glint data and (Right) robust median filter      [14] F. G. Meyer and R. R. Coifman, “Biorthogonal brushlet bases for direc-
estimate. Bottom: (Left) Nonrobust and (right) robust wavelet-based estimate          tional image compression,” in Proc. 12th Int. Conf. Anal. Optimization
(with local cosine packet).                                                           Syst., 1996.
                                                                                 [15] S. Sardy, “Regularization Techniques for Linear Regression with a Large
                                                                                      Set of Carriers,” Ph.D. dissertation, Univ. Washington, Seattle, 1998.
median filter’s estimate still shows a highly oscillating denoised               [16] M. Kojima, N. Megiddo, and S. Mizuno, “A primal-dual exterior point
                                                                                      algorithm for linear programming,” Math. Progr., vol. 61, pp. 261–280,
signal.                                                                               1993.
                                                                                 [17] D. Xiang and G. Wahba, “A generalized approximate cross validation
                           VI. CONCLUSION                                             for smoothing splines with nongaussian data,” Statistica Sinica, vol. 6,
                                                                                      pp. 675–692, 1996.
   We have proposed a robust version of basis pursuit by re-                     [18] M. S. Crouse, R. G. Baraniuk, and R. D. Nowak, “Signal estimation
                                                                                      using wavelet-markov models,” in Proc. IEEE Int. Conf. Acoust.,
placing the nonrobust loss function by the so-called Huber                            Speech, Signal Process., vol. 5, 1997, pp. 3429–3432.
loss function. We solved the corresponding nontrivial optimiza-
tion problem for a given smoothing parameter                and a
given cutpoint           with an efficient and converging block                                         Sylvain Sardy received the Maîtrise d’Ingénièrie
coordinate relaxation method that is empirically faster than an                                         Mathématique degree from the University of
                                                                                                        Franche-Comté, Besancon, France, in 1989. He
interior point competitor. The two techniques are available in                                          then received the M.S. degree in mathematics in
the wavelet module of the Splus statistical software; the BCR                                           1991 and the M.S. degree in statistics in 1992,
algorithm can otherwise be easily implemented. We proposed a                                            both from Utah State University, Logan. After
                                                                                                        doing his military service for France at the Nuclear
rule to choose the smoothing parameter and the cutpoint of                                              Engineering Department of the Massachusetts
the Huber loss function . We showed on a particular simula-                                             Institute of Technology, Cambridge, he received
tion that robust basis pursuit has a good predictive performance                                        the Ph.D. degree in statistics from the University of
                                                                                                        Washington, Seattle.
with both Gaussian and long-tailed symmetric additive noise; in                   He is now a Postdoctoral Assistant with the Swiss Federal Institute of Tech-
particular, we recommend using a cutpoint of at least          for              nology (EPFL), Lausanne, in the Mathematics Department (DMA).
the Huber loss function. As illustrated with two applications, ro-
bust basis pursuit has a definite advantage over both a nonrobust
wavelet-based estimator and a median filter estimator.                          Paul Tseng received the B.Sc. degree in engineering mathematics from Queen’s
                                                                                University, Kingston, ON, Canada, in 1981 and the Ph.D. degree in operations
                                                                                research from the Massachusetts Institute of Technology (MIT), Cambridge, in
                         ACKNOWLEDGMENT                                         1986.
                                                                                   From 1986 to 1987, he was a University Research Fellow with the Depart-
  The authors wish to thank the associate editor and five anony-                ment of Management Science, University of British Columbia, Vancouver, BC,
mous referees for their careful review and A. Davison for pro-                  Canada. From 1987 to 1990, he was a research associate with the Laboratory
viding help with Fig. 2.                                                        for Information and Decision Systems at MIT. He joined the Department of
                                                                                Mathematics, University of Washington, Seattle, in 1990, where he is currently
                                                                                a Professor. His research interests are in optimization and algorithms.
                              REFERENCES
  [1] D. L. Donoho and I. M. Johnstone, “Ideal spatial adaptation via wavelet
      shrinkage,” Biometrika, vol. 81, pp. 425–455, 1994.                       Andrew Bruce, photograph and biography not available at time of publication.

Mais conteúdo relacionado

Mais procurados

Image Denoising Using Wavelet Transform
Image Denoising Using Wavelet TransformImage Denoising Using Wavelet Transform
Image Denoising Using Wavelet TransformIJERA Editor
 
1998 characterisation of multilayers by x ray reflection
1998 characterisation of multilayers by x ray reflection1998 characterisation of multilayers by x ray reflection
1998 characterisation of multilayers by x ray reflectionpmloscholte
 
1989 optical measurement of the refractive index, layer thickness, and volume...
1989 optical measurement of the refractive index, layer thickness, and volume...1989 optical measurement of the refractive index, layer thickness, and volume...
1989 optical measurement of the refractive index, layer thickness, and volume...pmloscholte
 
Ds31568573
Ds31568573Ds31568573
Ds31568573IJMER
 
Modern features-part-2-descriptors
Modern features-part-2-descriptorsModern features-part-2-descriptors
Modern features-part-2-descriptorszukun
 
Accelarating Optical Quadrature Microscopy Using GPUs
Accelarating Optical Quadrature Microscopy Using GPUsAccelarating Optical Quadrature Microscopy Using GPUs
Accelarating Optical Quadrature Microscopy Using GPUsPerhaad Mistry
 
Gravitational lensing characteristics of the transparent sun
Gravitational lensing characteristics of the transparent sunGravitational lensing characteristics of the transparent sun
Gravitational lensing characteristics of the transparent sunSérgio Sacani
 
Paramètres X - Extension Long Term Dynamic Memory Effects
Paramètres X - Extension  Long Term Dynamic Memory EffectsParamètres X - Extension  Long Term Dynamic Memory Effects
Paramètres X - Extension Long Term Dynamic Memory Effectsmlecouve
 
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?Sebastian De Haro
 
Gravitation al field_equations_and_theory_of_dark matter_and_dark_energy
Gravitation al field_equations_and_theory_of_dark matter_and_dark_energyGravitation al field_equations_and_theory_of_dark matter_and_dark_energy
Gravitation al field_equations_and_theory_of_dark matter_and_dark_energySérgio Sacani
 

Mais procurados (20)

Microphone arrays
Microphone arraysMicrophone arrays
Microphone arrays
 
Image Denoising Using Wavelet Transform
Image Denoising Using Wavelet TransformImage Denoising Using Wavelet Transform
Image Denoising Using Wavelet Transform
 
1998 characterisation of multilayers by x ray reflection
1998 characterisation of multilayers by x ray reflection1998 characterisation of multilayers by x ray reflection
1998 characterisation of multilayers by x ray reflection
 
1989 optical measurement of the refractive index, layer thickness, and volume...
1989 optical measurement of the refractive index, layer thickness, and volume...1989 optical measurement of the refractive index, layer thickness, and volume...
1989 optical measurement of the refractive index, layer thickness, and volume...
 
05503983
0550398305503983
05503983
 
Recent advances of MEIS for near surface analysis
Recent advances of MEIS for near surface analysis Recent advances of MEIS for near surface analysis
Recent advances of MEIS for near surface analysis
 
Ds31568573
Ds31568573Ds31568573
Ds31568573
 
Modern features-part-2-descriptors
Modern features-part-2-descriptorsModern features-part-2-descriptors
Modern features-part-2-descriptors
 
Wavelet
WaveletWavelet
Wavelet
 
Accelarating Optical Quadrature Microscopy Using GPUs
Accelarating Optical Quadrature Microscopy Using GPUsAccelarating Optical Quadrature Microscopy Using GPUs
Accelarating Optical Quadrature Microscopy Using GPUs
 
Ak32253257
Ak32253257Ak32253257
Ak32253257
 
Gravitational lensing characteristics of the transparent sun
Gravitational lensing characteristics of the transparent sunGravitational lensing characteristics of the transparent sun
Gravitational lensing characteristics of the transparent sun
 
Projet Ma2
Projet Ma2Projet Ma2
Projet Ma2
 
Paramètres X - Extension Long Term Dynamic Memory Effects
Paramètres X - Extension  Long Term Dynamic Memory EffectsParamètres X - Extension  Long Term Dynamic Memory Effects
Paramètres X - Extension Long Term Dynamic Memory Effects
 
Compressed Sensing In Spectral Imaging
Compressed Sensing In Spectral Imaging  Compressed Sensing In Spectral Imaging
Compressed Sensing In Spectral Imaging
 
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?
Instantons and Chern-Simons Terms in AdS4/CFT3: Gravity on the Brane?
 
Crystallite size nanomaterials
Crystallite size   nanomaterialsCrystallite size   nanomaterials
Crystallite size nanomaterials
 
Hq3114621465
Hq3114621465Hq3114621465
Hq3114621465
 
Sparse representation and compressive sensing
Sparse representation and compressive sensingSparse representation and compressive sensing
Sparse representation and compressive sensing
 
Gravitation al field_equations_and_theory_of_dark matter_and_dark_energy
Gravitation al field_equations_and_theory_of_dark matter_and_dark_energyGravitation al field_equations_and_theory_of_dark matter_and_dark_energy
Gravitation al field_equations_and_theory_of_dark matter_and_dark_energy
 

Destaque

Destaque (6)

IVFC Signal Denoising
IVFC Signal DenoisingIVFC Signal Denoising
IVFC Signal Denoising
 
Signal Processing Course : Denoising
Signal Processing Course : DenoisingSignal Processing Course : Denoising
Signal Processing Course : Denoising
 
Signal Processing Course : Wavelets
Signal Processing Course : WaveletsSignal Processing Course : Wavelets
Signal Processing Course : Wavelets
 
Wavelet
WaveletWavelet
Wavelet
 
ECG Basics
ECG BasicsECG Basics
ECG Basics
 
Ecg Signal Processing
Ecg Signal ProcessingEcg Signal Processing
Ecg Signal Processing
 

Semelhante a Robust wavelet denoising

Localization of Objects Using Cross-Correlation of Shadow Fading Noise and Co...
Localization of Objects Using Cross-Correlation of Shadow Fading Noise and Co...Localization of Objects Using Cross-Correlation of Shadow Fading Noise and Co...
Localization of Objects Using Cross-Correlation of Shadow Fading Noise and Co...Rana Basheer
 
WAVELET THRESHOLDING APPROACH FOR IMAGE DENOISING
WAVELET THRESHOLDING APPROACH FOR IMAGE DENOISINGWAVELET THRESHOLDING APPROACH FOR IMAGE DENOISING
WAVELET THRESHOLDING APPROACH FOR IMAGE DENOISINGIJNSA Journal
 
A Review on Image Denoising using Wavelet Transform
A Review on Image Denoising using Wavelet TransformA Review on Image Denoising using Wavelet Transform
A Review on Image Denoising using Wavelet Transformijsrd.com
 
An efficient approach to wavelet image Denoising
An efficient approach to wavelet image DenoisingAn efficient approach to wavelet image Denoising
An efficient approach to wavelet image Denoisingijcsit
 
Fuzzieee-98-final
Fuzzieee-98-finalFuzzieee-98-final
Fuzzieee-98-finalSumit Sen
 
2001 Tf Lcmv Signal Enhancement Using Beamforming And Nonstationarity With Ap...
2001 Tf Lcmv Signal Enhancement Using Beamforming And Nonstationarity With Ap...2001 Tf Lcmv Signal Enhancement Using Beamforming And Nonstationarity With Ap...
2001 Tf Lcmv Signal Enhancement Using Beamforming And Nonstationarity With Ap...luca
 
Thresholding eqns for wavelet
Thresholding eqns for waveletThresholding eqns for wavelet
Thresholding eqns for waveletajayhakkumar
 
Paper id 21201498
Paper id 21201498Paper id 21201498
Paper id 21201498IJRAT
 
Image Compression Using Wavelet Packet Tree
Image Compression Using Wavelet Packet TreeImage Compression Using Wavelet Packet Tree
Image Compression Using Wavelet Packet TreeIDES Editor
 
Automatic high order absorption layers for advective-diffusive systems of equ...
Automatic high order absorption layers for advective-diffusive systems of equ...Automatic high order absorption layers for advective-diffusive systems of equ...
Automatic high order absorption layers for advective-diffusive systems of equ...Storti Mario
 
Image Fusion Based On Wavelet And Curvelet Transform
Image Fusion Based On Wavelet And Curvelet TransformImage Fusion Based On Wavelet And Curvelet Transform
Image Fusion Based On Wavelet And Curvelet TransformIOSR Journals
 
Waveguide beamprop
Waveguide beampropWaveguide beamprop
Waveguide beampropeiacqer
 
用R语言做曲线拟合
用R语言做曲线拟合用R语言做曲线拟合
用R语言做曲线拟合Wenxiang Zhu
 
Accelerometers to Augmented Reality
Accelerometers to Augmented RealityAccelerometers to Augmented Reality
Accelerometers to Augmented Realityjblocksom
 

Semelhante a Robust wavelet denoising (20)

Localization of Objects Using Cross-Correlation of Shadow Fading Noise and Co...
Localization of Objects Using Cross-Correlation of Shadow Fading Noise and Co...Localization of Objects Using Cross-Correlation of Shadow Fading Noise and Co...
Localization of Objects Using Cross-Correlation of Shadow Fading Noise and Co...
 
WAVELET THRESHOLDING APPROACH FOR IMAGE DENOISING
WAVELET THRESHOLDING APPROACH FOR IMAGE DENOISINGWAVELET THRESHOLDING APPROACH FOR IMAGE DENOISING
WAVELET THRESHOLDING APPROACH FOR IMAGE DENOISING
 
A Review on Image Denoising using Wavelet Transform
A Review on Image Denoising using Wavelet TransformA Review on Image Denoising using Wavelet Transform
A Review on Image Denoising using Wavelet Transform
 
An efficient approach to wavelet image Denoising
An efficient approach to wavelet image DenoisingAn efficient approach to wavelet image Denoising
An efficient approach to wavelet image Denoising
 
Lc3618931897
Lc3618931897Lc3618931897
Lc3618931897
 
Fuzzieee-98-final
Fuzzieee-98-finalFuzzieee-98-final
Fuzzieee-98-final
 
2001 Tf Lcmv Signal Enhancement Using Beamforming And Nonstationarity With Ap...
2001 Tf Lcmv Signal Enhancement Using Beamforming And Nonstationarity With Ap...2001 Tf Lcmv Signal Enhancement Using Beamforming And Nonstationarity With Ap...
2001 Tf Lcmv Signal Enhancement Using Beamforming And Nonstationarity With Ap...
 
Thresholding eqns for wavelet
Thresholding eqns for waveletThresholding eqns for wavelet
Thresholding eqns for wavelet
 
Paper id 21201498
Paper id 21201498Paper id 21201498
Paper id 21201498
 
Study and Analysis of Multiwavelet Transform with Threshold in Image Denoisin...
Study and Analysis of Multiwavelet Transform with Threshold in Image Denoisin...Study and Analysis of Multiwavelet Transform with Threshold in Image Denoisin...
Study and Analysis of Multiwavelet Transform with Threshold in Image Denoisin...
 
Image Compression Using Wavelet Packet Tree
Image Compression Using Wavelet Packet TreeImage Compression Using Wavelet Packet Tree
Image Compression Using Wavelet Packet Tree
 
Dx32764769
Dx32764769Dx32764769
Dx32764769
 
10.1.1.2.9988
10.1.1.2.998810.1.1.2.9988
10.1.1.2.9988
 
Automatic high order absorption layers for advective-diffusive systems of equ...
Automatic high order absorption layers for advective-diffusive systems of equ...Automatic high order absorption layers for advective-diffusive systems of equ...
Automatic high order absorption layers for advective-diffusive systems of equ...
 
Image Fusion Based On Wavelet And Curvelet Transform
Image Fusion Based On Wavelet And Curvelet TransformImage Fusion Based On Wavelet And Curvelet Transform
Image Fusion Based On Wavelet And Curvelet Transform
 
Waveguide beamprop
Waveguide beampropWaveguide beamprop
Waveguide beamprop
 
1873 1878
1873 18781873 1878
1873 1878
 
1873 1878
1873 18781873 1878
1873 1878
 
用R语言做曲线拟合
用R语言做曲线拟合用R语言做曲线拟合
用R语言做曲线拟合
 
Accelerometers to Augmented Reality
Accelerometers to Augmented RealityAccelerometers to Augmented Reality
Accelerometers to Augmented Reality
 

Último

SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESSALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESmohitsingh558521
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxLoriGlavin3
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxLoriGlavin3
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionDilum Bandara
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteDianaGray10
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfAddepto
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Mark Simos
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfMounikaPolabathina
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Commit University
 

Último (20)

SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESSALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An Introduction
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test Suite
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdf
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdf
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!
 

Robust wavelet denoising

  • 1. 1146 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 6, JUNE 2001 Robust Wavelet Denoising Sylvain Sardy, Paul Tseng, and Andrew Bruce Abstract—For extracting a signal from noisy data, waveshrink cients for to have a good mean squared error and basis pursuit are powerful tools both from an empirical and asymptotic point of view. They are especially efficient at estimating spatially inhomogeneous signals when the noise is Gaussian. Their MSE E performance is altered when the noise has a long tail distribution, for instance, when outliers are present. where the expectation is taken over . We propose a robust wavelet-based estimator using a robust loss Waveshrink uses orthonormal wavelets, which has two im- function. This entails solving a nontrivial optimization problem and appropriately choosing the smoothing and robustness param- portant consequences: First, the least squares estimate is simply eters. We illustrate the advantage of the robust wavelet denoising , where is the matrix of discretized , and de- procedure on simulated and real data. notes the transpose of ; second, is an unbiased estimate Index Terms—Basis pursuit, block coordinate relaxation, inte- of , and its covariance matrix is so that the estimated least rior point, robustness, wavelet, waveshrink. squares coefficients are independent if the noise is Gaussian. For a smaller mean squared error at the cost of introducing some bias, Donoho and Johnstone [1] apply the hard or the soft function I. INTRODUCTION S UPPOSE we observe a signal ated from gener- sign (3) (1) where is the identity function on , and where is for and zero otherwise. For Gaussian noise, the where the equally spaced sampling locations are points shrinkage can be applied to component-wise because its on the line for one dimensional (1-D) signals or on a grid for components are independent. images. For now, we assume that the s are identically and The hard and soft estimates are, interestingly, the closed-form independently distributed Gaussian random variables with solution to two optimization problems that are, in general, dif- mean zero and variance one. Our goal is to denoise the signal ficult to solve unless is orthonormal. , i.e., to find a good estimate of the underlying signal • Best Subset: is “the best subset of size ” with . The hat on top of a letter is the as the th smallest element of in the sense notation used throughout this paper to indicate the estimate of that it minimizes the residual sum of squares among all the corresponding parameter. Waveshrink [1] and basis pursuit sets with nonzero wavelet coefficients. [2] are two nonparametric expansion based estimators. They • -Penalized Least Squares: is the closed-form assume that can be well represented by a linear combination solution to the following optimization problem: of wavelet basis functions , namely (4) (2) This property leads to the relaxation algorithm of Sec- tion II-A and to the definition of basis pursuit. where are the wavelet coefficients. When is no longer orthonormal but overcomplete, the least Waveshrink is defined for orthonormal wavelets only (i.e., squares estimate no longer has independent components, and the ), whereas basis pursuit can also use an “overcomplete” shrinkage idea cannot be applied as such. Basis pursuit general- basis (i.e., ). The advantage of an overcomplete izes to that situation using the optimization problem formulation wavelet dictionary is discussed by Chen et al. [2]. The goal of (4), whose solution is not trivial when is not orthonormal. Waveshrink and basis pursuit is to estimate the wavelet coeffi- The selection of the smoothing parameter is important. Several ways of selecting have been proposed for Gaussian Manuscript received February 1, 2000; revised February 22, 2001. This work noise. They are based on a minimax argument (see Donoho was supported in part by an SBIR Phase I contract with the Naval Air Warfare and Johnstone [1] for real-valued noise and Sardy [3] for com- Center at China Lake, CA. The associate editor coordinating the review of this plex-valued noise) or on minimizing the Stein unbiased risk es- paper and approving it for publication was Prof. Paulo S. R. Diniz. S. Sardy is with the Department of Mathematics, Swiss Federal Institute of timate (SURE) (see Donoho and Johnstone [4]). Nason [5] se- Technology, Lausanne, Switzerland. lects the smoothing parameter by cross validation. P. Tseng is with the Department of Mathematics, University of Washington, The predictive performance of waveshrink and basis pursuit Seattle, WA 98195 USA. A. G. Bruce is with the MathSoft, Inc., Seattle, WA 98109 USA. deteriorates when the noise is not Gaussian. It is because of the Publisher Item Identifier S 1053-587X(01)03879-X. loss function in (4). It arises naturally as the log-likelihood 1053–587X/01$10.00 © 2001 IEEE
  • 2. SARDY et al.: ROBUST WAVELET DENOISING 1147 where . We use the Huber loss function [10], which is a hybrid between for small residuals and for large residuals, namely (6) where is some cutpoint. Using an even loss function, we implicitly assume that the noise is symmetric around zero. Note that when and , it becomes the and loss functions, respectively, both of which have the advantage of being convex. Our proposal of a robust wavelet denoising procedure raises two issues. On the one hand, we must solve the nontrivial opti- mization problem defined in (5) for a given pair ; on the other hand, we must select the smoothing parameter and the cutpoint . In Section II, we propose two algorithms to solve the robust wavelet denoising optimization problem (5): a block coordinate relaxation algorithm and an interior point algorithm. In Section III, we discuss the problem of selecting the cutpoint Fig. 1. Nonrobust estimation of heavisine. (Top) True signal. (Bottom left) Noisy signal. (Bottom right) Nonrobust estimate. and the smoothing parameter . In Section IV, we give the result of a simulation to compare the efficiency of the two algo- rithms and to compare the denoising performance of the robust when the noise is Gaussian, but it is no longer appropriate when versus the nonrobust estimators. In Section V, we illustrate ro- the departure from Gaussianity is too strong. In such a case, the bust basis pursuit on two real data sets. We conclude the paper quadratic loss function pulls the estimated function toward the in Section VI. outliers. We illustrate this phenomenon on a 1-D signal in Fig. 1; the true and noisy signal (90% standard Gaussian noise and 10% Gaussian noise with a standard deviation of 4) are plotted on the II. TWO OPTIMIZATION ALGORITHMS left side; on the right side, basis pursuit gives poor estimation near A. Block Coordinate Relaxation (BCR) Algorithm the outliers. The aim of this paper is to develop a robust wavelet- The BCR algorithm relies on the following observation. The based estimator that is less affected by a long-tailed noise. Huber loss function (6) may be rewritten as Some work has already been done in this direction. Bruce et al. [6] preprocess the estimation of the wavelet coefficients (7) by a “fast and robust smooth/cleaner” at each multiresolution level to downweight the effect of the outliers in the estimation This nontrivial fact can be inferred from results on infimal con- of the wavelet coefficients. Kovac and Silverman [7] prepro- volution as discussed in Rockafellar [11, ch. 16]. Interestingly, cess the original signal to remove “bad” observations by means (5) becomes of a rough statistical test involving a running median smoother with a window of, say, 5; their procedure has the drawback of (8) losing information by throwing out “bad” observations. For the situation of a known symmetric long tail noise, Averkamp and The reformulated problem (8) has a separable structure to which Houdré [8] derive minimax rules to select the smoothing pa- the BCR algorithm of Sardy et al. [12] can be applied. It solves rameter. Krim and Schick [9] derive a robust estimator of the exactly a succession of subproblems using the soft shrinkage wavelet coefficients based on minimax description length; their function (3). assumption of independent noise in the wavelet domain is not realistic, however. In this paper, we propose a different approach that has the BCR Algorithm: advantages of having a simple definition, of assuming a re- 1) Choose an initial guess for , e.g., ; alistic independent contamination in the measurements , and 2) Partition into two matrices: of orthonormal of being able to deal with overcomplete wavelets as well. Its columns; of the remaining columns. Define and challenges are in finding an efficient algorithm (see Section II) as the corresponding coefficients in ; and in choosing appropriately two tuning parameters (see Sec- 3) Define the residual vector . Find the improved tion III). Since the nonrobust behavior is due to the loss func- by solving the subproblem tion, we simply replace it by a robust loss function and define the coefficient estimate of the robust wavelet denoising pro- arg cedure as the solution to using soft shrinkage; (5) 4) If convergence criterion is not met, go to step 1;
  • 3. 1148 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 6, JUNE 2001 The BCR algorithm assumes that the matrix (here ) is where is the log-barrier penalty that is chosen identically for the union of a finite number of orthonormal matrices . This all the penalty terms. Letting be the th canonical basis assumption is verified for many wavelet dictionaries including vector, the first-order optimality condition for the log-barrier nondecimated wavelets, wavelet packets, local cosine packets, subproblem is chirplets [13], and brushlets [14]. Sardy et al. [12] propose two strategies for choosing in step 2 of the BCR algorithm and prove convergence for real and complex-valued signals. B. Interior Point (IP) Algorithm The interior point algorithm has the advantage of not re- quiring to be the union of orthonormal blocks. It does not Letting apply to complex-valued signals, however, and is computation- ally less efficient than the BCR algorithm (see Section IV-A). 1) Transformation to Quadratic Programming: First, we rewrite the optimization problem (5) as subject to (9) and letting By attaching Lagrange multipliers to the linear constraints, this, in turn, can be written as and the first-order optimality condition is a set of nonlinear equa- The dual to this problem, which is obtained by exchanging the tions order of “min” and “max,” is (12) (10) where diag and diag with and . The variables , , and are called, respectively, the primal, the where is the th column of . dual, and the dual slack variables. This IP problem could alter- Since the objective function in (9) is convex, defined every- natively have been derived from (8) and its dual; for instance, where, and the constraint is linear, it is known from convex du- corresponds to , and corresponds to in (7). ality theory (see, e.g., Rockaffelar [11, th. 28.2 and 28.4]) that In an interior point approach, one typically takes a single the duality gap between the primal (9) and the dual (10) prob- Newton step to solve the nonlinear system (12) inexactly and lems is zero. Using (6) and some algebra, the dual problem (10) then decreases . More precisely, given , , , is and , one computes the Newton direction , which is obtained by solving the following system of linear with (11) equations: This is a quadratic programming problem. Notice that in the case of , the dual problem (11) is a linear programming problem, implying that the primal problem can be transformed (13) into a linear programming problem. For brevity, we omit the derivation (see Sardy [15]). and then, one updates the variables according to 2) Interior Point Algorithm: We solve the quadratic pro- gramming problem using a primal-dual log-barrier interior (14) point algorithm inspired by Chen et al. [2]. The log-barrier (15) subproblem corresponding to (11) is (16) where , , and are chosen to maintain and . A popular choice is (17) (18)
  • 4. SARDY et al.: ROBUST WAVELET DENOISING 1149 and empirically, the choice of has worked well. The In one situation, the pair can be tuned “by eye.” For parameter may be updated in many ways. For example, Chen instance, in the first application of Section V, the noise is non- et al. [2] suggested . Typically, Gaussian, the signal to recover is known to be an aircraft. and only a small number of interior-point iterations is required to the signal-to-noise ratio (SNR) is inherent to the infrared sensor obtain a solution of desired accuracy. used. In that situation, the smoothing parameter and the cut- 3) Conjugate Gradient Solver for the Newton Step: Most of point can be tuned on a training set of images and then used the computational effort is spent in computing the Newton direc- on future images. tion at each iteration. From (13), we have that is the solution In the other situation, the underlying signal is not known, of and neither is the SNR; therefore, an automatic selection of the pair is needed. Our radar application of Sec- (19) tion V is an example of this situation. Several procedures have been developed to select the smoothing parameter where is a diagonal matrix. The dual slack and for non-Gaussian noise. Nason [5] observes, on a simulation primal Newton directions are then obtained by using i.i.d., Student noise, that the -based cross validation and . We adopt the algorithm of Chen gives a better prediction than the minimax or SURE rules et al. [2] and use the conjugate gradient method to solve the derived for Gaussian noise. With smoothing splines, Xiang and dense system (19). Because multiplication by and Wahba [17] develop a generalized cross validation criterion are typically fast (on the order of or oper- for a differentiable smoothness penalty and for a known noise ations), the conjugate gradient method is attractive. In practice, distribution in the exponential family. The knowledge of a however, the number of conjugate gradient iterations required specific noise distribution is also required by Averkamp and to solve (19) accurately can become very large as ap- Houdré [8], who develop a minimax rule for specific long tail proaches a solution, thus degrading the performance of the IP noise distributions; their selection can only do so much to cope algorithm. with the problem of using the unnatural loss function. Crouse 4) Finding an Initial Point: The IP algorithm requires an ini- et al. [18] propose wavelet-Markov models for the dependence tial point satisfying and , which between scales, and they employ an EM algorithm to estimate ideally would not be too far from the solution. Let the ridge re- their parameters by maximum likelihood. EM is typically gression estimate (obtained by replacing in slow and gets trapped into local maxima, however [this cannot (4) by ) be an initial guess for the coefficients. Let happen with our convex cost function (5)]. and . With and In this paper, we propose a pragmatic approach that does not , let and . Then, the require the specific knowledge of the noise distribution. First, primal variables are positive. an estimate of scale is required. We use the median absolute In addition, let sign , and let deviation of the high-frequency wavelet coefficients of Donoho . Then, the dual variables and Johnstone [4]. This estimate of scale is robust to features of satisfy , , and the dual the underlying signal and to outliers especially if, as suggested slack variables are positive. by Kovac and Silverman [7], the Haar wavelet (which support 5) Convergence: Although there have been many conver- has length two) is used. For the cutpoint in (6), we follow gence studies of IP algorithms, the algorithms that work well in Huber [10] and choose . The default in software pack- practice, including the one described above, often have no guar- ages is often based on the following statistical con- antee of convergence. Specifically, convergence requires the ex- sideration: Suppose you observe i.i.d. data with mean . Then, istence of positive constants , , such that the asymptotic relative efficiency of the Huber estimate of and , for , at all iterations. is asymptotically 95% efficient with respect to the sample av- We can enforce convergence by updating in a more conserva- erage when the noise is Gaussian (G). We can read this value tive manner, but this would slow down its convergence in prac- on the continuous curve (G) of Fig. 2. The relative efficiency is tice. (See, e.g., Kojima et al. [16] for discussions of these issues also plotted as a function of for the distributions used in the in linear programming problems.) simulation. Fig. 2 gives a guideline for the selection of in our A stopping rule for the IP algorithm is when all of the fol- nonparametric regression problem: We see that a value of be- lowing conditions are satisfied for a small : tween one and three gives overall a good efficiency. Based on the simulation of Section IV, we recommend using a cutpoint Primal feasibility: of at least . Finally, for the smoothing parameter , we Dual feasibility: use the minimax developed by Donoho and Johnstone [1] since the residuals within do not depart dramatically from Duality gap: (20) Gaussian residuals. III. SELECTION OF AND IV. SIMULATIONS The selection of the smoothing parameter and the cutpoint A. Computational Efficiency is a difficult problem. Two different situations can be distin- Empirically, Sardy et al. [12] found the BCR algorithm to guished. be more efficient than the IP algorithm at solving basis pursuit.
  • 5. 1150 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 6, JUNE 2001 Fig. 2. Asymptotic relative efficiency as a function of c of the Huber estimate with respect to the sample average of an i.i.d. sample generated from the distributions used in the simulation: Gaussian (G), Gaussian mixture (C), and Student t with three degrees of freedom (T). Fig. 4. Robust estimation of heavisine. (Top) True signal. (Bottom left) Noisy signal. (Bottom right) Robust estimate to compare with Fig. 1. Fig. 3. Decrease in the robust basis pursuit objective function as a function of the CPU time (in seconds) for the BCR algorithm and the IP algorithm. Fig. 5. Four signals used in the Monte Carlo experiment. We observe the same advantage for solving its robust version (5). The reason is that the matrix now used only has to serves the two discontinuities and downweights the influence of be augmented by the columns of the identity matrix. Fig. 3 the outliers. illustrates, on four contaminated signals, the superiority of the We perform a Monte Carlo experiment to evaluate the relative BCR algorithm that is up to 10 times faster than the IP algorithm performance of the nonrobust and robust wavelet-based estima- in achieving the desired precision [ in (20)]. tors with three noise scenarios: • (G) standard Gaussian noise; B. Predictive Performance • (C) a mixture of 90% Gaussian(0, 1) and 10% Gaussian(0, To illustrate graphically on a 1-D signal the advantage of 16) at random locations; using a robust procedure, Fig. 4 shows the robust estimation of • (T) Student noise with three degrees of freedom. heavisine for the same contaminated data as in Fig. 1. The ro- We use the four test functions plotted in Fig. 5 and defined in bust procedure gives a better reconstruction; in particular, it pre- Donoho and Johnstone [1, Tab. 1]: blocks, bumps, heavisine, and
  • 6. SARDY et al.: ROBUST WAVELET DENOISING 1151 TABLE I 1 RELATIVE PREDICTIVE PERFORMANCE OF BASIS PURSUIT (c = ) AND ROBUST BASIS PURSUIT (c = 2:0; c = 1:345). ESTIMATED MEAN SQUARED ERROR 2 ( 100) WITH A STANDARD ERROR OF ROUGHLY 3%. (G: GAUSSIAN; C: CONTAMINATED; T: STUDENT t Doppler. We normalize them such that their “standard devia- tion” is equal to 7 where (21) We choose two sample sizes of and , and the “s8” wavelet packet dictionary with all but four levels. The minimax smoothing parameters are 2.232 and 2.594. Following the discussion of Section III, the smoothing parameter is set to and the cutpoint of the Huber loss function to with and . For each combination of noise (G, C, T) of underlying func- tion (blocks, bumps, heavisine, Doppler), of sample size ( , ), and of procedure (nonrobust, robust), we es- timate the MSE by averaging (40)(1024)/ model errors (i.e., 40 for and 10 for to get the same number of points is generated for the two sample sizes). Table I reports the estimated MSEs of the competing estimators for the 24 sce- narios. In light of Table I, a cutpoint of at least is advisable for robust basis pursuit; the standard value of derived in the parametric context from asymptotic considerations is not 2 Fig. 6. Top: (Left) Noisy infrared sensor data and (Right) 3 3 robust median large enough. With a cutpoint of , the gain in efficiency can filter denoised image. Bottom: (Left) Nonrobust and (Right) robust wavelet be dramatic for non-Gaussian noise using robust basis pursuit. based denoised images. Its counterperformance on the bumps signal is due to the na- ture of the signal whose features are difficult to distinguish with the 2-D nondecimated wavelet transform with the “s8” wavelet. noise in the upper tail when the sampling is light ( ); In this application, we know the underlying image we want to with an heavier sampling ( ), robust basis pursuit recover; therefore, we tried several values of and (which was again beats the nonrobust estimator for non-Gaussian noise. feasible in a finite time thanks to the BCR algorithm) and chose for the best visually appealing reconstruction of the aircraft. Using this parameter setting , the V. APPLICATIONS robust procedure can be used to clean future images. The first data is an image taken by a long-wavelength infrared The second data are radar glint observations and consist of sensor. Just visible above the center of the image is an A-37 angles of a target in degrees. The signal contains a trainer aircraft flying above the Sierra Nevada at some distance number of glint spikes, causing the apparent signal to behave from the sensor platform. The 128 by 128 original image plotted erratically. From physical considerations, a good model for the in the top left of Fig. 6 clearly shows some “outlier pixels.” A true signal is a low-frequency oscillation about . The esti- standard median filter (with a window) gets rid of the bad mated standard deviation is . To get a nice “noise- pixels but does not preserve the aircraft well (top right). The two free” visual display, we choose the universal threshold bottom plots of Fig. 6 show the denoised image using (left) basis , and for the robust version, we choose pursuit and (right) robust basis pursuit. The robust denoising . Fig. 7 shows the (top) original signal and (left: procedure has the definite advantage of cleaning the image of nonrobust; right: robust) the denoised estimates at the bottom. the bad pixels while preserving the outline of the airplane. To The robust estimate is a low-frequency oscillation, as expected, clean the image with the two wavelet-based techniques, we used whereas the nonrobust estimate remains jagged. Note that the
  • 7. 1152 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 6, JUNE 2001 [2] S. S. Chen, D. L. Donoho, and M. A. Saunders, “Atomic decomposition by basis pursuit,” SIAM J. Sci. Comput., vol. 20, no. 1, pp. 33–61, 1999. [3] S. Sardy, “Minimax threshold for denoising complex signals with waveshrink,” IEEE Trans. Signal Processing, vol. 48, pp. 1023–1028, Apr. 2000. [4] D. L. Donoho and I. M. Johnstone, “Adapting to unknown smoothness via wavelet shrinkage,” J. Amer. Statist. Assoc., vol. 90, pp. 1200–1224, Dec. 1995. [5] G. P. Nason, “Wavelet function estimation using cross-validation,” in Wavelets and Statistics, A. Antoniadis and G. Oppenheim, Eds. New York: Springer-Verlag, 1995, pp. 261–280. [6] A. G. Bruce, I. M. Donoho, and R. D. Martin, “Denoising and robust nonlinear wavelet analysis,” Wavelet Applicat., vol. 2242, Apr. 1994. [7] A. Kovac and B. W. Silverman, “Extending the scope of wavelet regres- sion methods by coefficient-dependent thresholding,” J. Amer. Statist. Assoc., vol. 95, pp. 172–183, 2000. [8] R. Averkamp and C. Houdré, “Wavelet thresholding for non (neces- sarily) gaussian noise: A preliminary report,” Georgia Inst. Technol., Atlanta, Tech. Rep., 1996. [9] H. Krim and I. C. Schick, “Minimax description length for signal de- noising and optimized representation,” IEEE Trans. Inform. Theory, vol. 45, pp. 898–908, May 1999. [10] P. J. Huber, Robust Statistics. New York: Wiley, 1981. [11] R. Rockafellar, Convex Analysis. Princeton, NJ: Princeton Univ. Press, 1970. [12] S. Sardy, A. Bruce, and P. Tseng, “Block coordinate relaxation methods for nonparametric wavelet denoising,” J. Comput. Graph. Statist., vol. 9, no. 2, pp. 361–379, 2000. [13] S. Mann and S. Haykin, “Adaptive "chirplet" transform: And adaptive generalization of the wavelet transform,” Opt. Eng., vol. 31, no. 6, pp. 1243–1256, June 1992. Fig. 7. Top: (Left) Noisy radar glint data and (Right) robust median filter [14] F. G. Meyer and R. R. Coifman, “Biorthogonal brushlet bases for direc- estimate. Bottom: (Left) Nonrobust and (right) robust wavelet-based estimate tional image compression,” in Proc. 12th Int. Conf. Anal. Optimization (with local cosine packet). Syst., 1996. [15] S. Sardy, “Regularization Techniques for Linear Regression with a Large Set of Carriers,” Ph.D. dissertation, Univ. Washington, Seattle, 1998. median filter’s estimate still shows a highly oscillating denoised [16] M. Kojima, N. Megiddo, and S. Mizuno, “A primal-dual exterior point algorithm for linear programming,” Math. Progr., vol. 61, pp. 261–280, signal. 1993. [17] D. Xiang and G. Wahba, “A generalized approximate cross validation VI. CONCLUSION for smoothing splines with nongaussian data,” Statistica Sinica, vol. 6, pp. 675–692, 1996. We have proposed a robust version of basis pursuit by re- [18] M. S. Crouse, R. G. Baraniuk, and R. D. Nowak, “Signal estimation using wavelet-markov models,” in Proc. IEEE Int. Conf. Acoust., placing the nonrobust loss function by the so-called Huber Speech, Signal Process., vol. 5, 1997, pp. 3429–3432. loss function. We solved the corresponding nontrivial optimiza- tion problem for a given smoothing parameter and a given cutpoint with an efficient and converging block Sylvain Sardy received the Maîtrise d’Ingénièrie coordinate relaxation method that is empirically faster than an Mathématique degree from the University of Franche-Comté, Besancon, France, in 1989. He interior point competitor. The two techniques are available in then received the M.S. degree in mathematics in the wavelet module of the Splus statistical software; the BCR 1991 and the M.S. degree in statistics in 1992, algorithm can otherwise be easily implemented. We proposed a both from Utah State University, Logan. After doing his military service for France at the Nuclear rule to choose the smoothing parameter and the cutpoint of Engineering Department of the Massachusetts the Huber loss function . We showed on a particular simula- Institute of Technology, Cambridge, he received tion that robust basis pursuit has a good predictive performance the Ph.D. degree in statistics from the University of Washington, Seattle. with both Gaussian and long-tailed symmetric additive noise; in He is now a Postdoctoral Assistant with the Swiss Federal Institute of Tech- particular, we recommend using a cutpoint of at least for nology (EPFL), Lausanne, in the Mathematics Department (DMA). the Huber loss function. As illustrated with two applications, ro- bust basis pursuit has a definite advantage over both a nonrobust wavelet-based estimator and a median filter estimator. Paul Tseng received the B.Sc. degree in engineering mathematics from Queen’s University, Kingston, ON, Canada, in 1981 and the Ph.D. degree in operations research from the Massachusetts Institute of Technology (MIT), Cambridge, in ACKNOWLEDGMENT 1986. From 1986 to 1987, he was a University Research Fellow with the Depart- The authors wish to thank the associate editor and five anony- ment of Management Science, University of British Columbia, Vancouver, BC, mous referees for their careful review and A. Davison for pro- Canada. From 1987 to 1990, he was a research associate with the Laboratory viding help with Fig. 2. for Information and Decision Systems at MIT. He joined the Department of Mathematics, University of Washington, Seattle, in 1990, where he is currently a Professor. His research interests are in optimization and algorithms. REFERENCES [1] D. L. Donoho and I. M. Johnstone, “Ideal spatial adaptation via wavelet shrinkage,” Biometrika, vol. 81, pp. 425–455, 1994. Andrew Bruce, photograph and biography not available at time of publication.