extensions to mdl denoising

Upload: jebilee

Post on 14-Apr-2018

219 views

Category:

Documents


0 download

TRANSCRIPT

  • 7/30/2019 Extensions to MDL Denoising

    1/4

    Extensions to MDL denoising

    Janne OjanenHelsinki University of Technology TKK

    Department of Biomedical Engineering and Computational ScienceP.O.Box 9203 (Tekniikantie 14), FI-02015 TKK, Finland

    [email protected]

    Jukka HeikkonenEuropean Commission - Joint Research Centre

    Institute for the Protection and Security of the Citizen (IPSC)

    G04 Maritime Affairs (Fishreg Sector) TP 051, Via Fermi 1, I-21020 Ispra (VA), [email protected]

    Abstract

    The minimum description length principle in wavelet denoising can be extended fromthe standard linear-quadratic setting in several ways. We describe briefly three exten-sions: soft thresholding, histogram modeling and a multicomponent approach. TheMDL hard thresholding approach based on the normalized maximum likelihood uni-versal modeling can be extended to include soft thresholding shrinkage, which can beconsidered to give better results in some applications. In MDL histogram denoising

    approach the assumptions of the parametric density models for the data can be relaxed.The informative and noise components of the data are modeled with equal bin widthhistograms. The method can cope with different noise distributions. In multicompo-nent approach more than one non-noise components are included in the model, becauseit is possible that in addition to the random noise there may be other disturbing signalelements, or that the informative signal is comprised of several different componentswhich we may want to observe, separate or remove. In these cases adding informa-tive components in the model may result result in better performance than in the NMLdenoising approach.

    1 Introduction

    The observed data is thought to be corrupted by additive noise, yn = xn + n, where the noise termn is often assumed to be comprised of i.i.d. Gaussians. Given the orthonormal regression matrixWhe discrete wavelet transform (DWT) of the noisy data is defined as cn = WTyn. The aim of waveletdenoising is to obtain modified coefficients cn representing the informative part in the data. In MDLsetting wavelet denoising is seen as a model selection task. The linear regression model can be rewrittenas a density function f(yn|cn,

    2, ) = 1(22)n/2

    exp 1

    22||yn Wcn||

    , where the structure index

    defines which rows of the regressor matrix are included in the model, or equivalenty, which elementsof cn are non-zero. We may now define the NML density function, compute the well-known maximumlikelihood estimates for parameters and calculate the normalizing factor by a renormalization scheme

  • 7/30/2019 Extensions to MDL Denoising

    2/4

    discussed in Rissanen [1], and the result is a denoising criterion

    k

    2 ln1

    ki c

    2

    i

    +

    n k

    2 ln 1

    n kj /

    c

    2

    j

    +1

    2ln k(n k) + L()

    (1)

    approximating the stochastic complexity. The selection of and the resulting coefficient vector cn areobtained by minimizing the criterion. Furthermore, the criterion is shown to be minimized by the kcoefficients with largest magnitudes. For the code length for the model class a code length functionL() = ln

    nk

    is recommended in [2].

    2 MDL soft thresholding

    An extension to wavelet denoising is to include soft thresholding shrinkage [3]. In essence, softthresholding the observed wavelet coefficients with a threshold parameter gives two subsets indexedby 1 and 2: c1 = (c1(1), . . . , c1(k)) corresponding to the shrinked k informative coefficientsci = sign(ci)(|ci| ), and c2 = (c2(1), . . . , c2(nk)) containing the n k unmodified noise coeffi-cients for which sign(ci)(|ci|) < 0. A useful analogy is to think the process as data transmission overa channel. The sender must transmit enough information over a channel to the receiver so that the receiveris capable of reconstructing the original data from the transmitted signal. In this case we transmit, withas short a code length as possible, k soft thresholded coefficients c1 and n k noise coefficients c2 ,so that when (which also must be transmitted) is known the receiver is able to reconstruct the originaldata.

    The code length for the wavelet coefficients is obtained by encoding the subsets c1 and c2 with separateNML codes LNML(c1 |1) and LNML(c2 |2), respectively. For computing the NML code length for anysequence, see [4]. The code length of the model class, L(1, 2, ), is also required for describing theparameter of the shrinkage function as well as the index sets 1 and 2. The code length may be furtherdivided into L(1, 2, ) = L(1, 2|) + L(), where L(1, 2|) = ln

    nk

    gives the code length for

    choosing the k coefficients into 1 out of a total of n coefficients when is fixed. L() is required todescribe the threshold parameter value. However, L() may be considered to be a constant that can beignored in the final criterion. Finally, the encoding is performed by a two-part encoding where the totalcode length is given by the sum LNML(c1 |1) + LNML(c2 |2) + L(1, 2, ). Applying the Stirlingsapproximation to Gamma functions and ignoring all terms constant with respect to k gives the criterionfor choosing the optimal parameter ,

    min

    k

    2ln

    1

    k

    ki=1

    c21(i)

    + n k2

    ln

    1n k

    nki=1

    c22(i)

    +1

    2ln k(n k) + L()

    . (2)

    The criterion (2) is almost identical to the original MDL denoising criterion (1): the difference is in thefirst term, where in the soft thresholding criterion there are shrinked wavelet coefficient values instead ofthe originals.

  • 7/30/2019 Extensions to MDL Denoising

    3/4

    3 MDL histogram denoising

    The NML approach is restricted to the quadratic-linear case in which noise is assumed to follow Gaussiandistribution. We obtain another denoising criterion by employing histogram models. The main idea is tomodel the wavelet coefficients representing the denoised signal, cn, by an equal bin width histogram ateach resolution level of the wavelet transform, and the coefficients representing the noise, cn by a singleequal bin width histogram. Minimization of the total code length yields the optimal way of dividing thecoefficients into ones representing informative signal and noise. For information on how to compute thestochastic complexity for the data string given the number of bins in the fixed bin width histogram, see[5, 6].

    The key parts of the MDL-histo denoising algorithm discussed in more detail in [7] are summarized asfollows:

    1. Obtain the set of wavelet coefficients cn = cn11 , . . . , cnrr through the r-level wavelet transform.

    2. Recursively on resolution levels i = 1, . . . , r fit a an m-bin histogram Hi to the coefficients cnii

    and select a tentative collection of bins Si, with the number of chosen bins mi = |Si|. Denote

    by ni,(j) the number of points falling in the bin of Hi having index (j). The bins in Si contain kiretained coefficients. The retained and residual coefficients at level i are written as two stringscnii and c

    nii , respectively.

    3. Fit a histogram with M bins to the residual coefficients cn = cn11 , . . . , cnii , c

    ni+1i+1 , . . . , c

    nrr where

    the first i residual strings are obtained by setting the already retained coefficients to zero.

    4. Find the optimal Si by minimizing the criterion

    minSi,M

    log2

    ni

    ni,(1), . . . , ni,(mi), (ni ki)

    +

    log2

    ni + mi + 1

    ni

    + log2

    n

    i1j=1 kj ki

    1, . . . , M

    + log2

    n i1

    j=1

    kj ki + MM

    + ki log2

    Rm

    + ki log2

    M

    Ri

    +

    i1

    j=1

    kj

    log2

    M

    Ri

    (n 1)log2 M + 2 log2 log2 M

    + n log2 Ri + log2 Ri + 2 log2 log2 Ri

    , (3)

    where j is the number of coefficients falling into the jth bin of the M-bin histogram fitted tothe residual string cn, R is the range of wavelet coefficients, Ri are the levelwise ranges of the

    coefficients andi1j=1 kj denotes the number of retained coefficients in the so far optimized sets

    Sj ,j < i. For the first level i = 1 this sum is zero.

    The denoised signal results from the inverse transform of the sequence of retained coefficients c = cn =cn11 , . . . , c

    nrr .

    4 Multicomponent denoising

    It is possible that in addition to the random noise there may be other disturbing signal elements, or that theinformative signal is comprised of several different components which we may want to observe, separate

  • 7/30/2019 Extensions to MDL Denoising

    4/4

    or remove. With more than one informative component in the noisy measured data a multicomponentapproach may result in better performance than the original MDL denoising method [8].

    Roos et al. [9, 10, 2] have shown that a criterion similar to the renormalization result can be obtained

    by a different derivation, details of which can be found in [10, 2]. In short, they define a model for thewavelet coefficients, in which each coefficient is distributed according to a zero-mean Gaussian densitywith variance 2I if it belongs to the set of informative coefficients indexed by , or according to a zero-mean Gaussian density with variance 2N if it represents noise, with the restriction

    2I

    2N. Again, the

    optimal denoising result is given by the minimizing the normalized maximum likelihood code lengthof the data given the model class defined by .

    This approach may be extended by using m Gaussian components and specifying the restriction for theirvariance parameters, 21 ...

    2m. The NML code length for this model can be calculated in a

    manner following the derivation in [10] for the two-component denoising criterion. The derivation turnsout to be straightforward since the normalizing integral factors into m parts, each depending only on thecoefficients determined by the respective index set i. We obtain a criterion

    mi=1

    ki2 ln

    1

    kiji

    c

    2

    j +

    1

    2 ln ki

    + L(1, . . . , m) + m log log2max2min

    + const ,

    (4)

    where const refers to terms constant with respect to the index sets and m, and 2max and 2min are hyper-

    parameters for the maximum and minimum variance, respectively. The last two terms can be ignored ifwe wish to find the optimal m-component result. However, if we want to compare the results for twoapproaches with different number of components, for example m1 = 3 and m2 = 4, we cannot removethe term involving the hyperparameters as it affects the code length.

    References

    [1] J. Rissanen. MDL denoising. IEEE Transactions on Information Theory, 46(7):25372543, 2000.[2] T. Roos. Statistical and Information-Theoretic Methods for Data-Analysis. PhD thesis, University

    of Helsinki, 2007.

    [3] J. Ojanen and J. Heikkonen. MDL and wavelet denoising with soft thresholding. Submitted to 2008Workshop on Information Theoretic Methods in Science and Engineering, 2008.

    [4] J. Rissanen, editor. Information and Complexity in Statistical Modeling. Springer, 2007.

    [5] P. Hall and E.J. Hannan. On stochastic complexity and nonparametric density estimation.Biometrika, 75(4):705714, December 1988.

    [6] J. Rissanen, T.P. Speed, and B. Yu. Density estimation by stochastic complexity. IEEE Transactionson Information Theory, 38(2):315323, March 1992.

    [7] V. Kumar, J. Heikkonen, J. Rissanen, and K. Kaski. Minimum description length denoising with

    histogram models. IEEE Transactions on Signal Processing, 54(8):29222928, August 2006.[8] J. Ojanen, J. Heikkonen, and K. Kaski. Towards the multicomponent MDL denoising. In Festschrift

    in Honor of Jorma Rissanen on the Occasion of his 75th Birthday . Tampere University Press, 2008.

    [9] T. Roos, P. Myllymaki, and H. Tirri. On the behavior of MDL denoising. In Robert G. Cowell andZoubin Ghahramani, editors, AISTATS05, pages 309316. Society for Artificial Intelligence andStatistics, 2005.

    [10] T. Roos, P. Myllymaki, and J. Rissanen. MDL denoising revisited. Submitted for publication, 2006.Preprint available at: http://www.arxiv.org/abs/cs.IT/0609138 .