Comparison between Mallat's and the 'à trous' discrete wavelet transform based algorithms for the fusion of multispectral and panchromatic images

Maria González-Audícana, X. Otazu, O. Fors, A. Seco

    Research output: Contribution to journalArticleResearchpeer-review

    177 Citations (Scopus)

    Abstract

    In the last few years, several researchers have proposed different procedures for the fusion of multispectral and panchromatic images based on the wavelet transform, which provide satisfactory high spatial resolution images keeping the spectral properties of the original multispectral data. The discrete approach of the wavelet transform can be performed with different algorithms, Mallat's and the 'à trous' being the most popular ones for image fusion purposes. Each algorithm has its particular mathematical properties and leads to different image decompositions. In this article, both algorithms are compared by the analysis of the spectral and spatial quality of the merged images which were obtained by applying several wavelet based, image fusion methods. All these have been used to merge Ikonos multispectral and panchromatic spatially degraded images. Comparison of the fused images is based on spectral and spatial characteristics and it is performed visually and quantitatively using statistical parameters and quantitative indexes. In spite of its a priori lower theoretical mathematical suitability to extract detail in a multiresolution scheme, the 'à trous' algorithm has worked out better than Mallat's algorithm for image merging purposes. © 2005 Taylor & Francis Ltd.
    Original languageEnglish
    Pages (from-to)595-614
    JournalInternational Journal of Remote Sensing
    Volume26
    Issue number3
    DOIs
    Publication statusPublished - 10 Feb 2005

    Fingerprint

    Dive into the research topics of 'Comparison between Mallat's and the 'à trous' discrete wavelet transform based algorithms for the fusion of multispectral and panchromatic images'. Together they form a unique fingerprint.

    Cite this