SciELO - Scientific Electronic Library Online

 
 número85EDITORIALMaintenance of generation units coordinated with annual hydrothermal scheduling using a hybrid technique índice de autoresíndice de assuntospesquisa de artigos
Home Pagelista alfabética de periódicos  

Serviços Personalizados

Journal

Artigo

Indicadores

Links relacionados

  • Em processo de indexaçãoCitado por Google
  • Não possue artigos similaresSimilares em SciELO
  • Em processo de indexaçãoSimilares em Google

Compartilhar


Revista Facultad de Ingeniería Universidad de Antioquia

versão impressa ISSN 0120-6230

Rev.fac.ing.univ. Antioquia  no.85 Medellín out./dez. 2017

https://doi.org/10.17533/udea.redin.n85a02 

Articles

Fusion of World View2 images using Contourlet, Curvelet and Ridgelet transforms for edge enhancement

Implementación de las transformadas Contourlet, Curvelet y Ridgelet aplicadas a una imagen satelital WorldView2 fusionada para el realce de bordes

Giselle Helena Toro-Garay1 

Rubén Javier Medina-Daza1  * 

1Facultad de Ingeniería, Universidad Distrital Francisco José de Caldas. Carrera 7 No. 40B - 53 C. P 110231. Bogotá, Colombia.


ABSTRACT

This article discusses the implementation of three transforms, namely Contourlet, Curvelet and Ridgelet, which are intended for image edge enhancement. These transforms were applied to fused Worldview 2 satellite images. The fusion was performed over the WorldView 2 satellite images applying various types of wavelet transforms, such as Daubechies, Bior, rbior, Coiflet and Symlet 5, with different levels of decomposition. The best results were obtained for the case of Symlet 5, level 5. Wavelet fused images and the generated images (using Contourlet, Curvelet and transformed Ridgelet) were evaluated and analyzed quantitatively. Quantitative methods in the present analysis include ERGAS, RASE, Universal Quality (Qu) and correlation coefficient (CC). Image merging and the implementation of transforms were performed with MatLab®, which supplies the following tools: Wavelet toolbox, Image processing toolbox, Contourlet toolbox, and Curvelet and Ridgelet source code. The results show that the Curvelet and Ridgelet transforms yield better results in terms of edge enhancement for both the merged image and the original image.

Keywords: Contourlet; Curvelet; Ridgelet; Wavelet; image fusion

RESUMEN

Este artículo se discute la aplicación de tres transformadas, a saber, Contourlet, Curvelet y Ridgelet, que están destinadas a la mejora de bordes en una imagen, las cuales se aplicaron a las imágenes de satélite Worldview 2 fusionadas. La fusión se realizó sobre las imágenes de satélite WorldView 2 mediante la aplicación de varias transformada wavelet, entre ellas, Daubechies, Bior, rbior, Coiflet y Symlet 5 con diferentes niveles de descomposición. Los mejores resultados se obtuvieron con la transformada Symlet 5 en el nivel de descomposición cinco. Las imágenes fusionadas con la transformada wavelet (symlet 5) y las imágenes generadas (usando las transformadas Contourlet, Curvelet y Ridgelet) se evaluaron y analizaron cuantitativamente. Los métodos cuantitativos en el presente análisis incluyen ERGAS, RASE, Calidad Universal (Qu) y coeficiente de correlación (CC). La fusión de imágenes y la implementación de las transformadas se realizaron con el software MatLab®, que proporciona los siguientes Toolbox: Toolbox de Wavelet, Toolbox de procesamiento de imágenes, Toolbox Contourlet y código fuente para Curvelet y Ridgelet. Las transformadas Curvelet y Ridgelet producen mejores resultados en términos de mejora de bordes tanto para la imagen fusionada como para la imagen original.

Palabras clave: Contourlet; Curvelet; Ridgelet; Wavelet; fusión de imágenes

1. Introduction

Digital image processing has several purposes, including data treatment aimed at image enhancement [1]. Images taken from satellite sensors also have broad applications, such as the identification and classification of changes in land use [2]. Although the decision to acquire images shall depend on the needs of each study, in some cases it becomes necessary to obtain an image with higher resolution [2].

The Worldview 2 sensor is an example of this situation. The sensor has a spatial resolution of 0.5m for panchromatic images, while its equivalent resolution in multispectral images is of 2m [3]. The two images provide enough information, which permits generating a new image with high spectral and spatial resolution [2]. In this context, image fusion is concerned with the acquisition and combination of the information provided from different sensors (or from the same sensor used in different measurement contexts). The analysis in the present work is based on multi-resolution fusion [4].

Image fusion is a common approach that achieves high spectral and spatial resolution data requirements in a single image. This type of fusion allows integrating multi-spectral and panchromatic images from different spatial resolutions and / or different remote sensing satellites. Merged images provide detailed information about urban and rural environments, which is useful for applications such as urban planning and management. Additionally, image fusion is widely used in the development of projects in various fields such as agriculture, hydrology, the environment, and natural-disaster monitoring services, among others.

Recent fusion techniques include Wavelet-based schemes, which use four- frequency decomposition. In these schemes, three out of four frequencies are high and carry the details of the original image; the remaining frequency is low and carries the original image approximation [2-5]. Also, other techniques have been implemented for resolution enhancement, especially for the enhancement of contours. Such is the case of the Countourlet transforms, which include elongated supports that capture the contours more smoothly [6]. Applications of Countourlet have been reported in [7, 8]. Likewise, Curvelet transforms allow image reconstruction with a lower number of coefficients when compared to the Wavelet [9] and Ridgelet transforms. In fact, Ridgelet constitutes one of the steps in the implementation of Curvelet and provides filtering to produce a fine-scale sub-band [10].

This paper presents an implementation of Contourlet, Curvelet and Ridgelet transforms aimed at defining the contours of the objects that appear in a WorldView-2 image. The image is fused using a Wavelet symlet 5, which yielded the best results when compared with other 25 predefined (Matlab) Wavelets. Performance assessment was based on indices like ERGAS, Universal Quality (Qu) and the coefficient of correlation (CC). The resulting image can be used with other techniques such as mathematical morphology-based object extraction and segmentation, among others. The present study provides tools for the analysis of change detection, environmental studies and multi-temporal studies. The best result was found with Curvelet-based image fusion, in which the Ridgelet transform is used as an intermediate step for the construction of the new multispectral (NRBG) image.

2. Previous concepts

2.1. Wavelet Transform

The Wavelet transform consists in applying signal decomposition by means of certain wavelet functions, which can be obtained from a more general wavelet function (mother function). Decomposition provides a series of coefficients, which should be interpreted and subsequently treated [11]. Formally, according to [12], the Wavelet transform is multi-scale decomposition developed for the analysis of n-dimensional functions, or signals. This transform has proved to be an optimal representation of the original signal as a set of piece-wise smooth signals with a finite number of discontinuities.

2.2 Wavelet Symlet 5

Wavelets can be classified in three groups according to their symmetry, namely symmetrical, almost symmetrical and asymmetrical [13]. The Wavelet Symlet 5 belongs to a family of Wavelets derived from the name “symmetrical Wavelets”. These Wavelets were created with the fundamental purpose of obtaining the greatest possible symmetry from the applicable functions [14].

These Wavelets are not perfectly symmetric, but they were designed to obtain a relatively compact support as well as the highest number of null moments [14].

As this type of Wavelet is not completely symmetric, it has a minimum support in the interval [-p+1, p], with p null moments; yet, it has symmetric filters with respect to the center of its support, where the wavelet concentrates the energy of the image [14]. Additionally, the minimum phase roots (which in other cases are chosen so that the energy concentrations lie near the support point) are chosen from a set having the nearest symmetry but with a complex linear phase [15].

2.3. Contourlet Transform

The main purpose of the Contourlet is to perform a disperse expansion of piecewise smooth images that lack completely smooth contours with a high degree of anisotropy and directionality [6]. The Contourlet offers directionality and contour-capture characteristics; it also has some Wavelet characteristics such as multiscale, time, frequency and localization features [16]. The biggest advantage of this transform is the possibility to improve the contours or provide a smooth representation of such contours using less coefficients than those required by Wavelets. The base functions are oriented in powers of two, according to the number of directions, and allow flexible aspect ratios [17]. The Contourlet transform is twofold, it performs multiscale decomposition and multiresolution decomposition. The Laplacian pyramid is applied from the Gaussian pyramid, decomposing the original image into half its original scale. Thus, the resulting image can be used to predict the values of the next image by means of a Gaussian kernel function [11]. This process is aimed at having lower resolution and point density in each iteration [18]. After every iteration, the difference between two levels provides the Laplacian pyramid levels.

  • The multidirectional decomposition is completed using directional filter banks (DFB), which are designed to capture high-frequency components. Consequently, low frequencies cannot be analyzed using this method. Because of this, a combination of this analysis with multiresolution schemes is necessary [6].

  • Multiscale decomposition applies the Gaussian pyramid to the pyramidal directional filter banks (orthogonal and subsampled of order two for each dimension). This scheme generates a sequence of hierarchical spaces that are invariant to shifting and scaling. This is possible since the created subspace is defined over a uniform distribution of intervals that characterizes the approximation of the image to the resolution. Additionally, the details to increase the resolution are stored into the Laplacian pyramid images [11].

Figure 1 shows the decomposition process performed over the input image; first multiscale-wise and then directional, by means of a fan filter. The latter can also be performed using several filter types such as diamond and square filters [17].

Figure 1 Multiscale and multi-direction decomposition schema for the Countourlet transform [6

2.4. Curvelet and Ridgelet transform

The Curvelet transform has several characteristics, such as frames indexed according to scale parameters, a high level of specificity in the construction pyramid, and multiscale properties.

Additionally, this transform offers an anisotropic, scalar principle that is not found in standard Wavelets [18]. The transform also shares special scaling characteristics Eq. (1), where the length of the support for the elements and the width are related by:

(1)

To conduct the Curvelet transform, several steps are involved, including the Ridgelet analysis. Specifically, the stages are as follows:

Sub-band decomposition

The image is divided in resolution bands Eq. (2), defining a filter bank for the sub-band.

(2)

Object 𝑓 is then filtered in sub-bands Eq. (3).

(3)

Each band contains specific details of different frequencies; P0 contains the low-pass filter.

Δ1, Δ2, … contain the pass-band filters and the details associated to the width [19, 20].

Soft surface partition

  • A collection of smooth windows (𝑥1 𝑥2) is defined. The windows are located around the dyadic grid.

  • In this stage, the window dissection is applied to each of the sub-bands, which were isolated in the previous stage of the algorithm. To yield the spatial windows, the convention of pixels in the range of 1 to 256 is considered for each coordinate. The coordinates are related based on their spatial position (𝑥1, 𝑥2), which corresponds to the (𝑖1, 𝑖2) position, following the (𝑖ℓ−1)/4096 equivalence [19].

Renormalization

In this stage, each resulting grid from the previous stage must be renormalized (i.e. scale-converted into the corresponding units [19]).

Ridgelet Analysis

Each grid is analyzed in the orthonormal Ridgelet system. This means that, if there is an object (say f) that represents the border, once the filtering is performed in each sub band, each process yields a fine-scale sub-band that should contain a map of the borders, thickened by a factor of according to the sub-band filter operator. This provides the sub-band with a collection of flat ridges. Such ridges are almost straight at fine scales (since the border is almost a straight line) and these fragments represent the ideal input data for the Ridgelet transform [19-21].

Once the coefficients have been used over the Curvelet, it is possible to complete a reconstruction of the Ridgelet to obtain the final transformed image. To this end, the following stages are required:

Step 1. Ridgelet synthesis. Each square is rebuilt upon an orthonormal Ridgelet system.

Step 2. Renormalization. Each resulting square, from the previous step, is renormalized to its own corresponding square.

Step 3. Border integration. This is the inverse of window dissection applied to each of the rebuilt windows in the previous step.

Step 4. Sub-band re-composition. Finally, the sub-band filter banks are undone [19].

Figure 2 shows a schematic representation of the processes involved in the Curvelet and Ridgelet transforms, using DBF for the multiscale decomposition and applying the Ridgelet transform.

Figure 2 Implementation schema for the Curvelet and Ridgelet transforms [19

3. Methodology

3.1 Study area

The study area was determined by the presence of spatial objects and the possibility to observe border enhancement. The area belongs to the city of Bogota (Colombia), it lies in the northwestern part of the city near El Dorado international airport. The image was taken by the WorldView-2 satellite. To preserve the dyadic condition for the analysis, a cut of 4096 x 4096 pixels was made. The image was acquired on November 15 th , 2011, and it was re-projected onto system UTM/WSG84.

The image contains 8 spectral bands from which only RGB (431) components were used. The spatial resolution is 2m. Also, the associated panchromatic image has a resolution of 0.5m.

3.2. Fused image

From several existing methods, one of the most common methods for image fusion is the standard method, which is based on the transformation of the RGB multispectral channels into IHS [22]. Intensity (I) indicates the total brightness of the color, Hue (H) corresponds to the average wavelength of a color, and Saturation (S) refers to the purity of a color relative to a tone of gray [22].

The fused image was treated from the original RGB image with color composition (RGB-431). As mentioned above, the image was cropped to create the dyadic structure. Subsequently, a first fusion between the multispectral image and the panchromatic image was conducted according to the following stages, using the standard method mentioned above:

Step 1. The size of the pixel in the multispectral image is adjusted to the size of the panchromatic image, completing a transformation between color spaces, from RGB to HSV.

Step 2. The Value component is separated.

Step 3. The Wavelet transform is applied, creating a new image with NVALUE Value.

Step 4. The NVALUE components are concatenated according to the hue and saturation values obtained in step 1.

Step 5. The HSV space is transformed into RGB, using the HS-NVALUE components.

Step 6. A new RGB image is generated.

Step 7. The predefined Wavelets are implemented for 5 levels of decomposition using Matlab; Wavelets such as Daubechies 4 and 7, Coiflet 2 and 3, Haar and Symlet 5 and 8.

The best spectral and spatial assessment results were found for the Symlet 5 Wavelet with the 5 th level of decomposition. The evaluation was based on the spatial index of ERGAS, the universal quality index (Qu), and the spatial and spectral correlation index.

Figure 3 shows two zones corresponding to the northwestern side of the airport, where several airplanes placed on a surface of concrete can be observed; also, a nearby water body appears. In both images, spatial resolution enhancement is observable with the naked eye after applying the Contourlet, Curvelet and Ridgelet transforms.

Figure 3 Zones of the base image. Symlet 5 with decomposition level 5 

Satellite image fusion using Contourlet, Curvelet and Ridgelet

A standard procedure for the images was conducted according to the steps below. As a result, two images were obtained from the three transforms. This is because the Ridgelet transform is immerse into the Curvelet transform, as mentioned above. The general process was performed as shown in Figure 4.

Figure 4 General schema for the implementation of transforms Contourlet, Curvelet and Ridgelet. Made by the authors 

The steps outlined in Figure 4 are described as follows:

  • RGB color space transformation into HSV is conducted by separating the Value component over the Wavelet Symlet 5 image. This color space was chosen because, in terms of intensity, it provided better results while preserving the spectral and spatial image information.

  • Countourlet, Curvelet and Ridgelet transforms are applied to this component and to the panchromatic image.

  • The first component is extracted from the resulting coefficients of the panchromatic image and it is replaced on the Value.

Figure 5a shows the generated coefficients when implementing the transforms. The component replaced on Value corresponds to the L band from the PAN. In 5b, the first high-precision coefficient is taken from the PAN over the Value.

Figure 5 a) Coefficients generated by the Countourlet transform. b) Coefficients generated in the Curvelet-and-Ridgelet transform [23

  • A reconstruction is performed, obtaining new images from N-VAL Value for both the Countourlet and Ridgelet-Curvelet transforms.

  • Concatenation is conducted for the N-VAL images including the hue a saturation images.

  • The last step involves transformation into the RGB space for the HSN-VAL image, considering the images obtained from the Contourlet and Curvelet-Ridgelet transforms.

4. Results and discussion

The results presented below correspond to the fused Symlet 5 image with decomposition level 5. The spatial resolution was assessed using the Contoulet, Curvelet and Ridgelet transforms with respect to the fused image.

Figure 6, shows the results obtained from the fusions applied to the fused image using the Contourlet, Curvelet and Ridgelet transforms. As can be observed, there is improvement in terms of shape, size, color and borders when looking at the constructions and the airplanes in the images, especially when comparing these objects with the original image.

Figure 6 a) Original RGB (431) sub-image. b) Fused sub-image. c) Fused image obtained from the Countourlet transform. d) Fused image obtained from the Curvelet and Ridgelet transform 

4.1. Assessment using statistical methodsFigure 9

To assess the performance of the proposed image processing using statistical methods, a tool called “Image analysis index” [21], implemented in Matlab, was used. The tool estimates the difference in the spatial resolution of the fused images obtained with the Countourlet, Curvelet and Ridgelet transforms, and that of the fused image obtained with the Symlet 5 Wavelet. To this end, the ERGAS spatial indices, the universal quality index (Qu) and the correlation spatial index are used [24, 25].

ERGAS Index

Spectral ERGAS (for the French words “Erreur Relative Globale Adimensionalle de Synthèse”) is given by the following Eq. (4):

(4)

Where h and l represent the spatial resolution of the panchromatic and multispectral images [22]. N Bands is the number of bands of the fused image; and i PAN is the value of the radiance in the i-th band of the image [26, 27] (see Table 1].

Table 1 ERGAS index for the spatial analysis of the resulting images 

Table 1 shows the results after calculating the spatial ERGAS index. As can be observed, the best values, on average, correspond to the image obtained with the Curvelet and Ridgelet transform. For this image, the red canyon exhibits the best index (i.e. 8.12). Likewise, it was evident that the image obtained using the Contourlet transform showed the lowest values.

For the analysis, Figure 7 shows two planes taken from the images under assessment. Using the spatial ERGAS index, the results indicate that the best image is the one obtained with the Curvelet and Ridgelet transform. This result can be associated to the observable image as a shadow enhancement on the white color of the airplane, also as a better definition of the runway around the objects.

Figure 7 a) Fused sub-image using the Countourlet transform. b) Fused sub-image using the Curvelet-and-Ridgelet transform 

Universal Quality Index (Qu)

Universal Quality provides a model for a quality index Eq. (5), namely the index identifies any distortion as a combination of three factors: correlation loss, illuminance distortion and distortion contrast [26-29]. The best values on this index are obtained when the value is closer to one.

Let and , for an original signal or an original image, respectively. The proposed quality index is defined as:

(5)

The first component is the correlation coefficient between x and y , which measures the degree of linear correlation between x and y , and its dynamic range is [-1,1]. The highest value, 1, is obtained when for , where a and b are constant and . Despite the linear relation between x and y, some relative distortion may exist. Distortion can be evaluated in the two components, which are assessed in the second and third component. The second component, with a range of values in [0,1], provides a measurement of the amount of luminance between x and y. This value is equal to 1 if and only if and can be seen as an estimation of the difference between x and y , therefore the third component is a metric of the similarity between the contrast of the images. These values have also range [0,1], where the highest value of 1 can be achieved if and only if [26].

The results of the Qu index can be seen in Table 2. It is evident that the values associated to the generated images obtained with the Contourlet and Curvelet transforms preserve the spectral resolution and yield little loss of correlation with respect to the original image, where the values of the three RGB bands are close to 1, namely with indexes 0.84, 0.85 and 0.80 (see Table 2].

Table 2 Qu index for the spatial analysis of the generated images using Contourlet, Curvelet and Ridgelet transforms 

Figure 8 presents a body of water with aquatic vegetation inside the study zone. The results for the universal quality index applied to the fused images using Curvelet and Ridgelet are the best. In this image [Figure 8(b)), better defined coverages can be observed since the areas with less aquatic vegetation tend to mix.

Figure 8 a) Sub-image fused using the Contourlet transform b) Sub-image fused using Curvelet and Ridgelet transform 

Correlation Index

This index is estimated between the different fused bands and the original multispectral image Eq. (6). The main purpose is to obtain an index of preservation of the spectral resolution [6, 22].

(6)

Where are mean values from the corresponding images. This index can vary from -1 to 1. Table 3 shows the index per band as computed for both fused images (see Table 3].

Table 3 Correlation index for the spatial analysis of the resulting images 

Figure 9 a)  Fused sub-image using the Contourlet transform. b) Fused image using the Curvelet-and-Ridgelet transform 

Figure 9, shows a mixed picture corresponding to buildings and some vegetation. Although good results were found in both sub-images, the best results were obtained when using the Curvelet and Ridgelet transforms (i.e. an average of 0.87 for the assessed bands).

5. Conclusions

Previous research has shown that the fusion methods based on the Wavelet transform are more suitable for image fusion than conventional methods [5, 22, 24]. The implementation proposed using the Curvelet-Ridgelet mathematical development allowed obtaining the best fused image based on WorldView-2 images.

The results obtained from the qualitative and quantitative analyses show that the image created using the Countourlet transform yields good results in terms of border enhancement. Quantitatively, the value of the Correlation, ERGAS and Qu indices are very similar; however, they are slightly higher for the case of Curvelet-Ridgelet. The proposed methodology allows obtaining fused images that provide users with detailed information about urban and rural environments, which is useful for applications like urban management and planning. The usefulness of the present proposal may be extended to project development on diverse fields such as agriculture, hydrology, environmental and emergency management (associated to natural disasters like flooding, forest fires, etc.), and mapping generation, among others.

All the transforms involved in this study achieved significant results both visually and mathematically with respect to the different spatial objects under analysis. The indices applied indicate that there is preservation of the spatial resolution. However, it is important to consider that the original image used for this procedure corresponds to a fused image processed according to the Symlet 5 Wavelet with decomposition level 5, rated as the best choice among a series of Wavelets studied, which included Daubechies 4 and 7, Coiflet 2 and 3, Haar and Symlet 5 and 8, for 5 levels of decomposition. It can also be observed that different results are obtained depending on the fused Wavelet used in the first process. Finally, the findings indicate that the best outcome in terms of border enhancement on a WorldView-2 image, fused with Wavelet, is the one obtained when applying the Curvelet-and-Ridgelet transform.

6. References

1. J. Medina, C. Pinilla, and L. Joyanes, "Implementing Fast-Haar Wavelet transform on original Ikonos images to perform image fusion: qualitative assessment," Rev. Fac. Ing. Univ. Antioquia, no. 71, pp. 72-84, 2014. [ Links ]

2. J. Cheng, H. Liu, T. Liu, F. Wang, and H. Li, "Remote sensing image fusion via Wavelet transform and sparse representation," ISPRS J. Photogramm. Remote Sens., vol. 104, pp. 158-173, 2015. [ Links ]

3. A. Collin and S. Planes, "What is the value added of 4 bands within the submetric remote sensing of tropical coastscape? QuickBird-2 vs WorldView-2," in Geoscience and Remote Sensing Symposium (IGARSS), IEEE International, Vancouver, Canada, 2011, pp. 2165-2168. [ Links ]

4. G. Simone, A. Farina, F. C. Morabito, S. B. Serpico, and L. Bruzzone, "Image fusion techniques for remote sensing applications," Inf. Fusion, vol. 3, no. 1, pp. 3-15, 2002. [ Links ]

5. S. G. Mallat, "A theory for multiresolution signal decomposition: the Wavelet representation," IEEE Trans. Pattern Anal. Mach. Intell., vol. 11, no. 7, pp. 674-693, 1989. [ Links ]

6. R.J. Medina , L. Joyanes , C. Pinilla , O. Ardila, and F. Pineda, "Evaluación de la fusión de imágenes satelitales usando la Transformada rápida de Wavelet haar y contourlet," in 11th Latin American and Caribbean Conference for Engineering and Technology, Cancun, Mexico, 2013, pp. 1-10. [ Links ]

7. N. A. Al-Azzawi, H. A. Mat, and A. K. Wan, "An efficient medical image fusion method using contourlet transform based on PCM," in IEEE Symposium on Industrial Electronics Applications ISIEA., Kuala Lumpur, Malaysia, 2009, pp. 11-14. [ Links ]

8. J. Adu, J. Gan, Y. Wang, and J. Huang, "Image fusion based on nonsubsampled contourlet transform for infrared and visible light image," Infrared Phys. Technol., vol. 61, pp. 94-100, 2013. [ Links ]

9. M. Choi et al., "The curvelet transform for image fusion," Int. Soc. Photogramm. Remote Sens. ISPRS, vol. 35, pp. 59-64, 2004. [ Links ]

10. T. Chen, J. Zhang, and Y. Zhang, "Remote sensing image fusion based on ridgelet transform," in IEEE International Geoscience and Remote Sensing Symposium, 2005. IGARSS ’05. Proceedings. 2005 IEEE International, Seoul, South Korea, 2005, pp. 1150-1153. [ Links ]

11. C. A. Balbuena, "Método de protección con “marca de agua” para imágenes digitales, utilizando una técnica de representación orientada geometricamente: contourlet," M.S. thesis, Universidad Nacional Autonoma de Mexico, Ciudad de Mexico, Mexico, 2007. [ Links ]

12. J. E. Victorino, "Representación eficiente de una base de datos de imágenes 4D de resonancia magnética cardíaca," M.S. thesis, Universidad Nacional de Colombia, Bogotá, Colombia, 2011. [ Links ]

13. L. P. Pazmiño, "Aplicación de la Transformada Wavelet en dos dimensiones para el análisis y compresión de imágenes," M.S. thesis, ESPE, Sangolquí, Ecuador, 2012. [ Links ]

14. A. Fernández, "Estudio de técnicas basadas en la transformada Wavelet y optimización de sus parámetros para la clasificación por texturas de imágenes digitales," Ph.D. dissertation, Universidad Politecnica de Valencia, Valencia, España, 2007. [ Links ]

15. L. Chun-Lin, "A Tutorial of the Wavelet transform", NTUEE Taiwan, 2010. [ Links ]

16. P. M. Vitousek, "Beyond global warming: ecology and global change," Ecology, vol. 75, no. 7, pp. 1861-1876, 1994. [ Links ]

17. M. N. Do and M. Vetterli, "Contourlets," Stud. Comput. Math., vol. 10, pp. 83-105, 2003. [ Links ]

18. P. J. Burt and E. H. Adelson, "The Laplacian pyramid as a compact image code," IEEE Trans. on Commun., vol. 31, no. 4, pp. 532-540, 1983. [ Links ]

19. D. L. Donoho and M. R. Duncan, "Digital curvelet transform: strategy, implementation, and experiments," in AeroSense, Orlando, USA, 2000, pp. 12-30. [ Links ]

20. W. Junzheng, Y. Weidong, B. Hui, and N. Weiping, "A despeckling algorithm combining curvelet and wavelet transforms of high resolution SAR images," in International Conference on Computer Design and Applications (ICCDA), Qinhuangdao, China, 2010 . , vol. 1, pp. 302-305. [ Links ]

21. E. J. Candes, & D. L. Donoho, (2000). Curvelets: A surprisingly effective nonadaptive representation for objects with edges. Stanford Univ Ca Dept of Statistics. pp. 1-16, https://statweb.stanford.edu/~donoho/Reports/1999/curveletsurprise.pdfLinks ]

22. J. Nuñez et al., "Multiresolution-based image fusion with additive wavelet decomposition," IEEE transactions on geoscience and remote sensing, vol. 37, no. 3, pp. 1204-1211, 1999. [ Links ]

23. J. L. Starck, E. J. Candes , andD. L. Donoho , "The curvelet transform for image denoising," IEEE Transactions on Image Processing, vol. 11, no. 6, pp. 670-684, 2002. [ Links ]

24. A. D. Vaiopoulos, "Developing Matlab scripts for image analysis and quality assessment," Proceedings of the SPIE, vol. 8181, no. 1, pp. 10-12, 2011. [ Links ]

25. R. J. Medina , C. P. Ruiz, andL. Joyanes , "Two-dimensional fast Haar Wavelet transform for satellite-image fusion," J. Appl. Remote Sens., vol. 7, no. 1, pp002E 073698-073698, 2013. [ Links ]

26. J. Medina , L. Joyanes , andC. Pinilla , "Aplicativo Web para la Fusión de Imágenes Satelitales," RISTI - Revista Ibérica de Sistemas e Tecnologias de Informação, no. 11, pp. 17-30, 2013. [ Links ]

27. L. Wald, Data fusion: definitions and architectures: fusion of images of different spatial resolutions, Paris, France: Presses des MINES, 2002. [ Links ]

28. . R. Morales and J. H. S. Azuela , Procesamiento y Análisis Digital de Imágenes, 1st ed. México D.F.: Alfaomega Grupo Editor, S.A. de C.V., 2012 [ Links ]

29. Z. Wang and A. C. Bovik, "A universal image quality index," IEEE Signal Process. Lett., vol. 9, no. 3, pp. 81-84, 2002. [ Links ]

Received: May 04, 2017; Accepted: October 23, 2017

* Corresponding author: Rubén Javier Medina-Daza e-mail: rmedina@udistrital.edu.co

Creative Commons License This is an open-access article distributed under the terms of the Creative Commons Attribution License