SciELO - Scientific Electronic Library Online

 
vol.23 issue4Inversion Technique of Physical Parameters Based on Regularization Extension Depth ConstraintPaleo-tectonic Stress Field and Tectonic Evolution since the Mesozoic in the Eastern Mining Area of Pingdingshan author indexsubject indexarticles search
Home Pagealphabetic serial listing  

Services on Demand

Journal

Article

Indicators

Related links

  • On index processCited by Google
  • Have no similar articlesSimilars in SciELO
  • On index processSimilars in Google

Share


Earth Sciences Research Journal

Print version ISSN 1794-6190

Earth Sci. Res. J. vol.23 no.4 Bogotá Oct./Dec. 2019  Epub Apr 20, 2020

https://doi.org/10.15446/esrj.v23n4.84350 

Originals Articles

Fusion Methods Evaluation and Classification Suitability Study of Wetland Satellite Imagery

Evaluación del método de fusión y estudio de idoneidad de clasificación de imágenes satelitales de humedales

Danyao Zhu1  2  3 

Luhe Wan1  2  * 

Wei Gao1  2 

1Heilongjiang Province Key Laboratory of Geographical Environment Monitoring and Spatial Information Service in Cold Regions, Harbin Normal University, Harbin 150025, China

2College of Geographical Science, Harbin Normal University, Harbin 150025, China

3Department of Geography, Mudanjiang Normal University, Mudanjiang 157011, China


ABSTRACT

Based on HJ-1A HSI data and Landsat-8 OLI data, RS image fusion experiments were carried out using three fusion methods: principal component (PC) transform, Gram Schimdt (GS) transform and nearest neighbor diffusion (NND) algorithm. Four evaluation indexes, namely mean, standard deviation, information entropy and average gradient, were selected to evaluate the fusion results from the aspects of image brightness, clarity and information content. Wetland vegetation was classified by spectral angle mapping (SAM) to find a suitable fusion method for wetland vegetation information extraction. The results show that PC fusion image contains the largest amount of information, GS fusion image has certain advantages in brightness and clarity maintenance, and NND fusion method can retain the spectral characteristics of the image to the maximum extent; among the three fusion methods, PC transform is the most suitable for wetland information extraction. It can retain more spectral information while improving spatial resolution, with classification accuracy of 89.24% and Kappa coefficient of 0.86.

Keywords: HJ-1A HIS; Landsat-8 OLI; Fusion method; Wetland classification

RESUMEN

Con base en los datos HJ-1A HSI y Landsat-8 OLI, los experimentos de fusión de imágenes RS se llevaron a cabo utilizando tres métodos de fusión: transformación de componente principal (PC), transformación de Gram Schimdt (GS) y algoritmo de difusión vecina más cercana (NND). Se seleccionaron cuatro índices de evaluación, desviación estándar, entropía de información y gradiente promedio, para evaluar los resultados de fusión de los aspectos de brillo de imagen, claridad y contenido de información. La vegetación de humedales se clasificó por mapeo de ángulo espectral (SAM) para encontrar un método de fusión adecuado para la extracción de información de vegetación de humedales. Los resultados muestran que la imagen de fusión de PC contiene la mayor cantidad de información, la imagen de fusión GS tiene ciertas ventajas en el mantenimiento del brillo y la claridad, y el método de fusión NND puede retener las características espectrales de la imagen al máximo. Entre los tres métodos de fusión, la transformación de PC es la más adecuada para la extracción de información de humedales. Puede retener más información espectral al tiempo que mejora la resolución espacial, con una precisión de clasificación del 89,24% y un coeficiente Kappa de 0,86.

Palabras clave: HJ-1A HIS; Landsat-8 OLI; Método de fusión; Clasificación de humedales

Introduction

With the development of remote sensing technology, a large number of high-resolution satellites have appeared, such as IRS-1C/1D, HJ-1A/B, IKONOS, Quickbird, Orbview, GF-1/2. These satellites with different band ranges, spatial resolution and spectral resolution have led to an increasing number of remote sensing data of different types in the same region (Wang, Li, & Li, 2001) . How to integrate these remote sensing data and obtain the most applicable remote sensing image has become a difficulty in remote sensing technology. Image fusion technology, by integrating the spatial and spectral advantages of remote sensing images, has continuously expanded its application field, and researchers have introduced the fusion image concept into wetland vegetation classification. Wu, He and Wang (2015) applied the fusion image of CBERS-3 to the classification of wetlands by using four fusion methods, namely, GIHS transform, wavelet transform, Contourlet transform, nsct-gihs transform, and got better results. Dong, Lin, Sun, Qiu & Zhang (2013) studied the applicability of wetland classification by using Landsat-5 and SPOT-5 based on Brovey transform, gram-schmidt transform and principal component (PC) transform. The results show that the fusion of 543 bands of landsat-5 and the third principal component of SPOT-5 are most suitable for the extraction of wetland information. Wei, Li, Tan, & Xun (2011) used the fusion image of NDVI (normalized difference vegetation index) extracted -36°CHRIS high spectrum image and 0°CHRIS image, adopted the method of spectral Angle mapping (SAM) to extract wetland vegetation type information, the experimental results show that the accuracy of vegetation classification in hyperspectral images is improved. Zhu (2012) used the fusion image of hyperspectral image of HJ-1A and CCD image, and extracted the typical vegetation distribution in Zhalong region by using spectral Angle mapping, and obtained higher classification results.

The above mentioned image fusion methods are mainly focuses on the neighborhood processing of digital images. The focus of the research is how to improve the spatial resolution and spectral resolution of images through the improvement of algorithm. Most scholars who do digital image fusion research tend to study in the background of computer science, but there are not many studies from the perspective of geography, especially the application of hyperspectral fusion image to wetland vegetation classification, so this paper conducts the research from this perspective. In this study, the Wuyiling national nature reserve in Xiaoxing'an Mountains in northeast China was taken as the research area, and the HJ-1A remote sensing image in July 2015 and Landsat-8 remote sensing image in August 2015 were taken as the data sources. By analyzing the spatial resolution, spectral resolution, brightness, clarity, information content and the spectral features, the best fusion method applicable to the extraction of wetland information was found. The fuson image is classified by spectral Angle mapping method, and the accuracy of the fusion method is verified to compare the classification accuracy of typical wetland vegetation.

Methods

Study site

The images used in this study is located in Wuyiling national nature reserve, Geographic coordinates are 129°00'E-129°28'E,48°33'N-48°50'N, east of the Xiaoxing'an mountains peak, northeast Heilongjiang Province. Hyperspectral images can make up for the lack of distinguishing ground object information in wide-band remote sensing images (Yan, 2008). However, due to the low spatial resolution of hj-1a data source (100 m), by fusion with the high-resolution landsat-8 OLI full-color image (15 m), the spatial resolution of the data source was improved, so as to obtain more features of ground objects (Table 1). The auxiliary data include the vector boundary map of the research area and the measured data of the field sampling points.

Table 1 Data sources used in the study area 

Time Satellite designation Sensor type
20150719 HJ-1A HSI
20150815 Landsat-8 OLI

Figure 1 Study site on Xiaoxing'an Mountains in Heilongjiang province 

Data preprocessing

HJ-1A data. was preprocessed with ENVI5.3 software, which HJ-1A /1BTools tool is used to convert the data format, after calculating the signal to noise ratio of each band image, remove the first 20 bands with low signal to noise ratio. Due to the existence of severe tilted stripes and Bad lines in HSI data, the Replace Bad Line function is used to repair Bad lines and the global method is used to remove image stripes. Radiation calibration of HSI secondary data has been carried out, and FLAASH atmospheric correction is still needed, and Landsat8 OLI images of the same period were taken as reference, Select GCPS: Image to Image tool was used for geometric precision correction, finally, the image is Mosaic and cropped (Fig. 2).

Figure 2 Preprocessed HJ-1A HSI image (a. False color composite of band 111, 91, 71) and Landsat-8 OLI image (b. False color composite of band 5, 4, 3) 

Fusion methods

Due to the large number of bands in HJ-1A hyperspectral image, three fusion methods with unlimited band number are adopted in this paper for fusion operation.

(1) Principal component (PC) transform

PC transform is a multidimensional orthogonal linear transformation based on statistical characteristics (Ma, Feng, He, & Zhang, 2007). It concentrates the common information of multi-band images into the first principal component, while the unique spectral information that does not overlap with any other bands is put into other principal components, Then, the panchromatic band data are stretched to have the same mean and variance with the first principal component, and which are used to replace the first principal component for principal component inverse transformation to restore to the original space and generate multi-band fusion images with high spatial resolution.

(2) Gram Schimdt (GS) transform

GS transformation is essentially a phase recovery algorithm (Ma & Wang, 2016). The main steps of the algorithm are as follows: (1) Use the multi-spectral image with low spatial resolution to generate simulated panchromatic image with low spatial resolution; (2) The simulated image (as the first band) is superimposed on the multi-spectral image, and the reconstructed multi-spectral image is carried out GS positive transformation; (3) replace the first component of GS transform with high-resolution panchromatic image, then which was carried out GS inverse transformation on the multi-spectral image to get the final fusion image.

(3) Nearest neighbor diffusion (NND) algorithm

NND algorithm was proposed by Sun, Chen and David (2014). The main theory is that the new spectrum of high-resolution fusion images can be expressed as the weight combination of adjacent hyperpixel spectra of current pixels in low-resolution multispectral images. The fusion method based on nearest neighbor spreading can maintain accurate spectral information of multi-spectral images identified by spatial features, and is applicable to the application of accurate spectral information other than traditional visual inspection (Ocalan, Erdogan, Tunalioglu, & Durdag, 2016; Gao, 2015).

Evaluation methods

In this paper, four evaluation indexes including mean, standard deviation, information entropy and average gradient were selected to objectively evaluate the fusion image from the perspectives of image brightness, clarity and information content (Fusun, Saygin, Mustafa, & Filiz, 2017).

Mean, which is the gray mean value of pixels in remote sensing images, which can be expressed as:

Where: M, N are the width and height of the image respectively; F(i,j) is the DN value function of all pixels in the fusion image; i and j represent the row and column Numbers of raster cells.

Standard deviation, which reflects the dispersion of DN values of all pixels and their DN means, can be used to evaluate the brightness of images. The higher the value, the more information the image contains. The specific formula is:

Where: v is mean, M, N are the number of rows and columns of the image; F(i,j) is the DN value function of all pixels in the fusion image; i and j represent the row and column Numbers of raster cells.

Average gradient, with the ability to sensitively reflect the image contrast to small details and texture transformation expression. The larger the average gradient value is, the more layers the image has and the clearer the image will be. The expression is:

Where: M, N are the number of rows and columns of the image; ∆I x , ∆I y are the first-order difference of x direction and y direction, respectively; i, j are the line and column number of each pixel in the image of the same band.

Information entropy, an evaluation index of spatial detail information based on information quantity, reflects the amount of image information quantity. The larger the value is, the larger the information content of the fusion image be. The expression is:

Where: L represents the overall gray scale series in images; P i . represents the probability of gray value of pixels in images.

Results and analysis

The PCA transform, GS transform and NND algorithm results of HJ-1A HIS image are shown in the figure (Fig. 3.)

Figure 3 Results of different fusion methods (False color composite of band 111, 91, 71) 

Visual interpretation

It can be seen from the comparison between the fusion image and the original image that the fusion image maintains the spectral characteristics more greatly in terms of maintaining spectral information. PC fusion image and GS fusion image have moderate color contrast, and all kinds of ground objects in the fusion image show standard spectral detail features. Visually speaking, there is no significant difference in spectral features between the two fusion images; The color change of NND fusion image is serious, and spectral deformation of all kinds of ground objects has occurred, in which the construction land is green and the spectral deformation is serious. Sphagnum and its surroundings appear purplish, which is quite different from the original image.

It can be seen from the spatial information of the image that the fusion image is clearer and more distinguishable, with the same spatial resolution and clearer texture structure as the landsat-8 panchromatic image. In the NND fusion image, the texture of construction land is clearer, and the edges of ground objects are prominent, which makes it easier to identify. However, serious fuzzy phenomena appear at the boundaries of sphagnum and marsh, the fused image is not clear enough. The color of construction land in PC fusion image and GS fusion image is relatively bright and dazzling, and the vegetation features in GS fusion image are obvious.

Quantitative evaluation

According to the above quantitative evaluation indexes, with the support of ENVI/IDL8.5, The eigenvalues of fusion images obtained by the three fusion algorithms used in this study are extracted, and the calculated results are shown in Table 2.

Table 2 Fusion image evaluation index calculation results 

mean value Standard deviation Average gradient Information entropy
Original hyperspectral image 115.498 77.508 4.392 4.158
PC fusion image 111.143 92.902 12.358 4.652
GS fusion image 117.078 81.412 14.542 4.647
NND fusion image 117.971 77.215 6.142 4.515

In terms of the brightness information of the image, the difference between the mean value of the fused image and the original image is relatively small. Descending Means of fusion image: NND fusion image > GS fusion image > PC fusion image. The results show that the brightness of GS fusion image is the closest to the original image, indicating that GS fusion image has certain advantages in maintaining brightness information.

From the aspect of image sharpness, average gradient can be used to evaluate the effect of improving sharpness in objective evaluation indexes. It can be seen from the table that the average gradient of GS fusion image is the largest, followed by PC fusion image, indicating that GS fusion image and PC fusion image have advantages in terms of sharpness.

In terms of the information content of the image, the standard deviation and information entropy of the fused image can evaluate the improvement of the information content. The standard deviation and information entropy of PC fusion image are better than the other two fusion images. The information content of PC fusion image is the largest, followed by GS fusion image and NND fusion image.

Spectral variation analysis

In hyperspectral images, the radiation information of each pixel in each band can be extracted into a continuous spectral curve, reflect the spectral characteristics of ground objects, and realize the fine classification of ground objects, which is the most significant advantage of hyperspectral images (Moon, Choi, Lee, & An, 2017; Adam, Mutanga, Rugege, 2010). If the spectral curve of fusion image is distorted, the spectral information also changes. Therefore, to evaluate which fusion method can retain the spectral information of the original image to the greatest extent, it is necessary to analyze the spectral curve variation degree of fusion images. From the fusion images and the original images, typical features of 9 ground objects samples, which concluded construction land, lakes, farmland, Sphagnum, Carex tato, Salix brachypoda, White birch, Pinus koraiensis and Spruce, were selected to calculate the average spectral curve (Fig. 4.).

Figure 4 Comparison of spectral curves between fusion images and row image 

In the spectrum curve of construction land, the spectrum curve of PC, GS and NND fusion image is similar. Before 700 nm, the NND fusion image is lower than the original image spectrum curve, and after 720 nm, the spectrum curve of the three kinds of fusion image is higher than the original image. In comparison, the spectral fidelity of NND fusion image is better. In lake spectral curve, before 700nm, spectrum curves of three kinds of fusion image are almost the same, after 700 nm, spectrum curves of three kinds of fusion image are higher than the original image. It indicates that the spectral information of the water body of the three fusion images is seriously distorted, in which the change of the original spectral information by GS fusion is relatively small, while the change of the original spectral information by NND fusion is relatively large; The Sphagnum spectral curves of the three fusion images almost coincide with that of the original images, indicating that the fusion images have little influence on the Sphagnum spectra; The spectral curve of the Carex tato before and after fusion had the same trend before 700 nm, indicating that spectral fidelity was better. after 700 nm, The spectrum curve of NND fusion image is lower than that of the original image, while the spectrum curve of PC fusion image and GS fusion image is higher than that of the original image, and keep the same trend; The spectral curves of farmland before and after fusion, after 700 nm, the spectral curves of three fusion images and the original image presented significant differences and serious spectral deformation. The reflectance of the spectral curves of the three fusion images was much higher than that of the original image and show the same trend; The spectral curves of white birch, pinus koraiensis and spruce before and after fusion, the spectral curves of the three fusion images and the original images almost overlapped before 750 nm, indicating that the fusion images retained good spectral characteristics. After 750 nm, the original spectral curve of white birch was higher than that of the three fusion images, while the original spectral curve of pinus koraiensis and spruce was lower than that of the three fusion images, indicating that there was a certain degree of spectral deformation in the fused images.

Classification results and accuracy evaluation

In this study, the Spectral Angle Mapper (SAM) classification method with high accuracy was selected to extract the information of wetland types in Wuyiling National Nature Reserve. SAM is a hyperspectral classification algorithm designed based on the spectral shape characteristics of hyperspectral data. This algorithm has a simple design idea and does not require data compression of hyperspectral remote sensing data. It is a relatively mature classification algorithm in hyperspectral image classification. SAM classification processing is applied to fusion images of different methods (Fig. 5). Meanwhile, the accuracy of classification results is evaluated by using field sampling data in 2015 (Tables 3-5).

Figure 5 Spectral angle mapping classification results 

Table 3 PC Fusion image classification accuracy confusion matrix 

Table 4 GS Fusion image classification accuracy confusion matrix 

Table 5 NND Fusion image classification accuracy confusion matrix 

According to the classification results, SAM was used to classify the fusion image. The accuracy of PC fusion image classification was higher and that of GS fusion image classification was lower. Among them, the mapping accuracy of Pinus koraiensis is 79.23% to 83.56%, and that of spruce is 94.08% to 99.15%. There are often misclassification between Pinus koraiensis and spruce. In the commission errors, the values of spruce and white birch are relatively high. Among the three fusion images, the commission errors of spruce and white birch in PC fusion images are relatively low. In the Omission Errors, the values of Pinus koraiensis is larger, and the classification results of other ground objects are better because of the large difference in spatial details.

Conclusion and Discussion

In view of the low spatial resolution of hyperspectral remote sensing image of HJ-1A satellite, landsat-8 panchromatic image was applied to carry out fusion experiments on three fusion methods of maintaining image's band number, namely, PC transform, GS transform and NND algorithm, and the fusion results was evaluated. The main conclusions are as follows:

(1) Compared with the original image, the fusion image is clearer and more distinguishable. The three fusion images maintain the spectral characteristics to a large extent, among which PC fusion image and GS fusion image have moderate color contrast, while NND fusion image has serious color change. The texture of fusion image construction land is obvious, and the edges of construction land are prominent, while the edges of different vegetation types are seriously blurred. The high reflectivity of construction land in PC fusion image and GS fusion image makes it difficult to distinguish.GS image fusion significantly enhanced the vegetation characteristics. In conclusion, GS fusion image and PC fusion image have advantages in sharpness. PC fusion image contains the largest amount of information, followed by GS fusion image. GS fusion image has a better performance in brightness information.

(2) In the spectral variation analysis, except for the serious spectral information distortion of the lake, the spectral curves of other features are similar in the PC fusion image and the GS fusion image; The variation degree of spectrum curve in NND fusion result is slightly less than that of PC and GS fusion results; In a variety of typical ground objects, the spectral information of lake and farmland is seriously distorted in the near infrared band, and the spectral information of salix brachypoda, spruce and sphagnum is best preserved.

(3) The SAM was applied to extract typical vegetation types from fusion images, and the accuracy of classification results was evaluated. The optimal overall classification accuracy reached 89.24%, and the Kappa coefficient was 0.86. The classification accuracy of PC fusion image is the highest, and contains the largest amount of information; NND fusion image have certain advantages in spectral fidelity and higher classification accuracy. The classification accuracy of GS fusion image is relatively low.

Acknowledgements

Add a third grant fund: This study was supported by the National Natural Science Foundation of China (41671100), a grant from the Education Department of Heilongjiang Province (1352MSYZD003), and a grant from Mudanjiang Normal University (GG2017005).

References

Adam, E., Mutanga, O., Rugege, D. (2010). Multispectral and hyperspectral remote sensing for identification and mapping of wetland vegetation: A review. Wetlands Ecology and Management, 18, 281-296. [ Links ]

Dong, Q. L., Lin, H., Sun, H., Qiu, L., & Zhang, Y. (2013). Application of multi-source remote sensing data fusion method in wetland classification. Journal of Central South University of Forestry and Technology, 33(1), 52-57. [ Links ]

Fusun, B. S., Saygin, A., Mustafa, T. E., & Filiz, S. (2017). Evaluation of image fusion methods using PALSAR, RADARSAT-1 and SPOT images for land use/ land cover classification. Journal of the Indian Society of Remote Sensing, 45(4), 591-601. [ Links ]

Gao, L. (2015). Landsat8 OLI remote sensing image fusion based on the nearest neighbor diffusion method. Proceedings of the 2015 Annual Conference of Jiangsu Surveying and Mapping Geographic Information Society. [ Links ]

Ma, X. X. & Wang, J. L. (2016). The basic research of phase retrieval algorithm. Optik-International Journal for Light & Electron Optics, 127(4), 1561-1566. [ Links ]

Ma, Y. P., Feng, Z. K., He, Y. J., & Zhang, Z. H. (2007). Research on image fusion of quickbird based on ERDAS IMAGINE software. Journal of Beijing Forestry University, (S2), 181-184. [ Links ]

Moon, W., Choi, J., Lee, S., & An, K. (2017). In situ nutrient-spiking bioassays for determining phosphorus and light limitation in a wetland ecosystem. Journal of Environmental Biology, 38(4), 631-639. [ Links ]

Ocalan, T., Erdogan, B., Tunalioglu, N., & Durdag, U. M. (2016). Accuracy investigation of PPP method versus relative positioning using different satellite ephemerides products near/under forest environment. Earth Sciences Research Journal, 20(4), D1-D9. [ Links ]

Sun, W. H., Chen, B., & David, W. (2014). Nearest-neighbor diffusion-based pan- sharpening algorithm for spectral images. Optical Engineering, 53(1), 13-17. [ Links ]

Wang, Z. J., Li, D. R., & Li, Q. Q. (2001). Application of multiple wavelet theory in SPOT and TM image fusion. Journal of Wuhan University (Information Science Edition), 1, 24-28. [ Links ]

Wei, W., Li, Z. Y., Tan, B. X., & Xun, H. S. (2011). Research on remote sensing classification method of longbaotan wetland based on multi-angle hyperspectral CHRIS image. Forest Science Research, 24(2), 159-164. [ Links ]

Wu, R. J., He, X. F., & Wang, J. (2015). Evaluation of fusion quality of multi- spectral and panchromatic images of wetland on resource 3 satellit. High-Tech Communication, 25(2), 157-162. [ Links ]

Yan, Z. (2008). Research on multi-source remote sensing image fusion technology. Wuhan University of Technology. [ Links ]

Zhu, Z. X. (2012). Zhalong wetland vegetation classification based on hyperspectral remote sensing image. Harbin Normal University. [ Links ]

How to cite item Zhu, D., Wan, L., & Gao, W. (2019). Fusion Method Evaluation and Classification Suitability Study of Wetland Satellite Imagery. Earth Sciences Research Journal, 23(4), 339346. DOI: https://doi.org/10.15446/esrj.v23n4.84350

Received: May 20, 2019; Accepted: August 27, 2019

*Corresponding author: wanluhe@163.com

Creative Commons License This is an open-access article distributed under the terms of the Creative Commons Attribution License