Next Article in Journal
CNTR-YOLO: Improved YOLOv5 Based on ConvNext and Transformer for Aircraft Detection in Remote Sensing Images
Previous Article in Journal
Underwater Noise Modeling and Its Application in Noise Classification with Small-Sized Samples
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Spectral Enhancement Method Based on Remote-Sensing Images for High-Speed Railways

MOE Key Laboratory of Optoelectronic Imaging Technology and System, Beijing Institute of Technology, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(12), 2670; https://doi.org/10.3390/electronics12122670
Submission received: 14 March 2023 / Revised: 8 June 2023 / Accepted: 11 June 2023 / Published: 14 June 2023
(This article belongs to the Section Computer Science & Engineering)

Abstract

:
This paper proposes a pansharpening model in order to obtain remote-sensing images with high spatial resolution and high spectral resolution. Based on a generic component substitution (CS) fusion framework, the model utilizes the difference between the high-frequency component of the panchromatic (PAN) image and the high-frequency component of the luminance (L) image to express the missing spatial detail information of the ideal high-resolution multispectral (HRMS) image. A rolling guidance filter (RGF) is used in this framework to achieve the effective extraction of high-frequency information from remote-sensing images while reducing the spectral distortion of subsequent operations. The modulation transfer function (MTF) values of the sensor are also applied to the selection of adaptive weighting coefficients to further improve the spectral fidelity of the fused images. At the same time, the choice of suitable interpolation and gain coefficients improves the generalizability of the model while reducing spectral and spatial distortions. Finally, the use of a guided filter (GF) also greatly improves the quality of the fused image. The experimental results show that the model can effectively improve the spatial resolution for foreign objects at the perimeter of high-speed railways, while also ensuring the color fidelity of foreign objects such as colored steel tiles.

1. Introduction

China’s railway network is dense, and it is situated in a complex environment. With the growing development and popularity of the high-speed railway network, the safety of high-speed railway operations has become a major concern. In recent years, there have been many cases of high-speed railway delays or accidents caused by the intrusion of foreign objects (such as colored steel tiles, billboards, etc.) or natural disasters such as mudslides, so the development of early warning and detection technologies suitable for sensing foreign objects around high-speed railways has become an important research direction. The wide-area surveillance characteristics of remote-sensing satellites, especially with the increase in the number of satellite return cycles and ground resolution, make their application in wide-area monitoring and early warning along railway lines even more advantageous. The images obtained from remote-sensing satellites are usually panchromatic (PAN), multispectral (MS), and hyperspectral (HS) images; among these, the PAN images have higher spatial resolution but lack spectral information, while MS images contain spectral information of multiple bands, but they have lower spatial resolution than PAN images. As the spectral information of foreign bodies around high-speed railways is unique and this is equivalent to its own label, it is the key to the early warning of foreign bodies around high-speed railways. The integrity of the spatial information of foreign objects around the high-speed railway not only affects the comfort of our human eyes to watch the image, but also makes it easier for us to make sample identification labels. Therefore, remote-sensing early warning of foreign objects around high-speed railways requires both sufficient ground resolution and accurate multispectral information. In summary, combining the advantages of PAN and MS images, and obtaining high-spatial-resolution multispectral (HRMS) images using image processing methods, is of great significance for improving the surveillance and early warning capability for foreign bodies on the ground around high-speed railways.

2. Related Works

To date, a large number of pansharpening methods have been developed. According to the underlying logic of the image processing, pansharpening can be broadly classified into three categories [1]: methods based on multi-resolution analysis (MRA), methods based on component substitution (CS), and methods based on variational optimization (VO).
MRA-based methods consider that MS images lose a lot of spatial detail information compared to HRMS images. The basic idea of the MRA-based methods is to add the spatial detail information of PAN images to MS images to achieve sharpening enhancement of MS images [2]. In extracting the detail information of PAN images, the principle of multi-scale decomposition of remote-sensing images is normally used, with methods such as undecimated wavelet transform (UDWT) [3], nonsubsampled contourlet transform (NSCT) [4], generalized Laplacian pyramid (GLP) [5], “à trous“ wavelet transform [6], edge-preserving filters [7], etc. In terms of injecting detailed information, common models include the context-based decision (CBD) model [8], space data mining (SDM) model [9], high-pass modulation (HPM) model [10], enhanced context-based (ECB) model [5], etc. As MRA-based methods are based on MS images, the color fidelity of multispectral fused images is higher while also improving spatial resolution, which is conducive to effective identification of ground markers, but the spatial quality of the fusion often depends on the multi-scale decomposition method and the choice of injection model [11].
CS-based methods start with the idea that PAN images can be considered as high-frequency features in HRMS images [12], so they use PAN images to replace spatial information features in the MS image feature space to achieve sharpening enhancement of multispectral images. Because the sharpened image contains all the high-frequency features of the PAN image, its spatial resolution is high, and the visual effect of the image edge features is good. Due to their simple operation and easy implementation, many of these algorithms are widely used in various types of remote-sensing image processing software, and they can be divided into intensity–hue–saturation (IHS) [13], principal component analysis (PCA) [14], Gram–Schmidt (GS) [15], and so on, depending on the feature space formed by the MS image. Since CS-based methods replace high-frequency features of MS images, resulting in the degree of correlation between the replaced components and PAN images often determining the final sharpening effect, most subsequent research improvements have also focused on how to increase that correlation. Examples of this are the partial replacement adaptive CS (PRACS) algorithm [16], designed by simulating the spectral response through applying a linear regression strategy that applies regression coefficients obtained at low-spatial-resolution scales to weighted parameters; the Brovey algorithm [17], designed by using adaptive gain coefficients; and the band-dependent spatial detail (BDSD) algorithm [18], designed by minimizing the mean square error of the degraded model. In recent years, many scholars have combined the CS-based framework with many new models to make further innovations. For example, in 2016, Lari et al. combined the IHS algorithm in the CS-based model with the “à trous” wavelet (ATW) transform to design the IHS-ATW algorithm [19], which uses the I image generated by the IHS transform and the panchromatic image for ATW fusion to generate a new intensity component, reducing the spectral distortion of the IHS algorithm. In 2017, Zhong et al. combined the MRA-based model with the BDSD algorithm in the CS-based model, which designed the generalized BDSD (GBDSD) algorithm [20] by considering the low-pass weighted components of panchromatic images into the generalized expression of the CS-based model, improving the spectral resolution of the fused images of the BDSD algorithm. In 2019, Liu et al. applied the transform-based gradient transfer model to the synthesis of replacement components of panchromatic images to design pansharpening with the transform-based gradient transferring (PTBGT) algorithm [21], which improves the retention of spectral information in multispectral images. In 2021, Dong et al. used the binary partition tree (BPT) image segmentation algorithm to further estimate the calculation of the replacement component and the injection gain of the panchromatic image in the CS-based model [22], reducing the spectral distortion by improving the spectral similarity between the panchromatic image and its replacement component. However, due to the different spectral response of remote-sensing satellite PAN image detectors and MS image detectors, spectral distortion when using CS-based methods is inevitable, so how to reduce this distortion is one of the main directions of research.
VO-based methods focus on the application of constrained optimization to the solution of the system of disequations between PAN images, MS images, and HRMS images [23]. It formulates the process of panchromatic sharpening as an inverse problem of mapping high-resolution multispectral images to low-resolution multispectral images using reasonable assumptions and a priori knowledge. For the solution of a morbid problem, regularization is usually required to obtain the best results. Many subsequent improvements to this class of methods have been based on improvements to regularization methods. For example, Li and Leung [24] adaptively adjusted the weight of panchromatic images in the sharpening process by constraining the regularization parameters, and Zhang [25] adaptively adjusted the regularization parameters using the gradient descent method. Typically, VO-class-based methods require the input of multiple panchromatic and multispectral images from the same scene. As a result, the fusion results of this method are poor when large spatial and spectral errors are inherent in the input data. Benefiting from the development of sparse representation theory in signal science, its theory has also been applied in the field of remote-sensing image fusion. Unlike the rest of the VO class of methods, the sparse representation approach considers that HRMS images can be obtained by constructing an overcomplete dictionary of high-resolution multispectral images and their sparse coefficients [26]. Most of the subsequent studies have focused on how to construct overcomplete dictionaries. This class of methods often requires a large amount of runtime and runspace, and small improvements in fusion quality require significantly higher running costs. Nowadays, VO-based methods cannot yet be applied to engineering projects.
In order to improve the fusion image effect and enhance the surveillance and early warning capability for foreign objects on the ground around the railway, this paper investigates an edge-preserving adaptive CS algorithm (EPACS). The original CS-based method has been optimized by solving two problems: (1) how to increase the correlation between the PAN image and each channel of the MS image to reduce the spectral distortion of the fused image caused by the global variability of the relative spectral response functions and low correlation, and (2) how to increase the detail injection of the PAN image to improve the spatial resolution of the MS image.

3. Edge-Preserving Adaptive CS Algorithm

Since HRMS images of actual ground scenes are not directly available, there is no direct criterion for evaluating the quality of reconstructed HRMS images. Theoretically, it would be desirable for the HRMS image to have the same rich spectral information as the MS image, for which the upsampled image ( MS ~  image) of the MS image can be considered to replace the low-frequency part of the HRMS image; at the same time, it would be desirable for the HRMS image to have the same spatial detail features as the PAN image, and the difference between the amount of detail in each channel of the HRMS image should be similar to that of the MS image, so the high-frequency part of the HRMS image can be considered to be replaced by the relationship between the high-frequency part of the PAN image  I P A N  and the MS image  I M S .
The HRMS image  I H R M S  is represented as
I H R M S n = I H R M S h i g h n + I H R M S l o w n : = w ( I P A N , I M S H n ) + I M S ˜ n
where “:=” is the defining symbol (indicating that it does not exist by convention in nature);  I H R M S h i g h n  and  I H R M S l o w n  are the high- and low-frequency portions of the nth channel of the HRMS image, respectively;  I M S ~ n  is the nth channel of the  MS ~  image; and  w ( I P A N , I M S H n )  is the high-frequency information extracted between the PAN image and the high-frequency part  I M S H n  of the nth channel of the MS image.
As shown in Figure 1, the CS-based fusion framework is often defined as
w ( I P A N , I M S H i g h n ) = g n ( I P A N I L )
I L = w n I M S ˜ n
where  I L  is a luminance image similar to the PAN image synthesized using information from each band of the  MS ~  image; gn is the injection gain of the nth channel band detail quantity; and wn is the weighting factor for the nth channel of the L image.
The EPACS framework studied in this paper is shown in Figure 2. To solve problem 1 in Section 1, the  I P A N I L  in the traditional CS framework is replaced by the difference image ( I P A N H I L H ) between the high-frequency component  I P A N H  of the PAN image and the high-frequency component  I L H  of the L image (as shown in Figure 2, ①), and the linear regression strategy for solving the L image is optimized (as shown in Figure 2, ②), while an adaptive gain coefficient g is chosen. To solve problem 2, the model details image (D) is adaptively enhanced using a guided filter (GF) (as shown in Figure 2, ③), while the use of a rolling guidance filter (RGF) also enhances the representation of spatial information in PAN images (as shown in Figure 2). The EPACS algorithm is described in detail below.

3.1. Fused Image Detail Information Extraction

Analyzing the physical meaning of Equations (1) and (2) together, it is found that  g ( I P A N I L )  expresses the high-frequency component of the HRMS image, i.e., the detailed features such as contours and edges of the image. The L image can be seen as an approximate PAN image synthesized using an MS image, i.e., a synthetic approximation to the PAN image on the  MS ~  image scale, which ideally has the same spectral response as the PAN image. If the HRMS image is considered to have the same amount of detail as a panchromatic image,  I P A N I L  represents the amount of detail D that is missing from the MS image relative to the HRMS image;  g n ( I P A N I L ) + I M S n  represents how this relatively missing detail is added to each MS image channel.
Since the actual spectral response relationships in each band are different for each sensor panchromatic image sensor and multispectral image sensor, the synthetic L image is only an estimate and is not equivalent to a panchromatic image on the MS image scale. That is,  I P A N I L  may introduce an error which causes instability in the spectral information, or even a lack of spatial information, affecting the quality of the reconstructed image. To reduce this part of the error, the process of solving for detail D is changed to  I P A N H I L H  using the edge-preserving filter. The representation of the spectral and spatial information of the fused image is increased by reducing the interference of detail-independent pixels, i.e., by reducing the error due to the subtraction of low-frequency components between images. Problem 1 in Section 1 is thus transformed into a question of how to obtain  I P A N H  and  I L H .

3.1.1.  I P A N H  Acquisition

For the acquisition of  I P A N H , this paper uses the RGF to process images. Its advantage over other multi-scale decomposition methods is that it can more accurately separate spatial information at large scales, edge information at medium scales, and texture information at fine scales, allowing it to blur images while preserving edge texture features. This feature helps to reduce the ringing and blending effects during the fusion process, thus making the fusion effect more in line with human visual perception.
The filtering process of the RGF consists of two main steps:
  • Acquisition of the guided image using a Gaussian filter.
Using a Gaussian filter to eliminate fine structures in the original image T with a structure size smaller than  l s , the expression for the operation on the central pixel a of the input image is as follows:
G ( a ) = ( b S exp ( | | a b | | 2 2 l s 2 ) T ( b ) ) / H
H = b S exp ( | | a b | | 2 2 l s 2 )
where S is the set of surrounding pixels of the central pixel a; b is an element in S; and the structural scale parameter  l s  can be regarded as the minimum standard deviation of the Gaussian kernel.
2.
Edge recovery using a joint bilateral filter.
The previously obtained filtered image G(k1) is used as the guide image, and joint bilateral filtering is performed for the input image to obtain the filtered image k2. Then, the filtered image k2 is regarded as the new guide image, and the previous step operation is repeated; the output image O(ki+1) after the ith iteration operation can be expressed as
O ( a ) = k ( a ) i + 1 = 1 b S U b S U T ( b )
U = exp ( | | a b | | 2 2 σ s 2 ) ( | | k ( a ) i k ( b ) i | | 2 2 σ r 2 )
where T is the input image; O is the output image; S is the set of surrounding pixels of the central pixel a; b is an element of S σ s  and  σ r  are two hyperparameters that control the structural scale weight and range weight, respectively; and ki+1 is the result of the ith iteration.
In summary, the acquisition of PAN image detail information  I P A N H  can be formulated as
I P A N H = I P A N R G F ( K , I P A N , σ s , σ r )
where K is the number of iterations.

3.1.2.  I L H  Acquisition

The  I L H  is estimated using the linear regression strategy employed by PRACS, which assumes that there is some linear relationship between the HRMS image to be derived and the PAN image, i.e., the PAN image can be synthesized with information from each channel of the HRMS image information.
I P A N = i = 1 N w i I H R M S i + c i
From Equation (2), the image upsampled to the PAN image scale can be interpreted as the low-pass image of the HRMS image, and then the low-pass filtering of both sides of Equation (9) yields
I P L = i = 1 N w i I M S ˜ i + d i
where  I P L  is the low-pass images of PAN images, and c and d are the linear regression parameters.
Equation (10) is solved using the least squares method:
min w 1 , w 2 w n | | I P L i = 1 n w i I M S ˜ i | | 2
The weight parameter wi can be obtained using Equation (11). Up to this point, wi I P L , and  I M S ~ i  are known. Then, solve for di in Equation (10) using the pending coefficient method to find the synthetic approximation L of the PAN image on the  MS ~  image scale.
I L = i = 1 n w i I M S ˜ i + d i
The fusion process does not introduce spectral distortion if the low-pass image of the PAN image happens to be perfectly linear in relation to the channels of the MS image. However, in practice, the spectral response ranges of panchromatic and multispectral imaging from remote-sensing satellites do not coincide, making the L image not representative of a PAN image on the  MS ~  image scale, so  I P A N H I L H  is used instead of  I P A N I L .
As shown in Figure 3, for the GF-2 remote-sensing satellite, the wavelength range of the PAN image is larger than the overall wavelength range of the MS image (red band, green band, blue band, near-infrared band), and when viewed in the frequency domain, it is the PAN image and the MS image that have different frequencies of light. As shown in Figure 4, the MTF curves [27] of each sensor indicate the passable frequency range of the signal. In order to make Equation (10) better satisfy the linear relationship and let both sides of it have the same frequency range, we designed a low-pass filter for PAN images based on the MTF curves of the MS image sensors, i.e., by letting both sides of Equation (10) have the same frequency range, so as to obtain a better spectral weight w to improve the quality of the fused image. Since the MTF values for the different spectral channel cut-off frequencies of the  MS ~  images  I M S ~  vary slightly, but the  I P L  images are single channel, the average MTF output for each spectral channel of the  MS ~  images is used.
I P L = G M T F ( M S ˜ ) ( I P A N )
Here,  G M T F ( M S ~ )  denotes the low-pass filter designed using the average of the deadline frequency of the MTF curves of each channel of  MS ~ . The cutoff normalized Nyquist frequency of the MTF curve can be obtained by the edged edge method [28]. The MTF curve can be fitted using a Gaussian curve.
In summary, we first use the RGF filter to obtain the detailed image  I M S H  of the MS image.
I M S ˜ H i = I M S i R G F ( K , I M S ˜ i , σ s , σ r )       i = 1 , 2 N
Then, using the MTF filter  G M T F ( M S ˜ H )  designed through the upsampled image  I M S ~ H  of the  I M S H , low-pass filtering is performed for the  I P A N H  image to obtain  I P A N H L . Finally,  I L H  is obtained by the linear regression strategy of PRACS.
I L H = i = 1 n w i I M S ˜ H i + d i

3.2. Gain Factor Selection

As described in [13], the injected gain gi should satisfy  ( k g i ) 1 = 1  if the effect of differences in the spectral response functions of different sensors is ignored. However, the relationship between the panchromatic band range and the multispectral image band range is often not fully linear, and in order to increase this nonlinear relationship and to make the algorithm more adaptive for different images and different channels, the gain coefficients g1,g2…gn are defined as
g i = cov ( I M S ˜ H i , I L H ) cov ( I L H , I L H )       i = 1 , 2 N
where cov(A, B) denotes the covariance between the images A and B.
The amount of detail injected into the different spectral channels of a multispectral image is
D 1 i = g i ( I P A N H I L H )

3.3. Enhancement of Fusion Image Detail Information

To solve problem 2 in Section 1, the spatial detail information is enhanced using the structure transfer properties of the GF [29]. The GF is an edge-preserving filter that filters the input image p by the guided image T so that the final output image q is largely similar to the input image p, but the texture part is similar to the guided image T. When the input image of the guided filter and the guided image are the same image, the guided filter will smooth the input image. When the input image and the guided image are different images, the guided filter will extract the structural information from the guided image and merge the structural information into the input image.
The expression for the GF is
q i = j W i j ( T ) p j
where i, j are the indexes of the pixel points in the image, respectively, and  W i j ( T )  is the polynomial associated with the guided image T.
The guided filter theory assumes that the output image q and the guided image T are locally a linear model. In the local window  M k , the linear relationship can be expressed as
q i = a k T i + b k       i M k
where ak, bk are linear model parameters.
The cost function is designed so that the mean square error between the output image q and the input image p is minimized; its cost function in the window  M k  is
E ( a k , b k ) = i M k ( ( a k T i + b k p i ) 2 + ε a k 2 )
where  ε  is the regularization coefficient of the L2 norm.
Let N be the number of pixels in the window  M k . According to the condition that the partial derivative is 0 at the minimum value, Equation (20) for ak, bk for the partial derivative has
E a k = 0 ,       E b k = 0
Solving Equation (21) yields
a k = 1 | w | i w k T i p i μ k p k ¯ σ k 2 + ε
b k = p ¯ k a k μ k
where w is the number of pixels in the window  M k μ k  is the mean value of the guided image T over the window  M k σ k  is the variance of the guided image T over the range of the window  M k ; and  p k ¯  is the mean value of the input image p over the window  M k .
In summary, the process of guided filtering can be expressed as
q = G F ( T , p , r , ε )
where r is the radius of the guided filter window.
The structure-shifting nature of the guided filter allows it to extract features that are present in the guided image but not in the input image, similar to our idea of extracting missing details from MS images. Therefore, the  I P A N H  is used as the guide image, and each channel of the  I M S ~ H  is used as the input image to further extract the amount of detail D2 in order to increase the representation of the spatial information of the image as well as the spectral information, the expression of which is shown below.
D 2 i = I P A N H G F ( I P A N H , I M S ˜ H i , r , ε )       i = 1 , 2 N
It follows that for different spectral channels of the  I M S ˜ H  image, we will have different degrees of information enhancement using guided filtering. This corresponds to the fact that, in practice, the ability to express detail between different channels of an image should vary. Thus, the use of the guided filter not only enhances the representation of image detail but also improves the adaptivity between different channels.

3.4. HRMS Image Acquisition

In summary, the amount of detail we injected into each channel of the HRMS image is
D i = D 1 i + D 2 i  
Finally, the HRMS images are synthesized using a generic CS-based fusion framework as
I H R M S i = D i + I M S ˜ i

4. Experimental Results and Analysis

4.1. Selection of Data and Evaluation Indicators

Two sets of images from different satellites were used to test the validity of the EPACS method in this paper. Figure 5, Figure 6, Figure 7 and Figure 8 show those from the DEIMOS-2 remote-sensing satellite (containing common features such as trees and houses). The PAN image has a spatial resolution of 1 m and a size of 100 × 128; the multispectral image contains four channels of R, G, B, and NIR with a spatial resolution of 4 m and a size of 25 × 32 × 4. Figure 9 was derived from our remote-sensing image dataset of the high-speed railway perimeter (including the railway, nearby colorful steel tiles, and other foreign objects) based on GF-2 satellite remote-sensing images. The PAN image has a spatial resolution of 1 m and a size of 512 × 512; the multispectral image contains four channels of R, G, B, and NIR with a spatial resolution of 4 m and a size of 128 × 128 × 4.
For better visual assessment, only the B, G, and R channels were used to display the images in this paper. In order to objectively assess the imaging quality of remote-sensing images, in addition to the visual assessment of the fusion results, a quantitative assessment of the images obtained by panchromatic sharpening was also required. Since the HRMS reference image could not be obtained directly from the sensor, we used the Wald criterion to make an assessment at the degradation scale (using the fused image of the four-fold downsampled PAN image and the MS image for comparison with the initial MS image).
The evaluation metrics included ERGAS [30] (REF = 0), SAM [31] (REF = 0), RASE [32] (REF = 0), RMSE [32] (REF = 0), and UIQI [33] (REF = 1). Furthermore, in order to be able to quantitatively characterize the spectral and spatial distortions produced during the synthesis process, we assessed them on the original image scale using the QNR criterion. The metrics evaluated included DS (REF = 0),  D λ  (REF = 0), and QNR (REF = 1), which are defined in the article [34].
Because there is no uniform evaluation index for the quality of remote-sensing image fusion, and the spectral response functions of different satellite multispectral image sensors and panchromatic image sensors are not exactly the same, this leads to a certain amount of subjectivity in the choice of hyperparameters. In this paper, three parameter variations of EPGAS, SAM, and QNR were used as the basis for selecting hyperparameters for different sensor images; the EPGAS and SAM represented the integrated image quality and spectral quality, respectively, under the Wald criterion, and the QNR represented the integrated image quality under the QNR criterion. Based on a number of comparative experiments, the following choices were made for hyperparameters: the hyperparameters of this framework in the DEIMOS-2 remote-sensing image fusion process were K = 4,  σ s  = 3,  σ r  = 0.8, r = 3, and  ε  = 0.1; the hyperparameters of this framework in the GF-2 remote-sensing image fusion process were K = 3,  σ s  = 3,  σ r  = 0.8, r = 2, and  ε  = 0.3. The adjustment range for each hyperparameter is as follows: the number of iterations K ranges from two to six, with a modification step of 2; RGF scale weights  σ s  ranges from two to six, with a modification step of 1; RGF range weights  σ r  ranges from 0.2 to 0.8, with a modification step of 0.2; GF window radius r ranges from two to eight, with a modification step of 2; GF regularization factor ε ranges from 0.1 to 0.4, with a modification step of 0.1. In order to demonstrate the effect of variations in the hyperparameters on image quality, some data are provided in Appendix A for easy reference.
Meaning of evaluation indicators:
  • Root Mean Square Error (RMSE):
R M S E = E [ ( I H R M S I M S ˜ ) 2 ]
where E is the mathematical expectation of the image. The lower the RMSE value, the more similar the multispectral data are to each other.
  • Relative Average Spectral Error (RASE):
R A S E = 100 i = 1 N μ i N i = 1 N E [ ( I H R M S i I M S ˜ i ) 2 ]
where  μ i  is the average value of the ith band. The lower the RASE value, the more similar the multispectral data are to each other.
  • Erreur Relative Globale Adimensionnelle de Synthèse (ERGAS):
E R G A S = 100 d h d l 1 N i = 1 N ( R M S E ( i ) μ ( i ) ) 2
R M S E = E [ ( I H R M S I M S ˜ ) 2 ]
where dh/dl is the ratio of the spatial resolution of the PAN and MS images;  μ ( i )  is the mean value of band i of the reference multispectral image; N is the number of bands of the reference MS image; and RMSE(i) is the mean square error of the reference multispectral image and the fused image at band i. ERGAS is a global error metric improved by the RASE. The lower the ERGAS value, the more similar the multispectral data are to each other.
  • Spectral Angle Mapper (SAM):
S A M ( V M S ˜ ( i ) , V H R M S ( i ) ) = arccos ( V M S ˜ ( i ) , V H R M S ( i ) V M S ˜ ( i ) V H R M S ( i ) )
where  V M S ~ ( i )  and  V H R M S ( i )  are the vectors constructed by concatenating the bands of the  MS ~  image at position i and the vectors constructed by concatenating the bands of the HRMS image at position i, respectively. <...> denotes a dot product operation; ||...|| denotes a vector two-norm operation; and SAM is the absolute value of the spectral angle of two vectors and reflects the degree of spectral distortion. If the spectra of  MS ~  and HRMS are the same, then the SAM value is 0.
  • Universal Image Quality Index (UIQI)
The UIQI reflects the correlation, average brightness, and contrast between the two images and takes values in the range [−1, 1], with a value of 1 when the fused image is identical to the reference image.
U I Q I = σ A , B σ A σ B 2 A ¯ B ¯ [ ( A ¯ ) 2 + ( B ¯ ) 2 ] 2 σ A σ B ( σ A 2 + σ B 2 )
Here,  A ¯  and  B ¯  are the pixel means of the reference image and the fused image, respectively;  σ A  and  σ B  are the standard deviations of the reference image and the fused image, respectively; and  σ A , B  is the covariance of the reference image and the fused image.
  • Quality with No Reference (QNR):
This criterion consists of two metrics, spectral distortion and spatial distortion, which are combined to produce a unified quality metric, QNR; the higher the value, the better the result of the fusion.
Q N R = ( 1 D λ ) α ( 1 D s ) β
The values of  α  and  β  are usually set to 1.
D λ = 1 N ( N 1 ) i = 1 , r = 1 N i r N | U I Q I ( I M S ˜ i , I M S ˜ r ) U I Q I ( I H R M S i , I H R M S r ) | p p
D s = 1 N i = 1 N | U I Q I ( I M S ˜ i , I P A N ) U I Q I ( I H R M S i , I P A N ) | q q
Here, N is the number of multispectral image channels; i and r are the number of channels in the image; p and q are image deflation factors with a default value of 1; and  D λ  and DS are the spectral distortion and spatial distortion, respectively, which are calculated using the difference in UIQI values between the original image and the fused image.

4.2. Comparison of Fusion Effects

In order to verify the effectiveness of the improved framework of the ERGAS algorithm used in this paper, algorithms were compared for ①, ②, ③ of Figure 2, and various excellent remote-sensing image fusion algorithms were selected for a cross-sectional comparison of fusion algorithms.

4.2.1. Comparison of the Improvement Effects of Different Fusion Image Methods

(1)
Comparison of the effects of  I P A N H I L H  versus  I P A N I L
To verify that  I P A N H I L H  can increase the fused image spectral information, the IHS algorithm was used as an example. First, the RGF filter was used to obtain the high-frequency images  I M S ~ H  and  I P A N H  of the  MS ~  image and the PAN image, respectively, and then the IHS transform was performed for  I M S ~ H  to obtain the image  I L H , after which the subsequent steps of the IHS algorithm were performed. A comparison of the reconstructed images with the original IHS algorithm is given in Figure 5, and a comparison of the corresponding evaluation parameters is given in Table 1. It can be seen that the improved IHS method can effectively reduce the spectral distortion in the green part of the fused image, and the spectral-distortion-related indexes SAM and  D λ  are greatly reduced; however, the performance in the comprehensive evaluation index ERGAS was not as good as the IHS algorithm, which was due to the fact that some fine spatial structures were not shown (e.g., the segmentation line of the roof in the lower left corner). Firstly, because the L component of the IHS transform decomposition of the multispectral image was different from that of the PAN image, the fine structure of the image was not well expressed by  I P A N H I L H . Secondly, because the aim of any algorithm is a trade-off between spectral resolution and spatial resolution, the original IHS algorithm focused on the spatial information of the image, leaving the color information poorly represented. The improved algorithm does not enhance the spatial information of the image, but it does make a better trade-off between the two.
(2)
Comparison of MTF filter optimization status
To verify the performance of the low-pass filter that was designed based on the MTF of MS images in this framework, for purposes of comparison, a low-pass filter was designed with four-fold downsampling followed by four-fold upsampling of panchromatic images using bilinear interpolation, as is commonly used in the MRA-based method, and the rest of the steps were the same as in the framework of this paper.
A comparison of the reconstructed Images with the bilinear interpolation is given in Figure 6, and a comparison of the corresponding evaluation parameters is given in Table 2. It can be seen that the low-pass filter designed based on the MTF has a somewhat higher green saturation compared to the bilinear interpolation, but the other differences are not significant; the MTF-based filter has greater improvement in both the spectral distortion index SAM and  D λ ; the comprehensive evaluation index ERGAS and the spatial distortion index DS also get a small improvement, which verifies the relevant theory in Section 3.1.2.
(3)
Comparison of image information with and without guided filtering enhancement
To verify that guided filtering enhances the representation of frame image information, the frames in this paper were compared with and without GF, and the remaining steps were identical. Figure 7 gives a comparison of the reconstructed images and Table 3 gives a comparison of the corresponding evaluation parameters. It can be seen that the reconstructed image performed better on tiny features (e.g., on the division line of the roof in the lower left corner) after using the guided filter to enhance the image information. The guided filter reconstructed images had large improvements in the metric RASE, which describes spatial distortion, and in the composite evaluation metric ERGAS; small improvements were also obtained in the metrics SAM and  D λ , which describe spectral distortion, indicating that the use of guided filtering can enhance the representation of detailed information, especially image detail, in the images framed in this paper.

4.2.2. Comparison of the Effects of the Complete Algorithm

To verify the overall effectiveness of the EPACS algorithm used in this paper, the GIHS algorithm, PCA algorithm, GS algorithm, Brovey algorithm, BDSD algorithm PRACS algorithm, IHS-ATW algorithm and PTBGT algorithm in the CS-based algorithm, and the HPF algorithm, MTF-GLP (GLP [5] with MTF [10] with unitary injection model) algorithm, MTF-GLP-CBD (MTF-GLP with CBD) algorithm [8] and MTF-GLP-ECB (MTF-GLP with ECB) [5] algorithm in the MRA-based algorithm were selected to compare the fusion effects.
Figure 9 gives a comparison of the fusion effects of these methods on the GF-2 satellite images. Table 4 gives objective evaluation indicators for each fusion method used on the GF-2 satellite images, where underlined numbers indicate they are sub-optimal, and bolded and underlined numbers indicate they are optimal. In terms of visual results, the algorithms performed well in expressing the edge features of simple foreign objects such as colorful steel tiles around the railway. However, the GIHS, PCA, GS, and Brovey algorithms showed significant color distortion in expressing the color of the tiles, especially for white tiles, which was mainly due to the inconsistent spectral response of the different sensors, leading to the offsetting of the injected radiation during the fusion process. Both the IHS-ATW algorithm and the PTBGT algorithm improve the spectral fidelity of the CS-based algorithms, but the edge detail is significantly under-injected; in addition, the MTF-GLP-CBD algorithm and the MTF-GLP-ECB algorithm showed intermittent local edge features when describing white-colored steel tiles at railway perimeters, indicating that this type of algorithm has different sensitivities for different regions of the image. On the other hand, the EPACS, PRACS, and MTF-GLP algorithms were visually better in this paper. In terms of the evaluation metrics, the EPACS algorithm achieved optimal performance in terms of RMSE, UIQI,  D λ , Ds, QNR, and many other metrics. Compared with the traditional CS algorithm, the EPACS algorithm showed stronger spectral fidelity, and the fused image could more accurately describe the color information of dangerous features such as colorful steel tiles near the railway, while improving the spatial resolution compared to the original upsampled multispectral image.
Figure 8 gives a comparison of the fusion effects of each algorithm for the DEIMOS-2 satellite images. Table 5 gives objective evaluation metrics for each of the fusion algorithms used on the DEIMOS-2 satellite images. From the visual effect, the GIHS, GS, and PCA algorithms in the CS class showed different degrees of spectral distortion when expressing the green region in the lower right corner; the Brovey, the PRACS, and the IHS-ATW algorithms in the CS class offered some improvement for the spectral distortion phenomenon, but compared with the fusion results of the EPACS algorithm used in this paper, both the color saturation and brightness of the lower-right-corner region were far inferior. The HPF, MTF-GLP, and MTF-GLP-CBD algorithms in the MRA-based algorithms produced a grainy feel to the edges of the image, although there was no significant distortion in the green area in the lower right corner. This was mainly due to the insufficient detail injection in the panchromatic images. The MTF-GLP-ECB, the BDSD, and the PTBGT algorithms all performed well in terms of color and detail information in local areas, but the green area in the upper left corner of the MTF-GLP-ECB algorithm appeared locally blurred, and the red area in the upper right corner of the BDSD algorithm was significantly too bright and blurred at the edges, indicating that both algorithms exhibited good local adaptation but not good global adaptation. In contrast, the EPACS algorithm had no significant spectral distortion and showed good edge features for complex features, indicating that the algorithm can effectively reduce spectral distortion while ensuring effective extraction of detailed information.
As can be seen from the table, the CS algorithms such as GIHS, PCA, and GS had much higher global error indicators such as ERGAS, SAM, RMSE, and RASE than the MRA methods, while the UIQI was much lower than for the MRA methods, indicating that the images synthesized by these methods were less similar to the MS image spectral data. The root cause of this was that the global differences and low correlation of the relative spectral response functions were not considered in the fusion process, resulting in the addition of some redundant information to the MS images. In contrast, the Brovey algorithm with the design of adaptive gain coefficients, the PRACS algorithm with its adaptive weight coefficients, the IHS-ATW algorithm with its combining the ATW transform with the IHS algorithm, and the PTBGT algorithm with its design of a variational gradient transfer model all reduced the impact of the low correlation between the images of the CS-based algorithms, but this single improvement still did not allow them to outperform the MRA-based algorithms. The MRA-based algorithms, due to the design of the underlying logic and the MTF of the sensor, resulted in more similarity between the fusion results and the target multispectral data, and outperformed the CS class in terms of the global evaluation metrics.
In this paper, the EPACS algorithm achieved optimal performance on the metrics ERGAS, RMSE, RASE, and UIQI evaluated using the Wald criterion; it performed second best on SAM and was not far from the optimal, and it had better improvement compared to the other CS-based algorithms. The EPACS algorithm performed best on the metric  D λ  evaluated using the QNR criterion; on the metric Ds, the EPACS algorithm performed second best and the BDSD algorithm performed best. In summary, the experimental results showed that the adaptive weighting coefficients, the choice of gain coefficients, and the adaptive enhancement of model details by guided filtering through the MTF optimization of the EPACS algorithm greatly reduced the effect of the inevitable spectral distortion caused by the global variability and low correlation in the relative spectral response function, so that the fused image showed a spectral resolution no weaker than that of MRA-based images. Furthermore, in this design, we used  I P A N H I L H  to replace the  I P A N I L  of the generic CS framework, and the use of the GF filter as well as the RGF filter also enhanced the representation of the spatial information of the PAN images, effectively improving the spatial resolution for foreign objects such as color steel tiles at the perimeter of the high-speed railway, while also ensuring the color fidelity of the foreign objects.

5. Discussion

For the remote-sensing image enhancement of high-speed railways, this paper focuses on improving the traditional CS-based algorithm model from the following two perspectives. (1) Improving the spectral resolution of CS-based algorithms. To solve the problem, this paper first shows that the  I P A N I L  in the generic CS framework brings about severe spectral distortion due to the different spectral response of panchromatic and multispectral image sensors, for which the  I P A N H I L H  design is used instead. In Figure 5 and Table 1, it is clear that the improved image color fidelity has been improved considerably. In order to improve the similarity between PAN images and L images, this paper takes the multispectral image MTF curve into account in the CS-based fusion framework, designs a panchromatic image low-pass filter based on the multispectral image MTF curve, and applies it to the synthesis of L images. In Figure 6 and Table 2, it can be found that the degree of improvement is not as obvious as that of the  I P A N H I L H  design, but it is effective in the saturation of green and red of the image as well as in the evaluation index. This also shows that MTF filters not only improve the effectiveness of MRA-based methods [6] but also work in CS-based frameworks. (2) Improving the spatial resolution of CS-based methods. To solve this problem, the RGF filter is used in this paper in the improved design of  I P A N H I L H . The amount of detail added by any sharpening algorithm will have an impact on the fusion results: adding too much detail may lead to a decrease in spectral resolution; adding too little detail may lead to a decrease in spatial resolution. The RGF filter allows manual control of the scale of the high-frequency component of the panchromatic image by controlling the number of iterations K, thus enabling manual control of the amount of detail D. This allows us to increase the amount of detail as much as possible while maintaining the color information. In Table 1, we can find that  I P A N H I L H  has not only improved in the spectral performance index, but also in the spatial performance index. In this paper, GF is used to further enhance the information of each channel. In Figure 7 and Table 3, we can find that the image is effectively enhanced in terms of effective representation of minute feature details after the GF enhancement. Finally, the fusion effect is compared on GF-2 and DEIMOS-2 images, respectively, in this paper, proving that the proposed framework in this paper can well preserve the colour features and edge detail features of objects, which is crucial to improve the visualization and recognition of foreign objects around high-speed railways. The framework proposed in this paper also can be applied in the fields of storm detection, sand and dust detection, and post-earthquake assessment due to the many adaptive components designed into it.

6. Conclusions

In this paper, an EPACS algorithm is proposed that uses  I P A N H I L H  to replace the design of  I P A N I L  in the generic CS framework, increasing the representation of the spatial and spectral information of the fused images; some adaptive details that depend on data spatial variations were also designed, such as MTF low-pass filter construction, adaptive weighting coefficients and adaptive gain coefficients, and adaptive enhancement of guided filtering for different channel details. These adaptations allow the EPACS algorithm to produce a good fusion effect for images from different sensors, different scales, and features with different spectral characteristics. The EPACS algorithm fuses images clearly in subjective vision, improving the spatial resolution of the MS image of foreign object features on the railway perimeter, while also maintaining their color characteristics well. In terms of objective evaluation metrics, the EPACS algorithm is significantly improved compared to the other CS-based algorithms, and it demonstrates a spectral fidelity that is not inferior to those of the MRA-based algorithms.
The use of high-spatial-resolution means for earth observation began in the military field and is gradually being adopted in the civilian sector. With the growing application of high-speed railway lines, it is of great practical significance to accurately detect and recognize the special features around a railway that might affect the safety of high-speed railway operation, and to do so in a timely and effective manner. In order to better improve the visual interpretation of remote-sensing images, it is essential to synthesize remote-sensing images with good edge characteristics and color features. A new design trend in the development of remote-sensing image fusion technology has emerged in recent years, namely, the use of constrained optimization algorithms to solve the panchromatic sharpening problem. For the time being, such improved algorithms based on the idea of constraint optimization tend to add a significant amount of computational complexity; however, it is believed that with the continuing development of computer science, their time complexity can be significantly optimized. Therefore, the incorporation of constrained optimization ideas into the present framework will be a major research direction to further improve the quality of the fused images. We have thus used the EPACS algorithm as a pre-processing link for eventual use in high-speed rail perimeter foreign object intrusion warning applications.

Author Contributions

Conceptualization, W.J.; formal analysis, Y.L.; methodology, D.Z.; validation, H.G. and S.Q.; writing—original draft preparation, D.Z.; writing—review and editing, D.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China under Grant No. 2020YFF0304104.

Data Availability Statement

https://github.com/Zuo-free/EPACS/tree/master, accessed on 1 May 2022. It has the code and image results for the thesis in it.

Acknowledgments

The authors thank Qi Wei for sharing the remote-sensing image on Github (qw245/BlindFuse: Blind Model-Based Fusion of Multi-band and Panchromatic Images (github.com)) (accessed on 1 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Effect of changes in some hyperparameters on image quality, where bolded indicates optimal.
Table A1. Effect of changes in some hyperparameters on image quality, where bolded indicates optimal.
EPGASSAMQNR
K = 4,  σ s  = 3,  σ r  = 0.8,  ε  = 0.1
r = 22.33563.53500.7952
r = 42.34343.60690.8015
r = 62.39303.77630.8070
r = 82.42323.86240.8091
K = 4,  σ s  = 3,  σ r  = 0.8, r = 3
ε = 0.12.33563.53500.7952
ε = 0.22.35943.50960.7921
ε = 0.42.36113.51030.7920
K = 4,  σ s  = 3, r = 3,  ε  = 0.1
σ r  = 0.22.50363.52090.7878
σ r  = 0.42.37543.50870.7910
σ r  = 0.62.36153.50860.7918
σ r  = 0.82.33563.53500.7952
K = 4,  σ r  = 0.8, r = 3 and  ε  = 0.1
σ s  = 2 2.70953.52530.8277
σ s  = 32.33563.53500.7952
σ s  = 42.37393.69450.7679
σ s  = 62.90945.00230.7435
σ s  = 3,  σ r  = 0.8, r = 3 and  ε  = 0.1
K = 22.33563.53500.7922
K = 42.33563.53500.7952
K = 62.33563.53500.7921
Figure A1. Original dimensional drawing of Figure 9.
Figure A1. Original dimensional drawing of Figure 9.
Electronics 12 02670 g0a1aElectronics 12 02670 g0a1b
Figure A2. Original dimensional drawing of Figure 8.
Figure A2. Original dimensional drawing of Figure 8.
Electronics 12 02670 g0a2

References

  1. Vivone, G.; Alparone, L.; Chanussot, J.; Dalla Mura, M.; Garzelli, A.; Licciardi, G.A.; Restaino, R.; Wald, L. A Critical Comparison Among Pansharpening Algorithms. IEEE Trans. Geosci. Remote Sens. 2014, 53, 2265–2586. [Google Scholar] [CrossRef]
  2. Vivone, G.; Dalla Mura, M.; Garzelli, A.; Restaino, R.; Scarpa, G.; Ulfarsson, M.O.; Alparone, L.; Chanussot, J. A New Benchmark Based on Recent Advances in Multispectral Pansharpening: Revisiting pansharpening with classical and emerging pansharpening methods. IEEE Geosci. Remote Sens. Mag. 2020, 9, 53–81. [Google Scholar] [CrossRef]
  3. Nason, G.P.; Silverman, B.W. The Stationary Wavelet Transform and Some Statistical Applications; Springer: New York, NY, USA, 1995; pp. 281–299. [Google Scholar] [CrossRef]
  4. Do, M.N.; Vetterli, M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 14, 2091–2106. [Google Scholar] [CrossRef] [Green Version]
  5. Aiazzi, A.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. MTF-tailored multiscale fusion of high-resolution MS and Pan imagery. Photogramm. Eng. Remote Sens. 2006, 72, 591–596. [Google Scholar] [CrossRef]
  6. González-Audícana, M.; Otazu, X.; Fors, O.; Seco, A. Comparison between Mallat’s and the ‘a trous’ discrete wavelet transform based algorithms for the fusion of multispectral and panchromatic images. Int. J. Remote Sens. 2005, 26, 595–614. [Google Scholar] [CrossRef]
  7. Ma, X.; Hu, S.; Liu, S.; Fang, J.; Xu, S. Remote Sensing Image Fusion Based on Sparse Representation and Guided Filtering. Electronics 2019, 8, 303. [Google Scholar] [CrossRef] [Green Version]
  8. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A. Context-Driven Fusion of High Spatial and Spectral Resolution Images Based on Oversampled Multiresolution Analysis. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2300–2312. [Google Scholar] [CrossRef]
  9. Garzelli, A.; Nencini, F. Interband structure modeling for Pan-sharpening of very high-resolution multispectral images. Inf. Fusion 2005, 6, 213–224. [Google Scholar] [CrossRef]
  10. Alparone, L.; Garzelli, A.; Vivone, G. Intersensor Statistical Matching for Pansharpening: Theoretical Issues and Practical Solutions. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4682–4695. [Google Scholar] [CrossRef]
  11. Jiao, J.; Wu, L. Pansharpening with a Gradient Domain GIF Based on NSST. Electronics 2019, 8, 229. [Google Scholar] [CrossRef] [Green Version]
  12. Jiao, J.; Wu, L.; Qian, K. A Segmentation-Cooperated Pansharpening Method Using Local Adaptive Spectral Modulation. Electronics 2019, 8, 685. [Google Scholar] [CrossRef] [Green Version]
  13. Carper, W.J.; Lillesand, T.M.; Kiefer, P.W. The use of intensity-hue-saturation transformations for merging SPOT panchromatic and multispectral image data. Photogramm. Eng. Remote Sens. 1990, 56, 459–467. [Google Scholar]
  14. Chavez, A.; Kwarteng, P. Extracting spectral contrast in Landsat Thematic Mapper image data using selective principal component analysis. Photogramm. Eng. Remote Sens. 1989, 55, 339–348. [Google Scholar]
  15. Aiazzi, B.; Baronti, S.; Selva, M. Improving Component Substitution Pansharpening Through Multivariate Regression of MS +Pan Data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  16. Choi, J.; Yu, K.; Kim, Y. A New Adaptive Component-Substitution-Based Satellite Image Fusion by Using Partial Replacement. IEEE Trans. Geosci. Remote Sens. 2010, 49, 295–309. [Google Scholar] [CrossRef]
  17. Wang, Z.; Ziou, D.; Armenakis, C.; Li, D.; Li, Q. A comparative analysis of image fusion methods. IEEE Trans. Geoence Remote Sens. 2005, 43, 1391–1402. [Google Scholar] [CrossRef]
  18. Garzelli, A.; Nencini, F.; Capobianco, L. Optimal mmse pan sharpening of very high resolution multispectral images. IEEE Trans. Geosci. Remote Sens. 2007, 46, 228–236. [Google Scholar] [CrossRef]
  19. Lari, S.N.; Yazdi, M. Improved IHS pan-sharpening method based on adaptive injection of à trous wavelet decomposition. Int. J. Signal Process. Image Process. Pattern Recognit. 2016, 9, 291–308. [Google Scholar] [CrossRef]
  20. Zhong, S.; Zhang, Y.; Chen, Y.; Wu, D. Combining component substitution and multiresolution analysis: A novel generalized BDSD pansharpening algorithm. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 2867–2875. [Google Scholar] [CrossRef]
  21. Liu, P. Pansharpening with transform-based gradient transferring model. IET Image Process. 2019, 13, 2614–2622. [Google Scholar] [CrossRef]
  22. Dong, W.; Yang, Y.; Qu, J.; Xiao, S.; Du, Q. Hyperspectral pansharpening via local intensity component and local injection gain estimation. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  23. Li, J.; Hu, Q.; Ai, M. Multispectral and panchromatic image fusion based on spatial consistency. Int. J. Remote Sens. 2018, 39, 1017–1041. [Google Scholar] [CrossRef]
  24. Li, Z.; Leung, H. Fusion of multispectral and panchromatic images using a restoration-based method. IEEE Trans. Geosci. Remote Sens. 2008, 47, 1482–1491. [Google Scholar] [CrossRef]
  25. Zhang, L.; Shen, H.; Gong, W.; Zhang, H. Adjustable model-based fusion method for multispectral and panchromatic images. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2012, 42, 1693–1704. [Google Scholar] [CrossRef] [PubMed]
  26. Fei, R.; Zhang, J.S.; Liu, J.; Du, F.; Chang, P.; Hu, J. Manifold regularized sparse representation of injected details for pansharpening. Int. J. Remote Sens. 2019, 40, 8395–8417. [Google Scholar] [CrossRef]
  27. Massip, P.; Blanc, P.; Wald, L. A Method to Better Account for Modulation Transfer Functions in ARSIS-Based Pansharpening Methods. IEEE Trans. Geosci. Remote Sens. 2011, 50, 800–808. [Google Scholar] [CrossRef] [Green Version]
  28. Yuwei, Z.; Pinglv, Y.; Qiang, C.; Quansen, S. Pan-sharpening model based on MTF and variational method. Acta Autom. Sin. 2015, 41, 342–352. [Google Scholar] [CrossRef]
  29. He, K.M.; Sun, J.; Xiaoou, T. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409. [Google Scholar] [CrossRef]
  30. Wald, L. Data Fusion: Definitions and Architectures-Fusion of Images of Different Spatial Resolutions; Les Presses de l’ École des Mines: Paris, France, 2002. [Google Scholar]
  31. Yuhas, R.H.; Goetz, A.F.H.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the Spectral Angle Mapper (SAM) algorithm. In Proceedings of the Summaries 3rd Annual JPL Air-Bone Geoscience Workshop, Pasadena, CA, USA, 1–5 June 1992; pp. 147–149. [Google Scholar]
  32. Thomas, C.; Ranchin, T.; Wald, L.; Chanussot, J. Synthesis of multispectral images to high spatial resolution: A critical review of fusion methods based on remote sensing physics. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1301–1312. [Google Scholar] [CrossRef] [Green Version]
  33. Zhou, W.; Bovik, A.C. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  34. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and Panchromatic Data Fusion Assessment Without Reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Generic framework for CS−based algorithms.
Figure 1. Generic framework for CS−based algorithms.
Electronics 12 02670 g001
Figure 2. EPACS algorithm framework.
Figure 2. EPACS algorithm framework.
Electronics 12 02670 g002
Figure 3. Spectral response of the GF-2 sensing band.
Figure 3. Spectral response of the GF-2 sensing band.
Electronics 12 02670 g003
Figure 4. MTF curves for panchromatic and multispectral image sensors.
Figure 4. MTF curves for panchromatic and multispectral image sensors.
Electronics 12 02670 g004
Figure 5. Comparison of IHS algorithm improvement effects. (a) Original multispectral image; (b) original IHS algorithm; (c) improved IHS algorithm.
Figure 5. Comparison of IHS algorithm improvement effects. (a) Original multispectral image; (b) original IHS algorithm; (c) improved IHS algorithm.
Electronics 12 02670 g005
Figure 6. Comparison of low-pass filtering methods. (a) Original multispectral image; (b) bilinear interpolation filtering; (c) low-pass filtering based on MTF.
Figure 6. Comparison of low-pass filtering methods. (a) Original multispectral image; (b) bilinear interpolation filtering; (c) low-pass filtering based on MTF.
Electronics 12 02670 g006
Figure 7. Comparison of image information with or without GF. (a) Original multispectral image; (b) GF is not used; (c) GF is used.
Figure 7. Comparison of image information with or without GF. (a) Original multispectral image; (b) GF is not used; (c) GF is used.
Electronics 12 02670 g007
Figure 8. Comparison of DEIMOS-2 image fusion results. (1)  MS ~ ; (2) PAN; (3) GIHS; (4) PCA; (5) GS; (6) Brovey; (7) PRACS; (8) BDSD; (9) IHS-ATW; (10) PTBGT; (11) HPF (12) MTF-GLP; (13) MTF-GLP-CBD; (14) MTF-GLP-ECB; (15) EPACS.
Figure 8. Comparison of DEIMOS-2 image fusion results. (1)  MS ~ ; (2) PAN; (3) GIHS; (4) PCA; (5) GS; (6) Brovey; (7) PRACS; (8) BDSD; (9) IHS-ATW; (10) PTBGT; (11) HPF (12) MTF-GLP; (13) MTF-GLP-CBD; (14) MTF-GLP-ECB; (15) EPACS.
Electronics 12 02670 g008
Figure 9. Comparison of GF-2 image fusion results. (1)  MS ~ ; (2) PAN; (3) GIHS; (4) PCA; (5) GS; (6) Brovey; (7) PRACS; (8) BDSD; (9) IHS-ATW; (10) PTBGT; (11) HPF (12) MTF-GLP; (13) MTF-GLP-CBD; (14) MTF-GLP-ECB; (15) EPACS.
Figure 9. Comparison of GF-2 image fusion results. (1)  MS ~ ; (2) PAN; (3) GIHS; (4) PCA; (5) GS; (6) Brovey; (7) PRACS; (8) BDSD; (9) IHS-ATW; (10) PTBGT; (11) HPF (12) MTF-GLP; (13) MTF-GLP-CBD; (14) MTF-GLP-ECB; (15) EPACS.
Electronics 12 02670 g009aElectronics 12 02670 g009b
Table 1. Objective evaluation of the improved IHS algorithm using  I P A N H I L H  and objective evaluation of the original IHS algorithm, where bolded indicates optimal.
Table 1. Objective evaluation of the improved IHS algorithm using  I P A N H I L H  and objective evaluation of the original IHS algorithm, where bolded indicates optimal.
ERGASSAM   D λ DS
Improved IHS3.88554.09690.09300.1943
Original IHS3.02504.97470.16920.3700
Table 2. Objective evaluation of different low-pass filtering methods, where bolded indicates optimal.
Table 2. Objective evaluation of different low-pass filtering methods, where bolded indicates optimal.
ERGASSAM   D λ DS
Bilinear interpolation filtering2.60515.07960.09300.1943
Low-pass filtering based on MTF2.33563.53500.07930.2961
Table 3. Objective evaluation of image information with or without GF, where bolded indicates optimal.
Table 3. Objective evaluation of image information with or without GF, where bolded indicates optimal.
ERGASSAMRASE   D λ
GF is not used3.04473.858823.78450.0865
GF is used2.33563.535018.49270.0793
Table 4. Comparison of objective evaluation indicators for each fusion method used on GF-2 images, where bolded and underlined indicates best, underlined indicates second best.
Table 4. Comparison of objective evaluation indicators for each fusion method used on GF-2 images, where bolded and underlined indicates best, underlined indicates second best.
Underlying LogicReferenceERGASSAMRMSERASEUIQI   D λ DsQNR
CSGIHS0.77121.63320.03937.05540.96430.18010.32830.5507
PCA0.78431.81820.04117.37270.96340.13220.28790.6180
GS0.77921.74420.04037.23310.96240.12760.29080.6188
Brovey0.92161.67280.04037.23860.96230.16340.31370.5741
BDSD0.59391.74950.03075.48300.98100.09570.25450.6741
PRACS0.52971.67320.03865.12880.98200.09080.25120.6809
IHS-ATW0.58471.73560.03125.27480.97850.09660.24780.6795
PTBGT0.57981.74210.03475.47160.97360.09750.23160.6954
MRAHPF0.59081.52100.03115.58420.97870.09670.16440.7548
MTF-GLP0.57761.57590.03075.36760.98100.12200.23480.6724
MTF-GLP-CBD0.67921.62010.03566.37190.97400.12180.21390.6904
MTF-GLP-ECB0.59411.59250.03155.64280.97900.06770.14030.8015
EPACS0.59271.76760.03065.50180.98260.06010.11830.8287
Table 5. Comparison of objective evaluation indicators for each fusion method used on DEIMOS-2 images, where bolded and underlined indicates best, underlined indicates second best.
Table 5. Comparison of objective evaluation indicators for each fusion method used on DEIMOS-2 images, where bolded and underlined indicates best, underlined indicates second best.
Underlying LogicReferenceERGASSAMRMSERASEUIQI   D λ DsQNR
CSGIHS3.02504.97470.021125.34840.96350.16920.37000.5234
PCA2.98205.27470.020124.14210.96650.16650.36510.5292
GS3.04485.06170.020824.89840.96440.16320.36640.5302
Brovey2.89663.81720.020224.24600.96670.11040.31920.6056
BDSD2.54114.37360.017020.26310.97530.02240.18080.8008
PRACS2.89644.29000.017621.22830.97680.08300.24880.6888
IHS-ATW2.44874.38460.016520.43000.97660.07540.23560.7067
PTBGT2.48034.35000.017121.65300.97580.06330.24330.7088
MRAHPF2.87663.18190.018922.69730.97210.13270.24260.6569
MTF-GLP2.60464.04510.016920.24250.97870.15100.31030.5856
MTF-GLP-CBD2.80594.07820.018622.17150.97530.15270.28840.6029
MTF-GLP-ECB2.62303.82630.017020.37080.97850.07750.18760.7550
EPACS2.33563.53500.015418.49270.98220.02220.18670.7952
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zuo, D.; Li, Y.; Qiu, S.; Jin, W.; Guo, H. A Spectral Enhancement Method Based on Remote-Sensing Images for High-Speed Railways. Electronics 2023, 12, 2670. https://doi.org/10.3390/electronics12122670

AMA Style

Zuo D, Li Y, Qiu S, Jin W, Guo H. A Spectral Enhancement Method Based on Remote-Sensing Images for High-Speed Railways. Electronics. 2023; 12(12):2670. https://doi.org/10.3390/electronics12122670

Chicago/Turabian Style

Zuo, Dongsheng, Yingjie Li, Su Qiu, Weiqi Jin, and Hong Guo. 2023. "A Spectral Enhancement Method Based on Remote-Sensing Images for High-Speed Railways" Electronics 12, no. 12: 2670. https://doi.org/10.3390/electronics12122670

APA Style

Zuo, D., Li, Y., Qiu, S., Jin, W., & Guo, H. (2023). A Spectral Enhancement Method Based on Remote-Sensing Images for High-Speed Railways. Electronics, 12(12), 2670. https://doi.org/10.3390/electronics12122670

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop