Next Article in Journal
Deep Ensemble Learning Based on Multi-Form Fusion in Gearbox Fault Recognition
Previous Article in Journal
A Frontier Review of Semantic SLAM Technologies Applied to the Open World
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Graph-Based Superpixel Segmentation Approach Applied to Pansharpening

by
Hind Hallabia
1,2
1
UMR CNRS 7347—Materiaux, Microéléctronique, Acoustique, Nanotechnologies (GREMAN), Institut Universitaire de Technologie de Blois (IUT Blois), Tours University, 37000 Tours, France
2
Institut National des Sciences Appliquées Centre-Val de Loire (INSA CVL Campus Blois), 41000 Blois, France
Sensors 2025, 25(16), 4992; https://doi.org/10.3390/s25164992
Submission received: 31 December 2024 / Revised: 5 August 2025 / Accepted: 9 August 2025 / Published: 12 August 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

In this paper, an image-driven regional pansharpening technique based on simplex optimization analysis with a graph-based superpixel segmentation strategy is proposed. This fusion approach optimally combines spatial information derived from a high-resolution panchromatic (PAN) image and spectral information captured from a low-resolution multispectral (MS) image to generate a unique comprehensive high-resolution MS image. As the performance of such a fusion method relies on the choice of the fusion strategy, and in particular, on the way the algorithm is used for estimating gain coefficients, our proposal is dedicated to computing the injection gains over a graph-driven segmentation map. The graph-based segments are obtained by applying simple linear iterative clustering (SLIC) on the MS image followed by a region adjacency graph (RAG) merging stage. This graphical representation of the segmentation map is used as guidance for spatial information to be injected during fusion processing. The high-resolution MS image is achieved by inferring locally the details in accordance with the local simplex injection fusion rule. The quality improvements achievable by our proposal are evaluated and validated at reduced and at full scales using two high resolution datasets collected by GeoEye-1 and WorldView-3 sensors.

1. Introduction

A wide variety of Earth observation imaging sensors are available to deliver images about the covered scene (e.g., Pléiades, GeoEye-1, GaoFen-2, and WorldView-2/-3). Most of these satellites can produce two kinds of images: a panchromatic (PAN), image which is a greyscale image with high spatial resolution; and a multispectral (MS) image, which is a multi-band image, usually, with four or eight spectral bands and low spatial resolution. Hence, to generate a more comprehensive synthetic image for such a scene, remote sensing data fusion could be used to resolve this issue. Particularly, fusion of PAN and MS images, referred to as pansharpening [1,2,3], regarded as a sub-topic of the data fusion field, may be applied. Thus, a spatially enhanced MS image can be synthesized, which is at the same resolution as the PAN image, which ideally exhibits the spectral proprieties of the original MS image. The approach can be used in several applications, such as change and anomaly detection and land cover monitoring and mapping.

1.1. Background

Over the two last decades, numerous pansharpening algorithms have been proposed [4,5,6,7,8,9,10,11,12]. Based on the way details are extracted, they are classified into three categories: Multi-Resolution Analysis (MRA), Component Substitution (CS), and Model-Based Methods (MMs). The CS group operates in a transformed domain by replacing a component of the original MS image (intensity or first principal component) with the histogram-matched PAN image to generate a high-resolution MS image. The details are therefore presented as a residual image between the synthetic component and the histogram-matched PAN image. Instances include Brovey (BT) [13] and its optimized version with haze correction (BT-H) [14], generalized intensity-hue-saturation (GIHS) [15,16], principal component analysis (PCA) [15], the Gram–Schmidt adaptive (GSA) [17] method, partial replacement adaptive CS (PRACS) [18], band-dependant spatial detail (BDSD) [19], and the non-linear intensity-hue-saturation method (NL-IHS) [20]. Regarding the MRA group, it enables inference of details, which are extracted by means of spatial filtering the PAN image into the up-sampled MS bands. Some examples of this class are as follows: Indusion [21], the Generalized Laplacian Pyramid (GLP) tailored to the Modulation Transfer Function (MTF) with additive [22], multiplicative [7,23,24], and projective [25,26] injection models, the à-trous wavelet transform (ATWT) [27], the Filter Banks-based Pansharpening (FBP) [6] method, and the Half-Gradient Morphological Operator (MF-HG) [28]. Furthermore, with the intention of enhancing the quality of the fused products, the principles of MRA and CS are combined together and hybrid versions are developed [29,30,31], such that the MRA processing is performed for the intensity [29,30] or the first component [31]. A popular example is the Additive Wavelet Luminance Proportion (AWLP) [29], in which the wavelet transform is processed in the IHS domain.
The third group recasts the pansharpening task as an inverse problem, which can be solved with the application of regularization methods [32], convolutional neural networks (CNN) [9,33], and deep learning [34,35]. Considering the VO-based techniques categories, a variational optimization problem is defined and solved based on an acquisition or representation model and introducing regularization terms in some cases [32]. As examples, the model-based fusion using PCA and wavelets (PWMBF) [36] is proposed by modeling and solving the pansharpening via a maximum a posteriori problem. A variational approach based on filter estimation, known as FE-HPM [37], models the pansharpening as a deconvolution task by estimating a spatial degradation filter approximating the MTF of the MS sensors and relaying on the reduced- and full-scale spatial resolutions acquisitions. Pansharpening based on Total Variation (TV) [38] is proposed in which the fusion is modelled as a TV-regularized least squares problem, and solved according to the majoration-minimization optimization algorithm. The sparse representation of injected details (CS-D) [39] is an example of a pansharpening technique based on sparse representation theory. The spatial details to be injected into the up-sampled MS bands by a conventional MRA schema are generated through a dictionary of patches defined at reduced resolution taking advantage of the invariance property. Deep learning-based pansharpening methods [35] estimate the spectral and spatial properties of the high-resolution fused products using non-linear relationships estimated between the high-resolution PAN image and the low-resolution MS image. Consequently, advanced models, such as conventional neural networks (CNNs) [3,33,40], generative adversarial networks (GANs) [41] and auto-encoders (AEs) [42] have been applied in the field of pansharpening. Among CNN-based approaches, Masi et al [33,43] proposed a CNN network (model) introduced into the fusion process, and the method is referred to as PNN. The high-resolution MS image is synthesized using an ad hoc generation framework. It may be noted that CNN-based pansharpening methods are deeply related to the training procedure, a fundamental step for learning. The training stage for PNN is closely based on a resolution downgrading procedure, guided by Wald’s protocol [44]. For learning and updating the network parameters, a fused MS-PAN composite is generated and used as input to the CNN model. A more advanced variant of PNN is introduced in [40] by including a fine-tuning step in the training procedure.

1.2. Related Works

Most of the aforementioned pansharpening algorithms are processed on the original pixels (i.e., at pixel-level) [2]. In some cases, they neglect local contextual information about the targeted objects contained in the scenes which can not be sufficiently represented semantically, mainly for textured remote sensing imageries [24]. To reduce these problems, recent research studies focus on processing pansharpening at a region level. In [7,11,45], the authors argue that the fused results at feature level are better than those processed on pixels. In this context, several forms of local feature information are incorporated into the GLP framework, such as sliding windows [22], segments [45], clusters [11] or superpixels [7]. This technique has demonstrated its effectiveness in providing high quality fused outcomes at feature level by considering multiple injection gains. In [22], the authors defined them by computing local statistics over sliding windows, such that the model is referred to as Context-based Decision (CBD). Intuitively, the fixed structure of the latter causes overlap between the blocs. As a consequence, the spatial arrangement of the objects contained in the remote sensing data is ignored. To solve this issue, image clustering has been applied for pansharpening. For instance, in [11], the k-means clustering algorithm is adopted to generate several clusters of pixels showing the same semantic proprieties. Then, the injection coefficients are estimated for each cluster either as a solution of a multiple [10] or a robust regression problem [11]. Moreover, as in [45], the gain factors are computed using a statistical (i.e., projective injection model) through regions defined by performing the binary partition tree (BPT) segmentation method, which are incorporated in both the GLP and GSA techniques (referred as GLP-BPT and GSA-BPT, respectively). In addition, in another recent study [7], the pansharpening task is performed at superpixel level. In detail, the local information is defined over a set of superpixels extracted by means of a Simple Linear Iterative Clustering (SLIC) method [46] and the fusion process is performed using an optimized injection scheme that optimally calculates the injection gains through Shuffled Complex Evolution. Finally, as shown in [47], context-adaptive pansharpening methods can provide better fusion outcomes in terms of spectral fidelity than global techniques. This is justified by the fact that the local structural characteristics of regions could be better expressed by a set of pixels having the same proprieties as single pixels. In consequence, region-based fusion models are more effective in pansharpening by considering extracted features.

1.3. Motivation and Contributions

Existing pansharpening approaches are based on a unified fusion model. In some cases, spatial or spectral distortions could occur in fused products due to the diversity of sensor characteristics. Considering that remote sensing images contain heterogeneous pixels, and therefore diverse spectra, region-based fusion rules allows adaptive fusion of images to enhance the spatial and spectral resolutions in comparison with pixel-wise methods. Motivated by this idea, we introduced spatially variant injection gains estimated over multiple superpixels and its variant based on graph theory. In light of this issue, processing pansharpening on graph-based superpixels entities allows us to approximate very-high-resolution images. In this paper, a graph-driven pan-sharpening method is proposed, which is reformulated as a local regression problem with the purpose of preserving spectral fidelity and increasing the spatial quality of the MS image. The fusion process is performed on various regions defined by a graph-based SLIC technique applied on the MS image. The MS image is encoded as a graph to be used further as a guidance map for inference of spatial details into the expanded MS bands. In more detail, the proposed approach utilizes graph-based segmentation images by adopting a superpixel aggregation similarity concept instead of using explicit superpixel entities. The underpinning idea is to over-segment the MS image into a set of superpixels by means of SLIC [46], which represent initialization for the Region Adjacency Graph (RAG) [48]; then, a hierarchical merging step is performed. A graph-driven image is therefore generated with a number of regions lower than those obtained in the initial over-segmentation. Furthermore, a simplex-based fusion strategy is adopted in this study, where the injection gains are estimated over the proposed graphical regions. The pan-sharpening task is accomplished by regionally inferring details into their corresponding up-sampled MS regions defined by the graph-based segmentation map. The proposed approach does not inject redundant information into the MS image thanks to the graph-based detail injection procedure. The GLP technique is utilized for information analysis and further in fusion processing because it helps to preserve the edge information and consequently the details to be extracted.
In summary, our main contributions are three-fold:
  • Modeling the guidance image (for detail injection) as a graph through a region adjacency graph (RAG) performed over an initial over-segmentation obtained by means of SLIC and followed by a hierarchical merging step to obtain a meaningful regional map.
  • A graph-driven pansharpening method is proposed by considering the local simplex algorithm for gain estimation as well as the graph-based guidance image (map) for detail injection, with the aim of refining the fused outcomes at region level.
  • Extensive experimental comparisons between pixel- and region-wise pansharpening methods as well as between superpixel versus graph-driven superpixel guided injection schemes are presented. The performance of the proposed method is highlighted in comparison with traditional, variational optimization, and deep learning methods.
The paper is structured as follows. Section 2 gives an overview about the basic principles to be used in our study. The proposed graph-based pansharpening algorithm is presented in Section 3, including the generation of the guidance injection map using RAG and SLIC, the formulation of the graph-based fusion scheme within the Simplex model, and its introduction into the pyramidal framework. Section 4 describes the experimental results assessing the performance and robustness of the proposed fusion method. Finally, the conclusions and future perspectives are presented in Section 5.

2. Preliminaries and Background

2.1. Pansharpening as a Unified Model for Fusion

Classical pansharpening methods can be recast using a unified protocol, which can be defined by the following two phases: (i) detail extraction; and (ii) its subsequent inference into the MS channels [3,4]. Thus, the fusion process relies on the transfer of detail components, { D i } i = 1 N (where N is the total number of spectral bands) into the expanded MS channels, M S ˜ = { M S ˜ i } i = 1 N , preponderated by injection gains, and { G i } i = 1 N . This allows generating an enhanced MS image at fine-resolution, M S ^ = { M S ^ i } i = 1 N , which is at the same resolution as the original PAN image and has a high spectral-spatial quality [49].
From a mathematical point of view, the fusion algorithm can be expressed by the following general band-wise equation [1]:
M S ^ i = M S ˜ i + G i D i , i = { 1 , , N } ,
where ∘ is the pixel-wise multiplication operator.
For detail extraction, taking into account the multi-resolution (MRA) and composition substitution methods, { D i } i = 1 N can be derived as the image residual between the histogram-matched PAN image, { P i } i = 1 N , and the low-resolution PAN image, { P L i } i = 1 N (for MRA-based methods), or the intensity component, I (for CS-based methods). Concerning the VO- and DL-based methods, the high-frequency details can be extracted through a mathematical model, denoted by Φ , as a function of PAN and MS data before/and during the fusion procedure. Generally, the details can be defined as follows for each band ( i = { 1 , . . , N } ):
D i = ( P i P L i ) for MRA ( P i I ) for CS Φ ( P i , M S i ) for VO and DL
where Φ ( . , . ) denotes the mathematical model or function related to PAN and MS data fusion.
Concerning the detail injection stage, { D i } i = 1 N can be inferred into the expanded MS bands { M S ˜ i } i = 1 N modulated by means of gain coefficients { G i } i = 1 N [47]. The latter allows proportionally inferring the spatial details into the expanded MS bands taking into consideration several forms of injection strategies at different fusion levels [50] in a multiplicative or projective manner [1,7,11]. In this context, there are a large variety of pansharpening methods which use either a global or a context-adaptive injection model.

2.1.1. Global Injection Scheme

Fusion techniques that employ a global additive injection scheme consider a unique gain coefficient per spectral MS band or for the whole image. Instances include the GIHS [51], GLP with MTF-adjusted filters (GLP-HPF) (where HPF stands for high-pass filtering), AWLP [29], BDSD [19], and PRACS [18]. The two latter methods fall within the band-level fusion category since a unique gain factor is applied for each spectral channel.

2.1.2. Multiplicative Injection Scheme

This model is also known as High-Pass Modulation (HPM), which is generally performed at pixel level [1]. The gain factors are computed for each pixel in the image as the ratio between the expanded MS image ( M S ˜ ) and the low-resolution version of the PAN image ( P L ) as:
G i = M S ˜ P L .
This pixel-wise injection scheme is called the Spectral Distortion Minimisation (SDM) model, which is performed especially for the GLP-SDM [23] and FBP-SDM [6]. Other examples of pixel-based pansharpening techniques are the Brovey transform (BT) [13] and Smoothing Filter-based Intensity Modulation (SFIM) [52].

2.1.3. Projective Injection Scheme

This model consists of computing statistically the gain coefficients through a regression analysis [25] as:
G i = c o v ( M S ˜ i , P L i ) P L i , i = 1 , . . , N B .
where c o v ( . , . ) and v a r ( . ) stand for the covariance and the variance, respectively.
In the context of pansharpening, this strategy was initially applied to CS methods within the GSA approach [17] in a band-wise manner. Then, it was adopted for MRA approaches, particularly for the GLP technique, in which it is processed at band-to-band [25] and at region level [11,24,45].
To summarize, the estimation of injection gains is a crucial step for pansharpening. It can be performed either globally (i.e., the same coefficient is applied to the whole image or to the different spectral bands), or locally (i.e., spatially variant coefficients are applied computed locally on regions).

2.2. Simple Linear Iterative Clustering Method

The main goal of superpixel segmentation is to subdivide an image into a large number of regions, such that each region lies within object boundaries defined over the image to be segmented. This process is known as image over-segmentation [46]. Among the popular superpixel segmentation methods, examples include Simple Linear Iterative Clustering (SLIC) [46], which belongs to the gradient-ascent based category. The algorithm starts from initial rough groups of pixels (i.e., clusters) obtained by means of the local k-means clustering method and iteratively refines them to provide better segmentation until satisfying certain convergence criteria. The approach can be applied for both greyscale and color images. In our study, a colored image is considered. A five-dimensional vector [ l , a , b , x , y ] T is defined to represent the features of a cluster, where the superscript [ . ] T denotes the transpose operator. [ l , a , b ] T is the color feature vector corresponding to the CIELAB color space and [ x , y ] T denotes the spatial feature vector. The main steps of SLIC are as follows:
  • Initialize the k cluster centres defined by C k = [ I k , x k , y k ] T which denotes the pixel position. The pixel intensity I k is incorporated into the feature vector C k . The grid interval S = ( N k ) approximately and uniformly determines the superpixels size, where N is the number of pixels.
  • Each pixel at the position ( x i , y i ) is associated to the nearest cluster centre ( C k ) at the position ( x k , y k ) according to the distance ( D s ) estimated by the formula:
    D s = ( x i x k ) 2 + ( y i y k ) 2
    and the color CIELAB distance D c is defined as:
    D s = ( l i l k ) 2 + ( a i a k ) 2 + ( b i b k ) 2
    such that the clustering metric can be formulated:
    D s c = D c 2 + D s 2 S 2 m 2
    where m is the compactness parameter.
  • By specifying the clustering metric D s c as in Equation (7), the parameter m allows computation of the relative interest between the grayscale similarity and spatial proximity.
In summary, the clustering steps provide a number of N s superpixels. The higher the number of generated superpixels N s , the finer the over-segmentation. The SLIC superpixel segmentation procedure is summarized in Algorithm 1.
Algorithm 1: Pseudocode of SLIC method applied to MS image.
Input M S : MS image (R,G,B bands), k: number of initial clusters, m: compactness parameter.
Output G R G B : Label matrix result of the SLIC applied on RGB input image.
Represent the RGB image in the CIELAB space color: ( R , G , B ) ( l , a , b ) .
Initialize k cluster centres C k = I k , x k , y k by sampling pixels at a uniform grid S: S ( N k ) , where C k is the feature vector, I k is the pixel intensity and ( x k , y k ) are its coordinates.
for each cluster centres C k ( x k , y k ) and given a pixel coordiantes ( x i , y i ) do ( (
   Compute the spatial distance ( D s ) using Equation (5) as:
D s ( x i x k ) 2 + ( y i y k ) 2
   Compute the spectral (color) distance ( D c ) using Equation (6):
D c ( l i l k ) 2 + ( a i a k ) 2 + ( b i b k ) 2
   Compute the clustering distance ( D s c ) using Equation (7):
D s c D c 2 + D s 2 S 2 m 2
end (
Compute the new cluster centre C k according to the lowest distance D s c estimated in the grid ( 2 S × 2 S )
Update the SLIC map by computing the distances between the previous C k ( D s c ) and the newest cluster C k ( D s c ).
Repeat the algorithm steps (3)-(5) until the distance ϵ s c between two cluster centres is not a null-value: ϵ s c D s c D s c 0
Label the resulting semantic segmentation map, G R G B , assigning each superpixel with l = 1 , . . , L .

2.3. Image Graph Representation via Region Adjacency Graph

Given an image I , it can be partitioned into distinct non-overlapping and non-empty regions (i.e., segments). For a positive integer n > 1 , we define { R i } i = 1 n as the finite set corresponding to the resulting segmented image, where the subscript ( . ) i indicates that the segment is labeled with an index i { 1 , , n } , where n is the total number of segments. The segmentation process satisfies the following two relations:
I = i = 1 n R i ,
and
R i R j = for   i j .
Let G I = ( V , E ) denote an indirect image-driven graph of an image I , which is called the Region Adjacency Graph (RAG) [48]. By definition, the RAG represents the adjacency of regions with a graph, which can be represented by a set of vertices (nodes) v V and a set of edges e E V 2 . Mathematically, the graph can be formalized as:
G I = { ( v , e ) v V a n d e E }
In the context of image segmentation, V is the set of resulting regions (further the set of superpixels) from the image I and E is the set of all pairs of adjacent segments linked by a neighborhood relationship. Accordingly, they can be expressed as follows:
V = { R i } i = 1 n ,
and
E = { e e , ( v p , v q ) R 2 / e = ( v p , v q ) } .
The RAG steps are represented in Algorithm 2. An example of an image-driven graph is represented graphically and displayed in Figure 1. It contains seven nodes (i.e., regions) and 10 edges. The RAG [cf. Figure 1b] is constructed from an image segmentation [cf. Figure 1a] such that black dots are the nodes { v i } i = 1 7 and bloc lines are edges { e j } j = 1 10 .
Algorithm  2: Pseudocode of RAG method.
Input:
  I : A greyscale image: I G R G B .
Output:
  G I = ( V , E ) : RAG image-based graph: G I G R G B .
Define a set of vertices V from an over-segmented image I using Equation (11).
for each couple of labels (p,q)do ( (
   Define the respective vertices as:
( v p , v q ) ( R p , R q )
   Compute the the edge between ( R p , R q ) using Equation (12).
end (
Compute new vertices and edges E p q by repeating steps (1) et (2) until the distance between two edges is equal to zero, as: E p q ( e p e q ) = 0
Obtain the RAG segmented image defined by Equation (10).

3. Materials and Methods

The proposed graph-based pansharpening approach can be summarized in terms of the following three phases: image segmentation, gain estimation, and fusion, which are detailed in the subsections below. A schematic diagram of the proposed approach is presented in Figure 2.
Figure 2. Overall diagram of the proposed graph-based pansharpening approach. Note that Diagram 2 is displayed in Figure 3.
Figure 2. Overall diagram of the proposed graph-based pansharpening approach. Note that Diagram 2 is displayed in Figure 3.
Sensors 25 04992 g002
Figure 3. The steps involved in generation of the guidance image as a graph-based segmentation map.
Figure 3. The steps involved in generation of the guidance image as a graph-based segmentation map.
Sensors 25 04992 g003

3.1. RAG-Based SLIC Performed on MS Image

The main steps of the graph-based segmentation map are reported in Figure 3 and are experimentally illustrated in Figure 4. The image segmentation phase deals with the generation of graph-based superpixels using region similarities. The principle is to combine the superpixels, generated by means of SLIC, and the RAG concept to obtain novel graph-based superpixel entities. More specifically, it consists of merging each two adjacent regions in order to produce a novel region that contains all similar pixels from both of them. The process is iterated until fusing of similar region pairs occurs. The process is presented in more detail below.

3.1.1. SLIC Segmentation

An over-segmentation algorithm is applied to the MS image by means of SLIC (only the RGB bands are considered) [46]. This allows transforming the input image, denoted as S R G B , by aggregating homogeneous neighbouring pixels into a set of perceptually meaningful superpixels showing similar properties, which is the spectral information in our case study. A unique label is therefore assigned to each region.

3.1.2. RAG on SLIC

The MS image is over-segmented into a set of superpixels, S R G B . A RAG is built from the set of superpixels S R G B and denoted by G R G B = ( V R G B , E R G B ) , where V R G B = { R i } i n represents the set of resulting regions such that each region in the segmented image is a node in the graph and E R G B = ( R i , R j ) is obtained as the joint edge linking two adjacent regions.

3.1.3. Node Merging

This step enables improvement of the initial segmentation with the aid of the RAG. To this end, a hierarchical threshold cut method is performed. It enables removing the edges below a certain threshold and assigning a label to each connected component in the image-based graph. The threshold γ is determined as follows. Starting from an initial number of superpixels derived from the SLIC algorithm, merging between adjacent regions is performed iteratively in accordance with the greyscale pixels values (i.e., spectral information from the MS bands). γ is therefore defined as the value at which fusion between two adjacent vertices is completed. This procedure allows aggregating similar regions to obtain a uniform region.
In this study, the threshold is determined as follows. Given two vertices v 1 and v 2 linked by an edge e in the graph G R G B , the merging weight w e represents the probability of merging these two nodes as a region. First, the centroid of each region is determined. Then, edges linking every pair of adjacent regions are highlighted. Next, edge weights are computed between every two nodes, which are defined as the difference in the average color between two regions. The more similar the regions are, the lower the weight between them.
Figure 5 depicts an example for determining the threshold parameter using the GeoEye-1 (Captown-GE1) and WorldView-3 (Adelaide-WV3) datasets. It shows the variation in the number of regions (ie., superpixels initially generated by SLIC) versus the grayscale pixel values of the considered images. It is observed that the number of regions decreases after the node merging stage by applying the RAG concept. Using a number of initial clusters N c = 300 , the parameters for the resulting graph-based superpixels are equal to the following: N s = 205 , N g = 150 , and γ = 50 for the Captown-GE1 dataset; and N s = 200 , N g = 120 , and γ = 100 for the Adelaide-WV3 dataset. Moreover, we note that the second dataset gives a lower number of graph-based regions than the first dataset. This is explained by the fact that the urban dataset contains more textured and heterogeneous structures.
The proposed graph-based superpixel segmentation structure is shown graphically in Figure 4 using two samples of remote sensing datasets acquired from the GeoEye-1 [see Figure 4a–d] (GE1) and WorldView-3 (WV-3) [see Figure 4e–h] satellites. In this example, the corresponding boundaries, resulting from an initial segmentation method using SLIC are marked in blue color. The edges are marked as yellow lines and the centroids as red points. Comparing the obtained graph-based image with the segmentation results from directly applying SLIC [46], as in Figure 4d,h, we note that the presented process (RAG on SLIC over-segmentation) enables grouping of superpixels with similar sizes, as well as merging of small-sized superpixels. Recall that SLIC is considered an initial step for the proposed graph-based superpixel method. As depicted in Figure 5, the number of regions (i.e., graph-based superpixels) decreased from 220 to 160 for the Captown-GE1 dataset and from 200 to 105 for the Adelaide-WV3 dataset after the node merging step.

3.2. Graph-Based Superpixel Guided Regression Fusion Rule

Once the MS image is over-segmented into a set of graph-based superpixels, G F , the latter is super-imposed into the up-sampled MS and the detail images, yielding sets of regions R M S ˜ ( j ) and R D ( j ) for each respective image. Furthermore, from the graph G F , we computed the set of final regions (defined by N g ) to e involved in the local pansharpening process and consequently in the estimation of local injection factors. The pansharpening task consists in locally merging the PAN and MS regions in accordance with the GLP framework, and then computing the injection gains for each region defined by the graph-driven map G F . More especially, this region-dependent fusion problem can be defined by the following model [7]:
m s ^ i ( j ) = f ( x i ( j ) , g ^ i ( j ) ) , j = { 1 , , L } .
where f ( . , . ) assigns the fusion process in accordance with the pyramidal framework, and the superscript (j) and L denote the label numbers. x i ( j ) = [ m ˜ i ( j ) , d i ( j ) ] represents the up-sampled multispectral and detail segments, respectively. g ^ i ( j ) is the gain factor to be predicted.
The set of gain values are defined as a group of N g optimal solutions using the Simplex algorithm [7], which is given for each region of the label index j = { 1 , , N g } as:
g ^ i ( j ) = arg max g i ( j ) ( Q 2 n ( m ^ i ( j ) , m ¯ i ( j ) ) ) .
where Q 2 n is the Quaternion-based index [1,53] applied as an optimizer parameter used both for evaluation and fusion procedures, which are performed over all labeled segments in order to obtain the optimal gain values. arg max denotes the maximization solver. Notice that the Q 2 n formula (Equation (18)) is described in the “Quality Metrics” (Section 4.4). Equation (14) indicates that all pixels belonging to the same segment have the same injection gain ( g ^ ).
Next, for each spectral band, the set of gain values { g i ( j ) } i = 1 N g are spatially concatenated according to their label index to obtain meaningful gain maps:
G ^ i = c a t ( g ^ i ( 1 ) , . . . , g ^ i ( L ) ) .
The optimal injection coefficients are used to modulate the detail segments before inferences into the corresponding expanded MS segments.
The high-resolution MS segments are synthesized as the spatially arranged MS fused regions, { m s ^ i ( j ) } i = 1 N g , by applying Equation (13) locally:
M S ^ i = c a t ( m s ^ i ( 1 ) , . . . , m s ^ i ( L ) ) ,
and therefore Equation (16) can be written conventionally as follows:
M S ^ i = M S ˜ i + G ^ i D i , i = { 1 , . . , N } ,

4. Experiments and Results

The experimental results, including methods and implementation details, datasets, and quality metrics are presented and discussed in this section.

4.1. Benchmarks and Implementation Details

Both pixel- and region-based pansharpening approaches are evaluated, and the effectiveness of the proposal is demonstrated. As a yardstick, interpolation of the original MS image using a 23-tap polynomial filter (EXP) [22] is utilized for the tests. Among the CS-based techniques, the following methods are considered: GSA [17], BDSD [19], PRACS [18], GSA-BPT [45], and BT-H [14]. With respect to MRA methods, AWLP [29], MF-HG [28], the GLP technique that adopts additive injection schemes (GLP-MTF) [22], multiplicative injection (GLP-SDM) [23], projective injection schemes (GLP-OLS) [25], cluster injection (GLP-BPT) [45], and our proposed model GraphGLP are used in comparative experiments. For VO-based techniques, CS-D [39], PWMBF [36], TV [38], and FE-HPM [40] are applied. Deep learning approaches include PNN [33,43] and CNN [40] for comparison purposes. For all the mentioned pansharpening methods, both reduced, in agreement with the Wald [44] and Zou protocols [54], and full resolution, in accordance with the QNR protocol [55], were tested and experimentally evaluated.

4.2. Implementation Details

The number of clusters N c is equal to 500 generated from the graph-based superpixel method. The compactness parameter (m) is set to 10 4 . The value of the threshold γ is set to 400 for the two datasets, respectively. The resulting regions for the graph-based superpixels are equal to: N s = 205 , N g = 150 , and γ = 50 for the Captown-GE1 dataset; and N s = 200 , N g = 120 , and γ = 100 for the Adelaide-WV3 dataset.

4.3. Datasets

Two data sets were considered for the experiments in this study, in which two scenarios were acquired by two different sensors. The image sets were downloaded from the following website: “https://apollomapping.com/” (accessed on 2 April 2025). The data present two cases in terms of the number of spectral bands that characterize the original MS images (4 or 8 bands) and ensure diversity of the observed areas (e.g., vegetation and trees, roads, and buildings). The assessment of the two datasets is provided at reduced- and full-scale in agreement with the “quality metrics” described in Section 4.4. Details of these image sets are presented below.
1.
Captown dataset: This dataset is displayed in Figure 6a,b. It represents a sub-urban area of the city of Capetown (South Africa). It is collected by the GeoEye-1 (GE-1) satellite that works in the visible and near-infrared (NIR) spectrum range. It is composed of a PAN and a four-band MS image (red, green, blue, and NIR). The spatial sampling interval (SSI) is 2 m for MS and 0.5 m for PAN images. The size of the dataset is 256 × 256 × 4 for an MS image and 1024 × 1024 for PAN. The resolution ratio is equal to four. The radiometric resolution is 16 bits. The characteristics of the Captown set include buildings, vegetation, and grass soil.
2.
Adelaide dataset: This dataset is exhibited in Figure 6c,d. It displays an urban area of Adelaide city (Australia). It is acquired by the WorldView-3 (WV-3) satellite, working in the visible and near-infrared (NIR) spectrum range. The data contain an eight-band MS image covering four standard (blue, green, red, and NIR1) and four new (coastal, yellow, red edge, and NIR2) bands and a PAN image, with a 1.2 m and a 0.3 m SSI, respectively. The size of the PAN image is 1024 × 1024 pixels and that of the MS image is 256 × 256 × 8 pixels. The resolution ratio is equal to four. The radiometric resolution is 16 bits. The characteristics of the Adelaide set include dense buildings, trees, soil, and some cars.

4.4. Quality Metrics

The proposed approach was evaluated by carrying out the reduced [44] and the full-resolution [55] assessment procedures. The Spectral Angle Mapper (SAM) [56], the Erreur Relative Globale Adimensionnelle de Synthèse (ERGAS) [44], the Spatial Correlation Coefficient (SCC) [54], the Structural Similarity Index Measure (SSIM) [57], the Peak Signal-to-Noise Ratio (PSNR) [58] and the Quaternion-based index ( Q 2 n ) [55] are selected for the reduced scale assessment procedure and the Quality with no Reference (QNR) [55] for the full-scale quality assessment.
To quantify the quality of the fused products, we computed seven distinct measures:
1.
The Quaternion-based index, Q 2 n , is a global multiband extension [1] of the Universal image quality index [53]. The quality metric was first introduced for evaluating a four-band pansharpened MS image, and extended to MS images with 2 n bands. Taking an image with N spectral bands, a pixel is defined by a hypercomplex number having one real part and N 1 imaginary parts. Let us consider two hypercomplex numbers z = z ( u , v ) and z ^ = z ^ ( u , v ) , which represent the reference and the fused spectral images at pixel coordinates ( u , v ) . The Q 2 n index is written as follows:
Q 2 n = | σ z z ^ | σ z σ z ^ 2 σ z σ z ^ σ z 2 + σ z ^ 2 2 | z ¯ | | z ^ ¯ | | z ¯ | 2 | z ^ ¯ | 2
where σ z z ^ denotes the covariance between the two hypercomplex number z and z ^ . σ z (similarly σ z ^ ) denotes the standard deviation of the hypercomplex number z (similarly z ^ ). | z | (similarly | z ^ | ) denotes the mean value of the hypercomplex number z (similarly z ^ ). The Q 2 n metric is viewed as the product of three terms (as their ranking), which are the hypercomplex correlation coefficient, the contrast changes, and the mean bias between z and z ^ .
2.
The Spectral Angle Mapper (SAM) is a global spectral dissimilarity or distortion index. It was proposed for discriminating the materials based on their reflectance spectra [56]. Given two vectors, v = v 1 , , v N and v ^ = v ^ 1 , , v ^ N , which define the reference spectral pixel vector and the corresponding fused spectral pixel, respectively. The SAM index denotes the absolute value of the spectral angles between the two vectors, given by:
S A M = arccos ( v , v ^ v 2 . v ^ 2 )
The SAM metric is expressed in degrees (), and its optimal value is equal to zero if the reference and the fused vectors are spectrally identical.
3.
The Erreur Relative Globale Adimensionnelle de Synthèse (ERGAS) [44] is a normalized dissimilarity index estimated between the reference and the multi-band fused images. It is defined as:
E R G A S = 100 d h d l 1 N n N R M S E ( n ) 2 μ ( n )
where d h d l is the ratio between the pixel size of PAN and MS (generally equal to 1 4 for many sensors in the pansharpening case study). μ ( n ) is the mean (i.e., average) of the n t h band of the reference image. N is the number of bands. R M S E is the root mean square error of the n t h band. Low values of ERGAS indicate high similarity between the reference and the fused images.
4.
The Spatial Correlation Coefficient (SCC) is used for assessing the similarity between the spatial details of the pansharpened and the reference images. It uses a high-pass filter (a Laplacian filter) to extract the high-frequency information from both the reference and fused images, and computes the correlation coefficient between them. It is usually known as the Zou protocol [54]. The kernel of the Laplacian filter involved in this procedure is given by:
F = 1 1 1 1 8 1 1 1 1
and the Pearson correlation coefficient is computed as:
C C = i = 1 w j = 1 h ( X i j μ X ) ( Y i j μ Y ) i = 1 w j = 1 h ( X i j μ X ) 2 ( Y i j μ Y ) 2
where X and Y are the reference and fused images in pansharpening, respectively. w and h are the width and height of the image. μ represents the average of the image X (similarly the image Y). A high SCC value indicates that most of the spatial details of the PAN image are injected into MS bands during the fusion process.
5.
The Structural Similarity Index Measure (SSIM) [57] estimates the overall fusion quality by computing the first-order statistics (the mean, variance, and covariance) of the reference and its corresponding fused MS image. The SSIM index is the combination of three contrast modules, called brightness, contrast, and structure, given by:
S S I M ( X , Y ) = l ( X , Y ) α c ( X , Y ) β s ( X , Y ) γ
where
l ( X , Y ) = 2 μ x μ y + c 1 μ x 2 + μ y 2 + c 1 c ( X , Y ) = 2 δ x δ y + c 2 δ x 2 + δ y 2 + c 2 s ( X , Y ) = 2 δ x y + c 3 μ x μ y + c 3
and μ x , μ y , δ x 2 , δ y 2 , and δ x y are the means, variances, and covariances of X and Y, respectively.
The optimal value of SSIM is equal to one; a high value indicates high similarity between the images.
6.
The Peak Signal-to-Noise Ratio (PSNR) is originally introduced to estimate the noise quantity in images [57]. Then, it is used to calculate the Mean Squared Error (MSE) between the original and distorted images. It is expressed as:
P S N R = 10 log 10 M A X I 2 M S E
where M A X I denotes the maximum value representing the color of pixel of an image I.
In the context of image pansharpening, MSE is defined as:
M S E = 1 m n i = 0 m 1 j = 0 n 1 M S ( i , j ) M S ^ 2
where M S and M S ^ are the reference and pansharpened images of size m × n .
The high value of PSNR indicates that the reconstructed image presents less distortion compared to the reference MS image.
7.
The Quality with no Reference (QNR) protocol [55] is proposed for the assessment of fused images at full-resolution at the scale of the PAN band. This metric allows tackling the problem of unavailability of the true MS image for a full-scale PAN image.The QNR index is computed as the product of two normalized measurements of the spectral and spatial consistencies. In more detail, two distortions indexes, called the spectral distortion D λ and the spatial distortion D s , are combined to define the QNR index, as follows:
Q N R = ( 1 D λ ) α ( 1 D s ) β
where α = β = 1 (for most applications).
The optimal value of Q N R is equal to one.

4.5. Results and Discussion

4.5.1. General Fusion Results

The fusion results of the pansharpening methods on the Captown and Adelaide data at reduced resolution are displayed in Figure 7 and Figure 8. An enlarged area is displayed at full-scale evaluation (cf. Figure 9 and Figure 10). The comparative pansharpening pipelines are composed of several categories of methods: traditional, such as MRA and CS techniques (e.g., BDSD, PRACS, GSA, AWLP, MFHG, and GLP), second-generation methods based on physical models of sensors (e.g., FE-HPM, PWMBF, GLP-MTF, and TV), and de-hazing methods using radiative transfer through the atmosphere (e.g., BT-H).
The experiments evaluate the advantages of the adaptive multi-resolution pansharpening approach with injection rules using the ordinary least squares (known as regression) model and the optimized simplex method. Accordingly, the method is called GraphGLP-OLS for the first model, and GraphGLP-Simplex for the second case. The results were analysed in accordance with the statistical criteria and the spatial-spectral characteristics of the image sets to take advantage of the variability in the datasets in terms of pixel reflectances and textures [24,59]. GraphGLP-Simplex outperforms its statistical variants, GraphGLP-OLS and GLP-OLS, for most quality indexes because it takes into consideration the pixel neighbors of the segment in task. The variational-optimization (PMBF, TV, FE-HPM) and the deep learning (PNN and CNN) methods are included in quality assessment (cf. Figure 7, Figure 8, Figure 9 and Figure 10). It is clearly seen that the PNN results present a pixelization or blurring effect for both the GeoEye1 and WorldView-3 datasets. The CNN results display a high spectral quality compared to the up-sampled MS image, which could be explained by over-enhancement depending on the field of application. Analysis of the fused results confirms the diversity of the pansharpening methods which are crucial for validating the neW models. A compromise between visual analysis and quantitative assessment is established in this study, which may be explained by the fact that pansharpening can produce mathematically optimized results with respect to the computation complexity (such as in deep learning or sparse representation methods).

4.5.2. Full-Scale Image Results

The full-resolution experiments for the GeoEye-1 and WorldView-3 datasets are illustrated in Figure 9 and Figure 10, respectively. For the GeoEye-1 dataset, the results provided by GSA, PRACS, and BDSD, depicted in Figure 9b–d, suffer from spectral distortions. An unnatural color is exhibited in the red building. The MRA-based methods [AWLP and MF-HG, displayed in Figure 9f,g, respectively] provide close spatial quality. Considering the methods in which the fusion is performed at region level, such as GLP-BPT, GSA-BPT, GraphGLP-OLS, and GraphGLP-simplex, it is noted that these context-adaptive methods better preserve the spectral as well as the spatial quality when compared to the pixel-based methods. Numerical results are shown in Table 1. Considering the Adelaide dataset, the fused products of GSA, BDSD, and PRACS [Figure 10d,e] show approximately similar contrast. Moreover, it is clearly observed that the AWLP and MF-HG [Figure 10f,g] produce high-resolution MS images with satisfactory spectral preservation. Compared to the adaptive approaches, such as GLP-BPT and GLP-SDM, the proposed GraphGLP-OLS and GraphGLP-simplex provide the best fusion results. The obtained results are confirmed by the quantitative experiments reported in Table 2.
Figure 9. Enlarged close-ups of fused results at full-scale on the GeoEye-1 dataset. (a) PAN (b) MSup (c) BDSD (d) PRACS (e) GSA (f) GSA-BPT (g) AWLP (h) MFHG (i) GLP-MTF (j) GLP-SDM (k) GLP-BPT (l) GLP-OLS (m) PWMBF (n) TV (o) FE-HPM (p) BT-H (q) PNN (r) CNN (s) GraphGLP-simplex.
Figure 9. Enlarged close-ups of fused results at full-scale on the GeoEye-1 dataset. (a) PAN (b) MSup (c) BDSD (d) PRACS (e) GSA (f) GSA-BPT (g) AWLP (h) MFHG (i) GLP-MTF (j) GLP-SDM (k) GLP-BPT (l) GLP-OLS (m) PWMBF (n) TV (o) FE-HPM (p) BT-H (q) PNN (r) CNN (s) GraphGLP-simplex.
Sensors 25 04992 g009

4.5.3. Quantitative Performances

Table 1 and Table 2 summarize the fused results of the comparative evaluation of the proposed approach and the benchmarked pansharpening methods according to Wald’s protocol [e.g., ERGAS, SAM, SSIM, Q 2 n and PSNR], the Zou protocol (SCC), and the QNR protocol, for the two datasets. The proposed approach obtains satisfactory values in quality assessment. Furthermore, it is clearly shown that the performance of the fused results obtained with the regression scheme applied directly to the superpixels is lower than that obtained with the simplex-guided graph-based superpixel entities. Considering the level of detail injection, the fused products obtained at the region levels produce better results than those using an injection scheme at pixel level [23]. The proposed model gives enhanced fused results since it takes into account the spatial structural information defined by the graph structure of the data. The main limitation of the proposed approach is estimation of the injection weights at reduced resolution, in which the scale property is adopted according to Wald’s protocol [44]. Moreover, the injection coefficients are linearly estimated on segments, such that non-linear computation of gains could be advantageous for tracking cars or tackling the non-overlapping of MS bands [60].
Figure 10. Enlarged close-ups of fused results at full-scale on the World-view-3 dataset. (a) PAN (b) MS up-sampled (c) BDSD (d) PRACS (e) GSA (f) GSA-BPT (g) AWLP (h) MFHG (i) GLP-MTF (j) GLP-SDM (k) GLP-BPT (l) GLP-OLS (m) PWMBF (n) TV (o) FE-HPM (p) BT-H (q) PNN (r) CNN (s) GraphGLP-simplex.
Figure 10. Enlarged close-ups of fused results at full-scale on the World-view-3 dataset. (a) PAN (b) MS up-sampled (c) BDSD (d) PRACS (e) GSA (f) GSA-BPT (g) AWLP (h) MFHG (i) GLP-MTF (j) GLP-SDM (k) GLP-BPT (l) GLP-OLS (m) PWMBF (n) TV (o) FE-HPM (p) BT-H (q) PNN (r) CNN (s) GraphGLP-simplex.
Sensors 25 04992 g010

5. Conclusions

A novel region-based pansharpening method is proposed. The multispectral image is represented by a Region Adjacency Graph (RAG) structure. It is obtained from an initial over-segmentation by means of the SLIC method. The gains estimation and fusion phases are both performed over ensemble connected graph-based superpixels. The application of our proposed method involves first performing computation of the spectrally variant injection coefficients by means of a Simplex optimization model; then, introduction of a statistical ordinary least squares model during the fusion task.The fusion results are evaluated in comparison with traditional, variational optimization and deep learning-based pansharpening methods. The advantages of the proposed technique are analyzed and validated using two very-high-resolution remote sensing images collected by the GeoEye-1 and WorldView-3 satellites. Forthcoming research will consider application of the combination of gradient and graph-based superpixel segmentation to the pansharpening. Furthermore, estimation of object localization in the scene (e.g., cars, trees, or buildings) will require pixels spectra analysis.

Funding

The Author thanks the Editorial Office of Sensors Journal.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. All geospatial images involved in the experiments have been authorized by Mr. B. Adam McCarty.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AWLPAdditive wavelet luminance proportion
BDSDBand-dependent spatial detail
BT-HBrovey (BT) with haze correction
CSComponent substitution
CNNPansharpening based on convolutional neural network
CS-DPansharpening based on sparse representation of injected details
ERGASErreur Relative Globale Adimensionnelle de Synthèse
EXPExpanded multispectral image
FE-HPMFilter estimation based on a semi-blind deconvolution framework and HPM model
GLPGeneralized Laplacian pyramid
GLP-BPTGLP based on binary partition tree
GLP-MTFGLP tailored to the MTF
GLP-SDMGLP based on spectral distortion minimization
GLP-OLSGLP based on ordinary least squares model
GraphGLPPansharpening based on LP and graph model
GSAGram–Schmidt adaptive algorithm
GSA-BPTGSA with binary partition tree
MF-HGHalf-gradient morphological operator
MRAMulti-resolution analysis
MMsModel-based methods
MSMultispectral image
MTFModulation transfer function
NIRNear-infrared band
PANPanchromatic image
PRACSPartial replacement adaptive CS
PSNRPeak signal-to-noise ratio
PNNPansharpening based on neural networks
PWMBFModel-based fusion using PCA and wavelets
RAGRegion adjacency graph
SAMSpectral angle mapper
SCCSpatial correlation coefficient
SLICSimple linear iterative clustering
SSIMStructural similarity index measure
TVPansharpening based on total variation
Q 2 n Quaternion-based index
QNRQuality with no reference

References

  1. Vivone, G.; Alparone, L.; Chanussot, J.; Dalla Mura, M.; Garzelli, A.; Licciardi, G.A.; Restaino, R.; Wald, L. A Critical Comparison Among Pansharpening Algorithms. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2565–2586. [Google Scholar] [CrossRef]
  2. Vivone, G.; Mura, M.D.; Garzelli, A.; Pacifici, F. A Benchmarking Protocol for Pansharpening: Dataset, Preprocessing, and Quality Assessment. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6102–6118. [Google Scholar] [CrossRef]
  3. Vivone, G.; Mura, M.D.; Garzelli, A.; Restaino, R.; Scarpa, G.; Ulfarsson, M.O.; Alparone, L.; Chanussot, J. A New Benchmark Based on Recent Advances in Multispectral Pansharpening: Revisiting Pansharpening with Classical and Emerging Pansharpening Methods. IEEE Geosci. Remote Sens. Mag. 2021, 9, 53–81. [Google Scholar] [CrossRef]
  4. Hallabia, H. Advanced Trends in Optical Remotely Sensed Data Fusion: Pansharpening Case Study. Iris J. Astron. Satell. Commun. 2025, 1, 1–3. [Google Scholar]
  5. Alparone, L.; Garzelli, A. Benchmarking of Multispectral Pansharpening: Reproducibility, Assessment, and Meta-Analysis. J. Imaging 2025, 11, 1. [Google Scholar] [CrossRef]
  6. Hallabia, H.; Kallel, A.; Hamida, A.B.; Hégarat-Mascle, S.L. High Spectral Quality Pansharpening Approach based on MTF-matched Filter Banks. Multidimens. Syst. Signal Process. 2016, 27, 831–861. [Google Scholar] [CrossRef]
  7. Hallabia, H.; Hamam, H.; Hamida, A.B. An Optimal Use of SCE-UA Method Cooperated With Superpixel Segmentation for Pansharpening. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1620–1624. [Google Scholar] [CrossRef]
  8. Meng, X.; Shen, H.; Li, H.; Zhang, L.; Fu, R. Review of the pansharpening methods for remote sensing images based on the idea of meta-analysis: Practical discussion and challenges. Inf. Fusion 2019, 46, 102–113. [Google Scholar] [CrossRef]
  9. Benzenati, T.; Kessentini, Y.; Kallel, A.; Hallabia, H. Generalized Laplacian Pyramid Pan-Sharpening Gain Injection Prediction Based on CNN. IEEE Geosci. Remote Sens. Lett. 2019, 17, 651–655. [Google Scholar] [CrossRef]
  10. Restaino, R.; Vivone, G.; Addesso, P.; Chanussot, J. A Pansharpening Approach Based on Multiple Linear Regression Estimation of Injection Coefficients. IEEE Geosci. Remote Sens. Lett. 2019, 17, 102–106. [Google Scholar] [CrossRef]
  11. Vivone, G.; Marano, S.; Chanussot, J. Pansharpening: Context-Based Generalized Laplacian Pyramids by Robust Regression. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6152–6167. [Google Scholar] [CrossRef]
  12. Addesso, P.; Vivone, G.; Restaino, R.; Chanussot, J. A Data-Driven Model-Based Regression Applied to Panchromatic Sharpening. IEEE Trans. Image Process. 2020, 29, 7779–7794. [Google Scholar] [CrossRef]
  13. Gillespie, A.R.; Kahle, A.B.; Walker, R.E. Color enhancement of highly correlated images. II. Channel ratio and “chromaticity” transformation techniques. Remote Sens. Environ. 1987, 22, 343–365. [Google Scholar] [CrossRef]
  14. Lolli, S.; Alparone, L.; Garzelli, A.; Vivone, G. Haze Correction for Contrast-Based Multispectral Pansharpening. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2255–2259. [Google Scholar] [CrossRef]
  15. Chavez, P.S.; Sides, S.C.; Anderson, J.A. Comparison of Three Different Methods to Merge Multiresolution and Multispectral Data: Landsat TM and SPOT Panchromatic. Photogramm. Eng. Remote Sens. 1991, 57, 295–303. [Google Scholar]
  16. Tu, T.M.; Su, S.C.; Shyu, H.C.; Huang, P.S. A new look at IHS-like image fusion methods. Inf. Fusion 2001, 2, 177–186. [Google Scholar] [CrossRef]
  17. Aiazzi, B.; Baronti, S.; Selva, M. Improving Component Substitution Pansharpening Through Multivariate Regression of MS +Pan Data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  18. Choi, J.; Yu, K.; Kim, Y. A New Adaptive Component-Substitution-Based Satellite Image Fusion by Using Partial Replacement. IEEE Trans. Geosci. Remote Sens. 2011, 49, 295–309. [Google Scholar] [CrossRef]
  19. Garzelli, A.; Nencini, F.; Capobianco, L. Optimal MMSE Pan Sharpening of Very High Resolution Multispectral Images. IEEE Trans. Geosci. Remote Sens. 2008, 46, 228–236. [Google Scholar] [CrossRef]
  20. Ghahremani, M.; Ghassemian, H. Nonlinear IHS: A Promising Method for Pan-Sharpening. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1606–1610. [Google Scholar] [CrossRef]
  21. Khan, M.; Chanussot, J.; Condat, L.; Montavert, A. Indusion: Fusion of multispectral and panchromatic images using the induction scaling technique. IEEE Geosci. Remote Sens. Lett. 2008, 5, 98–102. [Google Scholar] [CrossRef]
  22. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A. Context-driven fusion of high spatial and spectral resolution images based on oversampled multiresolution analysis. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2300–2312. [Google Scholar] [CrossRef]
  23. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. MTF-tailored Multiscale Fusion of High-resolution MS and Pan Imagery. Photogramm. Eng. Remote Sens. 2006, 72, 591–596. [Google Scholar] [CrossRef]
  24. Hallabia, H.; Hamam, H.; Hamida, A.B. A context-driven pansharpening method using superpixel based texture analysis. Int. J. Image Data Fusion 2021, 12, 1–22. [Google Scholar] [CrossRef]
  25. Vivone, G.; Restaino, R.; Chanussot, J. A Regression-Based High-Pass Modulation Pansharpening Approach. IEEE Trans. Geosci. Remote Sens. 2018, 56, 984–996. [Google Scholar] [CrossRef]
  26. Vivone, G.; Restaino, R.; Chanussot, J. Full Scale Regression-Based Injection Coefficients for Panchromatic Sharpening. IEEE Trans. Image Process. 2018, 27, 3418–3431. [Google Scholar] [CrossRef] [PubMed]
  27. Nunez, J.; Otazu, X.; Fors, O.; Prades, A.; Pala, V.; Arbiol, R. Multiresolution-based image fusion with additive wavelet decomposition. IEEE Trans. Geosci. Remote Sens. 1999, 37, 1204–1211. [Google Scholar] [CrossRef]
  28. Restaino, R.; Vivone, G.; Mura, M.D.; Chanussot, J. Fusion of Multispectral and Panchromatic Images Based on Morphological Operators. IEEE Trans. Image Process. 2016, 25, 2882–2895. [Google Scholar] [CrossRef] [PubMed]
  29. Otazu, X.; González-Audícana, M.; Fors, O.; Nunez, J. Introduction of sensor spectral response into image fusion methods. Application to wavelet-based methods. IEEE Trans. Geosci. Remote Sens. 2005, 43, 2376–2385. [Google Scholar] [CrossRef]
  30. Kallel, A. MTF-Adjusted Pansharpening Approach Based on Coupled Multiresolution Decompositions. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3124–3145. [Google Scholar] [CrossRef]
  31. Liao, W.; Huang, X.; Coillie, F.V.; Gautama, S.; Pižurica, A.; Philips, W.; Liu, H.; Zhu, T.; Shimoni, M.; Moser, G.; et al. Processing of Multiresolution Thermal Hyperspectral and Digital Color Data: Outcome of the 2014 IEEE GRSS Data Fusion Contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2984–2996. [Google Scholar] [CrossRef]
  32. He, X.; Condat, L.; Bioucas-Dias, J.; Chanussot, J.; Xia, J. A New Pansharpening Method Based on Spatial and Spectral Sparsity Priors. IEEE Trans. Image Process. 2014, 23, 4160–4174. [Google Scholar] [CrossRef] [PubMed]
  33. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. CNN-based pansharpening of multi-resolution remote-sensing images. In Proceedings of the 2017 Joint Urban Remote Sensing Event (JURSE), Dubai, United Arab Emirates, 6–8 March 2017; pp. 1–4. [Google Scholar]
  34. Wei, Y.; Yuan, Q.; Shen, H.; Zhang, L. Boosting the Accuracy of Multispectral Image Pansharpening by Learning a Deep Residual Network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1795–1799. [Google Scholar] [CrossRef]
  35. Vivone, G.; Deng, L.J.; Deng, S.; Hong, D.; Jiang, M.; Li, C.; Li, W.; Shen, H.; Wu, X.; Xiao, J.L.; et al. Deep Learning in Remote Sensing Image Fusion: Methods, protocols, data, and future perspectives. IEEE Geosci. Remote Sens. Mag. 2025, 13, 269–310. [Google Scholar] [CrossRef]
  36. Palsson, F.; Sveinsson, J.R.; Ulfarsson, M.O.; Benediktsson, J.A. Model-Based Fusion of Multi- and Hyperspectral Images Using PCA and Wavelets. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2652–2663. [Google Scholar] [CrossRef]
  37. Vivone, G.; Restaino, R.; Mura, M.D.; Chanussot, J. Multi-band semiblind deconvolution for pansharpening applications. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 41–44. [Google Scholar]
  38. Palsson, F.; Sveinsson, J.R.; Ulfarsson, M.O. A New Pansharpening Algorithm Based on Total Variation. IEEE Geosci. Remote Sens. Lett. 2014, 11, 318–322. [Google Scholar] [CrossRef]
  39. Vicinanza, M.R.; Restaino, R.; Vivone, G.; Mura, M.D.; Chanussot, J. A Pansharpening Method Based on the Sparse Representation of Injected Details. IEEE Geosci. Remote Sens. Lett. 2015, 12, 180–184. [Google Scholar] [CrossRef]
  40. Scarpa, G.; Vitale, S.; Cozzolino, D. Target-Adaptive CNN-Based Pansharpening. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5443–5457. [Google Scholar] [CrossRef]
  41. Ma, J.; Yu, W.; Chen, C.; Liang, P.; Guo, X.; Jiang, J. Pan-GAN: An unsupervised pan-sharpening method for remote sensing image fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  42. Azarang, A.; Manoochehri, H.E.; Kehtarnavaz, N. Convolutional Autoencoder-Based Multispectral Image Fusion. IEEE Access 2019, 7, 35673–35683. [Google Scholar] [CrossRef]
  43. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by Convolutional Neural Networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef]
  44. Wald, L.; Ranchin, T.; Mangolini, M. Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images. Photogramm. Eng. Remote Sens. 1997, 63, 691–699. [Google Scholar]
  45. Restaino, R.; Mura, M.D.; Vivone, G.; Chanussot, J. Context-Adaptive Pansharpening Based on Image Segmentation. IEEE Trans. Geosci. Remote Sens. 2017, 55, 753–766. [Google Scholar] [CrossRef]
  46. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef]
  47. Aiazzi, B.; Baronti, S.; Lotti, F.; Selva, M. A Comparison Between Global and Context-Adaptive Pansharpening of Multispectral Images. IEEE Geosci. Remote Sens. Lett. 2009, 6, 302–306. [Google Scholar] [CrossRef]
  48. Bampis, C.G.; Maragos, P.; Bovik, A.C. Graph-Driven Diffusion and Random Walk Schemes for Image Segmentation. IEEE Trans. Image Process. 2017, 26, 35–50. [Google Scholar] [CrossRef]
  49. Xiao, J.L.; Huang, T.Z.; Deng, L.J.; Wu, Z.C.; Vivone, G. A New Context-Aware Details Injection Fidelity with Adaptive Coefficients Estimation for Variational Pansharpening. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  50. Selva, M.; Santurri, L.; Baronti, S. On the Use of the Expanded Image in Quality Assessment of Pansharpened Images. IEEE Geosci. Remote Sens. Lett. 2018, 15, 320–324. [Google Scholar] [CrossRef]
  51. Tu, T.M.; Huang, P.S.; Hung, C.L.; Chang, C.P. A fast intensity-hue-saturation fusion technique with spectral adjustment for IKONOS imagery. IEEE Geosci. Remote Sens. Lett. 2004, 1, 309–312. [Google Scholar] [CrossRef]
  52. Liu, J.G. Smoothing Filter-based Intensity Modulation: A spectral preserve image fusion technique for improving spatial details. Int. J. Remote Sens. 2000, 21, 3461–3472. [Google Scholar] [CrossRef]
  53. Wang, Z.; Bovik, A. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  54. Zhou, J.; Civco, D.L.; Silander, J.A. A wavelet transform method to merge Landsat TM and SPOT panchromatic data. Int. J. Remote Sens. 1998, 19, 743–757. [Google Scholar] [CrossRef]
  55. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and panchromatic data fusion assessment without reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar] [CrossRef]
  56. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the Spectral Angle Mapper (SAM) algorithm. In Proceedings of the JPL, Summaries of the Third Annual JPL Airborne Geoscience Workshop, Pasadena, CA, USA, 1–5 June 1992; pp. 147–149. [Google Scholar]
  57. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  58. Huynh-Thu, Q.; Ghanbari, M. Scope of validity of PSNR in image/video quality assessment. Electron. Lett. 2008, 44, 800–801. [Google Scholar] [CrossRef]
  59. Hallabia, H.; Hamam, H. A Graph-Based Textural Superpixel Segmentation Method for Pansharpening Application. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 2640–2643. [Google Scholar]
  60. Hallabia, H.; Hamam, H. An Enhanced Pansharpening Approach Based on Second-Order Polynomial Regression. In Proceedings of the 2021 International Wireless Communications and Mobile Computing (IWCMC), Harbin City, China, 28 June–2 July 2021; pp. 1489–1493. [Google Scholar]
Figure 1. (a) Image partitioning and (b) Region Adjacency Graph (RAG). As example, the image is represented by seven numbered regions and delimited by ten edges (borders). The corresponding graph-based structure, RAG, is represented by the nodes relationships estimated between neighbour regions or segments.
Figure 1. (a) Image partitioning and (b) Region Adjacency Graph (RAG). As example, the image is represented by seven numbered regions and delimited by ten edges (borders). The corresponding graph-based structure, RAG, is represented by the nodes relationships estimated between neighbour regions or segments.
Sensors 25 04992 g001
Figure 4. Illustration of generation of the guidance injection map. The first and second rows display the described process on GeoEye-1 and Adelaide datasets, respectively. The first column displays the expanded MS image. The second and third column show the initial RAG based on SLIC of the MS image and the segmented image, respectively. The final regions are overlaid on the source image.
Figure 4. Illustration of generation of the guidance injection map. The first and second rows display the described process on GeoEye-1 and Adelaide datasets, respectively. The first column displays the expanded MS image. The second and third column show the initial RAG based on SLIC of the MS image and the segmented image, respectively. The final regions are overlaid on the source image.
Sensors 25 04992 g004
Figure 5. Threshold computing based on the RAG concept for N c = 300 .
Figure 5. Threshold computing based on the RAG concept for N c = 300 .
Sensors 25 04992 g005
Figure 6. Datasets. (a,b) Captown dataset, (c,d) Adelaide dataset. The first column presents the expanded MS image, and the second column displays the PAN image.
Figure 6. Datasets. (a,b) Captown dataset, (c,d) Adelaide dataset. The first column presents the expanded MS image, and the second column displays the PAN image.
Sensors 25 04992 g006
Figure 7. Fused results at reduced-scale on GeoEye-1 dataset. (a) PAN (b) MSup (c) BDSD (d) PRACS (e) GSA (f) GSA-BPT (g) AWLP (h) MFHG (i) GLP-MTF (j) GLP-SDM (k) GLP-BPT (l) GLP-OLS (m) PWMBF (n) TV (o) FE-HPM (p) BT-H (q) PNN (r) CNN (s) GraphGLP-simplex.
Figure 7. Fused results at reduced-scale on GeoEye-1 dataset. (a) PAN (b) MSup (c) BDSD (d) PRACS (e) GSA (f) GSA-BPT (g) AWLP (h) MFHG (i) GLP-MTF (j) GLP-SDM (k) GLP-BPT (l) GLP-OLS (m) PWMBF (n) TV (o) FE-HPM (p) BT-H (q) PNN (r) CNN (s) GraphGLP-simplex.
Sensors 25 04992 g007
Figure 8. Fused results at reduced-scale on Worldview-3 dataset. (a) PAN (b) MS up-sampled (c) BDSD (d) PRACS (e) GSA (f) GSA-BPT (g) AWLP (h) MFHG (i) GLP-MTF (j) GLP-SDM (k) GLP-BPT (l) GLP-OLS (m) PWMBF (n) TV (o) FE-HPM (p) BT-H (q) PNN (r) CNN (s) GraphGLP-simplex.
Figure 8. Fused results at reduced-scale on Worldview-3 dataset. (a) PAN (b) MS up-sampled (c) BDSD (d) PRACS (e) GSA (f) GSA-BPT (g) AWLP (h) MFHG (i) GLP-MTF (j) GLP-SDM (k) GLP-BPT (l) GLP-OLS (m) PWMBF (n) TV (o) FE-HPM (p) BT-H (q) PNN (r) CNN (s) GraphGLP-simplex.
Sensors 25 04992 g008
Table 1. Performance indices of pansharpening algorithms for Captown images. Bold: the best.
Table 1. Performance indices of pansharpening algorithms for Captown images. Bold: the best.
MethodQ4SAMERGASSCCSSIMPSNRQNR
EXP0.7419.47410.3890.2430.734−39.3180.921
BDSD0.8798.7887.5260.6360.850-36.9850.926
PRACS0.8399.3808.5490.5860.819−37.7510.931
GSA0.8539.4567.8880.6370.837−37.4410.881
GSA-BPT0.9018.4986.8870.6570.884−36.0470.897
AWLP0.8988.3777.3990.6610.892−36.0850.919
MF-HG0.9118.7626.6560.6760.874−36.2220.846
CS-D0.86112.5577.9610.6610.810−38.3750.675
GLP-MTF0.9018.6896.8630.6580.880−36.2610.839
GLP-SDM0.9128.8126.4650.6820.875−36.0950.858
GLP-OLS0.9059.1846.7250.6520.869−36.5040.835
GLP-BPT0.9028.4926.8230.6560.886−35.9740.935
PWMBF0.87410.1337.4890.6450.836−37.5500.764
TV0.8629.3717.8740.4130.809−37.9100.942
FE-HPM0.9148.8046.4190.6810.876−36.0660.839
BT-H0.9108.1686.5910.6900.887−35.8770.913
PNN0.8898.0166.9050.6240.873−36.0130.919
CNN0.8998.2856.6840.6440.872−35.9920.941
GraphGLP-OLS0.9038.9996.7490.6550.873−36.3920.935
GraphGLP-Simplex0.9248.9546.3810.6650.895−36.0610.935
Table 2. Performance indices of pansharpening algorithms for Adelaide images. Bold: the best.
Table 2. Performance indices of pansharpening algorithms for Adelaide images. Bold: the best.
MethodQ8SAMERGASSCCSSIMPSNRQNR
EXP0.7068.2058.1210.2490.601−43.7130.915
BDSD0.8457.9576.1610.6460.770−41.5060.943
PRACS0.8478.1745.8710.6650.794−41.1670.917
GSA0.8388.1785.9700.6850.792−41.3260.884
GSA-BPT0.8857.8115.2660.7090.834−40.2100.847
AWLP0.8757.9505.4660.7040.820−40.4100.890
MF-HG0.8808.0405.5120.6980.817−40.6610.835
CS-D0.80311.5206.6280.7000.734−42.4860.660
GLP-MTF0.8807.9565.3910.7030.822−40.4570.838
GLP-SDM0.8818.0285.3550.7050.820−40.4410.845
GLP-OLS0.8878.1745.2500.6980.830−40.3640.850
GLP-BPT0.8867.7955.2360.7080.835−40.1640.879
PWMBF0.8558.9855.7880.7010.796−41.2390.753
TV0.8957.8535.2060.7080.839−40.2030.949
FE-HPM0.8828.0305.3470.7050.821−40.4290.839
BT-H0.8948.0495.2250.7130.835−40.1970.816
PNN0.8608.6117.0370.5400.785−42.1370.908
CNN0.7638.9107.2830.4030.651−43.0480.920
GraphGLP-OLS0.8868.1875.2600.6970.829−40.3810.922
GraphGLP-Simplex0.9098.1834.9350.8610.720−39.7920.922
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hallabia, H. A Graph-Based Superpixel Segmentation Approach Applied to Pansharpening. Sensors 2025, 25, 4992. https://doi.org/10.3390/s25164992

AMA Style

Hallabia H. A Graph-Based Superpixel Segmentation Approach Applied to Pansharpening. Sensors. 2025; 25(16):4992. https://doi.org/10.3390/s25164992

Chicago/Turabian Style

Hallabia, Hind. 2025. "A Graph-Based Superpixel Segmentation Approach Applied to Pansharpening" Sensors 25, no. 16: 4992. https://doi.org/10.3390/s25164992

APA Style

Hallabia, H. (2025). A Graph-Based Superpixel Segmentation Approach Applied to Pansharpening. Sensors, 25(16), 4992. https://doi.org/10.3390/s25164992

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop