Next Article in Journal
Deep Learning and Hydrological Feature Constraint Strategies for Dam Detection: Global Application to Sentinel-2 Remote Sensing Imagery
Previous Article in Journal
Rapid Response to Effusive Eruptions Using Satellite Infrared Data: The March 2024 Eruption of Fernandina (Galápagos)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Technical Note

A Novel Two-Stream Network for Few-Shot Remote Sensing Image Scene Classification

1
School of Electronic Information Engineering, Beihang University, Beijing 100191, China
2
Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, Xidian University, Xi’an 710071, China
3
Joint International Research Laboratory of Intelligent Perception and Computation, Xidian University, Xi’an 710071, China
4
International Research Center for Intelligent Perception and Computation, Xidian University, Xi’an 710071, China
5
Collaborative Innovation Center of Quantum Information of Shaanxi Province, Xidian University, Xi’an 710071, China
6
School of Artificial Intelligence, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(7), 1192; https://doi.org/10.3390/rs17071192
Submission received: 29 January 2025 / Revised: 14 March 2025 / Accepted: 25 March 2025 / Published: 27 March 2025
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
Recently, remote sensing image scene classification (RSISC) has gained considerable interest from the research community. Numerous approaches have been developed to tackling this issue, with deep learning techniques standing out due to their great performance in RSISC. Nevertheless, there is a general consensus that deep learning techniques usually need a lot of labeled data to work best. Collecting sufficient labeled data usually necessitates substantial human labor and resource allocation. Hence, the significance of few-shot learning to RSISC has greatly increased. Thankfully, the recently proposed discriminative enhanced attention-based deep nearest neighbor neural network (DEADN4) method has introduced episodic training- and attention-based strategies to reduce the effect of background noise on the classification accuracy. Furthermore, DEADN4 uses deep global–local descriptors that extract both the overall features and detailed features, adjusts the loss function to distinguish between different classes better, and adds a term to make features within the same class closer together. This helps solve the problem of features within the same class being spread out and features between classes being too similar in remote sensing images. However, the DEADN4 method does not address the impact of large-scale variations in objects on RSISC. Therefore, we propose a two-stream deep nearest neighbor neural network (TSDN4) to resolve the aforementioned problem. Our framework consists of two streams: a global stream that assesses the likelihood of the whole image being associated with a particular class and a local stream that evaluates the probability of the most significant area corresponding to a particular class. The ultimate classification outcome is determined by putting together the results from both streams. Our method was evaluated across three distinct remote sensing image datasets to assess its effectiveness. To assess its performance, we compare our method with a range of advanced techniques, such as MatchingNet, RelationNet, MAML, Meta-SGD, DLA-MatchNet, DN4, DN4AM, and DEADN4, showcasing its encouraging results in addressing the challenges of few-shot RSISC.

Graphical Abstract

1. Introduction

Remote sensing images (RSIs) offer a wealth of data related to land cover targets [1,2,3,4,5,6], finding extensive applications in diverse contexts, including road identification [7], ecological surveillance [8], disaster forecasting, and numerous other domains [9,10]. Classifying RSIs into different scenes based on the collected data is a key focus of the current research [11,12,13,14,15,16,17,18], aiming to classify new images into relevant classes [19,20,21,22,23]. The objective of few-shot scene classification is to accurately categorize images into their respective classes despite limited examples [24], a capability that is vital given the challenge of obtaining sufficient labeled data in numerous practical applications. The potential for few-shot classification of RSIs is significant because of its usefulness in many areas, thereby reducing the need for field investigation and manual labeling.
In the domain of conventional image classification, deep learning has shown great effectiveness [25,26] but usually demands extensive labeled datasets to achieve the optimal training results [27,28,29,30,31]. A neural network utilizing a well-designed architecture will achieve a good performance when equipped with comprehensive foundational knowledge [32,33]. Nevertheless, not having enough labeled data can cause deep models to overfit and incomplete feature representations. In diverse remote sensing applications, acquiring tagged examples is difficult and made harder by the time-consuming task of creating labeled datasets [34]. Moreover, training deep neural networks necessitates substantial computational power, and if the settings of the hyperparameters are not suitable, the process may need to be repeated [35,36,37,38].
Hence, as for remote sensing image scene classification (RSISC), the importance of few-shot learning cannot be overstated. The basic idea of few-shot learning is to create methods that rapidly adapt to a limited set of tagged examples, mainly to give models the ability to learn fast [39,40,41]. As for few-shot learning, it is crucial that the model learns deep features that are both well separated and easily distinguishable so that it can recognize novel classes despite having only a few tagged samples. Snell et al. [42] proposed Prototypical Networks for few-shot classification, learning a metric space for classification through computation of the distance to class prototypes, demonstrating that simple designs outperform complex architectures and meta-learning. Yan et al. [43] presented a straightforward and effective approach seamlessly integrable into both meta-learning and transfer-learning frameworks, isolating confusing negatives while preserving the impact of hard negatives, adjusting the gradients by their category affinity, and reducing the effects of false negatives without extra parameters. Zeng et al. [44] proposed a novel framework for few-shot RSISC which introduced a feature-generating model and prototype calibration, employing a self-attention encoder to filter irrelevant data, expanding the support set via generated features, and refining the prototypes to improve the category feature representation. Vinyals et al. [45] presented an episodic training method that dealt with the difficulties inherent in few-shot learning. Throughout the training phase, a support set is formed by randomly picking K images from each of the C classes within the dataset, yielding a collection of C × K images. After that, N images are taken from the rest of the dataset for each of the C chosen classes to make up a query set. Each episode combines one support set with one query set, and this process is repeated until the model converges. Many few-shot models use a method that focuses on learning distances, where they calculate the similarity measure between support and query instances directly to train the classifier based on this distance. However, these methods may not fully use the network’s ability to extract good features, which will significantly affect these methods’ discriminative ability.
Large variations in the object sizes are also a difficult problem in few-shot RSISC. In RSIs, the same target may look very different in different images because of factors like the camera’s angle and the height of the sensor. As depicted in Figure 1, scenes like airplanes, oil tanks, and power plants photographed at different heights show large size differences, even within the same type of object. Furthermore, because of a scene’s characteristics, the sizes of the objects in it can differ, as shown by the river scene in the figure, which has several different parts, like small rivers and streams. From Figure 1, it can be observed that the sizes of key objects differ, occupying just a small part of the whole picture and often being around many unrelated features and objects. Obviously, it does not make sense to use one big object to match pictures that contain many small ones.
To solve this problem, Chen et al. [46] developed the discriminative enhanced attention-based deep nearest neighbor neural network (DEADN4). Initially, the DEADN4 model employs a deep local–global descriptor (DLGD) to perform classification by combining localized and global characteristics, thereby facilitating the distinction between various classes. After this, to strengthen the similarity within the same class, DEADN4 uses a special loss to improve the overall details, bringing features in the same class closer together and reducing the differences within the class. Finally, DEADN4 enhances the accuracy of classification by incorporating a cosine margin into the Softmax loss, making the differences between features from different classes clearer. However, the DEADN4 method does not solve the problem of the large size differences among the objects in the RSI.
To address the abovementioned problem, this paper proposes a two-stream deep nearest neighbor neural network (TSDN4), accomplished by constructing a global–local two-stream network. This network has two parts: the global stream, which looks at the whole image to find out whether it belongs to a certain class, and the local stream, which focuses on the most important part of the image to see whether it falls into a particular class. The final classification result is obtained by combining the results from both parts. To find the most important areas in the image, the key area localization (KAL) method is used to link the two parts mentioned above. Section 2 discusses earlier work within this domain, while Section 3 introduces our proposed method. Next, Section 4 discusses the results of our experiments and analysis. Lastly, Section 6 highlights the main conclusions from our research.

2. Related Work

The method proposed in this paper builds upon DEADN4 [46]; therefore, this section mainly focuses on introducing DEADN4. The primary objective of DEADN4 is to tackle the challenge posed by few-shot RSISC, which comprises two key components: a deep embedding module that utilizes attention mechanisms, denoted as f ψ ( · ) , and a metric module, represented by f φ ( · ) . The core function of the f ψ ( · ) module is to derive representative details from the images and generate attention maps highly pertinent to the scene class. As for every input image X, the function f ψ ( X ) produces a feature map of the dimensions H × W × d , comprising H × W DLGD instances, each with d dimensions. A DLGD consists of a deep local descriptor (DLD) and a deep global descriptor (DGD). DLDs extract features from various regions of the image, while the DGD captures the global details, serving as a useful guide to enhance the distinctiveness of each local feature.
Moreover, the f ψ ( · ) module includes a component for acquiring class-distinctive attention capabilities. This component classifies the local features into pertinent and non-pertinent segments according to the scene classification. The main goal here is to minimize the impact of background noise and highlight features that are important to the scene class. Lastly, within the f φ ( · ) module, the classification of a query image is acquired by contrasting its features with those of the support images.
To summarize, DEADN4 employs the f ψ ( · ) module to derive features pertinent to scene classification from the images, and then the f φ ( · ) module assesses the feature resemblance between the query and support images to enable RSISC.
The DEADN4 loss function is defined as follows:
L D E A D N 4 = L s + λ L c
where the overall loss L D E A D N 4 includes contributions from the central loss L c and the class loss L s , with λ as a scalar that adjusts the influence of L c .
The central loss term L c in Equation (1) can be described as follows:
L c = 1 2 i = 1 C j = 1 K f i j f c i 2 2
where C denotes the overall count of classes, K indicates the quantity of instances per class in the support set, f i j is the global feature of the j-th instance within the support set’s i-th class, and f c i denotes the mean global feature across all instances associated with the i-th class in the support set.
The class loss L s , as mentioned in Equation (1), is defined as follows:
L s = 1 C × N i = 1 C j = 1 N log p i j = 1 C × N i = 1 C j = 1 N log e f φ f ψ q j , c i m k × M e f φ f ψ q j , c i m k × M + t i e f φ f ψ ( q j ) , c t
where p i j stands for the likelihood of accurately categorizing the query image, k represents the number of nearest neighbors, m denotes the DLGD’s count, and M serves as an additional margin enhancing the resemblance cos x i , x ^ i between descriptors.

3. Methods

3.1. Architecture

To tackle the challenge of significant variations in the objects within images, this paper proposes a method for scene image classification that combines global and local metrics. As shown in Figure 2, a two-stream network that combines both global and local measurements is designed. Within this framework, the network’s blue pathway represents the global stream for measuring the overall image, calculating the resemblance between images and classes. Meanwhile, the network’s green pathway represents the local stream, which focuses on measuring a particular area within the image. It emphasizes the most significant object by focusing on key local areas when calculating the similarities between areas. The two-stream network proposed in this paper allows for separate measurements of the entire image and specific areas from inputs of different sizes, with their classification results eventually being combined.
The global stream uses the DEADN4 model, while the local stream employs a prototype network. To identify the most important local areas within the whole image, this paper further introduces a weakly supervised key area detection method, called KAL, as shown in the path in Figure 2. KAL, using scene-related attention maps, is a clear method for finding key local areas, connecting the two streams. It aims to precisely identify the most significant area and find the coordinates of its bounding box B = [ x 1 , x 2 , y 1 , y 2 ] , where x 1 and x 2 are the left and right edges of the area’s width, and y 1 and y 2 represent the upper and lower edges of the area’s height.

3.2. Key Area Localization

In an image, finding key areas is the most important step, connecting the global and local streams. The flowchart for the KAL strategy is displayed in Figure 3. Specifically, using the attention maps related to the scene classes from DEADN4, a bounding box is created to help classify the local stream. KAL consists of two sub-steps: vector construction and a greedy boundary search.

3.2.1. Vector Construction

The attention maps related to the scene classes can measure how important each part of an image is. So, using this attention map, we can find areas with high energy, which are the key areas. Because searching in two-dimensional space is complex, in this paper, the attention map is simplified into two one-dimensional energy vectors, one along the height and one along the width of the space, which is formulated as follows:
V w = i = 0 H M M ( i , W ) V h = j = 0 W M M ( H , j )
where M M is the attention map we obtain, M M ( i , j ) is the value located at position ( i , j ) in M M , the total height is H, and the total width is W. By summing M M along its height and width, we obtain V w R W and V h R H , respectively.
To quickly find the most important one-dimensional area in an energy vector, this paper proposes a energy-based greedy boundary search method, which is formulated as follows:
E [ 0 : W ] = i = 0 W V w ( i ) E x 1 : x 2 = i = x 1 x 2 V w ( i )
where E [ 0 : W ] represents the total energy of all of the components in the width vector V w , and E [ x 1 : x 2 ] is the sum of the energies from spatial widths x 1 to x 2 .

3.2.2. Greedy Boundary Search

This paper defines the key areas in an image as the areas that occupy the minimum space while containing no less than a certain proportion of the overall energy, specifically where E x 1 : x 2 / E [ 0 : W ] > E T r , with E T r representing a hyperparameter for the energy ratio. Based on this criterion, a greedy-like algorithm can be employed to search for the most critical areas, which is outlined in Algorithm 1. The greedy boundary search algorithm is described as follows:
Algorithm 1 Greedy Boundary Search
Require: Width vector: V w R W
Ensure: Width boundaries of the most critical area: [ x 1 , x 2 ]
x 1 0
x 2 W / 2
for i = 0 W / 2 do
  if  E x 1 : x 2 < E [ i : i + W / 2 ]  then
    x 1 i
    x 2 i + W / 2
  end if
end for
if E x 1 : x 2 / E [ 0 : W ] > E T r  then
  while  E x 1 : x 2 / E [ 0 : W ] > E T r  do
   if  V w ( x 1 + 1 ) < V w ( x 2 1 )  then
     x 1 x 1 + 1
   else
     x 2 x 2 1
   end if
  end while
else
  while  E x 1 : x 2 / E [ 0 : W ] < E T r  do
   if  V w ( x 1 1 ) > V w ( x 2 + 1 )  then
     x 1 x 1 1
   else
     x 2 x 2 + 1
   end if
  end while
end if
x 1 I w x 1 / W
x 2 I w x 2 / W
Step 1: Initialize x 1 and x 2 .
Step 2: This involves repeatedly adjusting the limits of x 1 , x 2 to make the energy E x 1 : x 2 closer to E t r . After Step 1, two states are possible: the ratio of energy in the x 1 , x 2 area to the total energy in the width direction, E x 1 : x 2 / E [ 0 : W ] , is either higher or lower than E t r . When the ratio E x 1 : x 2 / E [ 0 : W ] is higher than E t r , the x 1 , x 2 area needs to contract in the direction of the slowest energy decrease until the ratio is no longer higher than E t r . On the other hand, when the ratio E x 1 : x 2 / E [ 0 : W ] is lower than E t r , the area must expand in the direction of the fastest energy increase until it is no longer below E t r . This paper’s greedy boundary search can identify the most compact yet informative area.
Step 3: The boundary coordinates are mapped from the feature map to the input image. Since the aforementioned boundary coordinates are located on the feature map, it is necessary to acquire the corresponding coordinates on the image, represented as x i = I w x f / W , where x i is the coordinate in the width direction on the image, x f is the coordinate on the feature map, I w represents the width size of the image, and W indicates the width size of the feature map. The width boundaries of the most critical area in the image, denoted as x 1 , x 2 , can be obtained using the three steps described above. By employing the same algorithm, the height boundaries of the most crucial area, represented as y 1 , y 2 , can similarly be obtained.

3.2.3. The Two-Stream Architecture

The key areas’ positions can be obtained using the KAL module, represented by B = [ x 1 , x 2 , y 1 , y 2 ] . The classification results for the global stream are obtained from query and support images inputted into the DEADN4 model, whereas the classification scores of the local stream are generated by utilizing bilinear interpolation to feed the key areas of the query and support images into the prototype network and obtain classification scores. The prototype network is designed to utilize a mapping function that maps input images into a lower-dimensional embedding space, and a prototype representation is learned for each class. Lastly, we compute the distances between the query image’s features and the prototype representation of each class using a standard distance measure for classification purposes. In the prototype network, convolutional neural networks are typically used as the mapping function, with the average of the embedding features for all images within a class being taken as the prototype representation for that class.
Let the mapping function f ϕ : R D R F be defined to transform sample data of the dimension D into an embedding space of the dimension F. The prototype representation for each class is computed as follows:
c c k = 1 S k x x i , y y i S k f ϕ x x i
where S k represents the dataset for class k, x x denotes a sample in S k , and y y is its corresponding class. The prototype representation c c k corresponds to the mean of all of the embedding features within the support set for the respective class.
For a given query image x x q , the likelihood that the query image falls into class k can be determined as follows:
p ϕ ( y y q = k x x q ) = exp d f ϕ ( x x q ) , c c k k exp d f ϕ ( x x q ) , c c k
where f ϕ ( x q ) denotes the mapping features obtained from the query image x x q through the mapping function, d ( · ) represents the distance function, and the cosine distance is used in this paper for computation. It should be noted that since the cosine distance offers advantages such as directional consistency, scale invariance, suitability for high-dimensional data, computational efficiency, and strong robustness and the global stream also employs the cosine distance for measurement, the cosine distance is adopted here as well for consistency.
Based on the actual labels of the images and the forecasted likelihood distribution, the local stream loss function L l is formulated as the negative log-probability, expressed using the formula below:
L l = 1 C × N c = 1 C q = 1 N log p ϕ y y q = c x x q
where N signifies the overall count of samples in the query set, and C denotes the overall count of classes in the query set.

3.2.4. The Loss Function

Here is the overall loss function for the network:
L T o t a l = L D E A D N 4 + L l
where L T o t a l represents the total loss, which is the combination of L D E A D N 4 from Equation (1) and L l from Equation (8).

4. Results

4.1. Dataset Description

In the experiments of this paper, the datasets utilized are the NWPU-RESISC45 [47] dataset, the UC Merced [48] dataset, and the WHU-RS19 [49] dataset. To facilitate comparison, the partitioning of the three datasets is the same as that for DLA-MatchNet [27].

4.1.1. The NWPU-RESISC45 Dataset

The NWPU-RESISC45 dataset, originating from Northwestern Polytechnical University in China, comprises a collection of 31,500 images sourced from Google Earth. The images encompass a diverse variety of over 100 countries and territories, created through a combination of satellite imagery, aerial photography, and geographical information. The dataset includes a variety of weather conditions, scene variations, sizes, seasonal settings, and lighting conditions. The images in this dataset typically have resolutions ranging from 30 m to 0.2 m and are presented in a full color palette, including red, green, and blue. Figure 4 shows 45 distinct scene classes, with each containing a set of 700 images with dimensions of 256 × 256 . In the experimental setup, the dataset is partitioned into three parts: a training subset with 25 classes, a validation subset with 10 classes, and a testing subset with another 10 classes.

4.1.2. The UC Merced Dataset

The UC Merced dataset, introduced in 2010, is based on detailed maps provided by the United States Geological Survey National Map, showcasing a variety of areas in the USA. The RGB dataset maintains a high image resolution of 0.3 m, spanning 21 varied scene categories. Within this dataset, each of the 21 categories comprises 100 detailed land use images, with each measuring 256 × 256 dimensions, as illustrated in Figure 5. To conduct the experiments, the dataset is partitioned into three distinct sets: a training subset with 10 classes, a validation subset with 6 classes, and a testing subset with another 5 classes.

4.1.3. The WHU-RS19 Dataset

Released by Wuhan University in China, the WHU-RS19 dataset is a collection of images sourced directly from the famous Google Earth platform. The RGB dataset employs a fine resolution of 0.5 m. Displayed in Figure 6, the dataset contains a diverse collection of 19 distinct scene classes. Each of these carefully selected classes contains a minimum of 50 representative samples, with each having a resolution of 600 × 600 pixels. The dataset includes a total of 1005 scene images. In the experimental setup, the dataset is partitioned into three distinct parts: a training subset for 9 specific classes, a validation subset for 5 selected classes, and a testing subset comprising another 5 classes.

4.2. The Experimental Setting

4.2.1. The Experimental Software and Hardware Environment

The detailed specifications of the software configurations and hardware setups used in this paper are thoroughly presented in Table 1.

4.2.2. The Experimental Design

We performed experimental evaluations of solving both 5-way 1-shot and 5-way 5-shot problems across the three aforementioned datasets. In order to assess the effectiveness of our method, we benchmarked it against eight well-known few-shot learning methods: MatchingNet [45], RelationNet [50], MAML [51], Meta-SGD [52], DLA-MatchNet [27], DN4 [53], DN4AM [54], and DEADN4 [46]. Given that our method builds upon the DEADN4 framework, we benchmarked it against DEADN4, utilizing an identical embedding network for a fair evaluation. We evaluated the classification results utilizing the mean top-1 accuracy, along with the 95 % confidence intervals (CIs), as described in [54].
The global stream of our method adopts the network from DEADN4, while the local stream uses a prototype network, and both streams utilize ResNet18 models for their embedding modules. Random cropping is applied to input images, resizing them to 224 × 224 pixels, and these resized images are then subjected to enhancements such as random horizontal flipping, brightness variations, color augmentation, and contrast enhancement. In the classification module, we specify the count of nearest neighbors as k = 3 , configure the loss function hyperparameter M to 0.01 , and set the parameter E T r to 70 % . The training set comprises 300,000 scenario sets in total, where in the 5-way 1-shot experiment, each scenario set includes 5 support images and 75 query images, while in the 5-way 5-shot experiment, each scenario set includes 25 support images and 50 query images. The Adam [55] algorithm is employed during network training, and the initial learning rate is set to 0.0001. The learning rate undergoes a decay every 100,000 scenarios, and for validation, 600 randomly generated scenario sets are utilized on the validation dataset. The top-performing model’s mean top-1 accuracy is used as the current network’s training outcome, and the model achieving the highest scores is saved as the final version. During the testing phase, we randomly sample 600 scenario sets from the test set five times, calculate the average top-1 accuracy for each iteration, and determine the final testing outcome as the mean of the results obtained from these iterations. Additionally, the 95 % CI is provided.

4.3. The Experimental Results

In this paper, corresponding experiments are performed on the three aforementioned datasets, and comparisons are made with other few-shot learning methods. Table 2, Table 3 and Table 4 show the experimental results of various few-shot classification methods on the three datasets.

5. Discussion

From the results in Table 2, Table 3 and Table 4, it is evident that our method outperforms the others across all three datasets, achieving the highest average accuracy. Additionally, when compared to DEADN4, our method demonstrates improvements of varying degrees across all three datasets, particularly a significant enhancement on the UC Merced dataset, with an average increase in accuracy of 0.85 % . The UC Merced test set comprises scene classes with multi-scale characteristics, such as rivers, sparse residential areas, and tennis courts, indicating that our method has effectively addressed the multi-scale challenges in RSISC.
To visually illustrate the efficacy of the KAL method, the localization results on the NWPU-RESISC45 dataset have been visualized, as depicted in Figure 7. Based on the visual outcomes, it is clear that our method accurately locate key areas within the scenes. In the airplane scene, targets of varying sizes can be accurately localized, while objects with slight scale variations in other scenes, such as rivers, storage tanks, thermal power stations, and parking lots, are also successfully located. This, in turn, benefits the classification processes of the local stream.

6. Conclusions

This paper introduces a few-shot scene classification model utilizing a dual-stream network. Our method constructs a global–local two-stream framework, integrating global and local streams, to calculate the likelihood of a sample belonging to a class by analyzing both the entire image and its most crucial regions. The KAL method is also designed to identify the most crucial areas within the entire image. Our method utilizes a greedy strategy to quickly identify the most critical areas within the overall image while linking the two aforementioned streams. To assess our method’s performance, tests are carried out on three public RSI datasets, and the visual outcomes of the KAL method on the NWPU-RESISC45 dataset are presented. Through a qualitative analysis and ablation experiments, our method can successfully tackle the multi-scale challenge in scene classification for few-shot RSIs. In the context of few-shot RSISC, mislabeled samples significantly impact the classification accuracy, and exploring strategies to mitigate their effects presents a promising direction for future research.

Author Contributions

Methodology, Y.L. (Yaolin Lei), Y.L. (Yangyang Li) and H.M. Resources, Y.L. (Yangyang Li). Software, H.M. and Y.L. (Yangyang Li). Writing, Y.L. (Yaolin Lei). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant No. 62476209, in part by the Key Research and Development Program of Shaanxi under Grant No. 2024CY2-GJHX-18, and in part by the Natural Science Basic Research Program of Shaanxi under Grant No. 2022JC-45.

Data Availability Statement

The NWPU-RESISC45 dataset can be retrieved from [47], the UC Merced dataset accessed via [48], and the WHU-RS19 dataset obtained through [49].

Acknowledgments

The authors express their sincere gratitude to all of the reviewers and editors for the invaluable feedback provided on this paper.

Conflicts of Interest

The authors assert the absence of any conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
Nsample count per class of the query set
Ksample count per class of the support set
Cclass number
f φ ( · ) metric module
f ψ ( · ) embedding module
ca specific class
M M feature map
sconstant
c o s cosine
L c central loss of DEADN4
L s class loss of EWADN4
x 1 left boundary of the key area width
x 2 right boundary of the key area width
y 1 upper boundary of the key area height
y 2 lower boundary of the key area height
V w total of the feature map across the width axis
V h total of the feature map across the height axis
Eenergy
L D E A D N 4 total loss of DEADN4
c c prototype representation
L l local stream loss of our method
L T o a l total loss of our method
S k total loss of our method
μ constant
λ constant
pclassification probability
Hfeature map’s height
Wfeature map’s width
dsize of the DLGD
Bbounding box

References

  1. Jiang, N.; Shi, H.; Geng, J. Multi-Scale Graph-Based Feature Fusion for Few-Shot Remote Sensing Image Scene Classification. Remote Sens. 2022, 14, 5550. [Google Scholar] [CrossRef]
  2. Xing, S.; Xing, J.; Ju, J.; Hou, Q.; Ding, X. Collaborative Consistent Knowledge Distillation Framework for Remote Sensing Image Scene Classification Network. Remote Sens. 2022, 14, 5186. [Google Scholar] [CrossRef]
  3. Xiong, Y.; Xu, K.; Dou, Y.; Zhao, Y.; Gao, Z. WRMatch: Improving FixMatch with Weighted Nuclear-Norm Regularization for Few-Shot Remote Sensing Scene Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5612214. [Google Scholar]
  4. Bai, T.; Wang, H.; Wen, B. Targeted Universal Adversarial Examples for Remote Sensing. Remote Sens. 2022, 14, 5833. [Google Scholar] [CrossRef]
  5. Muhammad, U.; Hoque, M.; Wang, W.; Oussalah, M. Patch-Based Discriminative Learning for Remote Sensing Scene Classification. Remote Sens. 2022, 14, 5913. [Google Scholar] [CrossRef]
  6. Chen, X.; Zhu, G.; Liu, M. Remote Sensing Image Scene Classification with Self-Supervised Learning Based on Partially Unlabeled Datasets. Remote Sens. 2022, 14, 5838. [Google Scholar] [CrossRef]
  7. Yao, X.; Feng, X.; Han, J.; Cheng, G.; Guo, L. Automatic Weakly Supervised Object Detection from High Spatial Resolution Remote Sensing Images via Dynamic Curriculum Learning. IEEE Trans. Geosci. Remote Sens. 2021, 59, 675–685. [Google Scholar] [CrossRef]
  8. Huang, X.; Han, X.; Ma, S.; Lin, T.; Gong, J. Monitoring ecosystem service change in the City of Shenzhen by the use of high-resolution remotely sensed imagery and deep learning. Land Degrad. 2019, 30, 1490–1501. [Google Scholar] [CrossRef]
  9. Zhu, Q.; Zhong, Y.; Zhang, L.; Li, D. Adaptive deep sparse semantic modeling framework for high spatial resolution image scene classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6180–6195. [Google Scholar]
  10. Fang, B.; Li, Y.; Zhang, H.; Chan, J.C.W. Semi-Supervised Deep Learning Classification for Hyperspectral Image Based on Dual-Strategy Sample Selection. Remote Sens. 2018, 10, 574. [Google Scholar] [CrossRef]
  11. Cheng, G.; Guo, L.; Zhao, T.; Han, J.; Li, H.; Fang, J. Automatic landslide detection from remote-sensing imagery using a scene classification method based on BoVW and pLSA. Int. J. Remote Sens. 2013, 34, 45–59. [Google Scholar] [CrossRef]
  12. Lv, Z.; Shi, W.; Zhang, X.; Benediktsson, J. Landslide inventory mapping from bitemporal high-resolution remote sensing images using change detection and multiscale segmentation. IEEE J. Sel. Top. Appl. Earth Observ. 2018, 11, 1520–1532. [Google Scholar]
  13. Longbotham, N.; Chaapel, C.; Bleiler, L.; Padwick, C.; Emery, W.; Pacifici, F. Very high resolution multiangle urban classification analysis. IEEE Trans. Geosci. Remote Sens. 2011, 50, 1155–1170. [Google Scholar] [CrossRef]
  14. Tayyebi, A.; Pijanowski, B.; Tayyebi, A. An urban growth boundary model using neural networks, GIS and radial parameterization: An application to Tehran, Iran. Landscape Urban. Plan. 2011, 100, 35–44. [Google Scholar] [CrossRef]
  15. Huang, X.; Wen, D.; Li, J.; Qin, R. Multi-level monitoring of subtle urban changes for the megacities of China using high-resolution multi-view satellite imagery. Remote Sens. Environ. 2017, 196, 56–75. [Google Scholar]
  16. Zhang, T.; Huang, X. Monitoring of urban impervious surfaces using time series of high-resolution remote sensing images in rapidly urbanized areas: A case study of Shenzhen. IEEE J. Sel. Top. Appl. Earth Observ. 2018, 11, 2692–2708. [Google Scholar]
  17. Li, X.; Shao, G. Object-based urban vegetation mapping with high-resolution aerial photography as a single data source. Int. J. Remote Sens. 2013, 34, 771–789. [Google Scholar]
  18. Rußwurm, M.; Körner, M. Multi-temporal land cover classification with sequential recurrent encoders. ISPRS Int. J. Geo-Inf. 2018, 7, 129. [Google Scholar] [CrossRef]
  19. Othman, E.; Bazi, Y.; Melgani, F.; Alhichri, H.; Alajlan, N.; Zuair, M. Domain adaptation network for cross-scene classification. Remote Sens. 2017, 55, 4441–4456. [Google Scholar] [CrossRef]
  20. Chaib, S.; Liu, H.; Gu, Y.; Yao, H. Deep feature fusion for VHR remote sensing scene classification. Remote Sens. 2017, 55, 4775–4784. [Google Scholar] [CrossRef]
  21. Wang, X.; Xu, H.; Yuan, L.; Dai, W.; Wen, X. A remote-sensing scene-image classification method based on deep multiple-instance learning with a residual dense attention ConvNet. Remote Sens. 2022, 14, 5095. [Google Scholar] [CrossRef]
  22. Gao, Y.; Sun, X.; Liu, C. A General Self-Supervised Framework for Remote Sensing Image Classification. Remote Sens. 2022, 14, 4824. [Google Scholar] [CrossRef]
  23. Zhao, Y.; Liu, J.; Yang, J.; Wu, Z. Remote Sensing Image Scene Classification via Self-Supervised Learning and Knowledge Distillation. Remote Sens. 2022, 14, 4813. [Google Scholar] [CrossRef]
  24. Alajaji, D.; Alhichri, H.S.; Ammour, N.; Alajlan, N. Few-Shot Learning For Remote Sensing Scene Classification. In Proceedings of the Mediterranean and Middle-East Geoscience and Remote Sensing Symposium, Tunis, Tunisia, 9–11 March 2020; pp. 81–84. [Google Scholar]
  25. Noothout, J.M.H.; De Vos, B.D.; Wolterink, J.M.; Postma, E.M.; Smeets, P.A.M.; Takx, R.A.P.; Leiner, T.; Viergever, M.A.; Išgum, I. Deep Learning-Based Regression and Classification for Automatic Landmark Localization in Medical Images. IEEE Trans. Med. Imaging. 2020, 39, 4011–4022. [Google Scholar] [CrossRef] [PubMed]
  26. Cen, F.; Wang, G. Boosting Occluded Image Classification via Subspace Decomposition-Based Estimation of Deep Features. IEEE Trans. Cybern. 2020, 50, 3409–3422. [Google Scholar] [CrossRef] [PubMed]
  27. Li, L.; Han, J.; Yao, X.; Cheng, G.; Guo, L. DLA-MatchNet for few-shot remote sensing image scene classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 7844–7853. [Google Scholar] [CrossRef]
  28. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556v6. [Google Scholar]
  29. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  30. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  32. Liu, Y.; Zhong, Y.; Fei, F.; Zhang, L. Scene semantic classification based on random-scale stretched convolutional neural network for high-spatial resolution remote sensing imagery. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Beijing, China, 10–15 July 2016; pp. 763–766. [Google Scholar]
  33. Wu, B.; Meng, D.; Zhao, H. Semi-Supervised Learning for Seismic Impedance Inversion Using Generative Adversarial Networks. Remote Sens. 2021, 13, 909. [Google Scholar] [CrossRef]
  34. Geng, J.; Deng, X.; Ma, X.; Jiang, W. Transfer Learning for SAR Image Classification via Deep Joint Distribution Adaptation Networks. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5377–5392. [Google Scholar] [CrossRef]
  35. Zhan, T.; Song, B.; Xu, Y.; Wan, M.; Wang, X.; Yang, G.; Wu, Z. SSCNN-S: A spectral-spatial convolution neural network with Siamese architecture for change detection. Remote Sens. 2021, 13, 895. [Google Scholar] [CrossRef]
  36. Du, L.; Li, L.; Guo, Y.; Wang, Y.; Ren, K.; Chen, J. Two-Stream Deep Fusion Network Based on VAE and CNN for Synthetic Aperture Radar Target Recognition. Remote Sens. 2021, 13, 4021. [Google Scholar] [CrossRef]
  37. Xu, P.; Li, Q.; Zhang, B.; Wu, F.; Zhao, K.; Du, X.; Yang, C.; Zhong, R. On-Board Real-Time Ship Detection in HISEA-1 SAR Images Based on CFAR and Lightweight Deep Learning. Remote Sens. 2021, 13, 1995. [Google Scholar] [CrossRef]
  38. Wang, X.; Wang, S.; Ning, C.; Zhou, H. Enhanced feature pyramid network with deep semantic embedding for remote sensing scene classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7918–7932. [Google Scholar] [CrossRef]
  39. Sun, X.; Wang, B.; Wang, Z.; Li, H.; Li, H.; Fu, K. Research Progress on Few-Shot Learning for Remote Sensing Image Interpretation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 637–648. [Google Scholar]
  40. Wang, Y.; Yao, Q.; Kwok, J.; Ni, L. Generalizing from a few examples: A survey on few-shot learning. ACM Comput. Surv. 2020, 53, 1–34. [Google Scholar]
  41. Li, X.; Sun, Z.; Xue, J.; Ma, Z. A concise review of recent few-shot meta-learning methods. Neurocomputing 2021, 456, 463–468. [Google Scholar] [CrossRef]
  42. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. Adv. Neural Inf. Process. Syst. 2017, 30, 4080–4090. [Google Scholar]
  43. Yan, B.; Lang, C.; Cheng, G.; Han, J. Understanding negative proposals in generic few-shot object detection. IEEE Trans. Circ. Syst. Vid. 2024, 34, 5818–5829. [Google Scholar]
  44. Zeng, Q.; Geng, J.; Huang, K.; Jiang, W.; Guo, J. Prototype calibration with feature generation for few-shot remote sensing image scene classification. Remote Sens. 2021, 13, 2728. [Google Scholar] [CrossRef]
  45. Vinyals, O.; Blundell, C.; Lillicrap, T.; Wierstra, D. Matching networks for one shot learning. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 3630–3638. [Google Scholar]
  46. Chen, Y.; Li, Y.; Mao, H.; Liu, G.; Chai, X.; Jiao, L. A Novel Discriminative Enhancement Method for Few-Shot Remote Sensing Image Scene Classification. Remote Sens. 2023, 15, 4588. [Google Scholar] [CrossRef]
  47. Cheng, G.; Han, J.; Lu, X. Remote sensing image scene classification: Benchmark and state of the art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
  48. Yang, Y.; Newsam, S. Bag-of-visual-words and spatial extensions for land-use classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems, San Jose, CA, USA, 2–5 November 2010; pp. 270–279. [Google Scholar]
  49. Sheng, G.; Yang, W.; Xu, T.; Sun, H. High-resolution satellite scene classification using a sparse coding based multiple feature combination. Int. J. Remote Sens. 2012, 33, 2395–2412. [Google Scholar] [CrossRef]
  50. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.; Hospedales, T. Learning to compare: Relation network for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1199–1208. [Google Scholar]
  51. Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the International Conference on Machine Learning (ICML), Sydney, Australia, 6–11 August 2017; pp. 1126–1135. [Google Scholar]
  52. Li, Z.; Zhou, F.; Chen, F.; Li, H. Meta-sgd: Learning to learn quickly for few-shot learning. arXiv 2017, arXiv:1707.09835v2. [Google Scholar]
  53. Li, W.; Wang, L.; Xu, J.; Huo, J.; Gao, Y.; Luo, J. Revisiting Local Descriptor Based Image-To-Class Measure for Few-Shot Learning. arXiv 2019, arXiv:1903.12290v2. [Google Scholar]
  54. Chen, Y.; Li, Y.; Mao, H.; Chai, X.; Jiao, L. A Novel Deep Nearest Neighbor Neural Network for Few-Shot Remote Sensing Image Scene Classification. Remote Sens. 2023, 15, 666. [Google Scholar] [CrossRef]
  55. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
Figure 1. Diagram of samples in the NWPU-RESISC45 dataset with large differences in the object scales.
Figure 1. Diagram of samples in the NWPU-RESISC45 dataset with large differences in the object scales.
Remotesensing 17 01192 g001
Figure 2. Diagram of global–local two-stream network.
Figure 2. Diagram of global–local two-stream network.
Remotesensing 17 01192 g002
Figure 3. Flowchart of key area localization method.
Figure 3. Flowchart of key area localization method.
Remotesensing 17 01192 g003
Figure 4. The NWPU-RESISC45 dataset comprises 45 different scene classes, with each containing 700 images measuring 256 × 256 pixels.
Figure 4. The NWPU-RESISC45 dataset comprises 45 different scene classes, with each containing 700 images measuring 256 × 256 pixels.
Remotesensing 17 01192 g004
Figure 5. The UC Merced dataset features 21 unique scene classes, with each containing 100 land use images measuring 256 × 256 pixels.
Figure 5. The UC Merced dataset features 21 unique scene classes, with each containing 100 land use images measuring 256 × 256 pixels.
Remotesensing 17 01192 g005
Figure 6. The WHU-RS19 dataset contains 19 scene classes, each with at least 50 images measuring 600 × 600 pixels.
Figure 6. The WHU-RS19 dataset contains 19 scene classes, each with at least 50 images measuring 600 × 600 pixels.
Remotesensing 17 01192 g006
Figure 7. Localization results for key areas (delineated by red bounding boxes) in some samples from the NWPU-RESISC45 dataset.
Figure 7. Localization results for key areas (delineated by red bounding boxes) in some samples from the NWPU-RESISC45 dataset.
Remotesensing 17 01192 g007
Table 1. The experimental setup included both software and hardware components.
Table 1. The experimental setup included both software and hardware components.
Hardware EnvironmentCPUIntel(R) Core(TM) i7-7800X CPU @ 3.50 GHz 32 GB
GPUNVIDIA Geforce RTX 2080Ti 11 GB
Software EnvironmentOSLinux Ubuntu 18.04 LTS
Programming LanguagePython 3.6
Deep Learning FrameworkPyTorch 1.4.0
CUDACuda 10.0
Table 2. Comparison of results of various methods tested on the NWPU-RESISC45 dataset.
Table 2. Comparison of results of various methods tested on the NWPU-RESISC45 dataset.
Method5-Way 1-Shot5-Way 5-Shot
MatchingNet54.46% ± 0.77%67.87% ± 0.59%
RelationNet58.61% ± 0.83%78.63% ± 0.52%
MAML37.36% ± 0.69%45.94% ± 0.68%
Meta-SGD60.63% ± 0.90%75.75% ± 0.65%
DLA-MatchNet68.80% ± 0.70%81.63% ± 0.46%
DN466.39% ± 0.86%83.24% ± 0.87%
DN4AM70.75% ± 0.81%86.79% ± 0.51%
DEADN473.56% ± 0.83%87.28% ± 0.50%
Our Method73.84% ± 0.80%87.86% ± 0.51%
Table 3. Comparison of results of various methods tested on the UC Merced dataset.
Table 3. Comparison of results of various methods tested on the UC Merced dataset.
Method5-Way 1-Shot5-Way 5-Shot
MatchingNet46.16% ± 0.71%66.73% ± 0.56%
RelationNet48.89% ± 0.73%64.10% ± 0.54%
MAML43.65% ± 0.68%58.43% ± 0.64%
Meta-SGD50.52% ± 2.61%60.82% ± 2.00%
DLA-MatchNet53.76% ± 0.62%63.01% ± 0.51%
DN457.25% ± 1.0179.74% ± 0.78%
DN4AM65.49% ± 0.72%85.73% ± 0.47%
DEADN467.27% ± 0.74%87.69% ± 0.44%
Our Method68.12% ± 0.81%88.57% ± 0.52%
Table 4. Comparison of results of various methods tested on the WHU-RS19 dataset.
Table 4. Comparison of results of various methods tested on the WHU-RS19 dataset.
Method5-Way 1-Shot5-Way 5-Shot
MatchingNet60.60% ± 0.68%82.99% ± 0.40%
RelationNet60.54% ± 0.71%76.24% ± 0.34%
MAML46.72% ± 0.55%79.88% ± 0.41%
Meta-SGD51.54% ± 2.31%61.74% ± 2.02%
DLA-MatchNet68.27% ± 1.83%79.89% ± 0.33%
DN482.14% ± 0.80%96.02% ± 0.33%
DN4AM85.05% ± 0.52%96.94% ± 0.21%
DEADN486.89% ± 0.57%97.63% ± 0.19%
Our Method87.34% ± 0.62%98.25% ± 0.15%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lei, Y.; Li, Y.; Mao, H. A Novel Two-Stream Network for Few-Shot Remote Sensing Image Scene Classification. Remote Sens. 2025, 17, 1192. https://doi.org/10.3390/rs17071192

AMA Style

Lei Y, Li Y, Mao H. A Novel Two-Stream Network for Few-Shot Remote Sensing Image Scene Classification. Remote Sensing. 2025; 17(7):1192. https://doi.org/10.3390/rs17071192

Chicago/Turabian Style

Lei, Yaolin, Yangyang Li, and Heting Mao. 2025. "A Novel Two-Stream Network for Few-Shot Remote Sensing Image Scene Classification" Remote Sensing 17, no. 7: 1192. https://doi.org/10.3390/rs17071192

APA Style

Lei, Y., Li, Y., & Mao, H. (2025). A Novel Two-Stream Network for Few-Shot Remote Sensing Image Scene Classification. Remote Sensing, 17(7), 1192. https://doi.org/10.3390/rs17071192

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop