Next Article in Journal
Detection of Surface Crevasses over Antarctic Ice Shelves Using SAR Imagery and Deep Learning Method
Previous Article in Journal
Study on High-Resolution Suspended Sediment Distribution under the Influence of Coastal Zone Engineering in the Yangtze River Mouth, China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Structural Attention Enhanced Continual Meta-Learning for Graph Edge Labeling Based Few-Shot Remote Sensing Scene Classification

1
Institute of Automation Chinese Academy of Sciences, 95 Zhongguancun East Road, Haidian District, Beijing 100190, China
2
School of Information, Central University of Finance and Economics, Shunsha Road, Shahe Higher Education Park, Changping District, Beijing 102206, China
3
College of Software, Beihang University, 37 Xueyuan Road, Haidian District, Beijing 100190, China
4
School of Mechanical Electronic and Information Engineering, China University of Mining and Technology, Ding No.11 Xueyuan Road, Haidian District, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(3), 485; https://doi.org/10.3390/rs14030485
Submission received: 15 November 2021 / Revised: 14 January 2022 / Accepted: 17 January 2022 / Published: 20 January 2022
(This article belongs to the Section AI Remote Sensing)

Abstract

:
Scene classification is one of the fundamental techniques shared by many basic remote sensing tasks with a wide range of applications. As the demands of catering with situations under high variance in the data urgent conditions are rising, a research topic called few-shot scene classification is receiving more interest with a focus on building classification model from few training samples. Currently, methods using the meta-learning principle or graphical models are achieving state-of-art performances. However, there are still significant gaps in between the few-shot methods and the traditionally trained ones, as there are implicit data isolations in standard meta-learning procedure and less-flexibility in the static graph neural network modeling technique, which largely limit the data-to-knowledge transition efficiency. To address these issues, this paper proposed an novel few-shot scene classification algorithm based on a different meta-learning principle called continual meta-learning, which enhances the inter-task correlation by fusing more historical prior knowledge from a sequence of tasks within sections of meta-training or meta-testing periods. Moreover, as to increase the discriminative power between classes, a graph transformer is introduced to produce the structural attention, which can optimize the distribution of sample features in the embedded space and promotes the overall classification capability of the model. The advantages of our proposed algorithm are verified by comparing with nine state-of-art meta-learning based on few-shot scene classification on three popular datasets, where a minimum of a 9% increase in accuracy can be observed. Furthermore, the efficiency of the newly added modular modifications have also be verified by comparing to the continual meta-learning baseline.

1. Introduction

Scene classification is one of the techniques fundamental in many remote sensing tasks with a wide range of applications, such as ecological and hydrological monitoring [1], urban planning [2], forest mapping and conservation [3], resource exploration [4], and agricultural assessments [5]. In recent years, with the development of satellite and aerial imaging technologies, the accumulation of free high quality remote sensing images grows to a huge quantity, and a number of famous domestic and foreign remote sensing research institutes including Wuhan University, Northwestern University of Technology, Beijing Normal University, INRIA, and DLR have published their own public scene classification datasets. Such prosperity in data sharing greatly encourages the development of deep learning-based methods, and the highest performance has constantly been renewed with the fast publication of novel algorithms [6]. Lately, the performances of algorithms based on traditional whole dataset range training with sufficient sample quantities have almost reached 100%, which is competitive even to human analyzers [7,8,9].
In the meantime, such advantageous performances come at high supervision costs, for every modification on the model requires a large amount of manual work in data collection and costly model retraining. Such characteristics result in several significant limitations on applications [10]. The first limitation takes place in transferability, for remote sensing scenery images have high intra-class variances caused by differences in regions, seasons, weathers, etc., models for recognizing the same land type also need to be retrained as to prevent large accuracy losses. The second limitation happens in the scalability, where any extension on the recognizable class set of the model requires thorough retraining over all the previous training samples. To make matters worse, if there are not enough samples for the extended class, heavy data collection and annotation labors are required. Many efforts have been made to alleviate these two issues under the traditional training framework, common methods include using the correct model parameter transfer methods to reduce the feature shift in cross domain model application [11], or to augment a dataset with artificial samples generated by generative adversarial network techniques [10].
Due to these problems, a new research topic for model training under data urgent conditions called Few-Shot Learning (FSL) has come into the view. FSL is designed for applicable scenarios where samples are naturally insufficient or very hard to collect, and most of the researches published under this topic adopted the concept of priori knowledge maximization, which happens both in datasets processing and model modifications. So far, the main categories of the state-of-art FSL solutions include: Meta-learning based [12,13], embedding learning-based [12,13,14], and generative modeling-based [15,16,17,18]. These three categories of methods focus on optimizations of the learning mechanism, feature embedding space, and training data samples specifically, and have achieved good applications results in areas that include classification [12,19], regression [20], detection [14,17], segmentation [21,22], etc.
The idea of using few-shot learning algorithms for the remote sensing scene classification problem dates back to 2018 [23,24,25,26], where researchers found severe data insufficiency problems in dealing with hyperspectral and SAR images. After that, researches using the similar principles spread quickly to other types of remote sensing images. Among the transfers of algorithms, the dataset extension is a straightforward migration, such as extending the quantity of samples using generative methods [27]. While for some other methods, a very different approach to construct the classifiers has been chosen, as most of have dropped the obsessiveness of making models as robust and accurate as those trained from large datasets, and pursued a tradeoff between their performances and usage scopes by extending the sample sets or building more flexible models. Based on such ideas, many ingenious methods have been devised. For instance, there are methods focused on optimizing the embedded feature space such as feature interpolation [28] and knowledge distillation [29,30] for better sample discrimination. Some methods utilized the heatedly studied attention mechanism such as attention fusion [31,32] and attention metric calculation [33] to improve discriminative power.
Among these FSL-based remote sensing scene classification methods, one category based on meta-learning has received much more attention in recent researches [34,35]. The core idea of meta-learning is learning the model adaptation patterns over specific data distributions, which is helpful for fast model convergence on a small dataset, and is fitted with the fundamental needs of few-shot learning. Categorized by the representative format of prior knowledge, meta-learning methods can be roughly divided into three categories: The meta-representation-based methods [36,37,38], the meta-optimization-based methods [39,40,41], and the meta-objective-based methods [42,43,44]. More recently, newly proposed researches also exploited the combination with metric learning [12,45,46], reinforcement learning, [47,48,49] and graph models [50,51,52], which has made a considerable number of technical breakthroughs in the field. As a result of this, an increasing number of few-shot remote sensing scene classification methods choose to use the meta-learning concepts. For instance, in [53,54], the efficiency of meta-learning methods have been explored on satellite and aerial images, where a novel balance loss and a new feature embedding cosine metric have been proposed specifically to improve the effectiveness of iterative model optimization. Article [55] fused the emerging frequency and imagery content into the process of embedding space optimization, and proposed a new loss objective based on a combination of contrast loss and cross-entropy loss, which greatly improved the convergence speed. Article [56] focused on the trade off between training dataset complexity and size of the model, then proposed a novel parameter trimming and fusion mechanism through parameter transferring, which effectively combine the meta-learning technique with an unsupervised domain adaptation training mode, and improved the training efficiency with fewer model parameters.
To sum up, meta-learning-based methods have shown their capability in fast model convergence on small datasets, but the performance gap between these methods and those trained with sufficient samples via the traditional way still exists. One of the major reasons is that the task-wise dynamic training and inferencing mechanism not only promotes the flexibility of the model, but also increases the isolation between samples in the sequence of tasks, which limit the further improvement in the accuracy of the model. To address this, a novel training mechanism called continual meta-learning [57] is adopted in this article to exploit the informative implicit correlations and alleviate the catastrophic forgetting phenomenon, which facilitates the transfer of historically-encoded node features for the recognition of newly input samples. Additionally, we further utilize structural attention to improve the node feature encoding via the graph transformer [58], which is further combined with an edge labeling Bayesian graph network [52] to further increase the categorical discriminative power.

2. Related Works

2.1. Continual Meta-Learning

The concept of continual meta-learning derives from standard meta-learning, and is built upon the assumption that the performance of the model will continually increase alongside accumulative meta-learning through tasks. Thus, the model can maintain a well-performing status on newly input samples without extra retraining. Such a mechanism is different from a standard meta-learning mechanism where the dynamic updating range is limited within a single meta-task, continual meta-learning utilizes historical knowledge from the whole learning process, which is especially beneficial under data urgent conditions. However, with the existence of a catastrophic forgetting phenomenon, model performance will not always grow monotonically or stay at a reasonable level, but suffer from significant regressions [59]. To address such a problem, many research efforts have been made. The representative methods include using an attention mechanism in continual learning [60], optimizing the feature-embedded space [36,61], or the formation process of it [62], tweaking the training mechanism [63].

2.2. Graph Neural Networks

The graph neural network (GNN) models composed of nodes and edges are natural for non-structural data modeling, since they can map data feature space into non-Euclidean metrics [64]. Thus they have been widely used in applications, including advertising analysis [65], social calculation [66], chemical molecular analysis [67], etc. Recently, GNNs have been frequently used to cater to challenging Few-shot Classification (FSC) problems. Typical examples include transforming the classification problem into edge labeling [57], or modeling complex hierarchical correlations [68]. Moreover, GNNs have also been widely used with meta-learning concepts, which effectively improve the deep model performance on small datasets. For instance, modeling inter-class relationships [51,52], improving feature embedding [51], or using it with continual meta-learning principles.

2.3. Self-Attention and Graph Transformer

Attention has always been considered an important discrimination feature in pattern recognition. Currently, researches using a transformer-based attention calculation technique have made significant progress in performances and grow to prosperity. The transformer modules were firstly used widely on natural language processing, equipped with a multi-head attention function, showing expertise in extracting a useful feature from sequentially-ordered inputs. Recent studies have managed to transfer them to vision tasks including classification [69,70], detection [71], segmentation [72], etc. successfully, and indicate a limitless potential in the future. Graph transformers are attention calculation models derived from the transformer models that focus on extracting the significant feature from non-structural data. Such characteristics facilitates the representative graph structure recognition, which is very valuable for applications involving complex scene recognition. Typical applications include paragraph chapters comprehension [58,73], objective relationship recognition [74,75], 2D and 3D spatial structure perception [76], time and space evolution process analysis [77], etc. Technically, the usage of transformer-based graph attention computation improves encoded graph features [58], optimization processes [74], updating mechanisms [75], etc.

3. Preliminary

Before diving into the details of the proposed algorithm, we will make a brief introduction to the core concepts of few-shot learning and continual meta-learning, as well as the mathematical definitions and symbols being used in these fields.
Few-shot classification is essentially a classification problem itself, where a classifier f : x y is to be trained on a given training set D t r a i n , with x being the input image, and the predicted categorical label y belongs to a set of classes predefined, where y C = 1 , 2 , , c n n = N . And under the normal training and testing setting, samples of the training and testing datasets are disjoint, used independently, but selected from the set of classes C. That is, D t r a i n , D t e s t = x i , y i i = 1 N , D t r a i n D t e s t = , as is shown in Figure 1a. The classifier is optimized and converged on the training set with a large and sufficient quantity of samples, which are the typical characteristics of deep learning methods. However, in few-shot classification problems, samples for training are limited or scarce, with a categorical sample quantity D t r a i n C being very small for each class C. Most few-shot learning problems follow a N-way K-shot specification, trained and tested in units called tasks, where K support samples and K query samples from N classes are selected for training and testing. The K small ranges from 1 to 10 usually, and the trained classifier can only be adjusted on the K samples from the N, before making accurate predictions on new test samples. This might seem impossible, but by modeling patterns in similarly distributed samples, or making more abstractive samples correlation modeling, these have helped the latest few-shot classification methods to achieve considerable accuracies, including the applications in remote sensing scenarios.
Meta-learning is one of the most developed and widely applied Few-shot Learning (FSL) solutions, its core concept is to learn the learning strategy on a set of tasks from a specific distributions, which makes the model capable of converging efficiently on only a small number of samples from unprecedented environments. In meta-learning, training and testing are organized in units of tasks T i N , each having its own training and testing samples called the support and query dataset, noted as S and T, and T = S Q . Under such dataset organization, for a N-way K-shot task, there is a support set S = x i , y i i = 1 N × K = n = 1 N x j , y j , y j = n j = 1 K with a size of N × K , and the query set has N × K or other quantities of samples from the same collection of N classes. The training and testing processes of the meta-learning are also termed as meta-training or meta-testing, where classes used for training D m e t a t r a i n and testing D m e t a t e s t are disjoint C m e t a t r a i n C m e t a t e s t = , and models have to deal with unseen classes with the assistance of very scarce supervision. To define such a classification prediction in the form of Bayesian inference, there is:
p y i Q x i Q , S ; Θ = p y i Q x i Q , ψ S ; Θ p ψ S S ; Θ .
In Equation (1), Θ represents the parameters denoting the priori knowledge learned from training tasks T p T with similar distributions as the testing ones, which usually kept static during the testing. ψ represents the task-dependent parameters correlated with the support subset S. Evidently, the transferability of a meta-learning model depends on two factors: Firstly, learning a set of more transferable priori parameters and secondly, finding the appropriate task-dependent parameter on the small support sample set.
Such a definition of meta-learning greatly encourages the flexibilities of training and testing, but also imports implicit manmade isolation between support sample sets fed sequentially to the model, which has virtually impacted the further improvement in meta-learning algorithm performances. Due to this, a branch of meta-learning research, named Continual Meta-learning (CML), have returned to researching interests, which can make continuous model updates while processing the input tasks sequence T u u = 1 , , U , and eventually affects the task-dependent parameter ψ j . Such a process is like a constant reutilization of the latest historical knowledge, noted as h i p h i S 1 : i 1 ; Θ , which gives the following equation:
p y i Q x i Q , S ; Θ = p y i Q x i Q , ψ S , h i ; Θ p h i S 1 : i 1 ; Θ p ψ S S ; Θ .
In Equation (2), with the continual optimization of the hidden historical parameters h i , the prediction accuracy query sample x i Q is improved by the support sets S 1 : i 1 in all previous processed tasks T 1 : i 1 . Based on such a learning concept, quite a few continual meta-learning algorithms explored the efficiencies of time series data modeling techniques including the Markov chain, LSTM for performance enhancement. Under such efforts, sharing between the learned task-dependent parameters is promoted, and significant algorithmic performance improvements have been seen in applications including scene classification, image segmentation, and object detection.
Finally, the performance advantages of continual learning methods does not come solely from the implicitly increased support samples, but depends more on the correlations being extracted across tasks, which have been used to enhance optimizing the generalization ability of the model on correlated yet unseen data samples. In fact, as being exemplified in our experiments, the incremental degree in overall accuracies surpasses the incremental ratio in sample quantities.

4. Overall Framework

As shown in Figure 2, the proposed algorithm architecture consists of three components: The encoding module, the continual learning module, and the Bayesian graph edge labeling module. In the proposed algorithm design, the support and query samples in tasks will be firstly encoded by the ResNet as node features, along with the structural attention features from the graph transformer in the encoding module. Then the node features will be fed to the gated recurrent unit (GRU) based continual meta-learning module for distributional optimization. After that, the distributionally stabilized node features and the structural attention are fed to the Bayesian graph labeling module for node correlation estimation via a Bayesian style edge weight estimation based on Gaussian distribution approximation. The functional design and characteristics of the three modules are described below.
The encoding module encodes the primitive features, including node features and structural attention features, for subsequent processing. In it, node features from task-wise support and query samples are primitively encoded by a Convolutional Neural Network (CNN) for multi-scale texture pattern modeling. Then the structural attention is calculated upon the distance adjacency matrix from the CNN features through a graph transformer unit, which helps to rectify the non-optimal categorical measures between different classes. Furthermore, the calculated structural attention is multiplied with the CNN node features for enhancement.
The continual learning module tightly connects the graph transformer from the encoding module and the subsequent Bayesian graph labeling module, being more like a macro structure adopting feature embedding improvement. Its core component is a GRU-based sequential data iterative optimization unit, using the long-short term memory-based historical feature fusion mechanism to recall and emphasize the intrinsic local correlations between categorical node features in the sequentially encoded node features. This module is also the central design to alleviate the catastrophic forgetting problem under task-dependent meta-learning framework.
The Bayesian graph edge labeling module weighs the node feature correlations and produce the final classification results. In it, an explicit fully connected adjacency graph upon the support and query samples is established. Then a Gaussian distributional approximation of the actual support and query nodes correlations via a Bayesian style inference from the combinatorial usage of the adjacency graph edge weights and the structural attention from the graph transformer, as the modeled estimates of the mean and deviation values of the resulting connection strengths. Finally, the rectified edge weights are transformed into the correlations between query and support samples, and turned into the estimated class labels of the query samples.

5. Algorithm Details

5.1. Encoding Module

Shown in Figure 3, the encoding module consists of two procedures: Node feature-encoding procedure, and the structural attention-encoding procedure. According to symbol definitions in Section 3, the support and query samples in a single task can be denoted as T = S , Q , and the encoding module extracts the primitive features and structural properties from them.
During the node feature encoding procedure, the N × K + M support and query samples are firstly fed to a CNN to get the encoded features X T = X S X Q , where X S = x i x i S and X Q = X j x j Q . Emperically, ResNet-18 is chosen as the CNN backbone, which is adequately complex for preserving the feature differential details and revealing the categorical structure.
Then, in the structural attention encoding procedure, an adjacency matrix A a d j is calculated straight-forwardly at first by getting differences d between encoded features X T in Equation (3). In practice, the numerical difference can be norm-1 or norm-2.
A a d j = d x 1 , x 1 d x 1 , x s d x 1 , x N + M d x t , x 1 d x t , x s d x t , x N + M d x N + M , x 1 d x N + M , x s d x N + M , x N + M , x t , x s X T × X T
As mentioned in [58], adjacency matrix A a d j derived from encoded feature value differences d x t , x s might have some problems representing correlations where a categorical structure is complex due to its discriminative power. Thus, a graph structure attention calculation module called a graph transformer is applied here, to enhance the encoding of correlations. In it, the attention matrix A a t t n are deducted from the linear transformation of node features x i , where the attention query q i , key k i , and value features v i are firstly calculated in Equation (4):
q i = x i · W q + b q k i = x i · W k + b k v i = x i · W v + b v .
Then, depending on the positional relationship of q i and k i in the attentional multiplication, a linear transformation operation is applied for constructing the directional correlation adjacency matrix:
R = A · W + b = r i j i , j = r 1 j , r 2 j , . . . , r M + N j j R = A · W + b = r i j i , j = r 1 j , r 2 j , . . . , r M + N j j .
The R = r i j i , j and R = r i j i , j represents the two kinds of directional correlations in between the support set S and query set Q. By this, directional correlations R and R are further fused with the query features q i by linear operations, which produces the structural attention matrix A a t t n as in below:
A a t t n = q i + r i 1 q i + r i 2 q i + r i M + N · q i + r i 1 q i + r i 2 q i + r i M + N .
With the structural attention matrix A a t t n , and the self-attention weighted vector v i , the structural attention strengthened node feature x i can be calculated by following:
x i = v i · A a t t n .
Lastly, the gotten attention weighted node feature x i will be adaptively combined with the CNN-encoded node features by means of skip connection, which helps the network to increase its flexibility and robustness.

5.2. Continual Meta-Learning by Online GRU-Based Feature Optimization

As introduced in the preliminaries in Section 3, continual learning contributes to the model performance not only by the implicit increments in sample quantity, but mainly by the informative sharing between task-wise encoded features, which can be represented by a hidden historical parameter iteratively updated in the process. Following the Bayesian structure, the hidden parameter correlated with the support and query samples in task sequence T u u = 1 U can be noted as h i p h i S 1 : i 1 ; Θ .
The Gated Recurrent Unit (GRU) is commonly used as a simplified LSTM module, which is computationally more efficient. In the GRU module, the attention strengthened node feature x i is serially encoded, as being shown in Figure 4:
r = σ W j r x i + b j r + W h r h i 1 + b h r z = σ W j z x i + b j z + W h z h i 1 + b h z n = tanh W j n x i + b j n + r W h n h i 1 + b h n h i = ( 1 z ) n + z h i 1 .
In Equation (8), z can be taken as a controlling parameter of h p 1 and n, where h p 1 is the historical hidden parameter, and n is the converted value from input node feature x i . If z grows larger, then more historical information will be used from the hidden variable h, otherwise the current value will be adopted more.
In the proposed algorithm, such GRU-based node feature polishing is performed through a task sequence sampled from the meta-training data subset, the longer this sequence is, the stronger long-short term memory historical information referencing effect it has. Moreover, the input features x i can be concatenated from two or more tasks to suite the conditions using a larger training batch size, and the node features being packed and fed in a single iteration is called a cell c i . For instance, for training settings with batch size S b a t c h = 32 and GRU iteration length L G R U = 16 , then the input feature to the GRU unit is packed from C = S b a t c h / L G R U = 2 tasks, and c i = k = 1 C x j x T i × C + k .
Then the sequentially cell-wise-encoded hidden state features h i , are rearranged into task-wise feature collections u = 1 U h i h i = f G R U x i , x i T u , which are fed to the succeeding Bayesian graph edge Labeling model.

5.3. Bayesian Graph Edge Labeling for Classification

In the Bayesian graph edge labeling module, distributionally converged node features are fused with the structural attention features from a graph transformer to produce the final estimates on the correlations between query and support samples for class prediction. This feature fusion takes the form of a Bayesian style inference upon the primitive adjacency matrix from the node feature distances and the structural attention matrix.
Final sample correlation inference takes place via two branches that later merges. The first branch produces the primitive adjacency matrix A a d j , which is calculated in a similar way as the adjacency matrix A a d j in Equation (3) for getting structural attention, as shown in the upper left part of Figure 5. In A a d j , the fundamental node features are the hidden state features h i outputted by the GRU-based continual meta-learning module, as in below:
A a d j = d h 1 , h 1 d h 1 , h N + M d h N + M , h 1 d h N + M , h N + M .
After that, the feature difference A a d j is fed to a Fully Convolutional Neural network (FCN) f s i m with 1 × 1 filter kernels to produce the mean correlation matrix A s i m = f s i m A a d j , as the final output of the first branch.
For the second branch in parallel, as has been shown in the lower left part of Figure 5, a set of calibration parameters μ i j W , δ i j W , μ i j b , μ i j b are estimated from the A a t t n outputted by a graph transformer by a set of FCNs, f * = f M W , f Δ W , f M b , f Δ b .
The second branch is the core of the Bayesian style correlations estimation concept, where the superscripts of the calibration parameters W t and b t , are supposed to follow a collection of Gaussian distributions as in the equations below:
W t N M W , Δ W 2 , b t N M b , Δ b 2 .
These calibration weights and biases are calculated by the FCNs in f * with a 1 × 1 filter kernels calculated with the following equations:
μ i j W = f M W r i j e n h c , δ i j W = f Δ W r i j e n h c μ i j b = f M b r i j e n h c , μ i j b = f Δ b r i j e n h c .
In Equation (11), there are r i j e n h c from A a t t n , and M W = μ i j W , Δ W = δ i j W , M b = μ i j b , Δ b = δ i j b , as the mean and deviation of the Gaussian-style mapping parameters W t , b t .
The attention-enhanced support and query sample correlation r i j a d j from A a d j will be mapped into the sample similarity matrix A f i n a l s i m , with the correlation r i j f i n a l from r i j a d j weights W t and biases b t .
Finally, for every correlation A f i n a l s i m of the resulting similarity matrix A f i n a l s i m , it is generated via the parameters W t , b t , with the equations:
r i j f i n a l N μ r i j f i n a l , δ r i j f i n a l μ r i j f i n a l = μ i j s i m = μ i j W · A i , j s i m + μ i j b δ r i j f i n a l = δ i j s i m = log e δ i j W · A i , j s i m 2 + e δ i j b .
Furthermore, in order to enhance the estimated inter-sampler correlations, annotated as multi-layered continual learning iteration in the lower right side of Figure 2, the correlations calculated by the graph transformer module, GRU module, and the Bayesian graph edge labeling module can be calculated in a multi-layered form. During this multi-layered correlation inference, several (denoted as V here) similarity maps A f i n a l s i m ( v ) v = 1 V are generated through V independently created chains of these modules. The final similarity matrix is produced via a FCN with 1 × 1 sized kernels, as in Equation (13). In it, A f i n a l s i m ( 1 ) ; ; A f i n a l s i m ( V ) D = 3 means stacking the similarity matrices at 3rd dimension:
A f i n a l s i m = f 1 × 1 A f i n a l s i m ( 1 ) ; ; A f i n a l s i m ( V ) D = 3 , A f i n a l s i m ( i ) A f i n a l s i m ( v ) v = 1 V .
Then the sample-wise correlation r i j f i n a l in A f i n a l s i m between the support and query set is picked out to be rectified by the logit function σ ( · ) and the query-to-support mask matrices M q and M s , where M q will preserve the query sample slots and M s preserves the support samples slots, as in the following equation:
p y ˜ ( k ) x ˜ , h ( k ) , ψ t = M q A f i n a l s i m M s .
The overall training process of proposed model is illustrated in Algorithm 1. During training, the accuracy of the posterior distribution of query samples inferred from query-to-support correlations, will be updated when comparing with the true query sample labels to construct the graph edge accuracy loss L E , which will be combined with the basic loss L B derived from the other components of the model.
L E = k = 1 K i = 1 | Q | y ˜ i log p y ˜ i ( k ) + 1 y ˜ i log 1 p y ˜ i ( k ) .
Θ * = arg min Θ L E + γ L B .
In the above equations, Θ * = θ R e s N e t , θ G r a p h T r a n s , θ G R U , θ h , θ E represents the parameters being used across the algorithmic framework, where θ R e s N e t and θ G r a p h T r a n s are from the encode module, θ G R U and θ h are from the GRU-based continual meta-Learning module, and θ E is from the Bayesian graph edge labeling module.
Algorithm 1: Continual Bayesian EGNN with Graph Transformer
Input :
  
  
A sequence of few-shot scene classification tasks T with distribution T p T , separated into sections of task periods S u , Q u u = 1 , , U .
Ouput :
  
  
Learnt parameters: ResNet Encoder θ R e s N e t , Graph Transformer θ G r a p h T r a n s . GRU module θ G R U and hidden parameter θ h . θ E of the Bayesian graph labeling.
Details :
  
1:
Initialize: Load the ImageNet pretrained ResNet-18 weights. Initialize the parameters θ R e s N e t , θ G r a p h T r a n s , θ G R U , θ h , θ E .
2:
for all T p T do
3:
for all k Ω do
4:
  Encoding Step:
5:
   Get features of input task samples as X T = X S X Q by using ResNet.
6:
  Continual Learning Step:
7:
   for cell c i in task splits c 1 , , c l  do
8:
    Calculate the structural attention A a t t n and attention enhanced feature x i from ResNet feature X T .
9:
    Iterative Continual Learning Step:
10:
     Calculate the historically enhanced feature x ˜ i from x i c i by the GRU module.
11:
    Graph Edge Labeling Step:
12:
     Calculate plain feature difference based raw adjacency A s i m from enhanced features x ˜ i , and get the Bayesian distributed adjacency parameters μ i j W , δ i j W , μ i j b and μ i j b by using CNN.
13:
     Calculate the Bayesian distributed parameters μ i j s i m and δ i j s i m , and generate the adjacency A f i n a l s i m , and get the final node correlations.
14:
   end for
15:
  Update Step:
16:
   Calculate the edge loss L E and node loss L B .
17:
   Update the model parameters θ R e s N e t , θ G r a p h T r a n s , θ G R U , θ h , θ E by L E + γ L B .
18:
end for
19:
end for

6. Experiments and Results

In this chapter, the first section will briefly introduce datasets UC Merced Landuse [78], NWPU-RESISC45 [79], and AID [80], which are used for training and evaluation, which are commonly used datasets with increasing complexities that ensure the objective analysis of the behavior of the model. Section 6.1 goes through the details of experimental parameters, including the initialization parameters and settings of some of the crucial modules. Section 6.2 provides performance comparisons between our proposed model and the SOTA methods, discussing the pros and cons of the current model design. Section 7 then makes the ablation studies on the main performance-critical parameters of the model containing the classification accuracy, training sample ratio, etc., thus forming the detailed quantitative analysis on the main functional components.

6.1. Experiment Datasets and Experiment Setup

6.1.1. Datasets Description

To fully testify the crucial algorithmic properties including robustness and generality, the proposed algorithm has been fully verified on three challenging open access datasets: UC Merced Landuse, NWPU-RESISC45, and AID. The characteristics of these datasets and their training splits are introduced.
UC Merced Landuse (UCM) [78] is an open access scene classification dataset published on 28 October 2010 by the Merced laboratory, University of California. The dataset contains 21 land use types, each containing at least 100 corresponding satellite images, all chosen from the USGS National Map Urban Area Imagery dataset, including a collection of urban and rural land types. In the dataset, each remote sensing image sized 256 × 256 with a ground sample distance of 0.3 m, with parts of the dataset image samples represented below.
In this article all datasets are split into three parts: Train, validation, and test, where the additional validation samples are for cross validation during the training iterations.
The splitting ratio of the partition on UC Merced dataset is 11:5:5, which is shown in Table 1 and Figure 6, with a categorical ratio of training at 52.38%. All three partitions contain natural and urban scenarios, which have imagery texture complexities ranging from simple to complex.
NWPU-RESISC45 [79] is a public scene classification dataset published in 2017 by Northwestern Polytechnical University (NWPU). There are 45 scene categories containing an airport, tennis field, residual places, snow mountains, and seas, with 700 images in each category collected from over 100 countries and regions all over the world, culminating to 41,500 images in total. Each image is in size 256 × 256 with a ground sample distance range from 0.2 m to 30 m.
The data sample split configuration of the NWPU-RESISC45 is shown in Table 2, the splitting ratios of the train, test, and validation partition are 25:10:10, which is shown in Table 2 and Figure 7, where 55.56% classes are chosen for training. Considering the complexity and data variances in NWPU-RESISC45, the scene categories selected for the train, test, and validation partitions are made as evenly as possible, by containing the typical natural and urban areas with similar image texture complexity distribution.
The Aerial Image Dataset (AID) [80] is a public remote sensing scene classification dataset published in 2017 by Wuhan University and the Huazhong University of Science and Technology. It contains 30 scene classes of typical natural and man-made scenarios including an airport, sport stadium, river, meadow, and desert. Every scenery class has 200 to 400 images and adds up to a total quantity of 10,000. All images sized 600 × 600 and are selected by professional remote sensing researchers from Google Earth remote sensing collections with a ground sampling distance range from 0.3 m to 40 m.
The train, test, and validate splits of the AID scene classes is shown in the table with a ratio categorical quantity of 15:8:7, which is shown in Table 3 and Figure 8. Thus, the ratio of the training classes is 50%. And the scene class types in test and validation dataset distribute evenly in natural and urban regions, and the image texture complexity distributions are also similar.

6.1.2. Experiment Settings

For the encode module parameter settings, all input images fed to the ResNet-18 are scaled to a uniformed size of 256 × 256 , and the size of the training image batch is 32 for comparisons in Section 6.2 for best performance, and 16 in Section 7. The encoded features from ResNet-18 is then mapped into 128 dimensional feature vectors which is then fed to the following graph transformer for attention oriented encoding. As to increase the diversity of the outputted module, the number of attention heads is set to 16 in the experiment, the dimension of the attention feature inside the graph transformer is set to 1024, as to ensure richer structural attention information will be introduced in to the new embedding space. In addition, the feature dimension of the attention-enhanced node feature outputted alongside the structural attention adjacency matrix is set to be 128, to be consistent with the input feature dimension.
For the continual meta-learning module parameter settings, in order to balance the computational cost and quality of encoded features, the iteration length of the GRU-based continual meta-learning is set to 16. Thus, there will be features from a single task with a batch size of 16, and features from 2 tasks for a batch size set to 32. Moreover, as the continual learning unit is an explicit distribution aggregation, the input and output feature dimensions are consistent with its direct predecessor and successor, which is 128 in the experimental settings.
Bayesian graph labeling module parameter settings: In the graph edge labeling module, the dimension of encoded edge-based correlation is set to 128, which is consistent with the dimension of features from the encoding module and the continual meta-learning module. Finally, the number of layers of the Bayesian edge labeling graph is set to 2, as the studies in ablation Section 7.3 suggested the best performance on all datasets.
Training and optimization settings: As the proposed algorithm contains multiple functional modules with varied feature encoding and correlating procedures, the Adam optimization method is adopted to cater with the high non-linearity characteristics. During training, the learning rate is initialized to 1 × 10 3 , with 1 × 10 6 as the weight decay for every 15,000 iterations, and the gradient clip is set to 5.

6.1.3. Evaluation Metrics

Following the experimental settings, we compare the proposed algorithm with state-of-the-art counter-parts on datasets UC Merced, NWPU-RESISC45, and AID. The selected nine counterparts algorithms covers almost all the few shot scene classification algorithms verified on the three datasets, and are highly representative in their periods. For convenience, names of algorithms are abbreviated as: RS-MetaNet Ref. [53], Few-Shot Multi-Atten Ref. [31], Few-Shot Aerial Ref. [54], RS-SSKD Ref. [30], Know. Distill. Ref. [29], Proto. Calib. Ref. [27], SAFFNet Ref. [32], AMN Ref. [33], and ParamTrans Ref. [56].
In addition, in order to objectively illustrate the functional effects of the newly-introduced graph transformer structural attention encoding module and the modified Bayesian graph edge labeling module in the proposed algorithm, the continual meta-learning algorithm without these major changes is also included in comparison as the baseline algorithm, and denoted as Orig. CML-BGNN Ref. [52].
The classification accuracies and the usage ratios of training data are all listed in the Table 4, Table 5 and Table 6, the best and second-best accuracies are emphasized in bold and underlined. The accuracy metric being adopted in experiment analysis follows the definition of averaged accuracy described in SAFFNet Ref. [32] in detail, and is consistent with all the other counterpart algorithms.

6.2. Main Results

Based on the comparisons with counterparts and baseline, as the continual meta learning framework makes a better utilization of the historical prior, the averaged classification accuracies of our proposed and baseline algorithms stay almost always among the top 2 positions when compared to counterparts. Our proposed algorithm showed a minimum 9% improvement in accuracy, and a minimum 5% accuracy improvement over its continual meta-learning baseline on NWPU-RESISC45 and AID datasets.
The complexity of the UC Merced dataset is slightly lower among the three datasets, as most algorithms produced a higher accuracy score on it. From Table 4, our proposed algorithm and its baseline occupied the first and second best positions on the list, with a leap of about 20% over algorithms using a standard meta-learning training mechanism. Besides, the accuracy of the baseline is even higher than our proposed version. This incidence is mainly caused by the dataset splits under our experiment settings, for there are only 5 classes in the test subset. By this, all tasks have the same collection of classes under the five-way one-shot meta-test setting, making the edge labeling graph structure almost static so that the flexibility structural pattern modeling effect of the graph transformer was weakened. Put it another way, such a phenomenon is also a convincing fact in that a structural attention calculation procedure is more beneficial in scenarios where a large number of novel classes and more complicated class relationships exists.
NWPU-RESISC45 is the biggest and most complex dataset among the three, with the highest number of classes and number of samples in each class, thus requiring higher inter and intra class discrimination modeling capabilities.
On it, due to the condensed categorical sample feature distributions by the feature rectification and smoothing effects through the continual meta-learning procedure, our proposed algorithm and its baseline achieved the top-1 and top-2 accuracies with a leap of more than 10% over the counterpart algorithms. Besides, the accuracies of almost every algorithm increased compared with scores on UC Merced, which mainly benefit from the richer supply of training samples in each class. Moreover, our proposed algorithm overpassed the baseline Orig. CML-BGNN Ref. [52], as the graphic structural attention calculation mechanism began to show its advantages in modeling more flexible graph structure caused by the increased quantity of categorical samples and number of classes, which can promote intrinsic intermediate class correlations through the attention enhanced edges in longer node connection hops.
Among the three datasets, AID has the median number of classes and median categorical sample size. Such characteristics reduces the complexity of class correlations and is beneficial for methods with strong categorical feature encoding capability, and the data distribution stability in each category is increased. As a result of this, the accuracies of algorithms Few-Shot Multi-Atten. Ref. [31], ParamTrans. Ref. [56], and Know. Distill. Ref. [29] even surpassed our baseline algorithm Orig. CML-BGNN Ref. [52] using a continual meta-learning mechanism. Specifically, compared with Orig. CML-BGNN Ref. [52], Few-Shot Multi-Atten. Ref. [31] is advantageous in its adoption of an attention calculation mechanism, which is more suitable for occasions where greater inner class diversities exist. For ParamTrans. Ref. [56] and Know. Distill. Ref. [29], their superiorities came from the richness of samples for their offline knowledge transfer mechanism, which helps greatly in strengthening the stability and generalization capability from larger base training datasets and sophisticated teacher models. For our proposed algorithm, its main characteristics is the combinatorial usage of structural attention encoding via the graph transformer, and the online feature smoothing via a continual meta-learning mechanism. These characteristics facilitate the smoothing of the fluctuations in distributions of categorical features in embedded space, which boosts the performance of the proposed algorithm to surpass most of the state-of-art counterparts.

6.3. Knowledge Transition Efficiency Analysis

For meta-learning algorithms, data-to-knowledge transition efficiency is an important performance index, and we approximate it here via an empirical measurement as a ratio between the averaged accuracy of the model and the dataset usage. The dataset usage is defined as the ratio between the number of classes used in meta-training and the overall dataset class quantity, which is conceptually consistent with most counterpart algorithms.
Based on the definition, the knowledge transition efficiency of algorithms can be visually shown on a scatter plot, with the X-axis marking data usage and the Y-axis marking the algorithm evaluation accuracy. In this plot, algorithms with a higher accuracy and lower data usage would stay close to the upper left side.
In Figure 9, the efficiency evaluation result on a different dataset is marked with a different marker style. As shown, our proposed algorithm and its continual meta-learning baseline stayed at positions (within the dashed line circle) within the upper left region of the plot, indicating a good trade off between the training data quantity and overall performance on all three datasets. Such high efficiency comes from the effective online knowledge transfer through the continual meta-learning from sequences of meta-test tasks. Besides, Know. Distill Ref. [29] and ParamsTrans Ref. [56]. also achieved a highly knowledge transition ratio among the counterparts. These two algorithms showed its superiority of offline knowledge transfer as they surpassed the baseline Orig. CML-BGNN Ref. [52] on dataset AID, which proved to be efficient enough within a moderate class relationship and sample distribution complexity range. Through such a comparison, a continual meta-learning mechanism as well as its combination with graph structural attention calculation, proved to be a promising novel meta-learning technique for extending the flexibility of the model on application scenarios with a greater number of classes and higher quantities of sample distributions.

6.4. Classification Accuracy Details

Detailed classification result of our proposed algorithm is illustrated by the confusion matrices, as in Figure 10, showing the confusion matrices on datasets UC Merced, NWPU-RESISC45, and AID.
Interestingly, accuracies of the classes are rather close, which have a very small difference between each other. One major reason is the usage of an edge labeling graph for class discrimination and classification. In an edge labeling graph, classification depends on the correlation modeling through the weight of the edges. So suppose there is a class A with a high classification accuracy, then under the N-way K-shot meta-learning settings, that means some of the remaining N 1 categorical correlations given by the edges are considerably correct. Moreover, node correlations in graph models can be inferred through an intermediate connection at more distant edge connections, so such an accuracy can be propagated onto classes closer in probability distributions.
In Figure 10a, on dataset UC Merced, Chaparral is the second best class as it is the only natural scenario compared with others containing man-made objects. The land type storagetanks has the lowest accuracy as it can be found near a harbor, airport, and dense urban regions, thus a classifier made a high false positive rate on other land types except chaparral.
In Figure 10b, on dataset NWPU-RESISC45, accuracy differences between classes are slightly higher than on UC Merced, as there are more classes involved in the test, and some of them are not within the 2-hop length correlation with the best classified one. The best classified class golf_course reached an accuracy of 91.82%, as the composition of golf_course is generally simple, mainly consisting of sand pot and large meadows. While the storage_tank still remained at a lower level with large false positive confusion ratios with other land types.
In Figure 10c, on dataset AID, the best classified land type is resort with an accuracy of 88.35%, whose contents are plants, water bodies, and buildings mainly exists in park-like regions, whose texture patterns and shapes are very different from those in natural regions or normal residential areas, thus the false positive ratio on this land type is at a rather low rate.

6.5. Training Stability Analysis

In this subsection, we analyze the generalization and stability pattern of our proposed model on the three datasets, by comparing the changes in evaluation accuracies and evaluation losses during training to that of the continual meta-learning baseline model Orig. CML-BGNN Ref. [52]. Generally speaking, a robust model tends to have a more stable evolution pattern during training, where the changes in evaluation losses and evaluation are gradual and slow, which have very small fluctuations through the training process.
Figure 11 shows a comparison of changes in accuracies and losses between our proposed model and its baseline Orig. CML-BGNN Ref. [52]. From the figure, it can obviously be seen that our proposed model has smaller evaluation accuracy oscillations during the training process, and are mostly higher than the scores produced by the baseline algorithm except at the UC Merced dataset. In training losses, changes in the losses of our proposed algorithm are also smaller than those of the baseline algorithm, and always stay at a lower level except in the UC Merced dataset. The main reasons for the superiority of the baseline algorithm has been explained in Section 6.2, which is mainly caused by the fitness between data split and meta-test settings. So, according to the performance comparisons and the overall accuracies constrasting in Table 4, Table 5 and Table 6, the proposed algorithm had made better classification performances on NWPU-RESISC45 and AID, and shows a comparative accuracy on the UC Merced dataset.

7. Ablation Studies

7.1. Graph Transformer Heads

For the graph transformer-based graphic structure attention encoding module, the number of attention heads determine the quantity of significant structures it is capable to encode; and the higher the number of encoded structural pattern quantity, the greater the complexities in the graphic structure and categorical correlations it is able to describe.
Table 7 compares the classification performances of our proposed model on three datasets with different attention head quantities. From the table, an obvious equilibrium between the head quantity and dataset complexity can be seen. For UC Merced and NWPU-RESISC45 with fewer samples in each class, the classification accuracies declined as the number of heads increases, but at a rather small rate within 3%. While on AID, which has much more samples in each class, the increase of head quantity greatly boosts the performance with a rate of over 6%. So increasing the head numbers is an effective model tuning method when the inner class sample variances are large.
In Figure 12 and Figure 13, we show the classification confusion matrices and their difference with graph attention heads number equaling 4 and 16. In each figure, the left two sub-figures are confusion matrices with different head numbers, and the third one on the right is the difference matrix between the second and first confusion matrices. In the confusion matrices, differences matrix of the third sub-figure, categorical accuracies, and the false positive rates can be easily seen.
In Figure 12, on dataset UC Merced, as there are only five test classes and the edge labeling graph structure is fixed, using more attention heads will make the model more complex, thus becoming more prone to overfitting during training. Therefore, the false positive rates between complicated man-made scenarios such as mobile-home-park to airplane, storage-tanks to mobile-home-parks increases greatly, while the accuracy loss on a natural scenario chaparral is smaller.
In Figure 13, on dataset AID, an increased dataset complexity is fit for the graphic structural attention encoding mechanism where a greater modeling flexibility is required, thus the discriminative power increased in distinguishing complex scenery differences including river vs forest, river vs medium residential, viaduct vs medium residential etc. The added structural attention shown fitted for modeling the local texture patterns, making confusing classes easier for recognition.

7.2. Length of Continual Meta-Learning Iterations

The Continual Meta-learning Iterations Length (CMLL) is a strong control variable for the capability of online historical meta-testing knowledge utilization, where a longer iteration length ensures greater connectivity and more active information sharing between tasks previously processed in the continual meta-learning sequence.
Table 8 shows the classification accuracies on the three datasets with different continual meta learning iteration lengths. As the iteration goes longer, the changes in classification accuracies is roughly linear. Great leaps in accuracy improvements can be observed as the iteration length changed from 4 to 16 on all datasets, while a small accuracy decrease appeared as the length grows from 4 to 8. Such performance degradation is caused by the hidden embedding feature smoothing effect in the continual meta-learning mechanism, which happened when the distribution of categorical features is shrunk by averaging, while not strong enough to enlarge inter-class discrepancies but condensed the distribution density of samples in the confusing interlaced region between classes. Obviously, the iteration length of 16 is more appropriate for concentrating the categorical features.
The detailed classification accuracy changes from length 4 to 8 and from length 4 to 16 are shown in Figure 14 and Figure 15. As can be seen, a large false positive rate increments and decrements mainly happened on classes including complex man-made land types, for instance playground vs medium residential and viaduct vs forest, where greater feature space interlacing exists between them caused by the shared complex ground structures and similar ground objects. Class pairs with significant structural differences are lesser influenced by the changes in continual meta learning length, as the semantic confusion between them is weaker.
Finally, the changes in class accuracies are illustrated as bar plots in Figure 16. It can be seen that classes having higher accuracies in length 4 setting tend to keep its rank in the 16 setting, having more increments in accuracy. Thus, enlarging the iteration length of continual learning module is more like an equal scaling of the classification discriminative power, which will not change the relation positions of categorical features in the embedded feature space.

7.3. Number of Layers in Bayesian Edge Labeling Graph

The number of layers in the Bayesian edge labeling graph module decides the modeling capabilities for complex node feature correlations, so there is also an equilibrium that needs to be balanced between the complexities of the dataset and model.
Table 9 shows the contrasts in classification accuracies on the three test datasets when different a Bayesian edge labeling graph layer number is chosen. It can be easily observed that the best performance of our model appears with the choice of 2 layers, where the data and model complexities matched well.
Figure 17 shows the confusion matrices on dataset AID when the graph labeling model layer quantities are set to 1, 2, and 3. With the number of layers increased, the false positive ratios increased almost monotonic in classes such as playground and beach, where differences in the scenario content exists in multiple spatial scales. While for classes sharing partial scenario contents or ground object types, such as a playground and resort, using a deeper graph labeling layer depth only increases the semantic confusions between classes, which jeopardize the overall performance.
Finally, the varied accuracies of all classes using different layer quantities is shown in Figure 18 for explicit comparisons. In it, land types with significant content differences from others, such as beach and viaduct, their classification accuracies tend to be less affected by the changes in the graph labeling layers.

8. Conclusions

As the demands of intelligent remote sensing imagery interpretation increases, the necessity of developing fast model adaption methods that is less dependent on human supervision and a smaller dataset grows significantly. Few-shot remote sensing scene classification is one such fundamental technique that is indispensable in a large variety of tasks. Among the popular few-shot learning methods, this article studies efficient algorithms based on the continual meta-learning principle, which is derived from meta-learning methods, but can alleviate the data isolation in standard meta-learning training schema by promoting the utilization of historical prior during the online meta-testing procedure. The proposed algorithm introduced a novel and efficient composition of latest techniques, including graph neural network-based classification and graph transformer-based structural attention encoding under the continual learning framework, which can significantly improve flexibility in modeling the complex categorical structure under data urgent conditions. In the proposed algorithm structure, node features encoded from support and query samples are further enhanced by the graph transformer calculating structural attention for better intra and inter class discrimination. The GRU-based continual meta-learning module aggregates the distribution of node features to the class centers and enlarges the categorical discrepancies. In addition, the node correlations modeled by edge weights are rectified by both the primitive node feature distances and the Bayesian-style Gaussian estimations. The advantages of the proposed model are clearly illustrated through the comparisons with other standard meta-learning-based counterparts, where a minimum of 9% leap in accuracy can be observed. The effectiveness of the novel combination structural attention computation being proposed is shown by the comparison with the baseline algorithm. Despite this, much work still needs to be done in further studies. Our future work will focus on improving the flexibility and scalability of the continual meta-learning-based tasks period sampling schema, thus making the algorithm fit for a wider range of dataset sizes and complexities, and is more useful for real life applications.

Author Contributions

Conceptualization, F.L. and H.C.; methodology, F.L. and S.L.; software, F.L. and S.L.; validation, F.L. and X.L.; formal analysis, F.L.; investigation, F.L. and S.L.; resources, F.L.; data curation, S.L.; writing—original draft preparation, F.L., S.L. and X.F.; writing—review and editing, F.L.; visualization, F.L.; supervision, H.C.; project administration, F.L.; funding acquisition, F.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (No. 61806199).

Data Availability Statement

Not applicable.

Acknowledgments

We would like to thank Section Managing Editor Chryssa Liu, Assistant Editor Trista Yang, Assistant Editor Hope Zhu, Assistant Editor Adriana Wang, for their patience and responsibility in processing our manuscript and sending notification emails. We would also like to thank the professors for using their valuable time during the end of semester to review our manuscript and give so many constructive comments.

Conflicts of Interest

The authors declare no conflict of interests.

Abbreviations

The following abbreviations are used in this manuscript:
AIDAerial Image Dataset
BGNNBayesian Graph Neural Network
CMLContinual Meta-Learning
CNNConvolutional Neural Network
DLRGerman Aerospace Center
FSCFew-Shot Classification
FSLFew-Shot Learning
GNNGraph Neural Network
GPNGated Propagation Network
GRUGate Recurrent Unit
HGNNHierarchical Graph Neural Network
INRIAThe National Institute for Research in Computer Science and Automation
LSTMLong Short-Term Memory
MAMLModel Agnostic Meta-Learning
MLMeta-Learning
NWPUNorthwestern Polytechnical University
RESISCRemote Sensing Image Scene Classification
RNNRecurrent Neural Network
SAMSelf-Attention Meta-Learner
UCMUC Merced Landuse Dataset

References

  1. Domeneghetti, A.; Schumann, G.J.P.; Tarpanelli, A. Preface: Remote Sensing for Flood Mapping and Monitoring of Flood Dynamics. Remote Sens. 2019, 11, 943. [Google Scholar] [CrossRef] [Green Version]
  2. Zhou, W.; Ming, D.; Lv, X.; Zhou, K.; Bao, H.; Hong, Z. SO–CNN based urban functional zone fine division with VHR remote sensing image. Remote Sens. Environ. 2020, 236, 111458. [Google Scholar] [CrossRef]
  3. Jahromi, M.N.; Jahromi, M.N.; Pourghasemi, H.R.; Zand-Parsa, S.; Jamshidi, S. Chapter 12—Accuracy assessment of forest mapping in MODIS land cover dataset using fuzzy set theory. In Forest Resources Resilience and Conflicts; Jahromia, M.N., Jahromib, M.N., Pourghasemic, H.R., Zand-Parsab, S., Jamshidibd, S., Eds.; Elsevier: Amsterdam, The Netherlands, 2021; pp. 165–183. [Google Scholar] [CrossRef]
  4. Avtar, R.; Sahu, N.; Aggarwal, A.K.; Chakraborty, S.; Kharrazi, A.; Yunus, A.P.; Dou, J.; Kurniawan, T.A. Exploring Renewable Energy Resources Using Remote Sensing and GIS—A Review. Resources 2019, 8, 149. [Google Scholar] [CrossRef] [Green Version]
  5. Li, Y.; Yang, J. Meta-learning baselines and database for few-shot classification in agriculture. Comput. Electron. Agric. 2021, 182, 106055. [Google Scholar] [CrossRef]
  6. Cheng, G.; Xie, X.; Han, J.; Guo, L.; Xia, G.S. Remote Sensing Image Scene Classification Meets Deep Learning: Challenges, Methods, Benchmarks, and Opportunities. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3735–3756. [Google Scholar] [CrossRef]
  7. Zhang, G.; Xu, W.; Zhao, W.; Huang, C.; Yk, E.N.; Chen, Y.; Su, J. A Multiscale Attention Network for Remote Sensing Scene Images Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 9530–9545. [Google Scholar] [CrossRef]
  8. Akkul, E.S.; Arıcan, B.; Töreyin, B.U. Bi-RDNet: Performance Enhancement for Remote Sensing Scene Classification with Rotational Duplicate Layers. In Proceedings of the International Conference on Computational Collective Intelligence, Rhodos, Greece, 29 September–1 October 2021; pp. 669–678. [Google Scholar]
  9. Deng, P.; Xu, K.; Huang, H. When CNNs Meet Vision Transformer: A Joint Framework for Remote Sensing Scene Classification. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  10. Zheng, J.; Wu, W.; Yuan, S.; Zhao, Y.; Li, W.; Zhang, L.; Dong, R.; Fu, H. A Two-Stage Adaptation Network (TSAN) for Remote Sensing Scene Classification in Single-Source-Mixed-Multiple-Target Domain Adaptation (S²M²T DA) Scenarios. IEEE Trans. Geosci. Remote Sens. 2021. Early Access. [Google Scholar] [CrossRef]
  11. Zhang, J.; Liu, J.; Pan, B.; Chen, Z.; Xu, X.; Shi, Z. An Open Set Domain Adaptation Algorithm via Exploring Transferability and Discriminability for Remote Sensing Image Scene Classification. IEEE Trans. Geosci. Remote Sens. 2021. Early Access. [Google Scholar] [CrossRef]
  12. Rusu, A.A.; Rao, D.; Sygnowski, J.; Vinyals, O.; Pascanu, R.; Osindero, S.; Hadsell, R. Meta-Learning with Latent Embedding Optimization. In Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  13. Sun, Q.; Liu, Y.; Chua, T.S.; Schiele, B. Meta-Transfer Learning for Few-Shot Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  14. Kang, B.; Liu, Z.; Wang, X.; Yu, F.; Feng, J.; Darrell, T. Few-shot object detection via feature reweighting. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 8420–8429. [Google Scholar]
  15. ZHANG, R.; Che, T.; Ghahramani, Z.; Bengio, Y.; Song, Y. MetaGAN: An Adversarial Approach to Few-Shot Learning. In Advances in Neural Information Processing Systems; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
  16. Verma, V.K.; Brahma, D.; Rai, P. Meta-Learning for Generalized Zero-Shot Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 34, pp. 6062–6069. [Google Scholar] [CrossRef]
  17. Mishra, A.; Verma, V.K.; Reddy, M.S.K.; Subramaniam, A.; Rai, P.; Mittal, A. A Generative Approach to Zero-Shot and Few-Shot Action Recognition. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision, Lake Tahoe, NV, USA, 12–15 March 2018; pp. 372–380. [Google Scholar] [CrossRef] [Green Version]
  18. Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast and Flexible Image Augmentations. Information 2020, 11, 125. [Google Scholar] [CrossRef] [Green Version]
  19. Chen, W.; Liu, Y.; Kira, Z.; Wang, Y.; Huang, J. A Closer Look at Few-shot Classification. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  20. Teshima, T.; Sato, I.; Sugiyama, M. Few-shot Domain Adaptation by Causal Mechanism Transfer. In Proceedings of the 37th International Conference on Machine Learning, Vienna, Austria, 13–18 July 2020; Volume 119, pp. 9458–9469. [Google Scholar]
  21. Zhang, C.; Lin, G.; Liu, F.; Yao, R.; Shen, C. CANet: Class-Agnostic Segmentation Networks with Iterative Refinement and Attentive Few-Shot Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  22. Liu, W.; Zhang, C.; Lin, G.; Liu, F. CRNet: Cross-Reference Networks for Few-Shot Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–18 June 2020. [Google Scholar]
  23. Chen, G.; Zhang, X.; Tan, X.; Cheng, Y.; Dai, F.; Zhu, K.; Gong, Y.; Wang, Q. Training Small Networks for Scene Classification of Remote Sensing Images via Knowledge Distillation. Remote Sens. 2018, 10, 719. [Google Scholar] [CrossRef] [Green Version]
  24. Gao, K.; Liu, B.; Yu, X.; Qin, J.; Zhang, P.; Tan, X. Deep Relation Network for Hyperspectral Image Few-Shot Classification. Remote Sens. 2020, 12, 923. [Google Scholar] [CrossRef] [Green Version]
  25. Liu, B.; Yu, X.; Yu, A.; Zhang, P.; Wan, G.; Wang, R. Deep Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2290–2304. [Google Scholar] [CrossRef]
  26. Rostami, M.; Kolouri, S.; Eaton, E.; Kim, K. Deep Transfer Learning for Few-Shot SAR Image Classification. Remote Sens. 2019, 11, 1374. [Google Scholar] [CrossRef] [Green Version]
  27. Zeng, Q.; Geng, J.; Huang, K.; Jiang, W.; Guo, J. Prototype Calibration with Feature Generation for Few-Shot Remote Sensing Image Scene Classification. Remote Sens. 2021, 13, 2728. [Google Scholar] [CrossRef]
  28. Xiong, W.; Gong, Y.; Niu, W.; Wang, R. Hierarchically Structured Network with Attention Convolution and Embedding Propagation for Imbalanced Few-Shot Learning. IEEE Access 2021, 9, 93916–93926. [Google Scholar] [CrossRef]
  29. Liu, Y.; Zhang, L.; Han, Z.; Chen, C. Integrating Knowledge Distillation with Learning to Rank for Few-Shot Scene Classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–12. [Google Scholar] [CrossRef]
  30. Zhang, P.; Li, Y.; Wang, D.; Wang, J. RS-SSKD: Self-Supervision Equipped with Knowledge Distillation for Few-Shot Remote Sensing Scene Classification. Sensors 2021, 21, 1566. [Google Scholar] [CrossRef] [PubMed]
  31. Yuan, Z.; Huang, W.; Li, L.; Luo, X. Few-Shot Scene Classification with Multi-Attention Deepemd Network in Remote Sensing. IEEE Access 2021, 9, 19891–19901. [Google Scholar] [CrossRef]
  32. Kim, J.; Chi, M. SAFFNet: Self-Attention-Based Feature Fusion Network for Remote Sensing Few-Shot Scene Classification. Remote Sens. 2021, 13, 2532. [Google Scholar] [CrossRef]
  33. Li, X.; Pu, F.; Yang, R.; Gui, R.; Xu, X. AMN: Attention Metric Network for One-Shot Remote Sensing Image Scene Classification. Remote Sens. 2020, 12, 4046. [Google Scholar] [CrossRef]
  34. Vanschoren, J. Meta-Learning: A Survey. arXiv 2018, arXiv:1810.03548. [Google Scholar]
  35. Hsu, K.; Levine, S.; Finn, C. Unsupervised Learning via Meta-Learning. In Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  36. Javed, K.; White, M. Meta-Learning Representations for Continual Learning. In Advances in Neural Information Processing Systems 32, Proceedings of the Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; pp. 1818–1828. [Google Scholar]
  37. Tripuraneni, N.; Jin, C.; Jordan, M.I. Provable Meta-Learning of Linear Representations. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, Virtual Event, 18–24 July 2021; Volume 139, pp. 10434–10443. [Google Scholar]
  38. Saunshi, N.; Gupta, A.; Hu, W. A Representation Learning Perspective on the Importance of Train-Validation Splitting in Meta-Learning. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, Virtual Event, 18–24 July 2021; Volume 139, pp. 9333–9343. [Google Scholar]
  39. Reif, M.; Shafait, F.; Dengel, A. Meta-learning for evolutionary parameter optimization of classifiers. Mach. Learn. 2012, 87, 357–380. [Google Scholar] [CrossRef] [Green Version]
  40. Rajeswaran, A.; Finn, C.; Kakade, S.M.; Levine, S. Meta-Learning with Implicit Gradients. In Advances in Neural Information Processing Systems 32, Proceedings of the Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019; Wallach, H.M., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E.B., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; pp. 113–124. [Google Scholar]
  41. Franceschi, L.; Frasconi, P.; Salzo, S.; Grazzi, R.; Pontil, M. Bilevel Programming for Hyperparameter Optimization and Meta-Learning. In Proceedings of the 35th International Conference on Machine Learning, Stockholmsmässan, ICML 2018, Stockholm, Sweden, 10–15 July 2018; Volume 80, pp. 1563–1572. [Google Scholar]
  42. Bechtle, S.; Molchanov, A.; Chebotar, Y.; Grefenstette, E.; Righetti, L.; Sukhatme, G.S.; Meier, F. Meta Learning via Learned Loss. In Proceedings of the 25th International Conference on Pattern Recognition, Milan, Italy, 10–15 January 2021; pp. 4161–4168. [Google Scholar] [CrossRef]
  43. Alet, F.; Lozano-Pérez, T.; Kaelbling, L.P. Modular meta-learning. In Proceedings of the 2nd Annual Conference on Robot Learning, CoRL 2018, Zürich, Switzerland, 29–31 October 2018; Volume 87, pp. 856–868. [Google Scholar]
  44. Yu, T.; Finn, C.; Xie, A.; Dasari, S.; Zhang, T.; Abbeel, P.; Levine, S. One-Shot Imitation from Observing Humans via Domain-Adaptive Meta-Learning. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  45. Yoon, J.; Kim, T.; Dia, O.; Kim, S.; Bengio, Y.; Ahn, S. Bayesian Model-Agnostic Meta-Learning. In Advances in Neural Information Processing Systems 31, Proceedings of the Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, Montréal, Canada, 3–8 December 2018; Bengio, S., Wallach, H.M., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Montreal, QC, Canada, 2018; pp. 7343–7353. [Google Scholar]
  46. Chen, J.; Zhan, L.; Wu, X.; Chung, F. Variational Metric Scaling for Metric-Based Meta-Learning. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020; pp. 3478–3485. [Google Scholar]
  47. Alet, F.; Schneider, M.F.; Lozano-Pérez, T.; Kaelbling, L.P. Meta-learning curiosity algorithms. In Proceedings of the 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  48. Houthooft, R.; Chen, Y.; Isola, P.; Stadie, B.C.; Wolski, F.; Ho, J.; Abbeel, P. Evolved Policy Gradients. In Advances in Neural Information Processing Systems 31, Proceedings of the Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, Montréal, Canada, 3–8 December 2018; Bengio, S., Wallach, H.M., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Montreal, QC, Canada, 2018; pp. 5405–5414. [Google Scholar]
  49. Xu, K.; Ratner, E.; Dragan, A.D.; Levine, S.; Finn, C. Learning a Prior over Intent via Meta-Inverse Reinforcement Learning. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 6952–6962. [Google Scholar]
  50. Liu, L.; Zhou, T.; Long, G.; Jiang, J.; Zhang, C. Learning to Propagate for Graph Meta-Learning. In Advances in Neural Information Processing Systems 32, Proceedings of the Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, 8–14 December 2019, Vancouver, BC, Canada; Wallach, H.M., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E.B., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; pp. 1037–1048. [Google Scholar]
  51. Zhou, F.; Cao, C.; Zhang, K.; Trajcevski, G.; Zhong, T.; Geng, J. Meta-GNN: On Few-shot Node Classification in Graph Meta-learning. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, CIKM 2019, Beijing, China, 3–7 November 2019; pp. 2357–2360. [Google Scholar] [CrossRef]
  52. Luo, Y.; Huang, Z.; Zhang, Z.; Wang, Z.; Baktashmotlagh, M.; Yang, Y. Learning from the Past: Continual Meta-Learning with Bayesian Graph Neufral Networks. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020; pp. 5021–5028. [Google Scholar]
  53. Li, H.; Cui, Z.; Zhu, Z.; Chen, L.; Zhu, J.; Huang, H.; Tao, C. RS-MetaNet: Deep Metametric Learning for Few-Shot Remote Sensing Scene Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6983–6994. [Google Scholar] [CrossRef]
  54. Zhang, P.; Bai, Y.; Wang, D.; Bai, B.; Li, Y. Few-Shot Classification of Aerial Scene Images via Meta-Learning. Remote Sens. 2021, 13, 108. [Google Scholar] [CrossRef]
  55. Li, Y.; Shao, Z.; Huang, X.; Cai, B.; Peng, S. Meta-FSEO: A Meta-Learning Fast Adaptation with Self-Supervised Embedding Optimization for Few-Shot Remote Sensing Scene Classification. Remote Sens. 2021, 13, 2776. [Google Scholar] [CrossRef]
  56. Ma, C.; Mu, X.; Zhao, P.; Yan, X. Meta-learning based on parameter transfer for few-shot classification of remote sensing scenes. Remote Sens. Lett. 2021, 12, 531–541. [Google Scholar] [CrossRef]
  57. Kim, J.; Kim, T.; Kim, S.; Yoo, C.D. Edge-Labeling Graph Neural Network for Few-Shot Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 11–20. [Google Scholar] [CrossRef] [Green Version]
  58. Cai, D.; Lam, W. Graph Transformer for Graph-to-Sequence Learning. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020; pp. 7464–7471. [Google Scholar]
  59. Hospedales, T.M.; Antoniou, A.; Micaelli, P.; Storkey, A.J. Meta-Learning in Neural Networks: A Survey. arXiv 2020, arXiv:2004.05439. [Google Scholar] [CrossRef] [PubMed]
  60. Sokar, G.; Mocanu, D.C.; Pechenizkiy, M. Self-Attention Meta-Learner for Continual Learning. In Proceedings of the 20th International Conference on Autonomous Agents and Multiagent Systems, AAMAS’21, Virtual Event, 3–7 May 2021; pp. 1658–1660. [Google Scholar]
  61. Vuorio, R.; Cho, D.; Kim, D.; Kim, J. Meta Continual Learning. arXiv 2018, arXiv:1806.06928. [Google Scholar]
  62. Jerfel, G.; Grant, E.; Griffiths, T.; Heller, K.A. Reconciling meta-learning and continual learning with online mixtures of tasks. In Advances in Neural Information Processing Systems 32, Proceedings of the Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019; Wallach, H.M., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E.B., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; pp. 9119–9130. [Google Scholar]
  63. Harrison, J.; Sharma, A.; Finn, C.; Pavone, M. Continuous Meta-Learning without Tasks. In Advances in Neural Information Processing Systems 33, Proceedings of the Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, Virtual, 6–12 December 2020; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2020. [Google Scholar]
  64. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 4–24. [Google Scholar] [CrossRef] [Green Version]
  65. Wang, X.; He, X.; Wang, M.; Feng, F.; Chua, T. Neural Graph Collaborative Filtering. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2019, Paris, France, 21–25 July 2019; pp. 165–174. [Google Scholar] [CrossRef] [Green Version]
  66. Fan, W.; Ma, Y.; Li, Q.; Wang, J.; Cai, G.; Tang, J.; Yin, D. A Graph Neural Network Framework for Social Recommendations. IEEE Trans. Knowl. Data Eng. 2020. Early Access. [Google Scholar] [CrossRef]
  67. Hao, Z.; Lu, C.; Huang, Z.; Wang, H.; Hu, Z.; Liu, Q.; Chen, E.; Lee, C. ASGN: An Active Semi-supervised Graph Neural Network for Molecular Property Prediction. In Proceedings of the 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD’20, Virtual Event, 23–27 August 2020; Gupta, R., Liu, Y., Tang, J., Prakash, B.A., Eds.; ACM: New York, NY, USA, 2020; pp. 731–752. [Google Scholar] [CrossRef]
  68. Chen, C.; Li, K.; Wei, W.; Zhou, J.T.; Zeng, Z. Hierarchical Graph Neural Networks for Few-Shot Learning. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 240–252. [Google Scholar] [CrossRef]
  69. Yuan, L.; Chen, Y.; Wang, T.; Yu, W.; Shi, Y.; Tay, F.E.H.; Feng, J.; Yan, S. Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet. arXiv 2021, arXiv:2101.11986. [Google Scholar]
  70. Zhang, D.; Zhang, H.; Tang, J.; Wang, M.; Hua, X.; Sun, Q. Feature Pyramid Transformer. In Computer Vision—ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Lecture Notes in Computer, Science; Part, XXVIII; Vedaldi, A., Bischof, H., Brox, T., Frahm, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2020; Volume 12373, pp. 323–339. [Google Scholar] [CrossRef]
  71. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. arXiv 2021, arXiv:2103.14030. [Google Scholar]
  72. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for Semantic Segmentation. arXiv 2021, arXiv:2105.05633. [Google Scholar]
  73. Choi, E.; Xu, Z.; Li, Y.; Dusenberry, M.; Flores, G.; Xue, E.; Dai, A.M. Learning the Graphical Structure of Electronic Health Records with Graph Convolutional Transformer. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020; pp. 606–613. [Google Scholar]
  74. Qu, M.; Gao, T.; Xhonneux, L.A.C.; Tang, J. Few-shot Relation Extraction via Bayesian Meta-learning on Relation Graphs. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, Virtual Event, 13–18 July 2020; Volume 119, pp. 7867–7876. [Google Scholar]
  75. Nguyen, H.D.; Vu, X.; Le, D. Modular Graph Transformer Networks for Multi-Label Image Classification. In Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, 2–9 February 2021; pp. 9092–9100. [Google Scholar]
  76. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.; Koltun, V. Point Transformer. arXiv 2020, arXiv:2012.09164. [Google Scholar]
  77. Yu, C.; Ma, X.; Ren, J.; Zhao, H.; Yi, S. Spatio-Temporal Graph Transformer Networks for Pedestrian Trajectory Prediction. In Computer Vision—ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Lecture Notes in Computer Science; Part XII; Vedaldi, A., Bischof, H., Brox, T., Frahm, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2020; Volume 12357, pp. 507–523. [Google Scholar] [CrossRef]
  78. Yang, Y.; Newsam, S. Bag-of-Visual-Words and Spatial Extensions for Land-Use Classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems, GIS’10, San Jose, CA, USA, 3–5 November 2010; pp. 270–279. [Google Scholar] [CrossRef]
  79. Cheng, G.; Han, J.; Lu, X. Remote Sensing Image Scene Classification: Benchmark and State of the Art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef] [Green Version]
  80. Xia, G.S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L.; Lu, X. AID: A Benchmark Data Set for Performance Evaluation of Aerial Scene Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3965–3981. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Illustrative comparisons of different training modes. (a) Traditional training mode, monolithic train and test periods. (b) The typical meta-learning training and testing in units of tasks. (c) The typical continual meta-learning training and testing procedure adopted in this paper.
Figure 1. Illustrative comparisons of different training modes. (a) Traditional training mode, monolithic train and test periods. (b) The typical meta-learning training and testing in units of tasks. (c) The typical continual meta-learning training and testing procedure adopted in this paper.
Remotesensing 14 00485 g001
Figure 2. Overview: The proposed continual meta-learning few-shot scene classification algorithm is composed of three main components: The encoding module, the GRU-based continual meta-learning module, and the Bayesian graph neural network-based graph edge labeling module.
Figure 2. Overview: The proposed continual meta-learning few-shot scene classification algorithm is composed of three main components: The encoding module, the GRU-based continual meta-learning module, and the Bayesian graph neural network-based graph edge labeling module.
Remotesensing 14 00485 g002
Figure 3. Illustration of the processing procedure of the encoding module.
Figure 3. Illustration of the processing procedure of the encoding module.
Remotesensing 14 00485 g003
Figure 4. Illustration of the GRU-based sequential encoding of features in tasks in the continual meta-learning module.
Figure 4. Illustration of the GRU-based sequential encoding of features in tasks in the continual meta-learning module.
Remotesensing 14 00485 g004
Figure 5. Illustration of the calculation workflow in the Bayesian graph labeling module.
Figure 5. Illustration of the calculation workflow in the Bayesian graph labeling module.
Remotesensing 14 00485 g005
Figure 6. The train, validate, and test partitions of dataset UC Merced.
Figure 6. The train, validate, and test partitions of dataset UC Merced.
Remotesensing 14 00485 g006
Figure 7. The train, validation, and test partitions of dataset NWPU-RESISC45.
Figure 7. The train, validation, and test partitions of dataset NWPU-RESISC45.
Remotesensing 14 00485 g007
Figure 8. The train, validation, and test partitions of dataset AID.
Figure 8. The train, validation, and test partitions of dataset AID.
Remotesensing 14 00485 g008
Figure 9. Data-to-knowledge transition efficiency comparison scatter plot. The positions of proposed algorithms are marked with red dashed circle.
Figure 9. Data-to-knowledge transition efficiency comparison scatter plot. The positions of proposed algorithms are marked with red dashed circle.
Remotesensing 14 00485 g009
Figure 10. Classification confusion matrix on datasets. (a) Classification confusion matrix on dataset UC Merced; (b) Classification confusion matrix on dataset NWPU-RESISC45; (c) Classification confusion matrix on dataset AID.
Figure 10. Classification confusion matrix on datasets. (a) Classification confusion matrix on dataset UC Merced; (b) Classification confusion matrix on dataset NWPU-RESISC45; (c) Classification confusion matrix on dataset AID.
Remotesensing 14 00485 g010
Figure 11. Evaluation accuracies and losses comparisons with the baseline model during training. (a) UC Merced dataset; (b) NWPU-RESISC45 dataset; (c) AID dataset.
Figure 11. Evaluation accuracies and losses comparisons with the baseline model during training. (a) UC Merced dataset; (b) NWPU-RESISC45 dataset; (c) AID dataset.
Remotesensing 14 00485 g011
Figure 12. Confusion matrices and their difference on dataset UC Merced with graph transformer attention heads equaling 4 and 16.
Figure 12. Confusion matrices and their difference on dataset UC Merced with graph transformer attention heads equaling 4 and 16.
Remotesensing 14 00485 g012
Figure 13. Confusion matrices and their difference on dataset AID with graph transformer attention heads equaling 4 and 16.
Figure 13. Confusion matrices and their difference on dataset AID with graph transformer attention heads equaling 4 and 16.
Remotesensing 14 00485 g013
Figure 14. Confusion matrices and their difference on dataset AID with continual meta-learning iteration lengths 4 and 8.
Figure 14. Confusion matrices and their difference on dataset AID with continual meta-learning iteration lengths 4 and 8.
Remotesensing 14 00485 g014
Figure 15. Confusion matrices and their difference on dataset AID with continual meta-learning iteration lengths 4 and 16.
Figure 15. Confusion matrices and their difference on dataset AID with continual meta-learning iteration lengths 4 and 16.
Remotesensing 14 00485 g015
Figure 16. Comparison of class accuracies when different Continual Meta-learning Iteration Lengths (CMLL) are 4, 8, and 16.
Figure 16. Comparison of class accuracies when different Continual Meta-learning Iteration Lengths (CMLL) are 4, 8, and 16.
Remotesensing 14 00485 g016
Figure 17. Confusion matrices comparison between settings using a different quantity of Bayesian edge labeling graph layers.
Figure 17. Confusion matrices comparison between settings using a different quantity of Bayesian edge labeling graph layers.
Remotesensing 14 00485 g017
Figure 18. Comparison of class accuracies when a different number of Bayesian edge labeling graph layers is used.
Figure 18. Comparison of class accuracies when a different number of Bayesian edge labeling graph layers is used.
Remotesensing 14 00485 g018
Table 1. Class segmentation for the UCM dataset.
Table 1. Class segmentation for the UCM dataset.
TrainValidateTest
agricultural,
beach,
denseresidential,
freeway,
golfcourse,
intersection,
mediumresidential,
parkinglot, river,
runway,
sparseresidential
baseballdiamond,
buildings,
forest,
overpass,
tenniscourt
airplane,
chaparral,
harbor,
mobilehomepark,
storagetanks
Table 2. Class segmentation for the NWPU-RESISC45 dataset.
Table 2. Class segmentation for the NWPU-RESISC45 dataset.
TrainValidateTest
airplane, basketball_court,
bridge, church,
circular_farmland,
dense_residential,
forest, freeway,
ground_track_field,
industrial_area,
intersection, island,
meadow, medium_residential,
mountain, overpass,
palace, railway_station,
rectangular_farmland,
roundabout, runway,
sea_ice, sparse_residential,
tennis_court,
terrace, wetland
baseball_diamond,
chaparral,
cloud,
desert,
mobile_home_park,
palace,
railway,
ship,
stadium,
thermal_power_station
airport,
beach,
commercial_area,
golf_course,
harbor,
lake,
parking_lot,
river,
snowberg,
storage_tank
Table 3. Class segmentation for the AID dataset.
Table 3. Class segmentation for the AID dataset.
TrainValidateTest
Airport,
BaseballField, Center,
Commercial,
DenseResidential,
Farmland, Meadow,
Park, Parking, Pond,
RailwayStation, School,
SparseResidential,
Stadium, StorageTanks
Bareland,
Bridge,
Church,
Desert,
Industrial,
Mountain,
Port,
Square
Beach,
Forest,
Medium Residual,
Playground,
Resort,
River,
Viaduct
Table 4. Classification accuracies on UC Merced.
Table 4. Classification accuracies on UC Merced.
MethodTraining RatioBackboneAccuracy
RS-MetaNet [53]80.00%ResNet5053.57%
Few-Shot Multi-Atten. [31]76.19%ResNet1861.16%
SAFFNet [32]52.38%ResNet1865.89%
ParamTrans. [56]40.00%ResNet1262.96%
Orig. CML-BGNN [52]52.38%ResNet1889.13%
Proposed52.38%ResNet1888.56%
Best and 2nd best accuracies are emphasize with bold font style and underline.
Table 5. Classification accuracies on NWPU-RESISC45.
Table 5. Classification accuracies on NWPU-RESISC45.
MethodTraining RatioBackboneAccuracy
RS-MetaNet [53]80.00%ResNet5046.32%
Few-Shot Aerial [31]55.56%ResNet1269.68%
RS-SSKD [30]55.56%ResNet1270.86%
Know. Distill. [29]62.22%Conv-473.86%
Proto. Calib. [27]55.56%-72.80%
SAFFNet [32]51.11%ResNet1864.63%
AMN [33]73.33%ResNet1874.25%
ParamTrans. [56]40.00%ResNet1267.14%
Orig. CML-BGNN [52]55.56%ResNet1885.63%
Proposed55.56%ResNet1890.71%
Best and 2nd best accuracies are emphasize with bold font style and underline.
Table 6. Classification accuracies on AID.
Table 6. Classification accuracies on AID.
MethodTraining RatioBackboneAccuracy
RS-MetaNet [53]80.00%ResNet5054.26%
Few-Shot Multi-Atten. [31]80.00%ResNet1874.52%
Know. Distill. [29]43.33%Conv-478.47%
SAFFNet [32]50.00%ResNet1867.88%
ParamTrans. [56]40.00%ResNet1277.15%
Orig. CML-BGNN [52]50.00%ResNet1871.35%
Proposed50.00%ResNet1887.60%
Best and 2nd best accuracies are emphasize with bold font style and underline.
Table 7. Influences of the number of attention heads of the graph transformer on the performance.
Table 7. Influences of the number of attention heads of the graph transformer on the performance.
Num. HeadsClassification Accuracy
UC MercedNWPU-RESISC45AID
Num. Heads 486.59%87.53%71.67%
Num. Heads 884.88%85.08%77.77%
Num. Heads 1681.79%83.34%85.03%
Table 8. Influences of the number of iterations of the GRU-CML module on performance.
Table 8. Influences of the number of iterations of the GRU-CML module on performance.
Num. Iter.(s)Classification Accuracy
UC MercedNWPU-RESISC45AID
Num. Iter.(s) 476.56%69.84%75.02%
Num. Iter.(s) 872.80%75.40%73.36%
Num. Iter.(s) 1681.79%83.34%85.03%
Table 9. Influences of the number of Bayesian edge labeling graph layers on performance.
Table 9. Influences of the number of Bayesian edge labeling graph layers on performance.
Num. LayersClassification Accuracy
UC MercedNWPU-RESISC45AID
Num. Layers 176.76%82.91%83.73%
Num. Layers 288.56%90.71%87.60%
Num. Layers 381.79%83.34%85.03%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, F.; Li, S.; Fan, X.; Li, X.; Chang, H. Structural Attention Enhanced Continual Meta-Learning for Graph Edge Labeling Based Few-Shot Remote Sensing Scene Classification. Remote Sens. 2022, 14, 485. https://doi.org/10.3390/rs14030485

AMA Style

Li F, Li S, Fan X, Li X, Chang H. Structural Attention Enhanced Continual Meta-Learning for Graph Edge Labeling Based Few-Shot Remote Sensing Scene Classification. Remote Sensing. 2022; 14(3):485. https://doi.org/10.3390/rs14030485

Chicago/Turabian Style

Li, Feimo, Shuaibo Li, Xinxin Fan, Xiong Li, and Hongxing Chang. 2022. "Structural Attention Enhanced Continual Meta-Learning for Graph Edge Labeling Based Few-Shot Remote Sensing Scene Classification" Remote Sensing 14, no. 3: 485. https://doi.org/10.3390/rs14030485

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop