Next Article in Journal
Disaggregating National Statistical Data to Assess the Sub-National SDG 6.4.2 Water Stress Indicator by Integrating Satellite Observations and Model Data
Next Article in Special Issue
Enhanced Wind Field Spatial Downscaling Method Using UNET Architecture and Dual Cross-Attention Mechanism
Previous Article in Journal
Identification of Dominant Species and Their Distributions on an Uninhabited Island Based on Unmanned Aerial Vehicles (UAVs) and Machine Learning Models
Previous Article in Special Issue
Evaluation of Ten Deep-Learning-Based Out-of-Distribution Detection Methods for Remote Sensing Image Scene Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Zero-Shot Sketch-Based Remote-Sensing Image Retrieval Based on Multi-Level and Attention-Guided Tokenization

1
Anhui Province Key Laboratory of Wetland Ecosystem Protection and Restoration, Anhui University, Hefei 230601, China
2
School of Resources and Environmental Engineering, Anhui University, Hefei 230601, China
3
Shanghai Ubiquitous Navigation Technology Co., Ltd., Shanghai 201702, China
4
The 54th Research Institute of China Electronics Technology Group Corporation, Shijiazhuang 050081, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(10), 1653; https://doi.org/10.3390/rs16101653
Submission received: 11 March 2024 / Revised: 1 May 2024 / Accepted: 2 May 2024 / Published: 7 May 2024

Abstract

:
Effectively and efficiently retrieving images from remote-sensing databases is a critical challenge in the realm of remote-sensing big data. Utilizing hand-drawn sketches as retrieval inputs offers intuitive and user-friendly advantages, yet the potential of multi-level feature integration from sketches remains underexplored, leading to suboptimal retrieval performance. To address this gap, our study introduces a novel zero-shot, sketch-based retrieval method for remote-sensing images, leveraging multi-level feature extraction, self-attention-guided tokenization and filtering, and cross-modality attention update. This approach employs only vision information and does not require semantic knowledge concerning the sketch and image. It starts by employing multi-level self-attention guided feature extraction to tokenize the query sketches, as well as self-attention feature extraction to tokenize the candidate images. It then employs cross-attention mechanisms to establish token correspondence between these two modalities, facilitating the computation of sketch-to-image similarity. Our method significantly outperforms existing sketch-based remote-sensing image retrieval techniques, as evidenced by tests on multiple datasets. Notably, it also exhibits robust zero-shot learning capabilities in handling unseen categories and strong domain adaptation capabilities in handling unseen novel remote-sensing data. The method’s scalability can be further enhanced by the pre-calculation of retrieval tokens for all candidate images in a database. This research underscores the significant potential of multi-level, attention-guided tokenization in cross-modal remote-sensing image retrieval. For broader accessibility and research facilitation, we have made the code and dataset used in this study publicly available online.

1. Introduction

The proliferation of remote-sensing sensors deployed on various carrier platforms has led to a continuous and rapid increase in the volume of observation data pertaining to the Earth’s surface. The data generated by these sensors possess the characteristics commonly associated with big data: it is voluminous, exhibits a wide variety, is generated at high velocity, and requires rigorous verification. While this abundance of data offers users unprecedented opportunities to discover and quantify underlying phenomena, it also presents many challenges [1,2,3]. The challenges in remote-sensing big data include data management, analysis, retrieval, and interpretation complexities. Among these challenges, data retrieval—accurately and efficiently finding the desired category images from a massive amount of remote-sensing images—is crucial for subsequent data mining processes [4]. Traditional query inputs typically involve image metadata values, such as bounding box coordinates, sensor names, and timestamps. Another commonly employed approach, which is content-based remote-sensing image retrieval (CBRSIR), involves querying remote-sensing warehouses using example images, often yielding promising results when coupled with state-of-the-art deep learning algorithms [5,6]. However, in numerous application scenarios, users encounter difficulty in providing desirable remote-sensing examples. Consequently, some cross-model retrieval methods, such as text–image retrieval [7] and sketch-based remote-sensing image retrieval (SBRSIR) [8], have captured the attention of some researchers. As demonstrated in Figure 1, SBRSIR offers users the ability to express the structure of a desired remote-sensing image from their mind through freehand sketches, which can then be employed as queries for retrieving images. This method is believed to be intuitive for users, easy to execute on touch-enabled devices, and capable of achieving a high level of expressiveness and flexibility [9,10,11,12]. From a pragmatic perspective, our research team has been working on a project known as “Habitat Yangtze”, which is one part of the broader Space Climate Observatory initiative (visit https://www.spaceclimateobservatory.org/habitat-yangtze for more information (accessed on 10 March 2024)). The objective of this project is to offer sophisticated remote sensing and mapping services to varied users, including wetland managers, bird watchers, and climate change researchers, many of whom possess limited expertise in remote-sensing technologies. These stakeholders have demonstrated a significant interest in a sketch-based remote-sensing image retrieval system, highlighting the limitations of traditional query inputs in representing their visions and imaginations. This expressed need has significantly inspired and directed the focus of our research.
Sketch-based image retrieval (SBIR) for common images has undergone extensive investigation in recent years and yielded promising results [13,14,15,16,17]. To bridge the two modalities, a common approach contains three stages: feature extraction from both modalities, feature enhancement, and image retrieval. The latest solutions, like ACNet [18], DAL [19], and ZSE-SBIR [13], often employ deep structures like ViT (vision transformer) and Resnet and use homogeneous [20], Siamese branch [9], or heterogeneous structures [21] for different modalities. Despite the inspiring and promising development of SBIR, only a small number of recent research publications have been dedicated to remote-sensing SBIR [8,22,23,24,25,26]. Current SBRSIR models underscore the imperative for novel strategies that can surpass the existing limitations, especially in terms of retrieval accuracy, zero-shot learning capability, and domain adaptation capability. Compared to ordinary images, remote-sensing images are often obtained from aerial viewpoints and cover a much larger geographical area. This leads to noticeable semantic and structural disparities with ordinary images. This constitutes the principal difference between SBIR and SBRSIR research and makes the direct application of SBIR questionable. Also, the benchmark training and testing datasets for common pictures are no longer pertinent for remote sensing. Consequently, researchers in the field of remote-sensing applications often find themselves compelled to design and train ad-hoc models. There are two primary challenges in the domain of SBRSIR. First, there is an extensive array of categories concerning objects and scenes within remote-sensing images. The sketch samples are largely scarce compared with ordinary photo-related sketches [8]. As a result, constructing a comprehensive training dataset for remote-sensing SBIR proves to be a formidable task. In contrast to the general image processing domain, well-annotated datasets for remote-sensing images and sketches remain insufficient for supervised learning [23], thus potentially affecting the performance of trained networks, particularly when confronted with unseen categories and images. Given the challenges associated with significantly expanding training datasets to encompass all potential categories and data sources in remote sensing, existing SBRSIR research underscores the importance of zero-shot learning performance [23,24]. Nevertheless, the existing models in this domain still exhibit very limited zero-shot capabilities. The second challenge is that remote-sensing images exhibit distinct characteristics depending on the sensor and their carrier platform. The SBRSIR model should have a good domain adaptation capability to support different sources of remote-sensing images. However, current work investigated very little in the domain adaptation capability of the SBRSIR model.
In response to the growing demand and challenges of SBRSIR, we propose a novel zero-shot cross-model deep learning network that leverages a new multi-level feature extraction and attention-guided tokenization mechanism. Additionally, we substantially expanded the RSketch SBRSIR benchmark dataset to serve as a comprehensive testbed, allowing for a comprehensive evaluation of the performance of our new method in various settings. Our test results reveal that our new method significantly outperforms existing SBRSIR algorithms. Our proposed algorithm exhibits excellent zero-shot capabilities, enabling precise retrieval of images from unseen categories, and demonstrates strong domain adaptation capabilities, as it can be trained on mixed-source remote-sensing image samples and retrieve images from previously unseen data sources, thereby offering extensive application flexibility.
The main contributions and innovations of our proposed method can be summarized in two key aspects:
  • We introduced a novel deep-learning categorical SBRSIR network equipped with a multi-level feature extraction, self-attention-guided tokenization and filtering, and cross-modality attention update across sketch and remote-sensing modalities. This approach simplifies the comparison process by focusing on a select number of significant patches from remote-sensing images relative to the query sketch. Moreover, our model eliminates the requirement for semantic knowledge input, significantly lowering the costs associated with constructing training datasets. This new approach has demonstrated substantial performance improvements compared to existing methods, particularly in terms of its good zero-shot capability. Furthermore, our proposed model showcases impressive domain adaptation capabilities as it can be trained using mixed-source remote-sensing datasets and can retrieve remote-sensing images from unseen sources.
  • We substantially expanded the SBRSIR benchmark RSketch dataset to the RSketch_Ext dataset as a testbed. This new testbed encompasses 20 categories, each containing 90 sketches and over 400 remote-sensing images from various datasets. We made this comprehensive dataset, as well as the code of our method, available online to facilitate the work of researchers in this field.
The rest of this article is organized as follows. Section 2 describes the related work of SBRSIR. Section 3 gives the details of our proposed method. Some experimental results are reported in Section 4. The discussion is given in Section 5, and a conclusion is drawn in Section 6.

2. Related Works

This section explores prior research in the field of SBIR, SBRSIR, and other application domains within image retrieval. The utilization of freehand-drawn sketches as input queries has attracted researchers in computer vision [27,28], human–computer interaction [29], and geographical information science [30] since its early stages. The inherent challenge in sketch-based retrieval lies in the multi-modality nature of such a task. Sketches significantly differ from photographic or remote-sensing images due to their abstract, symbolic, sparse, and stylistic nature, often containing rich topological and semantic information.
SBIR can be categorized into two levels in terms of granularity: categorical [31] and fine-grained [32]. Categorical SBIR focuses on recognizing the categorical information in sketches and retrieving images from the same category, often described by semantic notions like “plan”. In the remote-sensing domain, the development is currently at a categorical level [8,23]. On the other hand, fine-grained or instance-level SBIR delves into scene and stroke details, such as relative location and topology, for more precise image retrieval. Both categorical and fine-grained SBIR may frequently encounter images from unseen categories during testing. To address this, zero-shot learning (ZSL) algorithms are introduced, often employing assisting information such as word embeddings for semantic similarity measurement [23,33,34]. However, directly incorporating semantic information can present significant challenges in constructing training datasets and may also restrict the model’s generalizability. Recent works aim to enhance the algorithm’s generalizability, adopting features and local correspondence information for fine-grained yet generalizable algorithms [13,35]. The improvement of algorithms’ generalizability may also be helpful in cross-dataset retrieval [13]. Despite advancements in the computer vision domain, such algorithms have not been explored in the context of sketch-based retrieval in the remote-sensing domain.

2.1. Cross-Model Feature Extraction

Most contemporary SBIR and SBRSIR algorithms generally follow a three-part pipeline: cross-model feature extraction, feature enhancement, and image retrieval. The feature extraction process involves extracting informative and representative features from both sketches and images. Initially, hand-crafted features like gradient field HOG [27], edge maps [36], and histograms of edge local orientations [37] were applied. However, recent years have seen the widespread adoption of deep learning [9,38] in SBIR and SBRSIR [25,26], outperforming hand-crafted methods [39]. Various deep network structures, including FCN [40], CNN [41], RNN [28], VAE [42], GNN [43], transformers [44], and ViT [13], have been explored. The latest solutions often combine multiple deep structures [13,41] and employ homogeneous [20], Siamese branch [9], or heterogeneous structures [21] for different modalities.

2.2. Feature Enhancement

The feature enhancement part involves feature selection, aggregation, and embedding [16]. This step prepares the extracted feature for comparison and retrieval. Recent content-based image retrieval research [45,46,47] utilizes attention maps to weigh feature importance, with [48] further proposing feature elimination or merging for efficiency without significant accuracy loss—a novel approach in SBIR. For feature embedding, common methods include bag of features [49], bag of visual words [27], VLAD [50], and FV [51]. To enhance scalability, deep features could be pre-calculated and embedded in binary hash codes [52,53].

2.3. Image Retrieval

After encoding features in the embedding space, image retrieval typically involves a nearest neighbor search based on global similarity calculated through Euclidean distance [25] or Hamming distance for binary hash codes [54]. Deep metric learning is also prevalent in similarity calculation in SBIR [55,56]. Instead of global similarity calculation, some works focus on local matching pairs between sketches and images for similarity scoring, such as [57] using the number of matching pairs and [13] employing a cosine similarity matrix. Approaches like Deep Hashing [52], approximate nearest neighbor (ANN) using the k-means tree [58], and ANN using product-quantization [52] demonstrate scalability for efficient retrieval from large image databases.

2.4. Training and Data

Supervised and unsupervised deep learning approaches are both employed in SBIR [52,59,60,61], with most SBRSIR research using supervised training. In the supervised training process, adequately annotated sketch–remote-sensing datasets are crucial. However, only two dedicated SBRSIR benchmark datasets have been identified: RSketch [8] and Earth on Canvas [23]. They are relatively small compared to standard SBIR datasets like QuickDraw extend [28] and Sketchy [54]. Also, these two datasets only cover a small fraction of all potential sensors. The scarcity of large training data emphasizes the significance of zero-shot, semantic knowledge independent, and domain adaptation capabilities in SBRSIR—a key improvement in our proposed method.
In summary, compared to SBIR, research on SBRSIR remains limited and is mostly focused on category-based retrieval. The lack of comprehensive training and testing benchmark datasets means that zero-shot learning and domain adaptation capabilities are critical, yet the capabilities of existing models are not ideal. Research into SBRSIR urgently requires the development of new algorithms with better zero-shot learning and domain adaptation capabilities.

3. Methodology

Table 1 displays the symbols used in this chapter along with their corresponding.
Informed by the fundamental principles delineated in prior studies [13,62,63], the proposed model in this research consists of two stages: a self-attention for feature extraction and a cross-attention for similarity calculation. The structural design of this model is visually represented in Figure 2, which describes the model’s capacity for facilitating the retrieval process across two distinct modalities: remote-sensing imagery and sketches. Detailed explications of each stage’s specific functions and their role within the model are systematically presented in the following sections.

3.1. Self-Attention Feature Extraction

In the context of sketch-based remote-sensing image retrieval, the process typically begins with a query sketch S. The objective is to identify the corresponding remote-sensing image R from the remote-sensing image database D R . Traditional deep network methodologies transform hand-drawn sketches and remote-sensing images into a sequence of visual tokens. These methods predominantly capture local features of sketches, which are often characterized by sparse strokes, as noted in [13]. To address this limitation and effectively expand the receptive field of visual tokens to accommodate sparse strokes better, this study implements a multi-level feature extraction module applicable to sketches. This enhancement aims to preserve a more comprehensive set of feature information from sketches. The module is constructed by layering multiple convolutional layers with distinct kernel sizes and includes learnable parameters. Following each convolutional layer is a non-linear activation function. This configuration results in an n × d dimensional visual token embedding E = [ E 1 , E 2 ,   ,   E n ] , where each E represents a d-dimensional vector.
This module incorporates a transformer-based self-attention mechanism, an approach inspired by the human visual and cognitive systems. This mechanism enables neural networks to dynamically concentrate on the most informative segments within the input data. By integrating the self-attention mechanism, the neural networks in our model are designed to autonomously identify and focus on salient features in the input, whether it be a sketch or a remote-sensing image. This strategy can significantly enhance the network’s performance and its ability to generalize.
The initial step in our methodology involves processing the output from the convolutional layers. This output is first concatenated with a retrieval token [RT], a trainable d-dimensional embedding vector that embodies the global feature. This combination results in an augmented n + 1 dimensional visual token embedding E = [ R T , E 1 , E 2 , , E n ] . Then, the self-attention is realized by passing E through first the Multi-Head Self Attention (MSA) module and second the Multi-Layer Perceptron (MLP) module. The forward propagation formula of the model is as follows:
E 0 = E
E l = M S A L N E l 1 + E l 1 ,       l = 1 L
E l = M L P L N E l + E l ,       l = 1 L
Formulas (2) and (3) both incorporate residual connections, where L represents the number of hidden attention layers, and L N stands for layer normalization.
The MSA module is a core component of our deep network, designed to discern interrelations among various token vectors within a remote-sensing image or a sketch. The essential element of this module is the scaled dot-product attention mechanism. Initially, the layer-normalized visual token embedding is multiplied by three learnable matrices, W q , W k and W v , resulting in Q (queries), K (keys), and V (values) per the following equations:
Q = E · W q           K = E · W k           V = E · W v
Then, the scaled dot-product attention is calculated by the following equation:
A t t e n t i o n s e l f ( Q , K , V ) = s o f t m a x ( Q K T d ) V
In this equation, the product of Q and K assesses the similarity between the query and key. The result is then scaled by the square root of the dimension d of E i to mitigate the vanishing gradient problem. A softmax function normalizes the similarities across multiple keys relative to a query, ensuring their cumulative sum equals 1. The resulting similarity is used as a weight to compute the weighted average of the corresponding V, ultimately obtaining an attention head of ( n + 1 ) × d -dimension. This procedure is replicated h times to create h attention heads. These heads are subsequently integrated into ( n + 1 ) × d -dimensional MSA output by a dense network. The output of MSA is then followed by an MLP module, as described in [62].
The proposed method integrates feature filtering at specific layers of the self-attention stage. Given the varying degrees of information richness among local visual tokens generated through self-attention, selectively filtering out tokens with lesser feature information can reduce the number of tokens and improve efficiency for both training and referencing. The filtering is achieved by leveraging attention scores between [RT] and all other visual token embedding vectors. Specifically, the typical query of visual token embedding is replaced with the query of [RT]. The formula is as follows:
A t t e n t i o n f i l t e r i n g ( Q , K , V ) = s o f a m a x ( Q [ R T ] K T d )
Utilizing this equation enables the computation of attention scores between [RT] and all visual tokens. Based on the attention scores, only k visual token vectors are retained for further processing. Consequently, this leads to a more refined set of visual token embeddings for both the sketch image and remote-sensing images, denoted as E = R T f i n a l , E f i n a l 1 , , E f i n a l k   ( w h e r e   k < n ) .

3.2. Cross-Attention and Similarity Calculation

Our method employs cross-attention to establish cross-modal token embedding correspondences between sketches and remote-sensing images. This involves an interchange of the sketch query Q S and the candidate remote-sensing image query Q R . After the swap, the query, key, and value for the sketch and remote-sensing image become Q R , K S , V S and Q S , K R , V R , respectively. This interchange facilitates a direct connection between the visual token embedding sets of the sketches and the remote-sensing images. Taking Q S as an example, the cross-modal attention is obtained using the following formula:
A t t e n t i o n c r o s s Q S , K R , V R = s o f t m a x ( Q S K R T d ) V R
Through this attention mechanism, the visual token embeddings of both the sketch and the remote-sensing image, including the retrieval token [RT], are updated based on the pair-wise token information from each modality.
The final step in our methodology involves the use of Euclidean distance between [ R T ] S and [ R T ] R as a metric for measuring the similarity between the sketch input and a candidate remote-sensing image.

3.3. Model Training and Image Retrieval

In this study, the deep network is trained using the triplet loss function with [RT] derived from both sketches and remote-sensing images. Specifically, we consider a triplet ( S i ,   R i + ,   R i ) in the training set, where S i denotes a query sketch, R i + denotes a remote-sensing image with the same label as S i , and R i denotes a remote-sensing image with a different label. The primary aim of this loss function is to minimize the distance between correctly matched sketch–remote-sensing image pairs (positive examples) while ensuring that the distance between each sketch and incorrectly matched remote-sensing images (negative examples) exceeds a predefined margin. Here, [RT] is used as the global descriptor for sketches and remote-sensing images. The triplet loss is defined as the following equation:
L t r i = 1 T i = 1 T m a x   R T S i ,   R T R i + 2 R T S i ,   R T R i 2 + m , 0
In this equation, T represents the total number of triplets, and m denotes the margin that discriminates whether a sketch and a remote-sensing image are from the same class. If a sketch and a candidate remote-sensing image’s difference exceeds the margin value, they are not in the same class.
In the retrieval phase, the query sketch is processed through the network to obtain its retrieval token [ R T ] S . Similarly, all candidate remote-sensing images in the database are transformed into [ R T ] R . The similarity between [RT] of the sketch input and a candidate remote-sensing image is measured by calculating their Euclidean distance. A smaller Euclidean distance indicates a closer similarity. Subsequently, the remote-sensing images that exhibit the smallest distances from the query sketch—essentially, its k-nearest neighbors—are selected as the final output of the model’s retrieval process.
To further improve the response speed during actual retrieval operations, [ R T ] R is pre-computed and stored in the database by invoking the model in advance. During retrieval, the model is only required to compute the retrieval token for the input sketch. Following this, the pre-stored [ R T ] R from the remote-sensing image database are rapidly accessed, and the Euclidean distance between [ R T ] S and all [ R T ] R in the database can be directly calculated. This approach significantly accelerates the retrieval process, especially when dealing with big remote-sensing image collections. Moreover, the efficiency of the retrieval process can be further augmented by employing techniques such as approximate nearest neighbor algorithms or utilizing a vector database. This improvement is particularly useful as the volume of images in the remote-sensing database expands, underscoring the practical scalability and efficiency of the model in real-world retrieval scenarios. These mechanisms are crucial for optimizing the model’s performance, particularly in retrieval applications where speed and accuracy are paramount.

4. Experiments

4.1. Dataset

The development of effective cross-modal retrieval models, particularly for SBRSIR, is hindered by the scarcity of specialized datasets. While there are numerous cross-modal retrieval datasets for natural images, such as Sketchy, TUBerkin, and QuickDraw, the availability of similar datasets for remote-sensing images is very limited. This gap presents a significant challenge for training SBRSIR models.
This study used the RSketch dataset [8], the RSketch_Ext dataset (expanded based on the RSketch dataset), the Earth on Canvas dataset [23], the UCMerge Landuse dataset [64], and GF-1 image tiles of the Anhui Province section in the middle and lower basin of the Yangtze River, China. The RSketch dataset, a publicly available sketch–remote-sensing image dataset, comprises 20 categories, including airplanes, baseball fields, and bridges, each with 45 sketches and 200 remote-sensing images. In this study, we expanded the RSketch dataset to RSketch_Ext by augmenting each category with additional remote-sensing images sourced from various public datasets like AID [65], NWPU-RESISC45 [66], WHU_RS19 [67], and others. Furthermore, as part of our HABITAT YANGTZE project, we enlisted the help of 10 non-professional (in remote sensing) volunteers, none of whom possess a background in drawing, to engage in a sketching task. Each volunteer was instructed to create sketches in five categories based on their personal interpretation of the shapes associated with those categories. We then collected the sketches produced by the volunteers and obtained a specific number of sketches from each completed category to enrich the diversity of our augmented dataset’s sketch categories. Concurrently, we enhanced our sketch dataset further by producing simulate sketch from OpenStreetMap (OSM) data. We transformed OSM data into sketch-like images, and integrated these into appropriate sketch categories of our dataset. Post-expansion, each category includes 90 sketches and at least 400 remote-sensing images, significantly increasing the dataset’s size. The sketches are in both TIFF and JPEG formats. Figure 3 presents sample data from each category in the RSketch_Ext dataset. The Earth on Canvas dataset contains 14 categories, five of which are unique compared to those in the RSketch and RSketch_Ext dataset, with each category comprising 100 sketches and 100 remote-sensing images. The UCMerge Landuse dataset is a publicly available dataset for remote-sensing image classification. It features a rich set of image categories in comparison to other remote-sensing image datasets, comprising a total of 21 categories. Among these, 10 categories align with those in the RSketch dataset and RSketch_Ext dataset, each containing 100 remote-sensing images per category. Table 2 presents the classes in RSketch, RSketch_Ext, Earth on Canvas, and UCMerge Landuse dataset. Figure 3 illustrates the sample data corresponding to each category in this dataset. GF-1 images tiles specifically depicted a section of the Yangtze River in Anhui Province, China. There were 4842 image tiles, and each image tile was formatted to a resolution of 256 × 256 pixels.
In this study, we implemented a 4-fold cross-validation approach to evaluate our model. The 20 categories from both the RSketch and RSketch_Ext datasets were divided into four distinct sets of classes. These sets were designated as S1–S4. In each fold, 15 categories were assigned as seen classes and participated in the training process, while the remaining 5 categories were reserved as unseen classes, used exclusively for testing. The four sets of unseen classes are detailed in Table 3. Within each seen class, we adopted a split in each set whereby 50% of the remote-sensing images were utilized for training, and the remaining 50% were reserved for testing purposes. This partitioning provides a balanced approach for model training and evaluating, ensuring a comprehensive assessment across a diverse range of categories.

4.2. Implementation Details

In this research, the proposed method was implemented using the PyTorch framework on a single NVIDIA GeForce RTX 3080Ti GPU. Regarding the preprocessing of sketches and remote-sensing images for model training, several steps of preprocessing were undertaken. All images were cropped according to the bounding box of strokes and then uniformly scaled to a resolution of 224 × 224 pixels, conforming to the requirements of the pre-trained model used in this study. This preprocessing step was crucial to eliminate redundant information and allowed the self-attention stage to focus more effectively on pertinent feature information.
The kernel size of convolutional layers used in the multi-level feature extraction stage was set at 7 × 7 for the initial convolutional layer, while the subsequent three layers were set at 3 × 3. The stride parameter for all these convolutional layers was uniformly maintained at 2. This configuration resulted in the generation of 196 visual tokens, with each token represented as a 768-dimensional vector. In terms of architecture, the self-attention stage blocks were designed akin to those in the vision transformer (ViT), comprising 12 blocks. These blocks were pre-trained on the ImageNet-1K dataset [68]. The cross-modal attention stage had a single layer with 12 heads. During the training process, the AdamW [69] optimizer was employed, with the learning rate set to 2 × 10−5 to balance efficient training convergence with the need for network stability.

4.3. Evaluation Criteria

The evaluation criteria used in this study are as follows:
  • Mean average precision (mAP): This metric is a measure of retrieval accuracy, reflecting the mean area underneath the precision–recall curve in multiple queries.
  • Top-K Accuracy: This metric measures the proportion of correctly retrieved images within the top K results returned by the model. In this research, we specifically evaluate the model’s performance at three levels: the top 10, top 50, and top 100 retrieved images. These varying levels provide a comprehensive assessment of the model’s retrieval accuracy and its ability to rank relevant images effectively.
Both mAP and Top-K Accuracy are critical in evaluating the model’s performance in retrieving relevant images from the dataset, offering a multi-dimensional understanding of its effectiveness in various test scenarios.

4.4. Experiments on RSketch Dataset

To verify the effectiveness of the proposed method in this study, we compared the proposed model with the following baseline deep learning methods for cross-modal retrieval: MR-SBIR [8], DAL [19], DSM [70], DOODLE [71], DSCMR [72], CMCL [73], and ACNet [18].
Among the benchmark methods considered, only the initial baselines are specifically tailored for sketch-based remote-sensing image retrieval (SBRSIR). Due to the limited availability of methods dedicated to SBRSIR, we have included additional methodologies from the broader field of computer vision, which are typically applied to conventional images. Comparing our specialized method against these more generalized SBIR approaches may provide insights into the feasibility of adapting algorithms designed for ordinary images to the remote-sensing context.
In this study, both the proposed method and its counterparts were initially trained and tested using the RSketch dataset as a benchmark. The outcomes for each network were obtained by averaging the results across the four folds, as detailed in Table 4.
The data in Table 4 reveal that our proposed method attained a mean average precision (mAP) of 98.17% and a Top-100 accuracy of 95.93% for seen classes. Notably, the performance gap between our model and the baseline methods was more pronounced in unseen classes, with the proposed method achieving a mAP of 76.62% and a Top-100 accuracy of 70.01%. These results suggest that the method presented in this study significantly surpassed other baseline methods in terms of remote-sensing image retrieval accuracy. Also, the specific remote-sensing methods performed better than DOODLE, DSCMR, and ACNet, which are designed for ordinary images, highlighting the domain gap between the retrieval of ordinary images and remote-sensing images. The proposed method also achieved equivalent or slightly better mAP compared with the latest CBIR algorithms [5]. We detected no difference in the result between TIFF- and JPEG-encoded sketches.
We conducted a systematic investigation to determine how various hyperparameters influence the performance of the model. One of the key parameters adjusted was the number of heads in the multi-head attention mechanism. We observed that increasing the head count from 8 to 12 led to a notable improvement of 1.8% in the model’s mean average precision, and further increasing it from 12 to 16 led to a small additional improvement of 0.9% in mAP. Conversely, a reduction in the number of heads from 8 to 4 resulted in a mAP decrease of 2.2%. This finding suggests a direct correlation between the number of attention heads in the attention mechanism and the accuracy of retrieval. Appropriately increasing the number of attention heads can further improve the accuracy of retrieval. However, it should also be noted that too many attention heads will increase the complexity and computational cost of the model and may lead to overfitting of the model, especially the GPU memory consumption during training; it may also lead to overfitting of the model. In our test, the training time was prolonged when the number of heads increased. The training time for networks with 12 attention heads was about 2.8% longer than those with 4 attention heads, and 16 attention heads required an additional 0.9% longer time. Therefore, the appropriate number of attention heads should be determined by considering model performance, complexity, computational efficiency, and overfitting. Moreover, modifications to the learning rate of the AdamW optimizer also demonstrated significant impacts on the model’s retrieval accuracy. An increase in the learning rate from 1×10−5 to 2×10−5 yielded a 0.6% increase in the mAP value. In contrast, reducing the learning rate from 1×10−5 to 5×10−6 led to a substantial decline in mAP, a decrease of 5.9%. These insights highlight the sensitivity of the model to specific hyperparameter settings, underscoring the importance of careful tuning to optimize retrieval performance.

4.5. Experiments on RSketch_Ext Dataset

Evaluating our approach using the RSketch dataset facilitates direct benchmarking against existing methodologies. Nonetheless, the RSketch dataset presents considerable limitations in terms of category breadth, dataset size, and diversity. To more accurately assess the effectiveness and efficiency of our proposed method, we expanded our experimental framework to include the RSketch_Ext dataset, allowing for a more comprehensive evaluation. Compared to RSketch, this dataset maintains the same number of categories but with a substantially increased quantity of images and sketches in each. Notably, these additional images and sketches are sourced from a variety of different datasets and creators, providing a broader range of data inputs in both training and testing.
While keeping other settings unchanged, we trained the proposed model and three baseline models on the expanded RSketch_Ext dataset. Additionally, we also utilized the model trained on the RSketch dataset mentioned in this paper (marked Ours-RSketch). A total of five models were tested on the expanded RSketch_Ext dataset, and the results of the tests are shown in Table 5. This comparison aims to shed light on the relationship between data characteristics—particularly in terms of volume and source diversity—and the model’s efficacy.
From Table 5, it can be observed that the retrieval performance of the proposed model outperforms other baseline methods. Also, the model trained and tested using the extended dataset showed a little change in retrieval accuracy for seen classes compared to the model trained on the original RSketch dataset. This is mainly because the model trained on the original dataset already achieved high accuracy for seen classes, so there is not much room for improvement in retrieval accuracy. However, in the retrieval of unseen classes, the model trained on the expanded dataset exhibits a decrease of 6% in mAP value and a decrease of approximately 3% in Top-10 accuracy compared to the model trained on the original dataset. This result may be attributed to the expansion of the training datasets, with an increase in the diversity of remote-sensing sources per category, and the difficulty of the model to catch feature patterns also increases, leading to a decrease in mAP value for unseen classes. Meanwhile, the values for Top-50 showed minimal changes, while there was an increase of around 7% in Top-100 accuracy. The discrepancy between the performance change in mAP, Top10, and Top100, stating that better mAP and Top10 come with worse Top100, is possibly a common phenomenon in such a test, as the results of DOODLE and ACNet demonstrate a similar pattern. A possible explanation is that the more sensible model may miss some potential candidates within the borderline, resulting in a worse Top100. Such a result may indicate an inevitable tradeoff between the model’s capability for retrieving the most viable candidate in the first few attempts and the capability for catching more candidates in a larger candidate pool.
In addition to quantitative analysis, we conducted a manual examination of the retrieval results. Figure 4 and Figure 5 display selected results of our model for both seen and unseen classes within the RSketch_Ext dataset, where the proposed method achieves high accuracy in retrieving seen classes. Notably, in certain unseen categories, such as basketball courts and runways, the model also demonstrates high correct retrieval rates. Although there were two instances of incorrect retrievals in the crosswalk category, a closer inspection revealed that the incorrectly retrieved images did indeed feature a crosswalk, albeit categorized under “intersection” in the test dataset. While the retrieval results for the oil and gas field category were not accurate, there was a noticeable similarity in texture and structure between the sketches and the retrieved remote-sensing images.

4.6. Experiments on Earth on Canvas and UCMerge Landuse Datasets

To assess the proposed model’s zero-shot learning and domain adaptation capabilities, we designed an experiment in which our proposed model and three baselines, initially trained on the RSketch_Ext dataset, underwent testing on the distinct datasets of Earth on Canvas and UCMerge Landuse. For the UCMerge Landuse dataset, the experiment only incorporated data from the 10 categories that overlap with those in the RSketch_Ext dataset. The experiments were also categorized into seen and unseen classes based on whether the classes being tested were included in the model’s training dataset. The test results are detailed in Table 6 and Table 7. This approach of testing across differing datasets provides valuable insights into the model’s ability to generalize beyond its training dataset and its zero-shot capability in such situations, offering a stringent test of its practical applicability in real-world scenarios.
The results from Table 6 and Table 7 indicate that our model, even when trained on the RSketch_Ext dataset, still demonstrates very good and superior performance compared to other baseline methods when tested on the Earth on Canvas dataset and the UCMerge Landuse dataset. These results demonstrated the robust domain adaptation ability of the proposed model, particularly in its capacity to effectively retrieve relevant data from a dataset comprising previously unseen data and classes. This underscored the model’s adaptability and potential for practical application in diverse and novel retrieval scenarios.

4.7. Experiments on GF-1 Tiles

In this part of the study, we further evaluated the practical application of the model trained on the RSketch_Ext dataset by conducting retrieval experiments using real-world remote-sensing image tiles within the Habitat Yangtze project. As illustrated in Figure 6, the remote-sensing image tiles utilized for these experiments were from the GF-1 satellite, and there are a total of 4842 tiles, each with a size of 256 × 256 pixels. For the retrieval input, sketches from the RSketch_Ext dataset were employed. It is important to note that the GF-1 satellite imagery is not part of the training data. Consequently, this set of experiments is also a valuable test of the model’s capabilities in handling completely unseen data. The use of real satellite imagery in these tests provided a stringent assessment of the model’s retrieval accuracy and efficiency in realistic scenarios with remote-sensing image tiles of a large research region beyond the confines of the training dataset.
Figure 6 showcases the retrieval capabilities of the proposed model when applied to GF-1 remote-sensing image tiles, particularly focusing on river and bridge classes. By manually examining the top 10 retrieved images for river and bridge categories, we observed an accuracy of 68% for the top 10 images in the river category and 70% for the top 10 images in the bridge category. This result demonstrated the effectiveness of our proposed model in a practical setting, and the subjective feedback from the user was satisfied. Despite the model’s proficiency in retrieving relevant images, it is noteworthy that some retrieved images, while not precisely matching the query’s class, bear a resemblance in shape and texture to the input sketches. This is particularly evident in the retrieval attempts for the beach and tennis court classes, which are not presented in the selected research area. Although the images returned for these categories do not strictly belong to the specified classes, they share a similarity in shape with the sketches used for retrieval. We conducted retrieval tests on the same 100 query sketches, respectively, in the image tile database of GF-1 and the pre-calculated GF-1 retrieval tokens. With the same retrieval accuracy, the retrieval time in the GF-1 image tile database was 66.3 s, while the retrieval time in the pre-calculated retrieval tokens was only 11.7 s. From the comparison of the retrieval times of the two, it can be clearly seen that the retrieval time using pre-calculated retrieval tokens is much shorter than the time for retrieval using the original image tiles. This confirms that using pre-calculated retrieval tokens can effectively improve the efficiency of retrieval in practical applications of large-scale remote-sensing image retrieval, and retrieval efficiency is one of the key factors in retrieving large-scale remote-sensing image datasets.
This observation underlines a significant aspect of the proposed model’s functionality: its ability to discern and match shapes between sketches and remote-sensing images, exceeding mere categorical correspondence. This capability is particularly useful in instances where the category depicted in the queried sketch is absent from the dataset. In such cases, the model suggests alternative images that, while not categorically identical, are visually similar to the sketch in terms of shape. This outcome indicates the model’s potential for broader applications, where shape recognition plays a crucial role in retrieval processes.

5. Discussion

The method introduced in this study, incorporating multi-level feature extraction and attention-guided tokenization, offers a novel deep learning approach for sketch-based remote-sensing image retrieval. Our findings demonstrated the effectiveness of this method, which outperformed seven baseline methods, particularly under zero-shot conditions. Additionally, the method exhibited good domain adaptation capabilities. It could effectively handle remote-sensing images from sources not included in the training dataset, thus underscoring its robustness and flexibility. A notable feature of the proposed method is the pre-computation of retrieval tokens for each remote-sensing image in the database, enabling accelerated retrieval processes through vector search algorithms.
Despite ablation studies and the promising results from experiments across five datasets, further exploration of the network’s full potential remains pertinent. For instance, the current approach utilizes multi-level tokenization for sketches and applies identical token filtering mechanisms for both sketches and remote-sensing images, which could be optimized separately for each modality. Moreover, leveraging the latest pre-trained networks could enhance the self-attention model. Our current training process begins with a network pre-trained on the ImageNet-1K dataset. Exploring other cutting-edge pre-trained networks may yield additional improvements.
For the experiment dataset, the study’s expansion of the RSketch dataset to RSketch_Ext focuses primarily on increasing the volume and diversity of sketches and remote-sensing images per class rather than expanding the range of scenario classes. Given that the current 20 classes are insufficient to encompass all potential scenarios and their semantic categories, a significant expansion of the training dataset’s class diversity is necessary but poses challenges in terms of time and resources. An automated approach for extending the benchmark dataset would be invaluable for future research in this domain. Both the RSketch and RSketch_Ext datasets are optical remote-sensing data. We acknowledge that there are other remote-sensing data types, such as SAR data and hyperspectral data. These data have significantly different features compared to optical imagery and require extensive investigation. However, due to their complexity, exploring these datasets exceeds the scope of our current work. Nonetheless, they represent promising avenues for future research. We recognize the presence of several successful applications of Graph Convolutional Networks (GCNs) within the sketch-based image retrieval (SBIR) domain that leverage semantic information. Our proposed method has already demonstrated exceptional performance in terms of effectiveness and efficiency, even without the integration of GCNs. Being semantic knowledge independent is an advantage of our proposed method, as it can greatly reduce the burden of training data construction. However, exploring and incorporating a GCN-based design presents a promising avenue for future research, particularly when extensive training datasets encompassing a broad spectrum of scene semantics become available.
The sketches used in this study were limited to uniform line strokes, yet sketches can vary widely in form, including variations in line width, texture depiction, and annotations. Investigating how to adapt our model to accommodate these diverse sketch forms presents an intriguing avenue for future research. At the same time, the quality or style of sketches can also affect the effectiveness of retrieval. Generally, when using sketches to retrieve remote-sensing images, sketches that contain enough information and, at the same time, with simple outlines often achieve better results. Too many internal details of sketches may distract the model’s attention and may impair the effectiveness of categorical remote-sensing retrieval. Furthermore, the incorporation of sketch annotations could introduce a natural language modality, transitioning from the current two-modality to a three-modality SBRSIR framework. Building upon this, we can enhance retrieval by incorporating both sketches and semantically expressed information. This includes categories of sketches, potential colors, and named predicates indicating spatial relationships among sketches. Our tests with GF-1 image tiles also suggest the potential of our model for fine-grained retrieval, making further investigation and enhancement of the model for fine-grained remote-sensing image retrieval plausible.
For practical application in real-world settings, this method can be deployed by constructing a remote-sensing image retrieval website with our trained model. Users can upload or draw sketches using a mouse on the website interface, which are then sent to the backend of the website. The deployed model is called in the backend for retrieval from the pre-calculated remote-sensing image database, and the retrieval results are then returned to the user interface. Additionally, this retrieval method can be combined with traditional retrieval methods (such as by timestamp or image region) for joint retrieval.

6. Conclusions

In this paper, we introduced a novel categorical zero-shot, sketch-based, remote-sensing image retrieval method, leveraging multi-level feature extraction, self-attention-guided tokenization and filtering, and cross-modality attention update. The efficacy of this method has been thoroughly evaluated through experiments conducted on five remote-sensing datasets. The results clearly indicate that our method not only surpasses other baseline methods but also exhibits strong zero-shot learning and domain adaptation capabilities. Particularly noteworthy is the network’s ability to retrieve remote-sensing images from both unseen categories and unseen data sources. Also, our model is semantic knowledge independent, thus greatly reducing the complexity of constructing training datasets compared with other methods that leverage semantic knowledge. Our model is especially relevant for large-scale remote-sensing datasets, as it enables the pre-calculation of retrieval tokens for all images in a database, enhancing scalability. Another contribution of this research is the manual expansion of the RSketch dataset into the RSketch_Ext dataset. This expansion, which substantially increases both the volume and diversity of the dataset, provides valuable insights into the performance of SBRSIR algorithms. We made both the code of our method and the RSketch_Ext dataset publicly available online. This initiative aims to enable and encourage ongoing research and innovation in the area of sketch-based remote-sensing image retrieval.

Author Contributions

Conceptualization, C.W. and X.M.; methodology, B.Y. and C.W.; software, B.Y.; validation, B.Y. and B.S.; formal analysis, C.W.; resources, C.W.; data curation, B.S., Z.L. and F.S.; writing—original draft preparation, B.Y. and C.W.; writing—review and editing, B.Y., C.W. and F.S.; visualization, B.Y.; project administration, C.W. and X.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grants 41901410. Natural Science Research Project of Anhui Educational Committee under Grants 2023AH050103.

Data Availability Statement

We present the codes openly available at https://github.com/Snowstormfly/Cross-modal-retrieval-MLAGT (accessed on 10 March 2024) to encourage more extensive research and applications in the field of sketch-based remote-sensing image retrieval.

Acknowledgments

We would like to thank Anhui Province Key Laboratory of Wetland Ecosystem Protection and Restoration, Anhui University, for providing us with the hardware necessary for this project. We also thank the Space Climate Observatory Habitat Yangtze project, as this model is one of the project’s outcomes.

Conflicts of Interest

Zhuang Liu, from Shanghai Ubiquitous Navigation Technology Co. Ltd., and Fangde Sun, from The 54th Research Institute of China Electronics Technology Group Corporation, declare that there is no conflict of interest in the authorship and materials of the paper.

References

  1. Li, Y.; Ma, J.; Zhang, Y. Image retrieval from remote sensing big data: A survey. Inf. Fusion 2021, 67, 94–115. [Google Scholar] [CrossRef]
  2. Yuan, X.; Shi, J.; Gu, L. A review of deep learning methods for semantic segmentation of remote sensing imagery. Expert Syst. Appl. 2021, 169, 114417. [Google Scholar] [CrossRef]
  3. Zhang, X.; Zhou, Y.n.; Luo, J. Deep learning for processing and analysis of remote sensing big data: A technical review. Big Earth Data 2022, 6, 527–560. [Google Scholar] [CrossRef]
  4. Zhou, W.; Newsam, S.; Li, C.; Shao, Z. PatternNet: A benchmark dataset for performance evaluation of remote sensing image retrieval. ISPRS J. Photogramm. Remote Sens. 2018, 145, 197–209. [Google Scholar] [CrossRef]
  5. Li, X.; Yang, J.; Ma, J. Recent developments of content-based image retrieval (CBIR). Neurocomputing 2021, 452, 675–689. [Google Scholar] [CrossRef]
  6. Liu, C.; Ma, J.; Tang, X.; Liu, F.; Zhang, X.; Jiao, L. Deep hash learning for remote sensing image retrieval. IEEE Trans. Geosci. Remote Sens. 2020, 59, 3420–3443. [Google Scholar] [CrossRef]
  7. Abdullah, T.; Bazi, Y.; Al Rahhal, M.M.; Mekhalfi, M.L.; Rangarajan, L.; Zuair, M. TextRS: Deep bidirectional triplet network for matching text to remote sensing images. Remote Sens. 2020, 12, 405. [Google Scholar] [CrossRef]
  8. Xu, F.; Yang, W.; Jiang, T.; Lin, S.; Luo, H.; Xia, G.-S. Mental retrieval of remote sensing images via adversarial sketch-image feature learning. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7801–7814. [Google Scholar] [CrossRef]
  9. Yu, Q.; Liu, F.; Song, Y.-Z.; Xiang, T.; Hospedales, T.M.; Loy, C.-C. Sketch me that shoe. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 799–807. [Google Scholar]
  10. Xu, P.; Hospedales, T.M.; Yin, Q.; Song, Y.-Z.; Xiang, T.; Wang, L. Deep learning for free-hand sketch: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 285–312. [Google Scholar] [CrossRef] [PubMed]
  11. Chaudhuri, A.; Bhunia, A.K.; Song, Y.-Z.; Dutta, A. Data-Free Sketch-Based Image Retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 12084–12093. [Google Scholar]
  12. Chowdhury, P.N.; Bhunia, A.K.; Sain, A.; Koley, S.; Xiang, T.; Song, Y.-Z. SceneTrilogy: On Human Scene-Sketch and its Complementarity with Photo and Text. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 10972–10983. [Google Scholar]
  13. Lin, F.; Li, M.; Li, D.; Hospedales, T.; Song, Y.-Z.; Qi, Y. Zero-Shot Everything Sketch-Based Image Retrieval, and in Explainable Style. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 23349–23358. [Google Scholar]
  14. Sain, A.; Bhunia, A.K.; Chowdhury, P.N.; Koley, S.; Xiang, T.; Song, Y.-Z. Clip for all things zero-shot sketch-based image retrieval, fine-grained or not. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 2765–2775. [Google Scholar]
  15. Sain, A.; Bhunia, A.K.; Yang, Y.; Xiang, T.; Song, Y.-Z. Stylemeup: Towards style-agnostic sketch-based image retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Kuala Lumpur, Selangor, Malaysia, 20–25 June 2021; pp. 8504–8513. [Google Scholar]
  16. Chen, W.; Liu, Y.; Wang, W.; Bakker, E.M.; Georgiou, T.; Fieguth, P.; Liu, L.; Lew, M.S. Deep learning for instance retrieval: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 7270–7292. [Google Scholar] [CrossRef]
  17. Yu, D.; Liu, Y.; Pang, Y.; Li, Z.; Li, H. A multi-layer deep fusion convolutional neural network for sketch based image retrieval. Neurocomputing 2018, 296, 23–32. [Google Scholar] [CrossRef]
  18. Ren, H.; Zheng, Z.; Wu, Y.; Lu, H.; Yang, Y.; Shan, Y.; Yeung, S.-K. ACNet: Approaching-and-Centralizing Network for Zero-Shot Sketch-Based Image Retrieval. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 5022–5035. [Google Scholar] [CrossRef]
  19. Jiao, S.; Han, X.; Xiong, F.; Yang, X.; Han, H.; He, L.; Kuang, L. Deep cross-modal discriminant adversarial learning for zero-shot sketch-based image retrieval. Neural Comput. Appl. 2022, 34, 13469–13483. [Google Scholar] [CrossRef]
  20. Lei, J.; Song, Y.; Peng, B.; Ma, Z.; Shao, L.; Song, Y.-Z. Semi-heterogeneous three-way joint embedding network for sketch-based image retrieval. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 3226–3237. [Google Scholar] [CrossRef]
  21. Zhang, H.; Zhang, C.; Wu, M. Sketch-based cross-domain image retrieval via heterogeneous network. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  22. Chaudhuri, U.; Banerjee, B.; Bhattacharya, A.; Datcu, M. A simplified framework for zero-shot cross-modal sketch data retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 182–183. [Google Scholar]
  23. Chaudhuri, U.; Banerjee, B.; Bhattacharya, A.; Datcu, M. A zero-shot sketch-based intermodal object retrieval scheme for remote sensing images. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  24. Chaudhuri, U.; Bose, R.; Banerjee, B.; Bhattacharya, A.; Datcu, M. Zero-shot cross-modal retrieval for remote sensing images with minimal supervision. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  25. Jiang, T.-B.; Xia, G.-S.; Lu, Q.-K.; Shen, W.-M. Retrieving aerial scene images with learned deep image-sketch features. J. Comput. Sci. Technol. 2017, 32, 726–737. [Google Scholar] [CrossRef]
  26. Xu, F.; Zhang, R.; Yang, W.; Xia, G.-S. Mental retrieval of large-scale satellite images via learned sketch-image deep features. In Proceedings of the IGARSS 2019-2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 3356–3359. [Google Scholar]
  27. Hu, R.; Collomosse, J. A performance evaluation of gradient field hog descriptor for sketch based image retrieval. Comput. Vis. Image Underst. 2013, 117, 790–806. [Google Scholar] [CrossRef]
  28. Ha, D.; Eck, D. A neural representation of sketch drawings. arXiv 2017, arXiv:1704.03477. [Google Scholar]
  29. Huang, F.; Canny, J.F.; Nichols, J. Swire: Sketch-based user interface retrieval. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, UK, 4–9 May 2019; pp. 1–10. [Google Scholar]
  30. Bertolotto, M.; Carswell, J.D.; McLoughlin, E.; O’Sullivan, D.; Wilson, D. Using sketches and knowledge bases for geo-spatial image retrieval. Comput. Environ. Urban Syst. 2006, 30, 29–53. [Google Scholar] [CrossRef]
  31. Yelamarthi, S.K.; Reddy, S.K.; Mishra, A.; Mittal, A. A zero-shot framework for sketch based image retrieval. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 300–317. [Google Scholar]
  32. Bhunia, A.K.; Yang, Y.; Hospedales, T.M.; Xiang, T.; Song, Y.-Z. Sketch less for more: On-the-fly fine-grained sketch-based image retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9779–9788. [Google Scholar]
  33. Lampert, C.H.; Nickisch, H.; Harmeling, S. Attribute-based classification for zero-shot visual object categorization. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 453–465. [Google Scholar] [CrossRef]
  34. Zhang, Z.; Saligrama, V. Zero-shot learning via semantic similarity embedding. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4166–4174. [Google Scholar]
  35. Pang, K.; Li, K.; Yang, Y.; Zhang, H.; Hospedales, T.M.; Xiang, T.; Song, Y.-Z. Generalising fine-grained sketch-based image retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 677–686. [Google Scholar]
  36. Bhattacharjee, S.D.; Yuan, J.; Huang, Y.; Meng, J.; Duan, L. Query adaptive multiview object instance search and localization using sketches. IEEE Trans. Multimed. 2018, 20, 2761–2773. [Google Scholar] [CrossRef]
  37. Saavedra, J.M. Sketch based image retrieval using a soft computation of the histogram of edge local orientations (s-helo). In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 2998–3002. [Google Scholar]
  38. Dutta, T.; Biswas, S. s-sbir: Style augmented sketch based image retrieval. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Aspen, CO, USA, 1–5 March 2020; pp. 3261–3270. [Google Scholar]
  39. Zheng, L.; Yang, Y.; Tian, Q. SIFT meets CNN: A decade survey of instance retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 1224–1244. [Google Scholar] [CrossRef] [PubMed]
  40. Jiang, J.; Wang, R.; Lin, S.; Wang, F. Sfsegnet: Parse freehand sketches using deep fully convolutional networks. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar]
  41. Wang, F.; Li, Y. Spatial matching of sketches without point correspondence. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 4828–4832. [Google Scholar]
  42. Li, K.; Pang, K.; Song, Y.-Z.; Xiang, T.; Hospedales, T.M.; Zhang, H. Toward deep universal sketch perceptual grouper. IEEE Trans. Image Process. 2019, 28, 3219–3231. [Google Scholar] [CrossRef] [PubMed]
  43. Xu, P.; Joshi, C.K.; Bresson, X. Multigraph transformer for free-hand sketch recognition. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 5150–5161. [Google Scholar] [CrossRef] [PubMed]
  44. Lin, H.; Fu, Y.; Xue, X.; Jiang, Y.-G. Sketch-bert: Learning sketch bidirectional encoder representation from transformers by self-supervised learning of sketch gestalt. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6758–6767. [Google Scholar]
  45. Kim, W.; Goyal, B.; Chawla, K.; Lee, J.; Kwon, K. Attention-based ensemble for deep metric learning. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 736–751. [Google Scholar]
  46. Chen, B.; Deng, W. Hybrid-attention based decoupled metric learning for zero-shot image retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2750–2759. [Google Scholar]
  47. Li, X.; Wei, S.; Wang, J.; Du, Y.; Ge, M. Adaptive Multi-Proxy for Remote Sensing Image Retrieval. Remote Sens. 2022, 14, 5615. [Google Scholar] [CrossRef]
  48. Liang, Y.; Ge, C.; Tong, Z.; Song, Y.; Wang, J.; Xie, P. Not all patches are what you need: Expediting vision transformers via token reorganizations. arXiv 2022, arXiv:2202.07800. [Google Scholar]
  49. Liu, F.; Zou, C.; Deng, X.; Zuo, R.; Lai, Y.-K.; Ma, C.; Liu, Y.-J.; Wang, H. Scenesketcher: Fine-grained image retrieval with scene sketches. In Proceedings of the European Conference on Computer Vision (ECCV), Copenhagen, Denmark, 23–28 August 2020; pp. 718–734. [Google Scholar]
  50. Jgou, H.; Perronnin, F.; Douze, M.; Snchez, J.; Prez, P.; Schmid, C. Aggregating local image descriptors into compact codes. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1704–1716. [Google Scholar] [CrossRef] [PubMed]
  51. Spyromitros-Xioufis, E.; Papadopoulos, S.; Kompatsiaris, I.Y.; Tsoumakas, G.; Vlahavas, I. A comprehensive study over VLAD and product quantization in large-scale image retrieval. IEEE Trans. Multimed. 2014, 16, 1713–1728. [Google Scholar] [CrossRef]
  52. Xu, P.; Huang, Y.; Yuan, T.; Pang, K.; Song, Y.-Z.; Xiang, T.; Hospedales, T.M.; Ma, Z.; Guo, J. Sketchmate: Deep hashing for million-scale human sketch retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8090–8098. [Google Scholar]
  53. Shen, Y.; Liu, L.; Shen, F.; Shao, L. Zero-shot sketch-image hashing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3598–3607. [Google Scholar]
  54. Liu, L.; Shen, F.; Shen, Y.; Liu, X.; Shao, L. Deep sketch hashing: Fast free-hand sketch-based image retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2862–2871. [Google Scholar]
  55. Zhao, H.; Liu, M.; Li, M. Feature Fusion and Metric Learning Network for Zero-Shot Sketch-Based Image Retrieval. Entropy 2023, 25, 502. [Google Scholar] [CrossRef] [PubMed]
  56. Dai, G.; Xie, J.; Fang, Y. Deep correlated holistic metric learning for sketch-based 3D shape retrieval. IEEE Trans. Image Process. 2018, 27, 3374–3386. [Google Scholar] [CrossRef] [PubMed]
  57. Zhou, W.; Li, H.; Lu, Y.; Tian, Q. Large scale image search with geometric coding. In Proceedings of the 19th ACM International Conference on Multimedia, Scottsdale, AZ, USA, 28 November–1 December 2011; pp. 1349–1352. [Google Scholar]
  58. Matsui, Y.; Ito, K.; Aramaki, Y.; Fujimoto, A.; Ogawa, T.; Yamasaki, T.; Aizawa, K. Sketch-based manga retrieval using manga109 dataset. Multimed. Tools Appl. 2017, 76, 21811–21838. [Google Scholar] [CrossRef]
  59. He, J.-Y.; Wu, X.; Jiang, Y.-G.; Zhao, B.; Peng, Q. Sketch recognition with deep visual-sequential fusion model. In Proceedings of the 25th ACM International Conference on Multimedia, Silicon Valley, CA, USA, 23–27 October 2017; pp. 448–456. [Google Scholar]
  60. Xu, P.; Song, Z.; Yin, Q.; Song, Y.-Z.; Wang, L. Deep self-supervised representation learning for free-hand sketch. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 1503–1513. [Google Scholar] [CrossRef]
  61. Creswell, A.; Bharath, A.A. Adversarial training for sketch retrieval. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, Holland, 11–14 October 2016; pp. 798–809. [Google Scholar]
  62. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  63. Adil, R.; Kamel, B.; Amina, B. Deep Supervised Hashing by Fusing Multiscale Deep Features. Preprints 2023, 2023091699. [Google Scholar]
  64. Yang, Y.; Newsam, S. Geographic image retrieval using local invariant features. IEEE Trans. Geosci. Remote Sens. 2012, 51, 818–832. [Google Scholar] [CrossRef]
  65. Xia, G.-S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L.; Lu, X. AID: A benchmark data set for performance evaluation of aerial scene classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3965–3981. [Google Scholar] [CrossRef]
  66. Cheng, G.; Han, J.; Lu, X. Remote sensing image scene classification: Benchmark and state of the art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
  67. Sheng, G.; Yang, W.; Xu, T.; Sun, H. High-resolution satellite scene classification using a sparse coding based multiple feature combination. Int. J. Remote Sens. 2012, 33, 2395–2412. [Google Scholar] [CrossRef]
  68. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  69. Loshchilov, I.; Hutter, F. Fixing weight decay regularization in adam. In Proceedings of the 6th International Conference on Learning Representations (ICLR 2018), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  70. Radenovic, F.; Tolias, G.; Chum, O. Deep shape matching. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 751–767. [Google Scholar]
  71. Dey, S.; Riba, P.; Dutta, A.; Llados, J.; Song, Y.-Z. Doodle to search: Practical zero-shot sketch-based image retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2179–2188. [Google Scholar]
  72. Zhen, L.; Hu, P.; Wang, X.; Peng, D. Deep supervised cross-modal retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10394–10403. [Google Scholar]
  73. Jing, L.; Vahdani, E.; Tan, J.; Tian, Y. Cross-modal center loss for 3d cross-modal retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Kuala Lumpur, Selangor, Malaysia, 20–25 June 2021; pp. 3142–3151. [Google Scholar]
Figure 1. Demo of sketch-based remote-sensing image retrieval process and its advantages compared to content-based retrieval.
Figure 1. Demo of sketch-based remote-sensing image retrieval process and its advantages compared to content-based retrieval.
Remotesensing 16 01653 g001
Figure 2. Model overview: Self-attention feature extraction stage, multi-level extraction, and filtering of feature information for different modes. Cross-attention and similarity computation stage, establishing the correspondence of two modes and calculating the similarity. Model training stage, training the network using triplet loss.
Figure 2. Model overview: Self-attention feature extraction stage, multi-level extraction, and filtering of feature information for different modes. Cross-attention and similarity computation stage, establishing the correspondence of two modes and calculating the similarity. Model training stage, training the network using triplet loss.
Remotesensing 16 01653 g002
Figure 3. Samples in the RSketch_Ext dataset (presenting two sketches and two remote-sensing images for each category; one sketch is from original RSketch, and one sketch is from our extension).
Figure 3. Samples in the RSketch_Ext dataset (presenting two sketches and two remote-sensing images for each category; one sketch is from original RSketch, and one sketch is from our extension).
Remotesensing 16 01653 g003
Figure 4. Retrieval results of 5 seen categories in a test on the RSketch_Ext dataset with the top 5 retrieval results. The green checkmark in the lower right corner of the picture marks the correct retrieval, and the red X in the lower right corner marks the wrong retrieval. The five categories selected are baseball diamond, beach, bridge, football field, and river.
Figure 4. Retrieval results of 5 seen categories in a test on the RSketch_Ext dataset with the top 5 retrieval results. The green checkmark in the lower right corner of the picture marks the correct retrieval, and the red X in the lower right corner marks the wrong retrieval. The five categories selected are baseball diamond, beach, bridge, football field, and river.
Remotesensing 16 01653 g004
Figure 5. Retrieval results of 5 unseen categories in a test on the RSketch_Ext dataset with the top 5 retrieval results. The green checkmark in the lower right corner of the picture marks the correct retrieval, and the red X in the lower right corner marks the wrong retrieval. The five categories selected are basketball court, crosswalk, oil gas field, runway, and tennis court.
Figure 5. Retrieval results of 5 unseen categories in a test on the RSketch_Ext dataset with the top 5 retrieval results. The green checkmark in the lower right corner of the picture marks the correct retrieval, and the red X in the lower right corner marks the wrong retrieval. The five categories selected are basketball court, crosswalk, oil gas field, runway, and tennis court.
Remotesensing 16 01653 g005
Figure 6. Retrieval results of GF-1 image tiles.
Figure 6. Retrieval results of GF-1 image tiles.
Remotesensing 16 01653 g006
Table 1. Symbol correspondence table.
Table 1. Symbol correspondence table.
SymbolInterpretations
SQuery sketch
RRemote-sensing image
D R Remote-sensing image database
EVisual token embedding
[RT]Retrieval token
QQueries in transformers
KKeys in transformers
VValues in transformers
Table 2. Classes in RSketch, RSketch_Ext, Earth on Canvas, and UCMerge Landuse dataset.
Table 2. Classes in RSketch, RSketch_Ext, Earth on Canvas, and UCMerge Landuse dataset.
RSketch and RSketch_ExtEarth on CanvasUCMerge Landuse
AirplaneAirplaneAirplane
Baseball DiamondBaseball DiamondBaseball Diamond
Golf CourseGolf CourseGolf Course
IntersectionIntersectionIntersection
OverpassOverpassOverpass
RiverRiverRiver
RunwayRunwayRunway
Storage TanksStorage TanksStorage Tanks
Tennis CourtTennis CourtTennis Court
Basketball CourtBuildingsBuildings
BeachFreewayBeach
BridgeHarborAgricultural
Closed RoadMobile Home ParkChaparral
CrosswalkParking LotDense Residential
Football Field Forest
Oil Gas Field Freeway
Railway Harbor
Runway Marking Medium Residential
Swimming Pool Mobile Home Park
Wastewater Treatment Plant (WWTP) Parking Lot
Sparse Residential
Table 3. Unseen classes in each fold.
Table 3. Unseen classes in each fold.
S1S2S3S4
airplanebaseball diamondbasketball courtbeach
bridgeclosed roadcrosswalkfootball field
golf courseintersectionoil gas fieldoverpass
railwayriverrunwayrunway marking
storage tankswimming pooltennis courtWWTP
Table 4. Test result (%) with RSketch dataset.
Table 4. Test result (%) with RSketch dataset.
SeenUnseen
mAPTop10Top50Top100mAPTop10Top50Top100
MR-SBIR83.7588.7785.5977.2047.8657.7049.9642.44
DAL92.5694.9793.8189.1950.6761.9054.6246.02
DSM56.8071.6064.0655.0019.2921.9021.0619.71
DOODLE49.1156.6748.3323.6733.2435.0033.0031.50
DSCMR96.1296.3796.8294.8046.6027.0049.2641.71
CMCL95.1496.0795.5993.4141.6151.3043.2036.50
ACNet38.1144.2036.5432.4425.1428.1023.7321.32
Ours98.1798.5098.3195.9376.6285.0079.9670.01
Table 5. Model test results (%) on RSketch_Ext dataset.
Table 5. Model test results (%) on RSketch_Ext dataset.
SeenUnseen
mAPTop10Top50Top100mAPTop10Top50Top100
MR-SBIR75.4587.6085.1582.8635.2353.6046.7042.13
DOODLE48.4354.3946.8525.4730.1632.2026.5722.50
ACNet40.9037.6034.4030.4029.0831.6028.7226.40
Ours97.8898.7399.0599.0770.7982.2080.0477.80
Ours-RSketch98.1798.5098.3195.9376.6285.0079.9670.01
Table 6. Model test results (%) on Earth on Canvas dataset.
Table 6. Model test results (%) on Earth on Canvas dataset.
SeenUnseen
mAPTop10Top50Top100mAPTop10Top50Top100
MR-SBIR61.6363.9556.4548.9142.7252.1045.9841.06
DOODLE47.7852.0048.0040.0021.6228.0021.3316.00
ACNet34.8730.8031.6829.3230.9626.4023.8424.28
Ours83.3185.6182.5676.8363.3271.1064.1656.98
Table 7. Model test results (%) on UCMerge Landuse dataset.
Table 7. Model test results (%) on UCMerge Landuse dataset.
SeenUnseen
mAPTop10Top50Top100mAPTop10Top50Top100
MR-SBIR81.0188.3782.4771.9385.5485.1779.4771.73
DOODLE57.0373.3354.6730.6737.2340.0028.0024.00
ACNet38.4832.8031.1227.7231.7931.2028.0026.36
Ours98.0098.7898.6895.3589.2396.4291.8782.43
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, B.; Wang, C.; Ma, X.; Song, B.; Liu, Z.; Sun, F. Zero-Shot Sketch-Based Remote-Sensing Image Retrieval Based on Multi-Level and Attention-Guided Tokenization. Remote Sens. 2024, 16, 1653. https://doi.org/10.3390/rs16101653

AMA Style

Yang B, Wang C, Ma X, Song B, Liu Z, Sun F. Zero-Shot Sketch-Based Remote-Sensing Image Retrieval Based on Multi-Level and Attention-Guided Tokenization. Remote Sensing. 2024; 16(10):1653. https://doi.org/10.3390/rs16101653

Chicago/Turabian Style

Yang, Bo, Chen Wang, Xiaoshuang Ma, Beiping Song, Zhuang Liu, and Fangde Sun. 2024. "Zero-Shot Sketch-Based Remote-Sensing Image Retrieval Based on Multi-Level and Attention-Guided Tokenization" Remote Sensing 16, no. 10: 1653. https://doi.org/10.3390/rs16101653

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop