remotesensing-logo

Journal Browser

Journal Browser

Intelligent Remote Sensing Data Interpretation

A special issue of Remote Sensing (ISSN 2072-4292). This special issue belongs to the section "AI Remote Sensing".

Deadline for manuscript submissions: closed (15 July 2024) | Viewed by 12911

Special Issue Editors

Institute of Advanced Research in Artificial Intelligence (IARAI), Vienna, Austria
Interests: remote sensing image synthesis from text; AI security for earth observation; remote sensing image interpretation; natural hazard monitoring; cross-domain semantic segmentation
Special Issues, Collections and Topics in MDPI journals

grade E-Mail Website
Guest Editor
1.Helmholtz Institute Freiberg for Resource Technology, Helmholtz-Zentrum Dresden-Rossendorf (HZDR), D-09599 Freiberg, Germany
2. Institute of Advanced Research in Artificial Intelligence (IARAI), 1030 Wien, Austria
Interests: hyperspectral image interpretation; multisensor and multitemporal data fusion
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

With the successful launch of an increasing number of remote sensing satellites, the amount of Earth observation data is showing an explosive growth trend. Such a massive amount of data also makes it more challenging to achieve fast, accurate, and automated remote sensing data interpretation.

To address this challenge and boost the development of advanced artificial intelligence algorithms for the interpretation of remote sensing data, we would like to invite you to contribute to this Special Issue, which will gather new insights and contributions to the study of Intelligent Remote Sensing Data Interpretation. Original research articles and reviews are welcome. Topics can be related, but are not limited to, the following:

  • Intelligent interpretation algorithms for tasks such as scene classification, object detection, change detection, and semantic segmentation;
  • Earth-observation-oriented machine learning techniques such as weakly supervised learning, zero- and few-shot learning, and domain adaptation and transfer learning;
  • Advanced processing methods for hyperspectral/multispectral/RGB/LiDAR/synthetic aperture radar (SAR) data;
  • Multisensor and multitemporal remote sensing data fusion;
  • Vision and language models for Earth observation;
  • Artificial intelligence techniques for social good;
  • Intelligent natural hazard monitoring.

Dr. Yonghao Xu
Dr. Pedram Ghamisi
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Remote Sensing is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2700 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • remote sensing data interpretation
  • remote sensing data processing
  • artificial intelligence
  • machine learning
  • deep learning
  • scene classification
  • object detection
  • change detection
  • semantic segmentation
  • natural hazard monitoring

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Published Papers (9 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

22 pages, 11399 KiB  
Article
Automatic Aircraft Identification with High Precision from SAR Images Considering Multiscale Problems and Channel Information Enhancement
by Jing Wang, Guohan Liu, Jiaxing Liu, Wenjie Dong and Wanying Song
Remote Sens. 2024, 16(17), 3177; https://doi.org/10.3390/rs16173177 - 28 Aug 2024
Viewed by 321
Abstract
The SAR system possesses the ability to carry out all-day and all-weather imaging, which is highly valuable in the application of aircraft identification. However, aircraft identification from SAR images still faces great challenges due to speckle noise interference, multiscale problems, and complex background [...] Read more.
The SAR system possesses the ability to carry out all-day and all-weather imaging, which is highly valuable in the application of aircraft identification. However, aircraft identification from SAR images still faces great challenges due to speckle noise interference, multiscale problems, and complex background interference. To solve these problems, an efficient bidirectional path multiscale fusion and attention network (EBMA-Net) is proposed in this paper. It employs bidirectional connectivity to fuse the features of aircraft with different scales to perform the accurate detection of aircraft even when the background is highly complex. In the presented EBMA-Net, a module called efficient multiscale channel attention fusion (EMCA) and three parallel squeeze efficient channel attention (SECA) modules are proposed. In the EMCA module, the bidirectional paths are created by stacking upper and lower fusion modules, which effectively integrate shallow detailed features and deep semantic information. So, the detection performance of aircraft at different scales is improved. In the SECA module, the dependency relationship between feature channels is explicitly modeled, which can automatically learn the importance of different channels, prioritize key features, so as to improve the precision and robustness of aircraft identification. In the experiment, the public dataset of aircraft identification (i.e., SAR-AIRcraft-1.0, which is generated from the GF-3 satellite) from high-resolution SAR systems is used, and several other excellent target-detection networks are used for performance comparison, namely, YOLOv5s, YOLOv7, MGCAN, and EBPA2N. According to the results, the average aircraft detection accuracy of EBMA-Net is 91.31%, which is 4.5% higher than YOLOv7; and the false alarm rate is decreased by 5%. Its accuracy in the identification of aircraft can reach 95.6%, which is about 3.7% higher than YOLOv7. Therefore, the EBMA-Net obviously outperforms the other networks for aircraft detection and identification. The proposed EBMA-Net, which can capture the detailed information and better restrain the background interference, could also be used to perform the detection and identification of dense targets with different scales and background from SAR images. Full article
(This article belongs to the Special Issue Intelligent Remote Sensing Data Interpretation)
Show Figures

Graphical abstract

19 pages, 6224 KiB  
Article
Small Object Detection in Medium–Low-Resolution Remote Sensing Images Based on Degradation Reconstruction
by Yongxian Zhao, Haijiang Sun and Shuai Wang
Remote Sens. 2024, 16(14), 2645; https://doi.org/10.3390/rs16142645 - 19 Jul 2024
Viewed by 438
Abstract
With the continuous development of space remote sensing technology, the spatial resolution of visible remote sensing images has been continuously improved, which has promoted the progress of remote sensing target detection. However, due to the limitation of sensor lattice size, it is still [...] Read more.
With the continuous development of space remote sensing technology, the spatial resolution of visible remote sensing images has been continuously improved, which has promoted the progress of remote sensing target detection. However, due to the limitation of sensor lattice size, it is still challenging to obtain a large range of high-resolution (HR) remote sensing images in practical applications, which makes it difficult to carry out target monitoring in a large range of areas. At present, many object detection methods focus on the detection and positioning technology of HR remote sensing images, but there are relatively few studies on object detection methods using medium- and low-resolution (M-LR) remote sensing images. Because of its wide coverage area and short observation period, M-LR remote sensing imagery is of great significance for obtaining information quickly in space applications. However, the small amount of fine-texture information on objects in M-LR images brings great challenges to detection and recognition tasks. Therefore, we propose a small target detection method based on degradation reconstruction, named DRADNet. Different from the previous methods that use super resolution as a pre-processing step and then directly input the image into the detector, we have designed an additional degenerate reconstruction-assisted framework to effectively improve the detector’s performance in detection tasks with M-LR remote sensing images. In addition, we introduce a hybrid parallel-attention feature fusion module in the detector to achieve focused attention on target features and suppress redundant complex backgrounds, thus improving the accuracy of the model in small target localization. The experimental results are based on the widely used VEDAI dataset and Airbus-Ships dataset, and verify the effectiveness of our method in the detection of small- and medium-sized targets in M-LR remote sensing images. Full article
(This article belongs to the Special Issue Intelligent Remote Sensing Data Interpretation)
Show Figures

Figure 1

26 pages, 23255 KiB  
Article
DBI-Attack:Dynamic Bi-Level Integrated Attack for Intensive Multi-Scale UAV Object Detection
by Zhengyang Zhao, Buhong Wang, Zhen Wang and Xuan Yao
Remote Sens. 2024, 16(14), 2570; https://doi.org/10.3390/rs16142570 - 13 Jul 2024
Viewed by 537
Abstract
Benefiting from the robust feature representation capability of convolutional neural networks (CNNs), the object detection technology of intelligent high-altitude UAV remote sensing has been developed rapidly. In this field, the adversarial examples (AEs) pose serious security risks and vulnerabilities to deep learning-based systems. [...] Read more.
Benefiting from the robust feature representation capability of convolutional neural networks (CNNs), the object detection technology of intelligent high-altitude UAV remote sensing has been developed rapidly. In this field, the adversarial examples (AEs) pose serious security risks and vulnerabilities to deep learning-based systems. Due to the limitation of object size, image degradation, and scene brightness, adding adversarial disturbances to small and dense objects is extremely challenging. To study the threat of AE for UAV object detection, a dynamic bi-level integrated attack (DBI-Attack) is proposed for intensive multi-scale UAV object detection. Firstly, we use the dynamic iterative attack (DIA) method to generate perturbation on the classification level by improving the momentum iterative fast gradient sign method (MIM). Secondly, the bi-level adversarial attack method (BAAM) is constructed to add global perturbation on the decision level for completing the white-box attack. Finally, the integrated black-box attack method (IBAM) is combined to realize the black-box mislabeling and fabrication attacks. We experiment on the real drone traffic vehicle detection datasets to better evaluate the attack effectiveness. The experimental results show that the proposed method can achieve mislabeling and fabrication attacks on the UAV object detectors in black-box conditions. Furthermore, the adversarial training is applied to improve the model robustness. This work aims to call more attention to the adversarial and defensive aspects of UAV target detection models. Full article
(This article belongs to the Special Issue Intelligent Remote Sensing Data Interpretation)
Show Figures

Figure 1

22 pages, 5522 KiB  
Article
Spectral-Spatial Mamba for Hyperspectral Image Classification
by Lingbo Huang, Yushi Chen and Xin He
Remote Sens. 2024, 16(13), 2449; https://doi.org/10.3390/rs16132449 - 3 Jul 2024
Cited by 3 | Viewed by 1118
Abstract
Recently, transformer has gradually attracted interest for its excellence in modeling the long-range dependencies of spatial-spectral features in HSI. However, transformer has the problem of the quadratic computational complexity due to the self-attention mechanism, which is heavier than other models and thus has [...] Read more.
Recently, transformer has gradually attracted interest for its excellence in modeling the long-range dependencies of spatial-spectral features in HSI. However, transformer has the problem of the quadratic computational complexity due to the self-attention mechanism, which is heavier than other models and thus has limited adoption in HSI processing. Fortunately, the recently emerging state space model-based Mamba shows great computational efficiency while achieving the modeling power of transformers. Therefore, in this paper, we first proposed spectral-spatial Mamba (SS-Mamba) for HSI classification. Specifically, SS-Mamba mainly includes a spectral-spatial token generation module and several stacked spectral-spatial Mamba blocks. Firstly, the token generation module converts any given HSI cube to spatial and spectral tokens as sequences. And then these tokens are sent to stacked spectral-spatial mamba blocks (SS-MB). Each SS-MB includes two basic mamba blocks and a spectral-spatial feature enhancement module. The spatial and spectral tokens are processed separately by the two basic mamba blocks, correspondingly. Moreover, the feature enhancement module modulates spatial and spectral tokens using HSI sample’s center region information. Therefore, the spectral and spatial tokens cooperate with each other and achieve information fusion within each block. The experimental results conducted on widely used HSI datasets reveal that the proposed SS-Mamba requires less processing time compared with transformer. The Mamba-based method thus opens a new window for HSI classification. Full article
(This article belongs to the Special Issue Intelligent Remote Sensing Data Interpretation)
Show Figures

Figure 1

16 pages, 1748 KiB  
Article
Learning SAR-Optical Cross Modal Features for Land Cover Classification
by Yujun Quan, Rongrong Zhang, Jian Li, Song Ji, Hengliang Guo and Anzhu Yu
Remote Sens. 2024, 16(2), 431; https://doi.org/10.3390/rs16020431 - 22 Jan 2024
Cited by 2 | Viewed by 1515
Abstract
Synthetic aperture radar (SAR) and optical images provide highly complementary ground information. The fusion of SAR and optical data can significantly enhance semantic segmentation inference results. However, the fusion methods for multimodal data remains a challenge for current research due to significant disparities [...] Read more.
Synthetic aperture radar (SAR) and optical images provide highly complementary ground information. The fusion of SAR and optical data can significantly enhance semantic segmentation inference results. However, the fusion methods for multimodal data remains a challenge for current research due to significant disparities in imaging mechanisms from diverse sources. Our goal was to bridge the significant gaps between optical and SAR images by developing a dual-input model that utilizes image-level fusion. To improve most existing state-of-the-art image fusion methods, which often assign equal weights to multiple modalities, we employed the principal component analysis (PCA) transform approach. Subsequently, we performed feature-level fusion on shallow feature maps, which retain rich geometric information. We also incorporated a channel attention module to highlight channels rich in features and suppress irrelevant information. This step is crucial due to the substantial similarity between SAR and optical images in shallow layers such as geometric features. In summary, we propose a generic multimodal fusion strategy that can be attached to most encoding–decoding structures for feature classification tasks, designed with two inputs. One input is the optical image, and the other is the three-band fusion data obtained by combining the PCA component of the optical image with the SAR. Our feature-level fusion method effectively integrates multimodal data. The efficiency of our approach was validated using various public datasets, and the results showed significant improvements when applied to several land cover classification models. Full article
(This article belongs to the Special Issue Intelligent Remote Sensing Data Interpretation)
Show Figures

Figure 1

23 pages, 6052 KiB  
Article
DPSDA-Net: Dual-Path Convolutional Neural Network with Strip Dilated Attention Module for Road Extraction from High-Resolution Remote Sensing Images
by Like Zhao, Linfeng Ye, Mi Zhang, Huawei Jiang, Zhen Yang and Mingwang Yang
Remote Sens. 2023, 15(15), 3741; https://doi.org/10.3390/rs15153741 - 27 Jul 2023
Cited by 3 | Viewed by 1411
Abstract
Roads extracted from high-resolution remote sensing images are widely used in many fields, such as autonomous driving, road planning, disaster relief, etc. However, road extraction from high-resolution remote sensing images has certain deficiencies in connectivity and completeness due to obstruction by surrounding ground [...] Read more.
Roads extracted from high-resolution remote sensing images are widely used in many fields, such as autonomous driving, road planning, disaster relief, etc. However, road extraction from high-resolution remote sensing images has certain deficiencies in connectivity and completeness due to obstruction by surrounding ground objects, the influence of similar targets, and the slender structure of roads themselves. To address this issue, we propose a novel dual-path convolutional neural network with a strip dilated attention module, named DPSDA-Net, which adopts a U-shaped encoder–decoder structure, combining the powerful advantages of attention mechanism, dilated convolution, and strip convolution. The encoder utilizes ResNet50 as its basic architecture. A strip position attention mechanism is added between each residual block to strengthen the coherent semantic information of a road. A long-distance shortcut connection operation is introduced to preserve the spatial information characteristics of the original image during the downsampling process. At the same time, a pyramid dilated module with a strip convolution and attention mechanism is constructed between the encoder and decoder to enhance the network feature extraction ability and multi-scale extraction of road feature information, expand the model’s receptive field, and pay more attention to the global spatial semantic and connectivity information. To verify the reliability of the proposed model, road extraction was carried out on the Massachusetts dataset and the LRSNY dataset. The experimental results show that, compared with other typical road extraction methods, the proposed model achieved a higher F1 score and IOU. The DPSDA-Net model can comprehensively characterize the structural features of roads, extract roads more accurately, retain road details, and improve the connectivity and integrity of road extraction in remote sensing images. Full article
(This article belongs to the Special Issue Intelligent Remote Sensing Data Interpretation)
Show Figures

Graphical abstract

19 pages, 2971 KiB  
Article
LiteST-Net: A Hybrid Model of Lite Swin Transformer and Convolution for Building Extraction from Remote Sensing Image
by Wei Yuan, Xiaobo Zhang, Jibao Shi and Jin Wang
Remote Sens. 2023, 15(8), 1996; https://doi.org/10.3390/rs15081996 - 10 Apr 2023
Cited by 8 | Viewed by 2304
Abstract
Extracting building data from remote sensing images is an efficient way to obtain geographic information data, especially following the emergence of deep learning technology, which results in the automatic extraction of building data from remote sensing images becoming increasingly accurate. A CNN (convolution [...] Read more.
Extracting building data from remote sensing images is an efficient way to obtain geographic information data, especially following the emergence of deep learning technology, which results in the automatic extraction of building data from remote sensing images becoming increasingly accurate. A CNN (convolution neural network) is a successful structure after a fully connected network. It has the characteristics of saving computation and translation invariance with improved local features, but it has difficulty obtaining global features. Transformers can compensate for the shortcomings of CNNs and more effectively obtain global features. However, the calculation number of transformers is excessive. To solve this problem, a Lite Swin transformer is proposed. The three matrices Q, K, and V of the transformer are simplified to only a V matrix, and the v of the pixel is then replaced by the v with the largest projection value on the pixel feature vector. In order to better integrate global features and local features, we propose the LiteST-Net model, in which the features extracted by the Lite Swin transformer and the CNN are added together and then sampled up step by step to fully utilize the global feature acquisition ability of the transformer and the local feature acquisition ability of the CNN. The comparison experiments on two open datasets are carried out using our proposed LiteST-Net and some classical image segmentation models. The results show that compared with other networks, all metrics of LiteST-Net are the best, and the predicted image is closer to the label. Full article
(This article belongs to the Special Issue Intelligent Remote Sensing Data Interpretation)
Show Figures

Graphical abstract

24 pages, 43487 KiB  
Article
Interpretation of Latent Codes in InfoGAN with SAR Images
by Zhenpeng Feng, Miloš Daković, Hongbing Ji, Xianda Zhou, Mingzhe Zhu, Xiyang Cui and Ljubiša Stanković
Remote Sens. 2023, 15(5), 1254; https://doi.org/10.3390/rs15051254 - 24 Feb 2023
Cited by 3 | Viewed by 2132
Abstract
Generative adversarial networks (GANs) can synthesize abundant photo-realistic synthetic aperture radar (SAR) images. Some modified GANs (e.g., InfoGAN) are even able to edit specific properties of the synthesized images by introducing latent codes. It is crucial for SAR image synthesis since the targets [...] Read more.
Generative adversarial networks (GANs) can synthesize abundant photo-realistic synthetic aperture radar (SAR) images. Some modified GANs (e.g., InfoGAN) are even able to edit specific properties of the synthesized images by introducing latent codes. It is crucial for SAR image synthesis since the targets in real SAR images have different properties due to the imaging mechanism. Despite the success of the InfoGAN in manipulating properties, there still lacks a clear explanation of how these latent codes affect synthesized properties; thus, editing specific properties usually relies on empirical trials, which are unreliable and time-consuming. In this paper, we show that latent codes are almost disentangled to affect the properties of SAR images in a nonlinear manner. By introducing some property estimators for latent codes, we are able to decompose the complex causality between latent codes and different properties. Both qualitative and quantitative experimental results demonstrate that the property value can be computed by the property estimators; inversely, the required latent codes can be computed given the desired properties. Unlike the original InfoGAN, which only provides the visual trend between properties and latent codes, the properties of SAR images can be manipulated numerically by latent codes as users expect. Full article
(This article belongs to the Special Issue Intelligent Remote Sensing Data Interpretation)
Show Figures

Figure 1

15 pages, 2246 KiB  
Article
Adversarial Remote Sensing Scene Classification Based on Lie Group Feature Learning
by Chengjun Xu, Jingqian Shu and Guobin Zhu
Remote Sens. 2023, 15(4), 914; https://doi.org/10.3390/rs15040914 - 7 Feb 2023
Cited by 4 | Viewed by 1608
Abstract
Convolutional Neural Networks have been widely used in remote sensing scene classification. Since this kind of model needs a large number of training samples containing data category information, a Generative Adversarial Network (GAN) is usually used to address the problem of lack of [...] Read more.
Convolutional Neural Networks have been widely used in remote sensing scene classification. Since this kind of model needs a large number of training samples containing data category information, a Generative Adversarial Network (GAN) is usually used to address the problem of lack of samples. However, GAN mainly generates scene data samples that do not contain category information. To address this problem, a novel supervised adversarial Lie Group feature learning network is proposed. In the case of limited data samples, the model can effectively generate data samples with category information. There are two main differences between our method and the traditional GAN. First, our model takes category information and data samples as the input of the model and optimizes the constraint of category information in the loss function, so that data samples containing category information can be generated. Secondly, the object scale sample generation strategy is introduced, which can generate data samples of different scales and ensure that the generated data samples contain richer feature information. After large-scale experiments on two publicly available and challenging datasets, it is found that our method can achieve better scene classification accuracy even with limited data samples. Full article
(This article belongs to the Special Issue Intelligent Remote Sensing Data Interpretation)
Show Figures

Figure 1

Back to TopTop