remotesensing-logo

Journal Browser

Journal Browser

Intelligent Processing of Multimodal Remote Sensing Data

A special issue of Remote Sensing (ISSN 2072-4292).

Deadline for manuscript submissions: 20 May 2026 | Viewed by 1301

Special Issue Editors


E-Mail Website
Guest Editor
School of Electronics and Information Engineering, Harbin Institute of Technology, Harbin 150001, China
Interests: multimodal remote sensing; multi/hyper-spectral and lidar data processing; deep learning

E-Mail Website
Guest Editor
Department of Telecommunications and Information Processing, Ghent University, 9000 Ghent, Belgium
Interests: artificial intelligence in machine vision; multiresolution statistical image models; image and video reconstruction, restoration, and analysis
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

Earth observation via remote sensing enables accurate characterization and monitoring of objects on the surface from satellite and airborne platforms. Multimodal remote sensing data, acquired by both active (synthetic aperture radar, and LiDAR) and passive (optical, multispectral, hyperspectral, and infrared) sensors, became readily available, providing complementary information regarding the structure, elevation, and material composition of objects in the observed scenes. However, the high dimensionality, heterogeneity, diverse temporal–spatial–spectral resolutions, and intrinsic complexity of these multimodal data pose significant challenges for conventional processing methodologies. Recent breakthroughs in artificial intelligence—particularly transformer networks, graph neural networks, and domain-specific foundation models (e.g., Skysense, Fomo, RingMoE, and EarthGPT)—are revolutionizing the joint exploitation and intelligent processing of multimodal remote sensing data. This paradigm shift enables unprecedented capabilities to fuse complementary information and extract actionable insights.

This Special Issue aims to compile cutting-edge research on AI-driven methods for processing multimodal remote sensing data acquired from diverse sensors and platforms. Topics may cover novel sensor techniques, benchmark datasets, advanced algorithms (including multimodal fusion techniques), and practical applications. By fostering interdisciplinary collaboration at the nexus of AI, geoscience, and remote sensing, this Special Issue strives to accelerate progress in this field, bridging theoretical innovation and practical implementation, thereby advancing scalable, robust, and trustworthy Earth observation solutions in multimodal remote sensing.

We invite original articles, reviews, and perspectives on the following themes:

(1) Advanced deep learning architectures: Novel models for joint processing heterogeneous data.

(2) Multimodal data preprocessing: Innovative techniques for registration, calibration, and noise reduction of heterogeneous data.

(3) Foundation models for remote sensing: Development, fine-tuning, and application of large-scale pre-trained models tailored to multimodal data.

(4) High-level information extraction: AI-driven approaches for semantic segmentation, object detection, scene classification, change detection, and scene understanding using multimodal data.

(5) Spectral point cloud generation: Fusion point clouds and multispectral or hyperspectral imagery for 3D spatial–spectral information generation and analysis.

(6) Benchmark datasets and evaluation: Novel, publicly available multimodal datasets (e.g., combining satellite and UAV data) accompanied by standardized evaluation protocols and baselines.

We look forward to receiving your submissions.

Dr. Xian Li
Prof. Dr. Aleksandra Pizurica
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 250 words) can be sent to the Editorial Office for assessment.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Remote Sensing is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2700 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • multimodal remote sensing
  • intelligent processing
  • remote sensing
  • multisource remote sensing data processing
  • multi/hyperspectral and LiDAR remote sensing
  • deep learning
  • artificial intelligence in machine vision
  • multimodal feature fusion
  • heterogeneous feature fusion
  • space-intelligent information processing

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • Reprint: MDPI Books provides the opportunity to republish successful Special Issues in book format, both online and in print.

Further information on MDPI's Special Issue policies can be found here.

Published Papers (1 paper)

Order results
Result details
Select all
Export citation of selected articles as:

Research

25 pages, 6462 KB  
Article
YOLO-CMFM: A Visible-SAR Multimodal Object Detection Method Based on Edge-Guided and Gated Cross-Attention Fusion
by Xuyang Zhao, Lijun Zhao, Keli Shi, Ruotian Ren and Zheng Zhang
Remote Sens. 2026, 18(1), 136; https://doi.org/10.3390/rs18010136 - 31 Dec 2025
Viewed by 865
Abstract
To address the challenges of cross-modal feature misalignment and ineffective information fusion caused by the inherent differences in imaging mechanisms, noise statistics, and semantic representations between visible and synthetic aperture radar (SAR) imagery, this paper proposes a multimodal remote sensing object detection method, [...] Read more.
To address the challenges of cross-modal feature misalignment and ineffective information fusion caused by the inherent differences in imaging mechanisms, noise statistics, and semantic representations between visible and synthetic aperture radar (SAR) imagery, this paper proposes a multimodal remote sensing object detection method, namely YOLO-CMFM. Built upon the Ultralytics YOLOv11 framework, the proposed approach designs a Cross-Modal Fusion Module (CMFM) that systematically enhances detection accuracy and robustness from the perspectives of modality alignment, feature interaction, and adaptive fusion. Specifically, (1) a Learnable Edge-Guided Attention (LEGA) module is constructed, which leverages a learnable Gaussian saliency prior to achieve edge-oriented cross-modal alignment, effectively mitigating edge-structure mismatches across modalities; (2) a Bidirectional Cross-Attention (BCA) module is developed to enable deep semantic interaction and global contextual aggregation; (3) a Context-Guided Gating (CGG) module is designed to dynamically generate complementary weights based on multimodal source features and global contextual information, thereby achieving adaptive fusion across modalities. Extensive experiments conducted on the OGSOD 1.0 dataset demonstrate that the proposed YOLO-CMFM achieves an mAP@50 of 96.2% and an mAP@50:95 of 75.1%. While maintaining competitive performance comparable to mainstream approaches at lower IoU thresholds, the proposed method significantly outperforms existing counterparts at high IoU thresholds, highlighting its superior capability in precise object localization. Also, the experimental results on the OSPRC dataset demonstrate that the proposed method can consistently achieve stable gains under different kinds of imaging conditions, including diverse SAR polarizations, spatial resolutions, and cloud occlusion conditions. Moreover, the CMFM can be flexibly integrated into different detection frameworks, which further validates its strong generalization and transferability in multimodal remote sensing object detection tasks. Full article
(This article belongs to the Special Issue Intelligent Processing of Multimodal Remote Sensing Data)
Show Figures

Figure 1

Back to TopTop