remotesensing-logo

Journal Browser

Journal Browser

Photogrammetry Meets AI

A special issue of Remote Sensing (ISSN 2072-4292). This special issue belongs to the section "AI Remote Sensing".

Deadline for manuscript submissions: 31 March 2025 | Viewed by 31815

Special Issue Editors


E-Mail Website
Guest Editor
3D Optical Metrology (3DOM) Unit, Bruno Kessler Foundation (FBK), 38123 Trento, Italy
Interests: geomatics; mapping; UAV
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
1. Department of Civil, Environmental and Geodetic Engineering, The Ohio State University, Columbus, OH 43210, USA
2. Department of Electrical and Computer Engineering, The Ohio State University, Columbus, OH 43210, USA
Interests: data analytics; aerial/satellite photogrammetry; remote sensing; image processing; machine learning; 3D computer vision; 3D modeling/change detection; deformation analysis; unmanned aerial vehicles; image dense matching
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

For many years, photogrammetry has been the leading methodology to derive 3D metric and accurate information from imagery, at different scales (from satellite to aerial, terrestrial and under water) and from different sensors (linear, frame, panoramic). The inclusion of computer vision and robotics solutions has increased the level of automation in image processing and 3D data generation, leading to mainstream automatic solutions and massive 3D digitization processes. The recent advent of artificial intelligence methods based on machine and deep learning approaches is again changing the photogrammetric processes leading to unexpected automated solutions that can truly revolutionize the mapping and 3D documentation sector.

This Special Issue wants to focus on this recent change for 3D geometric tasks, and is seeking high-quality papers that explore all the potentialities offered by AI in photogrammetric problems. Papers should report progresses in supporting, integrating and boosting key areas of photogrammetry with AI-based methods. In particular, the following topics should be addressed in the proposed submissions:

  • Image matching and learning-based tie points extraction;
  • Outlier removal;
  • Structure from motion and bundle adjustment;
  • Camera project loss and calibration;
  • Simultaneous localization and mapping (SLAM) in the era of deep learning;
  • Monocular depth estimation;
  • Multi-view stereo (MVS) and dense point cloud generation with neural networks;
  • 3D representation and reconstruction with neural radiance field (NeRF);
  • Implicit methods for 3D representation from images and mesh reconstruction;
  • 3D fusion of heterogenous datasets;
  • Learning-based DSM inpainting;
  • Point clouds editing, cleaning and filtering;
  • Quantitative evaluations and analyses within applications.

Prof. Dr. Fabio Remondino
Dr. Rongjun Qin
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Remote Sensing is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2700 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • photogrammetry
  • machine/deep learning
  • structure from motion
  • 3D reconstruction
  • NeRF
  • data integration and fusion
  • quantitative analyses and comparisons

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Published Papers (6 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

Jump to: Other

20 pages, 11883 KiB  
Article
SIM-MultiDepth: Self-Supervised Indoor Monocular Multi-Frame Depth Estimation Based on Texture-Aware Masking
by Xiaotong Guo, Huijie Zhao, Shuwei Shao, Xudong Li, Baochang Zhang and Na Li
Remote Sens. 2024, 16(12), 2221; https://doi.org/10.3390/rs16122221 - 19 Jun 2024
Viewed by 855
Abstract
Self-supervised monocular depth estimation methods have become the focus of research since ground truth data are not required. Current single-image-based works only leverage appearance-based features, thus achieving a limited performance. Deep learning based multiview stereo works facilitate the research on multi-frame depth estimation [...] Read more.
Self-supervised monocular depth estimation methods have become the focus of research since ground truth data are not required. Current single-image-based works only leverage appearance-based features, thus achieving a limited performance. Deep learning based multiview stereo works facilitate the research on multi-frame depth estimation methods. Some multi-frame methods build cost volumes and take multiple frames as inputs at the time of test to fully utilize geometric cues between adjacent frames. Nevertheless, low-textured regions, which are dominant in indoor scenes, tend to cause unreliable depth hypotheses in the cost volume. Few self-supervised multi-frame methods have been used to conduct research on the issue of low-texture areas in indoor scenes. To handle this issue, we propose SIM-MultiDepth, a self-supervised indoor monocular multi-frame depth estimation framework. A self-supervised single-frame depth estimation network is introduced to learn the relative poses and supervise the multi-frame depth learning. A texture-aware depth consistency loss is designed considering the calculation of the patch-based photometric loss. Only the areas where multi-frame depth prediction is considered unreliable in low-texture regions are supervised by the single-frame network. This approach helps improve the depth estimation accuracy. The experimental results on the NYU Depth V2 dataset validate the effectiveness of SIM-MultiDepth. The zero-shot generalization studies on the 7-Scenes and Campus Indoor datasets aid in the analysis of the application characteristics of SIM-MultiDepth. Full article
(This article belongs to the Special Issue Photogrammetry Meets AI)
Show Figures

Figure 1

19 pages, 41835 KiB  
Article
Vision through Obstacles—3D Geometric Reconstruction and Evaluation of Neural Radiance Fields (NeRFs)
by Ivana Petrovska and Boris Jutzi
Remote Sens. 2024, 16(7), 1188; https://doi.org/10.3390/rs16071188 - 28 Mar 2024
Viewed by 1439
Abstract
In this contribution we evaluate the 3D geometry reconstructed by Neural Radiance Fields (NeRFs) of an object’s occluded parts behind obstacles through a point cloud comparison in 3D space against traditional Multi-View Stereo (MVS), addressing the accuracy and completeness. The key challenge lies [...] Read more.
In this contribution we evaluate the 3D geometry reconstructed by Neural Radiance Fields (NeRFs) of an object’s occluded parts behind obstacles through a point cloud comparison in 3D space against traditional Multi-View Stereo (MVS), addressing the accuracy and completeness. The key challenge lies in recovering the underlying geometry, completing the occluded parts of the object and investigating if NeRFs can compete against traditional MVS for scenarios where the latter falls short. In addition, we introduce a new “obSTaclE, occLusion and visibiLity constrAints” dataset named STELLA concerning transparent and non-transparent obstacles in real-world scenarios since there is no existing dataset dedicated to this problem setting to date. Considering that the density field represents the 3D geometry of NeRFs and is solely position-dependent, we propose an effective approach for extracting the geometry in the form of a point cloud. We voxelize the whole density field and apply a 3D density-gradient based Canny edge detection filter to better represent the object’s geometric features. The qualitative and quantitative results demonstrate NeRFs’ ability to capture geometric details of the occluded parts in all scenarios, thus outperforming in completeness, as our voxel-based point cloud extraction approach achieves point coverage up to 93%. However, MVS remains a more accurate image-based 3D reconstruction method, deviating from the ground truth 2.26 mm and 3.36 mm for each obstacle scenario respectively. Full article
(This article belongs to the Special Issue Photogrammetry Meets AI)
Show Figures

Figure 1

19 pages, 16005 KiB  
Article
Comparative Assessment of Neural Radiance Fields and Photogrammetry in Digital Heritage: Impact of Varying Image Conditions on 3D Reconstruction
by Valeria Croce, Dario Billi, Gabriella Caroti, Andrea Piemonte, Livio De Luca and Philippe Véron
Remote Sens. 2024, 16(2), 301; https://doi.org/10.3390/rs16020301 - 11 Jan 2024
Cited by 8 | Viewed by 3332
Abstract
This paper conducts a comparative evaluation between Neural Radiance Fields (NeRF) and photogrammetry for 3D reconstruction in the cultural heritage domain. Focusing on three case studies, of which the Terpsichore statue serves as a pilot case, the research assesses the quality, consistency, and [...] Read more.
This paper conducts a comparative evaluation between Neural Radiance Fields (NeRF) and photogrammetry for 3D reconstruction in the cultural heritage domain. Focusing on three case studies, of which the Terpsichore statue serves as a pilot case, the research assesses the quality, consistency, and efficiency of both methods. The results indicate that, under conditions of reduced input data or lower resolution, NeRF outperforms photogrammetry in preserving completeness and material description for the same set of input images (with known camera poses). The study recommends NeRF for scenarios requiring extensive area mapping with limited images, particularly in emergency situations. Despite NeRF’s developmental stage compared to photogrammetry, the findings demonstrate higher potential for describing material characteristics and rendering homogeneous textures with enhanced visual fidelity and accuracy; however, NeRF seems more prone to noise effects. The paper advocates for the future integration of NeRF with photogrammetry to address respective limitations, offering more comprehensive representation for cultural heritage preservation tasks. Future developments include extending applications to planar surfaces and exploring NeRF in virtual and augmented reality, as well as studying NeRF evolution in line with emerging trends in semantic segmentation and in-the-wild scene reconstruction. Full article
(This article belongs to the Special Issue Photogrammetry Meets AI)
Show Figures

Figure 1

23 pages, 8734 KiB  
Article
Motorcycle Detection and Collision Warning Using Monocular Images from a Vehicle
by Zahra Badamchi Shabestari, Ali Hosseininaveh and Fabio Remondino
Remote Sens. 2023, 15(23), 5548; https://doi.org/10.3390/rs15235548 - 28 Nov 2023
Cited by 3 | Viewed by 2523
Abstract
Motorcycle detection and collision warning are essential features in advanced driver assistance systems (ADAS) to ensure road safety, especially in emergency situations. However, detecting motorcycles from videos captured from a car is challenging due to the varying shapes and appearances of motorcycles. In [...] Read more.
Motorcycle detection and collision warning are essential features in advanced driver assistance systems (ADAS) to ensure road safety, especially in emergency situations. However, detecting motorcycles from videos captured from a car is challenging due to the varying shapes and appearances of motorcycles. In this paper, we propose an integrated and innovative remote sensing and artificial intelligence (AI) methodology for motorcycle detection and distance estimation based on visual data from a single camera installed in the back of a vehicle. Firstly, MD-TinyYOLOv4 is used for detecting motorcycles, refining the neural network through SPP (spatial pyramid pooling) feature extraction, Mish activation function, data augmentation techniques, and optimized anchor boxes for training. The proposed algorithm outperforms eight existing YOLO versions, achieving a precision of 81% at a speed of 240 fps. Secondly, a refined disparity map of each motorcycle’s bounding box is estimated by training a Monodepth2 with a bilateral filter for distance estimation. The proposed fusion model (motorcycle’s detection and distance from vehicle) is evaluated with depth stereo camera measurements, and the results show that 89% of warning scenes are correctly detected, with an alarm notification time of 0.022 s for each image. Outcomes indicate that the proposed integrated methodology provides an effective solution for ADAS, with promising results for real-world applications, and can be suitable for running on mobility services or embedded computing boards instead of the super expensive and powerful systems used in some high-tech unmanned vehicles. Full article
(This article belongs to the Special Issue Photogrammetry Meets AI)
Show Figures

Figure 1

22 pages, 5258 KiB  
Article
A Critical Analysis of NeRF-Based 3D Reconstruction
by Fabio Remondino, Ali Karami, Ziyang Yan, Gabriele Mazzacca, Simone Rigon and Rongjun Qin
Remote Sens. 2023, 15(14), 3585; https://doi.org/10.3390/rs15143585 - 18 Jul 2023
Cited by 24 | Viewed by 16377
Abstract
This paper presents a critical analysis of image-based 3D reconstruction using neural radiance fields (NeRFs), with a focus on quantitative comparisons with respect to traditional photogrammetry. The aim is, therefore, to objectively evaluate the strengths and weaknesses of NeRFs and provide insights into [...] Read more.
This paper presents a critical analysis of image-based 3D reconstruction using neural radiance fields (NeRFs), with a focus on quantitative comparisons with respect to traditional photogrammetry. The aim is, therefore, to objectively evaluate the strengths and weaknesses of NeRFs and provide insights into their applicability to different real-life scenarios, from small objects to heritage and industrial scenes. After a comprehensive overview of photogrammetry and NeRF methods, highlighting their respective advantages and disadvantages, various NeRF methods are compared using diverse objects with varying sizes and surface characteristics, including texture-less, metallic, translucent, and transparent surfaces. We evaluated the quality of the resulting 3D reconstructions using multiple criteria, such as noise level, geometric accuracy, and the number of required images (i.e., image baselines). The results show that NeRFs exhibit superior performance over photogrammetry in terms of non-collaborative objects with texture-less, reflective, and refractive surfaces. Conversely, photogrammetry outperforms NeRFs in cases where the object’s surface possesses cooperative texture. Such complementarity should be further exploited in future works. Full article
(This article belongs to the Special Issue Photogrammetry Meets AI)
Show Figures

Graphical abstract

Other

Jump to: Research

18 pages, 27623 KiB  
Technical Note
DDL-MVS: Depth Discontinuity Learning for Multi-View Stereo Networks
by Nail Ibrahimli, Hugo Ledoux, Julian F. P. Kooij and Liangliang Nan
Remote Sens. 2023, 15(12), 2970; https://doi.org/10.3390/rs15122970 - 7 Jun 2023
Cited by 3 | Viewed by 3248
Abstract
We propose an enhancement module called depth discontinuity learning (DDL) for learning-based multi-view stereo (MVS) methods. Traditional methods are known for their accuracy but struggle with completeness. While recent learning-based methods have improved completeness at the cost of accuracy, our DDL approach aims [...] Read more.
We propose an enhancement module called depth discontinuity learning (DDL) for learning-based multi-view stereo (MVS) methods. Traditional methods are known for their accuracy but struggle with completeness. While recent learning-based methods have improved completeness at the cost of accuracy, our DDL approach aims to improve accuracy while retaining completeness in the reconstruction process. To achieve this, we introduce the joint estimation of depth and boundary maps, where the boundary maps are explicitly utilized for further refinement of the depth maps. We validate our idea by integrating it into an existing learning-based MVS pipeline where the reconstruction depends on high-quality depth map estimation. Extensive experiments on various datasets, namely DTU, ETH3D, “Tanks and Temples”, and BlendedMVS, show that our method improves reconstruction quality compared to our baseline, Patchmatchnet. Our ablation study demonstrates that incorporating the proposed DDL significantly reduces the depth map error, for instance, by more than 30% on the DTU dataset, and leads to improved depth map quality in both smooth and boundary regions. Additionally, our qualitative analysis has shown that the reconstructed point cloud exhibits enhanced quality without any significant compromise on completeness. Finally, the experiments reveal that our proposed model and strategies exhibit strong generalization capabilities across the various datasets. Full article
(This article belongs to the Special Issue Photogrammetry Meets AI)
Show Figures

Graphical abstract

Back to TopTop