You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

24 February 2023

Physical Structure Expression for Dense Point Clouds of Magnetic Levitation Image Data

,
,
and
1
Department of Traffic Information and Control Engineering, Tongji University, Shanghai 200070, China
2
Shanghai Key Laboratory of Rail Infrastructure Durability and System Safety, Tongji University, Shanghai 200070, China
*
Author to whom correspondence should be addressed.
This article belongs to the Topic 3D Computer Vision and Smart Building and City

Abstract

The research and development of an intelligent magnetic levitation transportation system has become an important research branch of the current intelligent transportation system (ITS), which can provide technical support for state-of-the-art fields such as intelligent magnetic levitation digital twin. First, we applied unmanned aerial vehicle oblique photography technology to acquire the magnetic levitation track image data and preprocessed them. Then, we extracted the image features and matched them based on the incremental structure from motion (SFM) algorithm, recovered the camera pose parameters of the image data and the 3D scene structure information of key points, and optimized the bundle adjustment to output 3D magnetic levitation sparse point clouds. Then, we applied multiview stereo (MVS) vision technology to estimate the depth map and normal map information. Finally, we extracted the output of the dense point clouds that can precisely express the physical structure of the magnetic levitation track, such as turnout, turning, linear structures, etc. By comparing the dense point clouds model with the traditional building information model, experiments verified that the magnetic levitation image 3D reconstruction system based on the incremental SFM and MVS algorithm has strong robustness and accuracy and can express a variety of physical structures of magnetic levitation track with high accuracy.

1. Introduction

The construction of a high-speed magnetic levitation transportation system carries the important mission of developing state-of-the-art (SOTA) key technologies. As a large-scale transportation infrastructure, magnetic levitation transportation faces higher technical requirements in the development of a new generation of information technology. In recent years, the magnetic levitation transportation system followed up the pace of SOTA technologies such as artificial intelligence, smart city, smart manufacturing, digital twin (DT), etc. and became a forward-looking science and technology that can change the traditional transportation industry, which is the key choice facing intelligent magnetic levitation transportation.
Nowadays, magnetic levitation track transportation [1] has increasingly become a promising spot of research focus all over the world. Numerous studies have been performed, and many research scholars have committed to the key technologies [2,3,4] and have maintained continuous growth, such as the guiding method [5] and the track structure [6,7] of the magnetic levitation track. The change and development of the new generation of information technology have made 3D reconstruction technology of image data a new research focus. The 3D reconstruction based on image data primarily includes monocular vision, binocular stereo vision, and multistereo vision. The research on 3D reconstruction of image data mainly includes the structure from motion (SFM) algorithm and the multiview stereo (MVS) vision technology [8,9,10,11]. The SFM algorithm can be divided into many aspects, such as incremental, global, and hybrid, etc. The SFM algorithm takes photos as input information, reduces the camera parameters, and outputs sparse point clouds based on the pixel-matching relationship between the photos. And the incremental SFM algorithm [12,13,14] can recover the camera position and the 3D structural information directly from the overlapping images, which has become a common solution for the current aerial triangulation of unmanned aerial vehicle (UAV) images.
The task of MVS is to reconstruct the real scene closely, usually using a calibrated multiview map, such as getting dense point clouds or voxel representation of the scene. Then it can generate a 3D geometry of the object, and finally, recover the surface of the scene. The research of MVS is mainly divided into three aspects, which are also the three methods of 3D scene representation, that is, voxel-based algorithms, surface slices to cover scene surfaces, and depth map recovery and fusion algorithms. With the rapid development of deep learning, the depth map recovery and fusion algorithms adopted in this paper can make the calculation results more accurate and faster.
In recent years, with the cross-integration of UAV technology and various disciplines, UAV aerial photography technology has developed rapidly. Due to its advantages of high resolution, low cost, and convenient operation [15], UAV photography technology has become an important photogrammetry technology. Current large-scale infrastructure photogrammetry mainly includes UAV, street view cars, mobile phones, and panoramic photography, installed on the rails of vehicles for scanning track photogrammetry methods, of which the most widely used, and the most efficient and convenient, is UAV photography technology. The magnetic levitation image data obtained by oblique photography by UAV, with high definition and large data volume, can provide an effective research basis for magnetic levitation 3D reconstruction. And the 3D reconstruction of magnetic levitation image data can express the physical structure characteristics of magnetic levitation track with high accuracy, providing an effective research basis for cutting-edge technologies such as magnetic levitation digital twin.
The greatest contribution of specific work in this paper is that we realized the effective expression process of 3D reconstruction of magnetic levitation track image data by applying SFM and MVS algorithms. These algorithms make use of the advantages of image data to accurately express the current physical structure of the magnetic levitation track, and the 3D reconstructed magnetic levitation model can provide more important work for further research of DT. Moreover, this paper evaluated the 3D reconstruction system of the magnetic levitation track from the physical structure perspective, and experiments finally prove that the system has strong robustness and accuracy. The remainder of this article is organized as follows: Section 2 reviews the related studies. Section 3 describes the steps and methods of the SFM and MVS algorithms that make magnetic levitation image data to generate dense point clouds. Section 4 analyzes the magnetic levitation scenario. Section 5 expresses and analyzes the experiments and results. Finally, Section 6 and Section 7 summarize all of the contributions of this paper and propose future research directions.

3. Magnetic Levitation Image to Generate Dense Point Clouds Method

The incremental SFM algorithm is one of the commonly researched algorithms in the field of multivision 3D reconstruction. It is capable of recovering camera pose parameters of image data and generating sparse point clouds of image data based on the scene geometric relationships between image data feature points. In this paper, we applied feature extraction and matching, triangulation, and bundle adjustment (BA) optimization from the incremental SFM algorithm.
Moreover, we calculated the depth information and normal vector between images from multiple views of image data to recover a 3D stereo model based on the MVS algorithm.

3.1. Feature Extraction and Matching

In this paper, two seed images are randomly selected as initialized images from magnetic levitation image data. As shown in Figure 1, these two seed images are random images taken from a UAV at different shooting angles, and the SIFT operator is applied to complete the feature point detection on the seed image data. The SIFT operator is an image local feature extraction algorithm that can find the precise location and principal direction of feature points or key points in different image scale spaces by ensuring rotation invariance and constructing a 128-dimensional SIFT feature vector descriptor to extract features.
Figure 1. Two randomly selected magnetic levitation image data as seed images.
Building the Gaussian difference pyramid first.
Number of pyramid groups as expressed in Equation (1):
Octave = [log2min(M,N)] − 2,
where M and N are the number of rows and columns of the original image.
Next, calculate the local maximum or minimum values from the pixel values around the pixel point detected and judge whether it is a key point according to a preset threshold. Then, the main direction of the key point is confirmed, so that the key point feature vector descriptor can be constructed.
Where the scale space radius is expressed in Equation (2):
r = 3 σ o c t × 2 × d + 1 + 1 2
where σoct is the intragroup scale of the octave group of key-point.
To match the feature, key point feature vector descriptors are applied to match the seed image feature. We adopt the K nearest neighbor algorithm, which is a commonly used supervised learning algorithm by judging the pixels within the nearest neighbor of a key point belonging to a certain class and by setting the highest threshold so as to connect the feature matching between the key points. In the K nearest neighbor algorithm, the matching problem between objects is avoided by calculating the distance between each object as a nonsimilarity indicator for each object, and this process is based on the Euclidean distance algorithm as Equation (3).
d ( x , y ) = k = 1 n ( x k y k ) 2 ,
where x, y are the coordinates of the pixels.
The final feature point matching between the two seed images is obtained as shown in Figure 2.
Figure 2. Seed image feature point matching.
Next, we applied the RANSAC algorithm to eliminate the matching error, and the final experiment is shown in Figure 3. The feature matching after applying the RANSAC algorithm can effectively eliminate feature matching error and improve the accuracy of 3D reconstruction.
Figure 3. RANSAC algorithm to eliminate seed image feature matching error.

3.2. Triangulation

The triangulation principle of the incremental SFM algorithm focuses on the 3D scene structure information by recovering the camera pose parameters of the image data with key points and gradually adding image data to this system, thus achieving 3D reconstruction of the image data to generate sparse point clouds.
First, calculating and decomposing the essential matrix E by the two seed image data, with their feature matching point pairs, thus can the camera poses of the two seed images be obtained and the 3D points can be generated by triangulation. Next, by adding images to this system continuously and optimizing BA, the 3D reconstruction of the incremental SFM can be achieved.

3.3. BA Optimization

BA optimization, as shown in Figure 4, enables nonlinearly optimization of both 3D point locations and camera parameters. After adding a new image, the local BA can only be optimized on the partial images that are most associated with that image. When adding a certain percentage of the model, global BA can be optimized, which can reduce the computational cost.
Figure 4. BA optimization decomposition.
The Levenberg-Marquardt method is used to solve the minimization reprojection error, as expressed in Equation (4).
E ( M , X ) = i = 1 m j = 1 n D ( x i j , M i X j ) 2 ,
where xij are the real point coordinates and MiXj are the reconstructed point coordinates.
From Table 1, we can see that the interval between local BA optimization and pose refinement is performed, and each local BA optimization of the SFM system will be iteratively adjusted, so as to change the observations subtly. There are also global BA optimizations with a larger interval, though it is not listed here due to the large amount of data.
Table 1. BA optimization and Pose refinement data process.

3.4. Dense Point Clouds Reconstruction

The reconstruction process of the dense point clouds uses the depth map fusion method in the MVS technique [32]. First, a point that can be projected onto multiple images is filtered from the sparse point clouds as a seed point, one of the corresponding image data at that point is selected as the seed image, and the distance from that seed point to the origin of the camera coordinate system of the seed image is recorded as the initial depth of the seed point pixel, and the direction from the seed point to the origin of the camera coordinate system is the initial normal vector. Next, the pixel coordinates around the seed point are projected into space and then gradually expanded to all images to perform nonlinear depth optimization for each seed point, so the depth and normal vector information of the optimized pixel points can be formed to estimate the depth map and the normal map. Finally, according to the depth of each pixel of the depth map, the depth pixel points can be projected into the 3D space using the inverse projection matrix of the camera to obtain 3D dense point clouds.

3.5. Method Advantages and Disadvantages

In this paper, we first preprocessed the image data of the magnetic levitation transportation test line acquired by UAV oblique photography, the preprocessed data is performed sequentially to extract the image feature, match the feature, and eliminate matching errors, etc. Next, the 3D scene structure information of camera pose parameters and key points of the image data is recovered and optimized by the BA to output the sparse point clouds of the physical structure of the magnetic levitation track, and these processes are based on an incremental SFM algorithm. Moreover, we applied the MVS algorithm to estimate the depth map and normal map of the magnetic levitation track from the sparse point clouds before and reconstructed the dense point clouds that can express the physical structure of the magnetic levitation track with high accuracy. The whole system is based on COLMAP version 3.7 open-source software to handle the 3D reconstruction algorithm of magnetic levitation image data. The magnetic levitation image to generate the dense point clouds method structure diagram in this paper is shown in Figure 5.
Figure 5. Structure of magnetic levitation image to generate dense point clouds method.
The incremental SFM algorithm has strong system robustness. It can extract feature points from image data information and filter mismatched feature points by the RANSAC algorithm in the feature matching process, so the accuracy and precision of 3D sparse point cloud scenes are high. BA optimization continuously optimizes the sparse point cloud’s 3D structure after adding a new image, local BA only is optimized on the partial images that are most associated with that image. When adding a certain percentage of the model, global BA can be optimized which can reduce the computational cost. However, when the number of images reaches a certain upper limit, each global BA will take much time. The MVS algorithm is suitable for large scenes and massive images and can obtain a number of dense point clouds, however, it much depends on the choice of neighborhood image groups.

4. Magnetic Levitation Scenario Analysis

In this paper, we collected image data from the entire 1.5 km magnetic levitation test line of Tongji University Jiading campus in Shanghai, as shown in Figure 6. Our photographic range is above the magnetic levitation line, following the trend of the magnetic levitation line, with a width of 600 m and a length of 2 km of arc-shaped airspace. There were a total of 120 hectares of photographic area photos and the aerial photograph was about 143.3 m from the ground. The specific route map is shown in Figure 7. The collected image data includes the magnetic levitation track grider, bridge decks and piers, turnout, linear line, and turning structure of the magnetic levitation test line, etc. Due to the UAV equipment and the meeting of precision requirements, it has obtained other environmental image data around the magnetic levitation test line.
Figure 6. Magnetic levitation transportation test line at Tongji University’s Jiading campus.
Figure 7. Oblique photography route plan of the magnetic levitation test line.
The turnout structure of the magnetic levitation track, as shown in Figure 8a, starts from the sixth pier at the exit of the magnetic levitation assembly train test line, and it deviates from the magnetic levitation track in about a 5° direction, and the turnout track has six concrete piers as its support. The turning structure of the magnetic levitation track, as shown in Figure 8b, has a total length of 433.44 m, including a variety of linear structures such as gentle curve, flat curve, vertical curve, etc. It is one of the important physical structures of the magnetic levitation traffic test line and the 3D modeling study of it has great research significance for the intelligent magnetic levitation transportation experiment. The linear structure of the magnetic levitation track, shown in Figure 8c, includes a variety of physical structures such as the magnetic levitation track decks and piers, and the magnetic induction coils, which is one of the basic physical structures of the magnetic levitation track.
Figure 8. UAV images of the physical structure of the magnetic levitation track grider: (a) Turnout structure; (b) Turning structure; (c) Linear structure.
We obtained magnetic levitation image data by operating the Mini-Scan quadcopter with WIC-61MP camera equipment, adopted the method of oblique photography through the surrounding flight, and shot in five directions of left view, right view, front view, rear view, and downward view. Among them, the four viewing angles of the front and rear, left and right are 45 degrees, and the downward viewing angle is vertically downward. The focal length of the camera is 40 mm fixed focus, and the sensor size is 35.7 × 23.8 mm. Through the above preparations, we acquired a total of 14,590 photos, each with 6000 × 4000 pixels, and each photo occupied about 10 M storage space. In order to reduce the running time cost of this study, and without changing the accuracy of the dense point clouds of the magnetic levitation track, we first preprocessed the image data, manually removed the image data that did not capture the magnetic levitation track (i.e., fully captured the other surrounding environment) from the oblique photography, so that every preprocessed image data can extract the structural features of the magnetic levitation track. In the end, we obtained 3242 images after preprocessing.
This paper applied COLMAP open source software to reconstruct the image data of magnetic levitation track in 3D, applying the main steps of feature extraction, feature matching, sparse point clouds reconstruction, and dense point clouds reconstruction of this software, based on incremental SFM algorithm and MVS technology, and, finally, we reconstructed the dense point clouds that can express the physical structure of magnetic levitation track with high accuracy, such as with magnetic levitation turnout structure, the linear structure, turning structure, etc.

5. Experiment and Results

In this study, we preprocessed the magnetic levitation track image data collected by UAV oblique photography and applied COLMAP to successively extract and match features and reconstruct sparse point clouds to finally generate dense point clouds that can express the physical structure characteristics of the magnetic levitation track with high accuracy.

5.1. Feature Extraction and Matching

In this paper, we applied preprocessed 3242 magnetic levitation track images to extract features and set and modify initialization parameters such as the maximum number of features, resolution, maximum threshold, etc. Each image was extracted to 8000–12,000 feature points on average. Then we put the feature extracted image data to match their feature, set and modified initialization of the maximum number of matches, maximum ratio, block size, and other parameters, and divided the system feature points into 65 × 65 blocks to compute respectively. Finally, we matched pairs of 407,258 points to complete the data key points research.

5.2. Sparse Point Clouds Reconstruction

The image data after feature extraction and matching are reconstructed according to the incremental SFM algorithm to recover the camera pose parameters and 3D scene structure information of key points of the image data and carry out sparse point clouds reconstruction. Finally, this system generated a total of 2,424,880 points to form the sparse point clouds, and the exported point clouds can visualize the overall effect of the 3D reconstruction of the magnetic levitation track image data, as shown in Figure 9. we can see point clouds of physical facilities directly, such as magnetic levitation transportation test lines, residential buildings, factories, etc. and other environmental elements point clouds, such as shrubs, rivers, natural grass playgrounds, bridges, etc.
Figure 9. Overall effect of sparse point clouds of magnetic levitation track: (a) The color of points are empty; (b) The color of points with RGB.
This section may be divided into subheadings. It should provide a concise and precise description of the experimental results, their interpretation, as well as the experimental conclusions that can be drawn.

5.3. Dense Point Clouds Reconstruction

Based on the calculated depth and normal vector information of these pictures’ pixels, we set and modify initialization of max and min pixels number, max depth, normal error, and other parameters, and set the first 20 images with priority as the reference image for seed images are automatically selected during dense reconstruction. Then, we estimated the depth map and normal map as shown in Figure 10.
Figure 10. The original image and depth image and normal image of one picture.
Finally, the depth pixel points can be projected into the 3D space using the inverse projection matrix of the camera to obtain 3D dense point clouds and matched 149 million dense point clouds as shown in Figure 11.
Figure 11. Overall effect of dense point clouds of magnetic levitation track.

5.4. Dense Point Clouds Effect of the Whole Magnetic Levitation Track

For the dense point clouds of the 3D reconstruction of the magnetic levitation track grider, we performed point clouds data preprocessing manually by data segmentation and other operations to extract dense point clouds that can express the physical structure of the magnetic levitation track grider with high accuracy. The dense point clouds after data preprocessing are mainly divided into three kinds according to the physical structure of the magnetic levitation track grider: turnout structure, turning structure, and linear structure.
The dense point clouds of the turnout structure are shown in Figure 12. We can intuitively know that the turnout grider here is made of cement, which belongs to a typical single-type turnout, and the conversion grider in front of the turnout is made of steel material, which has a larger number of point clouds.
Figure 12. Illustration of the dense point clouds of the magnetic floating turnout structure.
Figure 13 shows the dense point clouds of the turning structure for a magnetic levitation flat curve radius of 400 m.
Figure 13. Illustration of the dense point clouds of the magnetic floating turning structure.
As shown in Figure 14, the 3D reconstructed dense point clouds can visually and clearly represent the physical information of the linear structure.
Figure 14. Illustration of the dense point clouds of the magnetic floating linear structure.

5.5. Dense Point Clouds Effect of the Details of Magnetic Levitation Track

We made a simulation model of a magnetic levitation track through BIM technology. This simulation model is based on the basic data of the magnetic levitation test line of Tongji University Jiading construction, such as track width, straight line length, track grider length, circular curve radius, and transition curve parameters, etc. Then, we mapped the basic model according to the image data and we relied on the material of the track grider.
For cement material and steel material, so as to build a BIM model that can restore the magnetic levitation track on the surface, the whole process is applied with professional software, such as AutoCAD, 3DSMAX, Context Capture, etc. This BIM model can be compared with our dense point clouds model. Figure 15 shows the local detail comparison of the magnetic levitation turnout structure.
Figure 15. Local details of turnout structure: (a) dense point clouds model; (b) BIM model; (c) standard cross-sectional turnout structure diagram.
As shown in Figure 14, it can be intuitively expressed that the difference between the traditional BIM model constructed by static drawings and the dense point clouds model reconstructed by applying image data to 3D. First of all, the BIM model reflects the data of the track surface at the beginning of engineering construction, however, the actual construction will have systematic errors. These data cannot be presented from BIM models, while the dense point clouds model is formed by image data, which is scanning the current track entity and can show physical differences. This magnetic levitation track test line has been in operation for more than two decades, which could not be estimated by the BIM model due to the changes caused by the track vibration test, physical friction, sensor aging, and acid rain corrosion, though the dense point clouds model has the current collected data, which can provide more economical and convenient feasible measures for aging sensors maintenance, and can also provide better real data for magnetic levitation DT applications. Therefore, the dense point clouds model has more objective research value than the BIM model from the perspectives of economy, application value, and future development prospects.
In order to verify the 3D reconstruction effect of the dense point clouds model, we compared the track width, track length, building height, turnout length, etc., as shown in Table 2 with the basic data of the BIM model, which is hardly affected by time.
Table 2. Comparison of data accuracy between the BIM model and dense point clouds model.
Table 2 shows that compared with the basic data of the BIM model, the accuracy rate of the dense point clouds model is higher than 97%. The error is controlled in the millimeter range, and the overall effect of the 3D reconstruction is very good, so it can be proved that the 3D reconstruction model of the image data based on the SFM and MVS algorithms has good robustness.

6. Discussion

In the field of large-scale transportation infrastructure and magnetic levitation transportation, we applied SFM and MVS algorithms to realize the 3D reconstruction process of image data collected by UAV oblique photography technology, which integrates engineering, computer science, applied physics, and other disciplines. In this paper, it is an integrated innovation between the technology and application levels. In addition, we overcame the application limitations of large data on the algorithm and can ensure the accuracy of the model and the robustness of the system. However, we also faced some difficulties with long calculation times. There was a dilemma with a long global BA optimization time and we hope that could be solved in follow-up experiments.
In general, our experiment can expand new fields for current ITS technology, break the traditional restrictions on the maintenance and overhaul of magnetic levitation large-scale traffic, provide basic research for magnetic levitation digital twin technology, and provide application value for smart cities.

7. Conclusions

In this paper, we extracted magnetic levitation image feature information and feature matching based on an incremental SFM algorithm, recovered camera pose parameters of image data and 3D scene structure information of key points, and optimized BA to generate sparse point clouds data, and then calculated and optimized depth map and normal map by MVS algorithm, and output a magnetic levitation track that can accurately express its physical structure information such as turnout, turning and linear structures. Finally, we compared the basic data of the BIM model with the modeling data of the dense point clouds model, and proved that the accuracy of the dense point clouds model is higher than 97%. Therefore, the model we reconstructed has a better expression effect, the system has strong robustness, and it can better show the real situation of the current magnetic levitation transportation infrastructure.
  • We reconstructed high-precision sparse point clouds of magnetic levitation track based on incremental SFM algorithm with COLMAP version 3.7 software;
  • We calculated dense point clouds based on the MVS algorithm that is able to express the physical structure of the magnetic levitation track, such as turnout, turning, and linear structures, with high accuracy;
  • The magnetic levitation dense point clouds model in this paper is easy to integrate with other 3D reconstruction models for further research, such as the laser point clouds model.
The research of intelligent magnetic levitation image 3D reconstruction can provide a significant scientific basis for the further development of ITS industry technologies such as magnetic levitation digital twin, smart city, etc.

Author Contributions

Conceptualization, Y.Z. and L.Z.; Data curation, Y.Z.; Formal analysis, Y.Z. and G.S.; Funding acquisition, L.Z.; Methodology, Y.Z.; Project administration, L.Z.; Resources, L.Z.; Software, Y.Z. and G.S.; Supervision, L.Z. and Q.X.; Validation, Y.Z. and L.Z.; Visualization, Y.Z.; Writing—original draft, Y.Z.; Writing—review & editing, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was sponsored by the Shanghai Science and Technology Innovation Action Program (No. 20511106400) and partly the by Shanghai Collaborative Innovation Research Center for Multi-network & Multi-modal Rail Transit.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Due to privacy restrictions, we cannot provide information on the use of data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xiong, J.; Deng, Z. Research progress of high-speed maglev rail transportation. J. Transp. Eng. 2021, 21, 177–198. [Google Scholar]
  2. Sotelo, G.G.; Oliveira, R.D.; Costa, F.S.; Dias, D.H.N.; de Andrade, R.; Stephan, R.M. A Full Scale Superconducting Magnetic Levitation (MagLev) Vehicle Operational Line. IEEE Trans. Appl. Supercond. 2015, 25, 1–5. [Google Scholar] [CrossRef]
  3. Han, B. Research on Key Technologies of Wireless Communication System for Vacuum Tube High-Speed Trains. Ph.D. Thesis, Beijing Jiaotong University, Beijing, China, 2021. [Google Scholar]
  4. Okano, M.; Iwamoto, T.; Furuse, M.; Fuchino, S.; Ishii, I. Running Performance of a Pinning-Type Superconducting Magnetic Levitation Guide. J. Phys. Conf. 2006, 43, 999–1002. [Google Scholar] [CrossRef]
  5. Tang, W.B.; Xiao, L.Y.; Wang, S.; Zhang, J.Y. A review of research on maglev guidance methods in maglev rail transportation. New Technol. Electr. Power 2022, 41, 45–60. [Google Scholar]
  6. Cai, W.F.; Zhang, W.F.; Feng, Y.; Hu, B.R. Study on vibration response and transmission characteristics of magnetic levitation track structure under speed increase. Railw. Stand. Meter 2022, 1–7. [Google Scholar]
  7. Cai, W.T.; Wang, C.J.; Teng, N.T.; Wen, Q. Cross-platform coupled vibration analysis of ultra-high-speed magnetic levitation track beam system. J. Shanghai Jiaotong Univ. 2021, 55, 1228–1236. [Google Scholar]
  8. Yao, Y.; Luo, Z.; Li, S.; Shen, T.; Fang, T.; Quan, L. Recurrent Mvsnet for High-Resolution Multi-View Stereo Depth Inference. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5525–5534. [Google Scholar]
  9. Yang, J.; Mao, W.; Alvarez, J.M.; Liu, M. Cost Volume Pyramid Based Depth Inference for Multi-View Stereo. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 4877–4886. [Google Scholar]
  10. Yao, Y.; Luo, Z.; Li, S.; Zhang, J.; Ren, Y.; Zhou, L.; Fang, T.; Quan, L. Blendedmvs: A Large-Scale Dataset for Generalized Multi-View Stereo Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  11. Liao, J. Research on High Precision 3D Reconstruction of Complex Scenes based on Multi-View Photographs. Ph.D. Thesis, Wuhan University, Wuhan, Hubei, 2021. [Google Scholar]
  12. Jiang, S.; Chen, W.; Li, Q.Q.; Jiang, W.S. Research progress on incremental motion recovery structures for unmanned aircraft images. J. Wuhan Univ. Inf. Sci. Ed. 2022, 1–19. [Google Scholar]
  13. Schonberger, J.L.; Hardmeier, H.; Sattler, T.; Pollefeys, M. Comparative Evaluation of Hand-Crafted and Learned Local Features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1482–1491. [Google Scholar]
  14. Tian, Y.; Fan, B.; Wu, F. L2-Net: Deep Learning of Discriminative Patch Descriptor in Euclidean Space. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 661–669. [Google Scholar]
  15. Ding, B. Aerial Triangulation and Accuracy Analysis of 3D Model for Oblique Photography of Consumer UAV. Master’s Thesis, Beijing University of Civil Engineering and Architecture, Beijing, China, 2020. [Google Scholar] [CrossRef]
  16. Lv, N.; Wang, C.; Qiao, Y.; Zhang, Y. Dense Robust 3D Reconstruction and Measurement for 3D Printing Process Based on Vision. Appl. Sci. 2021, 11, 7961. [Google Scholar] [CrossRef]
  17. Li, L.; Yang, F.; Zhu, H.; Li, D.; Li, Y.; Tang, L. An improved RANSAC for 3D point cloud plane segmentation based on normal distribution transformation cells. Remote Sens. 2017, 9, 433. [Google Scholar] [CrossRef]
  18. Hafeez, J.; Lee, J.; Kwon, S.; Ha, S.; Hur, G.; Lee, S. Evaluating feature extraction methods with synthetic noise patterns for image-based modelling of texture-less objects. Remote Sens. 2020, 12, 3886. [Google Scholar] [CrossRef]
  19. Huang, Y.; Liu, Y.; Han, T.; Xu, S.; Fu, J. Low Illumination Soybean Plant Reconstruction and Trait Perception. Agriculture 2022, 12, 2067. [Google Scholar] [CrossRef]
  20. Leroy, V.; Franco, J.S.; Boyer, E. Shape Reconstruction Using Volume Sweeping and Learned Photoconsistency. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 781–796. [Google Scholar]
  21. Riegler, G.; Ulusoy, A.O.; Bischof, H.; Geiger, A. Octnetfusion: Learning Depth Fusion from Data. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; pp. 57–66. [Google Scholar]
  22. Xu, Y.; So, Y.; Woo, S. Plane Fitting in 3D Reconstruction to Preserve Smooth Homogeneous Surfaces. Sensors 2022, 22, 9391. [Google Scholar] [CrossRef] [PubMed]
  23. Li, M.; Zheng, D.; Zhang, Y.; Zhang, R.; Zhang, F. An Improved Patch-Based Multi-View Stereo Algorithm for Large Image Sets. J. Comput. Theor. Nanosci. 2016, 13, 5755–5760. [Google Scholar] [CrossRef]
  24. Petrie, G. Systematic oblique aerial photography using multiple digital cameras. Photogramm. Eng. Remote Sens. 2009, 75, 102–107. [Google Scholar]
  25. Liu, M. Research on 3D Reconstruction Technology of based on UAV Mapping. Master’s Thesis, Shanghai Institute of Technology, Shanghai, China, 2021. [Google Scholar] [CrossRef]
  26. Zhou, X.; Zhang, X. Individual Tree Parameters Estimation for Plantation Forests Based on UAV Oblique Photography. IEEE Access 2020, 8, 96184–96198. [Google Scholar] [CrossRef]
  27. Zhang, R.; Li, H.; Duan, K.; You, S.; Liu, K.; Wang, F.; Hu, Y. Automatic Detection of Earthquake-Damaged Buildings by Integrating UAV Oblique Photography and Infrared Thermal Imaging. Remote Sens. 2020, 12, 2621. [Google Scholar] [CrossRef]
  28. Schonberger, J.L.; Frahm, J.M. Structure-From-Motion Revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar]
  29. Schönberger, J.L.; Zheng, E.; Frahm, J.M.; Pollefeys, M. Pixelwise View Selection for Unstructured Multi-View Stereo. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 501–518. [Google Scholar]
  30. Grieves, M. Digital Twin: Manufacturing excellence through virtual factory replication. White Pap. 2014, 1, 1–7. [Google Scholar]
  31. Li, F.; Wang, H.L. Application of digital twin technology to rail transit automatic operation system. Jiangsu Sci. Technol. Inf. 2021, 38, 3. [Google Scholar]
  32. Furukawa, Y.; Ponce, J. Accurate, dense, and robust multiview stereopsis. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 1362–1376. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.