Next Article in Journal
Impact of Climate Change on the Climatic Suitability of Oilseed Rape (Brassica napus L.) Planting in Jiangsu Province, China
Previous Article in Journal
Estimation of Wheat Leaf Water Content Based on UAV Hyper-Spectral Remote Sensing and Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

LiDAR-IMU Sensor Fusion-Based SLAM for Enhanced Autonomous Navigation in Orchards

1
Interdisciplinary Program in Smart Agriculture, College of Agriculture and Life Sciences, Kangwon National University, Chuncheon 24341, Republic of Korea
2
Department of Biosystems Engineering, College of Agriculture and Life Sciences, Kangwon National University, Chuncheon 24341, Republic of Korea
3
Horticultural Research Division, Gangwon Agricultural Research & Extension Services, Chuncheon 24203, Republic of Korea
*
Author to whom correspondence should be addressed.
Agriculture 2025, 15(17), 1899; https://doi.org/10.3390/agriculture15171899 (registering DOI)
Submission received: 11 August 2025 / Revised: 2 September 2025 / Accepted: 5 September 2025 / Published: 7 September 2025
(This article belongs to the Special Issue Advances in Precision Agriculture in Orchard)

Abstract

Labor shortages and uneven terrain in orchards present significant challenges to autonomous navigation. This study proposes a navigation system that integrates Light Detection and Ranging (LiDAR) and Inertial Measurement Unit (IMU) data to enhance localization accuracy and map stability through Simultaneous Localization and Mapping (SLAM). To minimize distortions in LiDAR scans caused by ground irregularities, real-time tilt correction was implemented based on IMU feedback. Furthermore, the path planning module was improved by modifying the Rapidly-Exploring Random Tree (RRT) algorithm. The enhanced RRT generated smoother and more efficient trajectories with quantifiable improvements: the average shortest path length was 2.26 m, compared to 2.59 m with conventional RRT and 2.71 m with A* algorithm. Tracking performance also improved, achieving a root mean square error of 0.890 m and a maximum lateral deviation of 0.423 m. In addition, yaw stability was strengthened, as heading fluctuations decreased by approximately 7% relative to the standard RRT. Field results validated the robustness and adaptability of the proposed system under real-world agricultural conditions. These findings highlight the potential of LiDAR–IMU sensor fusion and optimized path planning to enable scalable and reliable autonomous navigation for precision agriculture.
Keywords: autonomous navigation; LiDAR–IMU; sensor fusion; SLAM; path planning autonomous navigation; LiDAR–IMU; sensor fusion; SLAM; path planning

Share and Cite

MDPI and ACS Style

Choi, S.; Han, X.; Chang, E.; Jeong, H. LiDAR-IMU Sensor Fusion-Based SLAM for Enhanced Autonomous Navigation in Orchards. Agriculture 2025, 15, 1899. https://doi.org/10.3390/agriculture15171899

AMA Style

Choi S, Han X, Chang E, Jeong H. LiDAR-IMU Sensor Fusion-Based SLAM for Enhanced Autonomous Navigation in Orchards. Agriculture. 2025; 15(17):1899. https://doi.org/10.3390/agriculture15171899

Chicago/Turabian Style

Choi, Seulgi, Xiongzhe Han, Eunha Chang, and Haetnim Jeong. 2025. "LiDAR-IMU Sensor Fusion-Based SLAM for Enhanced Autonomous Navigation in Orchards" Agriculture 15, no. 17: 1899. https://doi.org/10.3390/agriculture15171899

APA Style

Choi, S., Han, X., Chang, E., & Jeong, H. (2025). LiDAR-IMU Sensor Fusion-Based SLAM for Enhanced Autonomous Navigation in Orchards. Agriculture, 15(17), 1899. https://doi.org/10.3390/agriculture15171899

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop