This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Open AccessArticle
L2: Accurate Forestry Time-Series Completion and Growth Factor Inference
1
School of Information Science and Technology, Beijing Forestry University (BFU), Beijing 100083, China
2
Engineering Research Center for Forestry-Oriented Intelligent Information Processing of National Forestry and Grassland Administration, Beijing Forestry University (BFU), Beijing 100083, China
3
Ministry of Education Key Laboratory of Silviculture and Conservation, Beijing Forestry University (BFU), Beijing 100083, China
4
State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
5
School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 101408, China
*
Author to whom correspondence should be addressed.
Forests 2025, 16(6), 895; https://doi.org/10.3390/f16060895 (registering DOI)
Submission received: 17 March 2025
/
Revised: 29 April 2025
/
Accepted: 22 May 2025
/
Published: 26 May 2025
Abstract
In forestry data management and analysis, data integrity and analytical accuracy are of critical importance. However, existing techniques face a dual challenge: first, sensor failures, data transmission interruptions, and human errors lead to the prevalence of missing data in forestry datasets; second, the multidimensional heterogeneity and environmental complexity of forestry systems not only increase the difficulty of missing value estimation, but also significantly affect the accuracy of resolving the potential correlations among data. In order to solve the above problems, we proposed the model using the aspen woodland as the experimental object. The model consists of a complementary model and a predictive model. The complementary model integrates low tensor tensor kernel norm minimisation (LRTC-TNN) to capture global consistency and local trends, and combines long and short-term memory and convolutional neural network (LSTM-CNN) to extract temporal and spatial features, which is effective in accurately reconstructing the missing values in forestry time-series data. We also optimised the LRTC-TNN model to handle multi-class data and incorporated a self-attention mechanism into the LSTM-CNN framework to improve performance in the case of complex missing data. The prediction model adopts a dual attention mechanism (temporal attention mechanism and feature attention mechanism) based on LSTM to construct a stem diameter prediction model, which achieves high-precision prediction of stem diameter variation. Then we further analyzed the effects of various factors on stem diameter using SHAP (Shapley Additive Explanations).Experimental results demonstrate that our significantly improves data completion accuracy while preserving the original structure and key characteristics of the data. Moreover, it enables a more precise analysis of the factors affecting stem diameter, providing a robust foundation for advanced forestry data analysis and informed decision making.
Share and Cite
MDPI and ACS Style
Jiang, L.; Yang, M.; Xi, B.; Meng, W.; Duan, J.
L2: Accurate Forestry Time-Series Completion and Growth Factor Inference. Forests 2025, 16, 895.
https://doi.org/10.3390/f16060895
AMA Style
Jiang L, Yang M, Xi B, Meng W, Duan J.
L2: Accurate Forestry Time-Series Completion and Growth Factor Inference. Forests. 2025; 16(6):895.
https://doi.org/10.3390/f16060895
Chicago/Turabian Style
Jiang, Linlu, Meng Yang, Benye Xi, Weiliang Meng, and Jie Duan.
2025. "L2: Accurate Forestry Time-Series Completion and Growth Factor Inference" Forests 16, no. 6: 895.
https://doi.org/10.3390/f16060895
APA Style
Jiang, L., Yang, M., Xi, B., Meng, W., & Duan, J.
(2025). L2: Accurate Forestry Time-Series Completion and Growth Factor Inference. Forests, 16(6), 895.
https://doi.org/10.3390/f16060895
Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details
here.
Article Metrics
Article Access Statistics
For more information on the journal statistics, click
here.
Multiple requests from the same IP address are counted as one view.