Next Article in Journal
Geospatial Sensing and Data-Driven Technologies in the Western Balkan 6 (Agro)Forestry Region: A Strategic Science–Technology–Policy Nexus Analysis
Previous Article in Journal
Tree Functional Identity Drives Soil Enzyme Stoichiometric Ratios and Microbial Nutrient Limitation Responses to Artificial Forest Conversion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Interpretable Approach to Wood Species Identification Based on Anatomical Features in Microscopic Images

School of Big Data and Intelligent Engineering, Southwest Forestry University, Kunming 650224, China
*
Author to whom correspondence should be addressed.
Current address: Yunnan Provincial Key Laboratory of Wood Adhesives and Glued Products, Southwest Forestry University, Kunming 650224, China.
Forests 2025, 16(8), 1328; https://doi.org/10.3390/f16081328
Submission received: 9 July 2025 / Revised: 12 August 2025 / Accepted: 13 August 2025 / Published: 15 August 2025
(This article belongs to the Section Wood Science and Forest Products)

Abstract

Wood recognition plays a vital role in the trade and conservation of rare wood species. However, the computer vision-based methods classify the wood species by the features that are not used within the framework of wood anatomy, leading to results that are not interpretable. This study proposes a novel wood recognition method that detects anatomical structures such as vessels, wood rays, and parenchyma in wood microscopic images. These structures are quantified and mapped to the International Association of Wood Anatomists (IAWA) features, which are then used for species classification. Experimental results on 32 wood species demonstrate the effectiveness of the approach, achieving an accuracy of 94.1%, precision of 92.6%, recall of 93.3%, and an F1-score of 92.7%. In addition to its recognition performance, the method may offer interpretable IAWA-based classification criteria in wood science. These findings suggest that the method could serve as an anatomically interpretable framework for wood species identification, contributing to the regulation of the rare timber trade and supporting the conservation of endangered tree species.

1. Introduction

Illegal logging is one of the most economically significant natural resource crimes worldwide, with approximately 15% to 30% of global timber being sourced through illicit channels each year [1]. The illegal wood trade has caused declines in tree populations, pushed some species to the brink of endangerment, and severely disrupted the balance of forest ecosystems [2,3]. To combat illegal logging, the international community has established a series of protective measures, including the Convention on International Trade in Endangered Species of Wild Fauna and Flora (CITES) [4], the European Union Timber Regulation (EUTR) [5] enacted in 2013, and the United States Lacey Act. The effective implementation of these measures relies heavily on reliable wood identification technologies.
Traditional manual wood identification is a complex process. At the macroscopic level, wood species can be identified by observing physical characteristics of the wood surface, such as color, texture, and luster. At the microscopic level, experts typically examine anatomical features in cross, radial, and tangential sections—such as vessels, wood rays, and parenchyma—to perform wood species identification. To standardize and guide this process, the International Association of Wood Anatomists (IAWA) has published systematic lists of microscopic wood identification features [6,7,8], which is well-established and widely recognized as an important reference for wood identification. However, identification based on anatomical observation remains a challenging task. Experts require specialized knowledge in wood anatomy, and quantitatively analyzing some anatomical features is time-consuming and inefficient, making this approach unsuitable for large-scale inspections that involve the analysis of numerous wood samples [9]. In recent years, computer vision-based classification methods have rapidly gained popularity across various fields. To address the low efficiency of traditional manual wood identification, researchers have explored computer vision approaches to automate the recognition process [10,11,12], an overview of representative studies on wood classification is provided in Table 1.
Machine learning and deep learning are two commonly used approaches in computer vision. Among them, machine learning-based wood species recognition methods typically treat wood image identification as a texture recognition task, classifying species by extracting texture features from images. For example, Martins [13] employed local binary pattern (LBP) texture features to classify 112 wood species, achieving an accuracy of 86%. Filho applied Gabor features to classify 41 Brazilian tree species, reaching an accuracy of 97.77% [14]. Rosa da Silva utilized local phase quantization (LPQ) texture features to classify 77 wood species, achieving an accuracy of 88% [15]. Several other studies have also demonstrated that machine learning methods can improve the efficiency of wood identification [16,17,18]. However, most machine learning approaches rely on manually selected feature types. Specifically, Martins, Filho, and Rosa da Silva extracted LBP, Gabor, and LPQ features respectively, which perform well on specific datasets but lack generalizability.
In recent years, deep learning has made significant advances in image recognition and has gradually become the main approach for wood species identification. Among deep learning architectures, convolutional neural networks (CNNs) are the most mature and widely used, owing to their superior capability in image feature extraction. CNNs have been extensively applied to wood image recognition tasks [19,20,21]. Unlike traditional machine learning methods that rely on shallow features such as texture or shape, CNNs are capable of learning abstract and hierarchical features from images, thereby more effectively capturing the subtle anatomical differences among species and significantly improving classification accuracy and generalization performance [22,23].
The above literature review indicates that computer vision-based methods for wood species identification have significantly improved efficiency while maintaining high accuracy. However, according to the IAWA feature lists, key classification features in wood anatomy are primarily related to anatomical structures such as vessels, rays, and parenchyma observed in wood section images [24]. Whether based on texture features in machine learning or high-dimensional abstract features in deep learning, the features utilized in those methods do not align with those employed by wood anatomy experts. As a result, the recognition outcomes often lack interpretability and credibility from a wood science perspective [25], which limits their applicability in professional wood identification contexts [26].
This study proposes an interpretable wood species recognition approach that detects anatomical structures in wood microscopic images, then measures these structures, and maps the measurements to IAWA features. Due to variations in shooting angles, cutting positions, and magnification levels, a single image typically contains limited IAWA features, making it insufficient for classification [27]. Therefore, our method classifies wood species using IAWA features from multiple anatomical images. Finally, a Bayesian fusion network is utilized for wood species classification. Different from traditional methods, our approach not only identifies wood species but also provides interpretable evidence in wood science. The contributions of this study include:
1.
An interpretable wood species classification method is proposed, which leverages IAWA features as the basis for species identification.
2.
Automatic quantification of IAWA features helps reduce the manual effort traditionally required by wood anatomists.
3.
The study demonstrates the importance of integrating multiple anatomical views, showing that feature fusion from different perspectives can improve classification performance.
Table 1. Summary of prior studies on wood classification, detailing the image types used, feature extraction methods, classification algorithms, dataset scales (in terms of number of images and species), and achieved recognition rates.
Table 1. Summary of prior studies on wood classification, detailing the image types used, feature extraction methods, classification algorithms, dataset scales (in terms of number of images and species), and achieved recognition rates.
ReferencesImage TypeFXMClassifiersImages/SpeciesRec. Rate (%)
Martins et al. [13]MicroLBPSVM2240/11298.6
Filho et al. [14]MicroGaborSVM2942/4197.77
Rosa da Silva et al. [15]MicroLPQ+LBPKNN1221/7788
Tou et al. [21]MacroCNN featuresResNet-5010,787/2591.8
Yang et al. [28]MicroLBP+GLCMBPNN+SVM3000/1997.54
Zhao et al. [29]MacroGLCMBPNN1000/590
Zamri et al. [30]MacroGLCMBPNN5200/5297.01
Yuliastuti et al. [31]MacroGaborMLP400/1095
Nasirzadeh et al. [32]MacroLBPKNN3700/3796.60
Yusof et al. [33]MacroGLCMMLBP3000/3095.44
Abbreviations: FXM—feature extraction methods; LBP—Local Binary Pattern; Gabor—Gabor Filter; LPQ—Local Phase Quantization; CNN—Convolutional Neural Network; GLCM—Gray-Level Co-occurrence Matrix; SVM—Support Vector Machine; BPNN—Backpropagation Neural Network; MLP—Multilayer Perceptron; KNN—k-Nearest Neighbors; MLBP—Multi-layer neural network based on Back Propagation.

2. Materials

The wood samples used in the study are acquired from the Inside Wood website [34] and the Wood Specimen Museum at Southwest Forestry University (SWFU). The data from Inside Wood website is used in accordance with Inside Wood’s copyright policy. These samples come from 32 wood species(see Table 2). Most of the species are of commercial importance in the global timber trade, and several are listed under the CITES appendices. In addition, the geographic origins of the majority of these species span three major regions—Africa, Asia, and the Americas.
The images are acquired from the cross section and the tangential section. The cross section image is acquired perpendicular to the growth rings, which contains vessels and parenchyma. The tangential section image is acquired parallel to the growth rings, which contains wood rays. The positions of cross section and tangential section in a wood log are as shown in Figure 1.
The workflow of wood microscopic image acquisition is shown in Figure 2. The microscopy imaging is conducted under consistent illumination conditions using brightfield reflected light with an LED light source to ensure clear anatomical structures. Prior to imaging, standard wood anatomical sample preparation procedures are followed, including specimen softening, slicing, staining, and Mounting. A mild safranin stain is applied to enhance the contrast of the anatomical structures. Overlapping acquisition regions are avoided during imaging to ensure non-redundant coverage. The microscope used for image acquisition is a Leica DM2000 LED, manufactured by Leica Microsystems in Wetzlar, Germany. equipped with objective lenses of 5×, 10×, and 20× magnifications. Depending on the size and clarity of the anatomical structures, images were captured at these three levels, corresponding to scale bars of 500 μm, 200 μm, and 100 μm, respectively.
Figure 3 presents representative images from the dataset, where subfigures (a)–(e) correspond to cross section images, and (f)–(i) correspond to tangential section images. The image has a resolution of 1205 × 1536 pixels and includes scale information, which provides a reference for converting pixel-based measurements into real-world units and is crucial for accurately measuring the size and quantity of anatomical structures.
A total of 3926 images are collected in this study, comprising 2104 cross section images and 1822 tangential section images, yielding a tangential-to-cross section ratio of approximately 1:1.15. For each of the 32 wood species, around 122 images are acquired to ensure class balance. All anatomical features in the images are annotated with polygonal masks using LabelMe software. The data are split into training and testing sets at a ratio of 8:2 according to section type within each wood species.

3. Methods

The proposed wood recognition method consists of three steps, as shown in Figure 4. First, anatomical feature detection models are employed to detect the contours of anatomical structures, such as vessels, wood rays, and parenchyma. These detection models were trained using the YOLOv8-seg [35] architecture. Second, a series of measurement methods are employed to quantify density, diameter, arrangement, etc., and map them to the IAWA feature codes according to the IAWA principle. Finally, a Bayesian network is adopted to classify wood species using the fused IAWA feature vector from multiple anatomical images.
The method outputs three parts, the first two parts are classification evidence, and the last is the classification result. The first output consists of annotated images, in which vessels, wood rays, and parenchyma are labeled with distinct colors. These color-labeled images serve as visual evidence to support the classification results and facilitate the interpretation of the detected IAWA features. The second output is a list of IAWA codes representing the anatomical features extracted from the images, providing interpretable classification evidence in wood science. The final output is the predicted wood species, derived from the detected IAWA features.

3.1. Anatomy Structure Detection

Anatomy structure detector (ASD) is a fine-tuned YOLOv8-seg model, which was trained on the wood section images dataset. The backbone of the ASD is a feature pyramid, which is used to extract the anatomical structure feature at different scales. The neck of ASD is a fusion module, which combines the features of each scale to enhance the structural feature. The heads of the ASD are used to predict the masks of anatomical structures, such as vessels, rays, and parenchyma. The top header is used to detect small targets, and the bottom is used to detect big targets.
The input of ASD is anatomy images, including cross section images and tangential section images, and the output of ASD is the label, position, and boundary of the anatomy structure. Assuming the image list is I m g = ( i 1 , i 2 , , i 3 ) , the output label is L, the output mask is M, and the output boundary is B, then the processing of the ASD can be represented as Equation (1).
L , B , M = ASD ( Img )
where B is the bounding box point list, M is a binary image matrix, and L is the type of anatomy structure.

3.2. Auto Measurement

IAWA features are important in the wood identification task. In our experiments, features related to the material and suitable for automatic computation were selected. The selected features are shown in Table 3, in which, the first column is the IAWA codes and the second column is the description of the IAWA codes. The measurement module calculates the features of each anatomical structure and maps the IAWA codes. The selected IAWA features contain 25 features related to vessel arrangement, vessel diameter, vessel density, parenchyma type, ray width, and ray density.
To determine vessel arrangement and grouping features (IAWA codes 6, 9, 10, and 11), a clustering-based approach is employed. As illustrated in Algorithm 1, the center points of the detected vessels are clustered using the DBSCAN algorithm [36]. For each cluster, a linear regression is performed on the center points to estimate the main orientation. The slope of the fitted line is used to identify the vessel arrangement: near-horizontal slopes indicate tangential arrangement (feature 6), whereas near-vertical slopes indicate radial arrangement (feature 10). Feature 11 is determined by evaluating whether the vessels in a cluster exhibit both tangential and radial dispersion, quantified by standard deviations in the horizontal and vertical directions. To assess feature 9, the method checks for overlapping vessels within each cluster. If over 90% of the vessels are not involved in any overlaps, the feature is considered present.
Algorithm 1 Determination of IAWA Features 6, 9, 10, and 11
Require:  
Vessel center points P = { p 1 , p 2 , , p n }
1:
Apply DBSCAN clustering on P, obtain clusters { C 1 , C 2 , , C k }
2:
for each cluster C i  do
3:
   if  | C i | 4  then
4:
    Fit a line L i to the center points in C i
5:
    Compute slope θ i of L i
6:
    if  | θ i | < θ t  then
7:
     Mark IAWA feature 6 | θ i | > θ r
8:
    else if | θ i | > θ r then
9:
     Mark IAWA feature 10
10:
  end if
11:
end if
12:
if  | C i | 3  then
13:
  Compute x s t d , y s t d of C i
14:
  if  τ 1 < x s t d / y s t d < τ 2  then
15:
   Mark IAWA feature 11
16:
  end if
17:
end if
18:
end for
19:
Count solitary vessels
20:
if solitary vessels/total vessels > = 90 %  then
21:
 Mark IAWA feature 9
22:
end if
IAWA features (40–43) relate to the vessel tangential diameter, which can be calculated by the vessel boundary box. Define the pixel width of the vessel d, the pixel width is converted to the real width by d s . The features (46–50) relate to the vessel density, which is calculated by Equation (2).
D = N A A = w h s
where N is the number of detected vessels, w is the image width, h is the image height, s is the scale of pixels to millimeters, and A represents the area of the image in millimeters squared.
The IAWA features 79, 83, and 85 are related to the shape of the parenchyma, which are derived from the ASD detection results. For parenchyma’s predicted label is ‘wing’, features 80 and 82 are classified based on the ratio of the bounding box height to its width.
The IAWA features 96–99 are related to wood rays and determined by the average of the ray width. According to the IAWA description, the width of a wood ray is defined as the number of cells at its widest part. As shown in Figure 5a, firstly, the rays detected by the model are converted into binary images. Next, a cell detection algorithm [37] is used to detect the cells. The middle part between 1 / 3 and 2 / 3 is used to compute the ray width. Equally dividing the middle part into multiple horizontal segments, and counting the ray cells per segment. The IAWA features 96–99 are calculated by the average width of segments.
The IAWA features 114–116 are related to the number of rays. The image is divided into n equal parts using horizontal lines, and the number of wood rays intersected by each horizontal line is counted. The principle is shown in Figure 5b, in which the anatomy image is divided by 5 lines. The average number of divided parts is used for calculating the rays per image width in pixels. Finally, using a scale to convert the rays per pixel to rays per millimeter.

3.3. Classification

Wood species classification is based on IAWA code ( I A C ) from both cross section and tangential section images. As illustrated in Figure 6, define the IAWA features of each image as A n . A union operation is used to integrate the IAWA features, which is shown in Equation (3). The presence and absence of the IAWA features are calculated using Equation (4). The result 1 means the feature is present, which is detected from the images. The result 0 means the feature is absent, which is not detected in images.
D = A 1 A 2 A n
f a ( x ) = 1 if   x D 0 if   x D
The IAC is calculated using Equation (5).
IAC = { x | f a ( x ) } , x { 6 , 9 , , 116 }
A Naive Bayes-based classifier was employed to classify the wood species. Define the x denotes the IAWA feature code matrix IAC , and Y denotes the wood species. The probability of x belonging to Y is calculated using Formula (6). Where the variable c represents the index of the wood species.
argmax c   P ( Y = c x ) = argmax c   P ( Y = c ) i ( x i Y = c ) c P ( Y = c ) i ( x i Y = c )

3.4. Evaluation

The performance of classification is evaluated by metrics accuracy, precision, recall, and F1-Score. Accuracy represents the overall correctness of a classification model, calculated as the ratio of correctly predicted samples (both true positives and true negatives) to the total number of samples. Precision is the proportion of true positives among all predicted positives. Recall is the proportion of true positives among the positives, indicating the model’s ability to detect all positive instances. F1-score is a balance of precision and recall. These metrics are computed using Equations (7)–(10).
Accuracy = TP + TF TP + FP + TN + FN
Precision = TP TP + FP
Recall = TP TP + FN
F1-Score = 2 × Precision × Recall Precision + Recall
where True Positives (TP) denotes the number of correctly predicted positives, False Positives (FP) represents the number of incorrectly predicted positives, True Negatives (TN) indicates the number of correctly predicted negatives, and False Negatives (FN) corresponds to the number of incorrectly predicted negatives.
The intersection over union (IoU) measures the overlap between the predicted and ground truth masks. It is defined as the ratio of their intersection to their union, as shown in Formula (11). The average precision (APx) is used to evaluate the detection performance of the anatomy structure under IoU threshold x, which is the mean of precisions calculated from images. The metric mAP is the mean of APx under different IoU thresholds, and is calculated using the Formula (12).
IoU = A pred A gt A pred A gt
where A pred is the predicted area and A gt is the ground truth area.
mAP = 1 m A P i ,   i ( 0.5 , 0.55 , 0.6 , , 0.95 ) ,   m = len ( P )
where P i denotes the precision calculated under the IoU threshold i.

4. Results

4.1. Classification Performance

The wood species classification performance of the proposed method is evaluated in the testing dataset, which consists of 32 species, and the results are shown in Table 4. Among all species, Rose wood, such as Dalbergia oliveri, Diospyros crassiflora, and Pterocarpus macrocarpus achieved perfect or near-perfect precision values (≥0.97), while maintaining high recall rates (>0.91), indicating it has good performance in these species. Although some species, such as Albizia ferruginea and Pericopsis elata have low F1-score, the method achieved a high F1-score on average. The average accuracy of all species reached 0.941, precision reached 0.926, the average recall reached 0.933, and the average F1-score reached 0.927. These indicate that our proposed method has high robustness in the testing dataset.

4.2. Interpretable Wood Recognition

This method not only predicts the wood species but also outputs image evidence and IAWA features for classification. The full outputs are shown in Figure 7, where three classification examples are presented. Among them, the first two species are correctly predicted, while the last one represents a misclassification case. In Output1, the anatomical structures are color masked with structure names, and different colors represent different structures. These colored masks serve as evidence of the classification in the context of wood anatomy. The cross section images provide vessel and parenchyma structures, and tangential section images provide wood ray features. The images illustrate the detection results for vessels, wood rays and parenchyma. Output2 represents the IAWA feature codes derived from statistical analysis of detected anatomical structures. It provides visible indicators to explain the IAWA features. Output3 is the classification result, which contains the wood species.

4.3. Anatomy Feature Detection

The model was trained on an NVIDIA RTX 4090 GPU, with hyperparameters summarized in Table 5. The training process, including the trends of loss and F1-score over epochs, is shown in Figure 8. Figure 9 presents the detected anatomies from the deep learning model. The first row illustrates the detection results for vessels. The vessels are colored with different colors and used to calculate the IAWA features. The second row exhibits the wood rays, which are covered by color polygons. The third row represents the parenchyma structure detection and classification. The parenchyma contains three types, including winged parenchyma (IAWA 80, 82), parenchyma confluent (IAWA 83), and banded parenchyma (IAWA 85). These images demonstrate that the proposed method highly detects the vessels, pharenchyma, and wood rays. The boundaries of the anatomical structure are accurately recognized using polygons, which provide precise data for calculating the IAWA features.
The number and density of vessels and rays are key parameters used in calculating IAWA features. To evaluate the accuracy of the proposed automated measurement method, 50 images were manually annotated to obtain reference counts of vessels and rays. The comparison between the results from the automated module and manual counting is presented in Figure 10. It can be clearly observed that the vessel and ray counts generated by the automated system are highly consistent with those obtained through manual annotation. The maximum and average measurement errors for vessel count are 7 and 1.16, respectively, while those for ray count are 5 and 1.6. Despite these levels of error, the measured IAWA features remain within the same categorical range, indicating that the measurement results still maintain a high degree of accuracy.

5. Discussion

5.1. Advantage of Multiple Outputs

Compared to traditional deep learning-based methods of wood species recognition, the proposed method enhances the interpretability of the recognition process in a certain field. Unlike deep learning models that only output a classification label, the proposed method outputs not only the wood species but also visually interpretable evidence that includes segmented anatomical structures and IAWA features. These evidences improve the credibility, transparency, and scientific relevance of the recognition results in wood science.
By combining computer vision techniques with established principles from wood science, the proposed method may help narrow the gap between computer science and botanical taxonomy. This interdisciplinary approach could contribute to the development of explainable AI systems tailored for analyzing wood anatomical structures and supporting automated wood identification. In addition, it may facilitate collaboration between data scientists and wood anatomists, offering the potential for more reproducible and interpretable recognition tools in wood science applications.

5.2. Performance of Different Models

Anatomical structure detection is crucial in this method, as it forms the foundation for both IAWA feature calculation and wood species classification. In the experiments, the detection performance of anatomy structures in YOLOv8-seg, CondInst [38], PointRend [39], and Mask R-CNN [40] models is evaluated to find the best structure detection model. The comparison is shown in Figure 11, in which, the first column displays the original image, and the other columns are detection results from image detection models. The first row is vessel images, the middle row is ray images, and the last row is parenchyma images. It is evident that the YOLOv8-seg model successfully detected nearly all the targets, whereas the other models missed many in the first row. In the detection results on other images, YOLOv8-seg consistently outperformed or matched other models. Overall, YOLOv8-seg demonstrates a stronger capability in detecting all target objects compared to other models, which is particularly important for the accurate calculation of IAWA features.
The anatomical feature detection capabilities of different models are detailed in Table 6. Notably, YOLOv8-seg achieves the highest vessel detection performance with an A P 50 score of 0.912, surpassing other models by approximately 0.2 points. Meanwhile, for the detection of wood rays and parenchyma, YOLOv8-seg reached the highest mAP of 0.716 and 0.572. Compared to other models, YOLOv8-seg demonstrates superior detection performance overall, which can be attributed to its efficient feature extraction and fusion mechanisms that enhance its ability to detect densely packed small targets.

5.3. Impact of Multiple Images

The IAWA features are regarded as the international standard for wood species identification. Due to a single microscopic image usually including partially anatomical characteristics, reliable wood identification typically requires multiple microscopic images to capture enough IAWA features. For finding the best combination of images, we evaluated the impact of cross section and tangential section images on classification performance metrics precision and recall, as illustrated in Figure 12. The combination of cross section and tangential section images yields significantly higher classification metrics compared to using either type of image alone. Specifically, the combined input (CS+TS) achieves both precision and recall about 80%. Furthermore, among the single-section inputs, cross-sectional images consistently outperform tangential section images in classification performance. Notably, as Figure 13 shows, the accuracy first peaks when the data includes 6 cross section images and 4 tangential section images. Considering efficiency, this combination is recommended, achieving an accuracy of 94%.
The above results can be attributed to the complementary nature of anatomical features captured by different section types. The combination of cross section and tangential section images provides a more comprehensive representation of wood anatomy. By integrating the two views, the model benefits from diverse features, leading to enhanced classification performance. When comparing single-view performance, cross section images outperform tangential section images. This outcome can be attributed to the more IAWA features present in the cross section. According to the IAWA feature table used in this study, a total of 18 features, accounting for 72% are related to the cross section. The limited features in the tangential section may restrict the model’s learning capacity and negatively affect classification performance. Overall, these findings highlight the importance of view integration in achieving reliable wood species classification.

5.4. Classification Methods

To evaluate the effectiveness of different machine learning algorithms for wood species classification based on IAWA features, we compared five commonly used classification methods: Support Vector Machine (SVM) [41], Naive Bayes, K-Nearest Neighbors (KNN) [42], decision tree [43], and Random Forest [44]. Their performance is summarized in Table 7. All five classification methods exhibited strong classification performance on the test dataset, with accuracy scores exceeding 0.90. Among them, the Naive Bayes model achieved the highest overall performance, with an accuracy of 94.1% on the test set. Notably, its accuracy on the training set was 94.5%, indicating a well-balanced performance between the training and test sets. In contrast, the Decision Tree model achieved an accuracy of 97.3% on the training set but dropped to 90.0% on the test set, suggesting a noticeable overfitting issue.
The performance of all five classifiers can be attributed to the two sections of images capturing a wide range of IAWA features, and these features exhibit high discriminability for wood species classification. Among the five classification methods, the Naive Bayes model achieved the best performance. The Naive Bayes model is based on a strong assumption that all features are conditionally independent. In this study, the anatomical features used for classification are clearly defined and relatively independent of each other. These features reduce the likelihood of strong correlations, making the Naive Bayes model especially suitable and effective for this task. Moreover, Naive Bayes exhibits inherent robustness in handling high-dimensional and sparse feature spaces, as well as missing values, since its probabilistic framework allows each feature to contribute independently to the final prediction [45]. Compared to more complex models like SVM or Random Forest, which may be more sensitive to feature interactions or require extensive hyperparameter tuning, Naive Bayes offers a simpler yet highly effective approach for classification. Its consistent performance on both training and testing datasets further suggests strong generalization capability under the given data distribution.
The importance and correlation of IAWA features with wood species were assessed using SHAP values and Pearson correlation coefficients [46] derived from a random forest model. The results were visualized in the form of a heat bubble chart shown in Figure 14. From a species-level perspective, features 42, 46, 47, 79, and 98 exhibit relatively high SHAP values and strong correlations in Swietenia macrophylla. From a feature-level perspective, features 50 and 114 generally display low SHAP values across most species, with only a few species exhibiting notable contributions from these features. Overall, among all IAWA features, those related to density and diameter—which are relatively easy to quantify—tend to have higher SHAP values and stronger correlations.

5.5. Limitations and Challenges

In our experiments, we observed that certain wood species exhibited relatively low classification accuracy. For example, as shown in the fourth classification result in Figure 7, the model misclassified Albizia ferruginea as Pericopsis elata. A comparison of the IAWA features of the two species revealed a high degree of similarity, with Albizia ferruginea possessing an additional characteristic—IAWA feature 98: “Larger rays commonly 4- to 10-seriate”—which the model failed to identify. Upon further inspection of the images, it was found that some samples suffered from low image quality, as illustrated in Figure 15. Although the model was able to detect the general outline of wood rays, after the binarization process, the boundaries of ray cells became indistinct or even disappeared, leading to a significant underestimation in cell count. In some extreme cases, such as when ray cells are overlapping or poorly contrasted, ray cells were difficult to distinguish even by human observation, making their identification and counting particularly challenging.
To address this issue, future work may consider incorporating measurable anatomical features, such as the actual width and height of wood rays, into the classification process—aligned with the IAWA standard, which often uses cell width as a discriminative feature. In addition, previous studies have shown that introducing a third viewing plane—the radial section—can reveal more diagnostic features and improve classification performance [27]. The radial section offers a new perspective of the wood sample, providing anatomical features that cannot be observed in the cross and tangential sections, enabling a more comprehensive and interpretable identification. Thus, integrating radial-section images into the current framework may help reduce uncertainty caused by the limitations of cross and tangential sections.
In this study, the model achieves an average inference time of approximately 80 ms per image, which is suitable for near real-time processing in laboratory conditions. However, this may pose challenges for deployment on resource-constrained devices. To improve efficiency and broaden applicability, future research could explore lightweight network architectures and model compression techniques to reduce computational requirements. Moreover, the current method still relies on high-quality images, particularly in terms of anatomical clarity under the microscope. The acquisition of microscopic images typically requires a certain level of technical expertise, including sample sectioning, staining, and microscopic imaging, which may limit its use in non-laboratory or field settings. As a future direction, we will explore the feasibility of using non-microscopic images, such as macroscopic section images, to extract IAWA features for species classification. This approach may help reduce the reliance on specialized equipment while maintaining the anatomical interpretability.

6. Conclusions

In this study, we proposed a novel wood recognition method that integrates a deep learning-based anatomical structure detector, quantitative feature extraction aligned with IAWA standards, and an interpretable Bayesian classifier for wood species identification. By incorporating both cross and tangential section images, the method achieved a classification performance with an accuracy of 94.1%. The results suggest that the method can achieve high classification accuracy while offering a degree of biological interpretability based on wood anatomical principles, which may help domain experts to understand and validate the model’s decisions.
Nevertheless, two main limitations of the proposed approach should be acknowledged. First, the method requires high-quality microscopic images, which rely on specialized equipment and precise sample preparation. This requirement may hinder its practical deployment in field environments or in contexts where such resources are unavailable. Second, although the model demonstrates near real-time performance under laboratory conditions, it may pose challenges for deployment on resource-constrained devices.
Future research will focus on reducing dependency on high-resolution microscopic inputs by exploring the use of macroscopic images and on optimizing the model for lightweight deployment. These directions are crucial for enhancing the practicality, scalability, and accessibility of AI-assisted wood identification systems in real-world applications.

Author Contributions

Methodology, Y.C.; Resources, J.Q.; Validation, Q.L.; Visualization, S.Q.; Writing—original draft, L.L.; Writing—review & editing, Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This document is the result of the research projects funded by the Agricultural Joint Project of Yunnan province (202301BD070001-241, 202501BD070001-105), the Science Research Project of Yunnan education department (2023J0699).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data from the InsideWood website is © InsideWood, North Carolina State University, all rights reserved. Data from the Wood Specimen Museum at Southwest Forestry University (SWFU) will be made available on request.

Acknowledgments

The authors gratefully acknowledge the following contributors for providing the images used in this study: Elisabeth Wheeler (North Carolina State University) for images from specimens BWCw 8556, BWCw 8332, BWCw 8618, BWCw 8516, PACw En.cyl, USw PR19, Hw 23557, BWCw 8058, FHOw 19992, and PACw 4673 (permission granted on 5 August 2025); Hans Beeckman (Royal Museum for Central Africa) for images from specimens Tw 17698, Tw 425, Tw 609, Tw 724, Tw 5225, Tw 6987, Tw 26885, Tw 52896, and Tw 5016 (permission granted on 12 August 2025); Peter Gasson (Royal Botanic Gardens, Kew) for images from specimens Kw Per.ela.Trade.1951, Kw Swi.mac, Kw 4353, Kw 70560, and FHOw 4291 (permission granted on 6 August 2025).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Smith, J.K.; Jones, L.M. Illegal logging and its impacts on global biodiversity. For. Policy Econ. 2019, 99, 12–19. [Google Scholar]
  2. Reboredo, F. Socio-economic, environmental, and governance impacts of illegal logging. Environ. Syst. Decis. 2013, 33, 295–304. [Google Scholar] [CrossRef]
  3. Bösch, M. Institutional quality, economic development and illegal logging: A quantitative cross-national analysis. Eur. J. For. Res. 2021, 140, 1049–1064. [Google Scholar] [CrossRef]
  4. United Nations. Convention on International Trade in Endangered Species of Wild Fauna and Flora. Appendices I, II and III. 2021. Available online: https://cites.org/eng/app/appendices.php (accessed on 20 October 2021).
  5. European Union Timber Regulation (EUTR). Regulation (EU) No 995/2010 of the European Parliament and of the Council. Available online: https://eur-lex.europa.eu/eli/reg/2010/995/oj (accessed on 20 October 2010).
  6. Wheeler, E.; Baas, P.; Gasson, P. IAWA List of Microcopie Features for Hardwood Identification. IAWA J./Int. Assoc. Wood Anat. 1989, 10, 219–332. Available online: https://www.researchgate.net/publication/294088872_IAWA_List_of_Microcopie_Features_for_Hardwood_Identification (accessed on 8 July 2025).
  7. Baas, P.; Blokhina, N.; Fujii, T.; Gasson, P.; Grosser, D.; Heinz, I.; Ilic, J.; Xiaomei, J.; Miller, R.; Newsom, L.; et al. IAWA List of Microcopie Features for Softwood Identification. IAWA J. 2004, 25, 1–70. Available online: https://www.researchgate.net/publication/286488911_IAWA_List_of_microscopic_features_for_softwood_identification (accessed on 8 July 2025).
  8. Angyalossy, V.; Evert, R.; Marcati, C.; Oskolski, A.; Terrazas, T.; Kotina, E.; Lens, F.; Mazzoni-Viveiros, C.; Angeles, G.; Machado, S.; et al. IAWA List of Microcopie Bark Features. IAWA J. 2016, 37, 517–615. [Google Scholar] [CrossRef]
  9. Silva, J.L.; Bordalo, R.; Pissarra, J.; de Palacios, P. Computer Vision-Based Wood Identification: A Review. Forests 2022, 13, 2041. [Google Scholar] [CrossRef]
  10. Tou, J.; Tou, P.; Lau, P.Y.; Tay, Y.H. Computer Vision-based Wood Recognition System. In Proceedings of the International Workshop on Advanced Image Technology, Bangkok, Thailand, 8–9 January 2007. [Google Scholar]
  11. Fabijańska, A.; Danek, M.; Barniak, J. Wood species automatic identification from wood core images with a residual convolutional neural network. Comput. Electron. Agric. 2021, 181, 105941. [Google Scholar] [CrossRef]
  12. Wu, F.; Gazo, R.; Haviarova, E.; Benes, B. Wood identification based on longitudinal section images by using deep learning. Wood Sci. Technol. 2021, 55, 553–563. [Google Scholar] [CrossRef]
  13. Martins, J.; Oliveira, L.S.; Nisgoski, S.; Sabourin, R. A database for automatic classification of forest species. Mach. Vis. Appl. 2013, 24, 567–578. [Google Scholar] [CrossRef]
  14. Filho, P.L.; Oliveira, L.S.; Nisgoski, S.; Britto, A.S. Forest species recognition using macroscopic images. Mach. Vis. Appl. 2014, 25, 1019–1031. [Google Scholar] [CrossRef]
  15. Rosa da Silva, N.; De Ridder, M.; Baetens, J.M.; Van den Bulcke, J.; Rousseau, M.; Martinez Bruno, O.; Beeckman, H.; Van Acker, J.; De Baets, B. Automated classification of wood transverse cross-section micro-imagery from 77 commercial Central-African timber species. Ann. For. Sci. 2017, 74, 30. [Google Scholar] [CrossRef]
  16. Barmpoutis, P.; Dimitropoulos, K.; Barboutis, I.; Grammalidis, N.; Lefakis, P. Wood species recognition through multidimensional texture analysis. Comput. Electron. Agric. 2018, 144, 241–248. [Google Scholar] [CrossRef]
  17. Wang, H.j.; Zhang, G.q.; Qi, H.n. Wood Recognition Using Image Texture Features. PLoS ONE 2013, 8, e76101. [Google Scholar] [CrossRef]
  18. Zhou, X.; Liu, Y.; Pan, S.; Luo, S. Wood texture detection based on optimized deep belief network model with multiple feature fusion. In Proceedings of the Fourth International Conference on Mechanical Engineering, Intelligent Manufacturing, and Automation Technology (MEMAT 2023), Guilin, China, 20–22 October 2023; Chen, W., Liu, X., Eds.; International Society for Optics and Photonics, SPIE: Bellingham, WA, USA, 2024; Volume 13082, p. 130823H. [Google Scholar] [CrossRef]
  19. Nguyen, T.K. Evaluation of Wood Species Identification Using CNN-Based Networks at Different Magnification Levels. Int. J. Adv. Comput. Sci. Appl. 2023, 14. [Google Scholar] [CrossRef]
  20. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  21. He, T.; Liu, S.; Lu, Y.; Zhang, Y.; Jiao, L.; Yin, Y. iWood: An Automated Wood Identification System for Endangered and Precious Tree Species Using Convolutional Neural Networks. Sci. Silvae Sin. 2021, 57, 152–159. [Google Scholar] [CrossRef]
  22. Zheng, Z.; Ge, Z.; Tian, Z.; Yang, X.; Zhou, Y. WoodGLNet: A multi-scale network integrating global and local information for real-time classification of wood images. J. Real-Time Image Process. 2024, 21, 147. [Google Scholar] [CrossRef]
  23. Hwang, S.W.; Sugiyama, J. Computer vision-based wood identification and its expansion and contribution potentials in wood science: A review. Plant Methods 2021, 17, 47. [Google Scholar] [CrossRef]
  24. Wheeler, E.A.; Baas, P. Wood Identification—A Review. IAWA J. 1998, 19, 241–264. [Google Scholar] [CrossRef]
  25. Lee, I.; Shin, Y.J. Machine learning for enterprises: Applications, algorithm selection, and challenges. Bus. Horizons 2020, 63, 157–170. [Google Scholar] [CrossRef]
  26. Zheng, C.; Liu, S.; Wang, J.; Lu, Y.; Ma, L.; Jiao, L.; Guo, J.; Yin, Y.; He, T. Opening the black box: Explainable deep-learning classification of wood microscopic image of endangered tree species. Plant Methods 2024, 20, 56. [Google Scholar] [CrossRef]
  27. Rosa da Silva, N.; Deklerck, V.; Baetens, J.M.; Van den Bulcke, J.; De Ridder, M.; Rousseau, M.; Bruno, O.M.; Beeckman, H.; Van Acker, J.; De Baets, B.; et al. Improved wood species identification based on multi-view imagery of the three anatomical planes. Plant Methods 2022, 18, 79. [Google Scholar] [CrossRef]
  28. Yang, X.; Jiang, H.; Ma, L.; Yang, W.; Zhao, X.; Hu, C.; Ge, Z. Micro image classification of 19 high-value hardwood species based on texture feature fusion. BioResources 2023, 18, 3373–3386. [Google Scholar] [CrossRef]
  29. Zhao, P.; Dou, G.; Chen, G.S. Wood species identification using feature-level fusion scheme. Optik 2014, 125, 1144–1148. [Google Scholar] [CrossRef]
  30. Zamri, M.I.; Aan, P.; Cordova, F.; Khairuddin, A.S.M.; Mokhtar, N.; Yusof, R. Tree species classification based on image analysis using improved-basic gray level aura matrix. Comput. Electron. Agric. 2016, 124, 227–233. [Google Scholar] [CrossRef]
  31. Yuliastuti, E.; Suprijanto; Sasi, S.R. Compact computer vision system for tropical wood species recognition based on pores and concentric curve. In Proceedings of the 3rd International Conference on Instrumentation, Control and Automation (ICA), Ungasan, Indonesia, 28–30 August 2013; pp. 198–202. [Google Scholar] [CrossRef]
  32. Nasirzadeh, M.; Khazael, A.A.; Bin Khalid, M. Woods recognition system based on local binary pattern. In Proceedings of the 2nd International Conference on Computational Intelligence, Communication Systems and Networks (CICSyN), Liverpool, UK, 28–30 July 2010; pp. 308–313. [Google Scholar] [CrossRef]
  33. Yusof, R.; Rosli, N.R.; Khalid, M. Using Gabor filters as image multiplier for tropical wood species recognition system. In Proceedings of the 12th International Conference on Computer Modelling and Simulation (UKSim), Cambridge, UK, 24–26 March 2010; pp. 289–294. [Google Scholar] [CrossRef]
  34. Wheeler, E.A. Inside Wood—A Web resource for hardwood anatomy. IAWA J. 2011, 32, 199–211. [Google Scholar] [CrossRef]
  35. Li, H.; Huang, J.; Gu, Z.; He, D.; Huang, J.; Wang, C. Positioning of mango picking point using an improved YOLOv8 architecture with object detection and instance segmentation. Biosyst. Eng. 2024, 247, 202–220. [Google Scholar] [CrossRef]
  36. Schubert, E.; Sander, J.; Ester, M.; Kriegel, H.P.; Xu, X. DBSCAN Revisited, Revisited: Why and How You Should (Still) Use DBSCAN. ACM Trans. Database Syst. 2017, 42. [Google Scholar] [CrossRef]
  37. Suzuki, S.; be, K. Topological structural analysis of digitized binary images by border following. Comput. Vis. Graph. Image Process. 1985, 30, 32–46. [Google Scholar] [CrossRef]
  38. Tian, Z.; Shen, C.; Chen, H. Conditional Convolutions for Instance Segmentation. In Proceedings of the Computer Vision—ECCV 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer: Cham, Switzerland, 2020; pp. 282–298. [Google Scholar] [CrossRef]
  39. Kirillov, A.; Wu, Y.; He, K.; Girshick, R. PointRend: Image Segmentation As Rendering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1–16. [Google Scholar]
  40. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  41. Chandra, M.A.; Bedi, S.S. Survey on SVM and their application in image classification. Int. J. Inf. Technol. 2021, 13, 1–11. [Google Scholar] [CrossRef]
  42. Zhang, S.; Li, X.; Zong, M.; Zhu, X.; Cheng, D. Learning k for knn classification. ACM Trans. Intell. Syst. Technol. (TIST) 2017, 8, 1–19. [Google Scholar] [CrossRef]
  43. Song, Y.; Lu, Y. Decision tree methods: Applications for classification and prediction. Shanghai Arch. Psychiatry 2015, 27, 130–135. [Google Scholar]
  44. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  45. Ojetunmibi, T.; Asagba, P.O.; Okengwu, U.A. Pneumonia disease detection and classification system using naive Bayesian technique. Sci. Afr. 2023, 22, 97–114. [Google Scholar] [CrossRef]
  46. Mu, Y.; Liu, X.; Wang, L. A Pearson’s correlation coefficient based decision tree and its parallel implementation. Inf. Sci. 2018, 435, 40–58. [Google Scholar] [CrossRef]
Figure 1. Anatomical image positions in wood samples using Acer rubrum as an example. The left image (specimen BWCw 8556) shows the cross section (perpendicular to the growth rings), where vessels and parenchyma are visible; the right image (specimen BWCw 8332) shows the tangential section (parallel to the growth rings), where wood rays can be seen.
Figure 1. Anatomical image positions in wood samples using Acer rubrum as an example. The left image (specimen BWCw 8556) shows the cross section (perpendicular to the growth rings), where vessels and parenchyma are visible; the right image (specimen BWCw 8332) shows the tangential section (parallel to the growth rings), where wood rays can be seen.
Forests 16 01328 g001
Figure 2. Workflow of wood microscopic image acquisition. The steps of wood microscopic image acquisition include softening, slicing, staining, and mounting, then image capture using a Leica DM2000 LED microscope.
Figure 2. Workflow of wood microscopic image acquisition. The steps of wood microscopic image acquisition include softening, slicing, staining, and mounting, then image capture using a Leica DM2000 LED microscope.
Forests 16 01328 g002
Figure 3. Example images in the wood sections dataset. Subfigures (ae) show cross-section images, where vessels and parenchyma are visible; Subfigures (fi) show tangential section images, where wood rays are visible. Images from specimens: (a) specimen BWCw 8556, (b) specimen BWCw 8618, (c) specimen Kw Per.ela.Trade.1951, (e) specimen PACw 4673, (f) specimen BWCw 8516, (g) specimen PACw En.cyl, (h) specimen Kw Swi.mac, (i) specimen USw PR19.
Figure 3. Example images in the wood sections dataset. Subfigures (ae) show cross-section images, where vessels and parenchyma are visible; Subfigures (fi) show tangential section images, where wood rays are visible. Images from specimens: (a) specimen BWCw 8556, (b) specimen BWCw 8618, (c) specimen Kw Per.ela.Trade.1951, (e) specimen PACw 4673, (f) specimen BWCw 8516, (g) specimen PACw En.cyl, (h) specimen Kw Swi.mac, (i) specimen USw PR19.
Forests 16 01328 g003
Figure 4. Interpretable wood recognition process. The wood recognition method involves three steps: anatomical feature detection using an anatomy structure detector, quantifying features and mapping them to IAWA codes, and wood species identification using a Bayesian network. Example images from specimens: BWCw 8556, Kw 70560, and USw PR19.
Figure 4. Interpretable wood recognition process. The wood recognition method involves three steps: anatomical feature detection using an anatomy structure detector, quantifying features and mapping them to IAWA codes, and wood species identification using a Bayesian network. Example images from specimens: BWCw 8556, Kw 70560, and USw PR19.
Forests 16 01328 g004
Figure 5. Principle of the wood ray cells count and rays density calculation. (a) The middle area of the ray is divided into multiple parts. Count the number of cells in each part and calculate the average. (b) Divide the image (specimen BWCw 8332) into n equal horizontal sections and count the wood rays intersected by each horizontal line.
Figure 5. Principle of the wood ray cells count and rays density calculation. (a) The middle area of the ray is divided into multiple parts. Count the number of cells in each part and calculate the average. (b) Divide the image (specimen BWCw 8332) into n equal horizontal sections and count the wood rays intersected by each horizontal line.
Forests 16 01328 g005
Figure 6. Multi-image IAWA feature fusion process. Image 1 (specimen Kw 4353) yields IAWA feature codes 42 and 47; Image 2 (specimen Kw 4353) yields code 79; Image n (specimen Kw Swi.mac) yields codes 98 and 115. The codes are fused into a combined feature code set D, which is then binarized to generate the final feature vector IAC.
Figure 6. Multi-image IAWA feature fusion process. Image 1 (specimen Kw 4353) yields IAWA feature codes 42 and 47; Image 2 (specimen Kw 4353) yields code 79; Image n (specimen Kw Swi.mac) yields codes 98 and 115. The codes are fused into a combined feature code set D, which is then binarized to generate the final feature vector IAC.
Forests 16 01328 g006
Figure 7. Interpretable recognition results: Output1 provides image evidence, Output2 corresponds to the IAWA feature codes. Output3 represents the wood classification result. The first (specimen Tw 52896) and the second (specimens Tw 5225, Tw 6987) predictions are correct; the last one (specimens Tw 609, Tw 724) is incorrect—the correct species is Albizia ferruginea.
Figure 7. Interpretable recognition results: Output1 provides image evidence, Output2 corresponds to the IAWA feature codes. Output3 represents the wood classification result. The first (specimen Tw 52896) and the second (specimens Tw 5225, Tw 6987) predictions are correct; the last one (specimens Tw 609, Tw 724) is incorrect—the correct species is Albizia ferruginea.
Forests 16 01328 g007
Figure 8. Training loss and F1-score across epochs for different anatomical structures. The three plots correspond to vessel, ray, and parenchyma, respectively, showing training loss (left axis) and F1-score (right axis) over epochs.
Figure 8. Training loss and F1-score across epochs for different anatomical structures. The three plots correspond to vessel, ray, and parenchyma, respectively, showing training loss (left axis) and F1-score (right axis) over epochs.
Forests 16 01328 g008
Figure 9. Anatomy structure detection results. The first row (specimens Hw 23557, BWCw 8058, PACw 4673, and USw PR19) represents the vessel detection result, the second row (specimens BWCw 8332, USw PR19, Tw 5016, and Kw Swi.mac) represents the wood rays detection result, and the third row (specimens Tw 26885, Tw 609, and Tw 425) represents the parenchyma detection result.
Figure 9. Anatomy structure detection results. The first row (specimens Hw 23557, BWCw 8058, PACw 4673, and USw PR19) represents the vessel detection result, the second row (specimens BWCw 8332, USw PR19, Tw 5016, and Kw Swi.mac) represents the wood rays detection result, and the third row (specimens Tw 26885, Tw 609, and Tw 425) represents the parenchyma detection result.
Forests 16 01328 g009
Figure 10. Counting errors of vessels and wood rays between the proposed method and manual method.
Figure 10. Counting errors of vessels and wood rays between the proposed method and manual method.
Forests 16 01328 g010
Figure 11. Detection results of different models: The first column displays the original images, while the subsequent columns show the model-predicted images. The colors represent different anatomical structures. The first row (specimen FHOw 19992) correspond to vessel images, the second row (specimen FHOw 4291) is wood ray images, and the last row (specimen Tw 17698) represent parenchyma images.
Figure 11. Detection results of different models: The first column displays the original images, while the subsequent columns show the model-predicted images. The colors represent different anatomical structures. The first row (specimen FHOw 19992) correspond to vessel images, the second row (specimen FHOw 4291) is wood ray images, and the last row (specimen Tw 17698) represent parenchyma images.
Forests 16 01328 g011
Figure 12. Impact of combining cross section and tangential section images on classification performance.
Figure 12. Impact of combining cross section and tangential section images on classification performance.
Forests 16 01328 g012
Figure 13. Accuracy of image combinations with different quantities. The accuracy reaches its highest when there are 6 cross section images and 4 tangential section images.
Figure 13. Accuracy of image combinations with different quantities. The accuracy reaches its highest when there are 6 cross section images and 4 tangential section images.
Forests 16 01328 g013
Figure 14. Visualization of feature correlation and SHAP importance for IAWA features (y-axis) across wood species (x-axis). SHAP: a model-agnostic interpretability method based on cooperative game theory. It assigns each feature an importance value by quantifying its marginal contribution to the prediction across all possible feature combinations. Higher SHAP values indicate features that play a more decisive role in distinguishing between classes.
Figure 14. Visualization of feature correlation and SHAP importance for IAWA features (y-axis) across wood species (x-axis). SHAP: a model-agnostic interpretability method based on cooperative game theory. It assigns each feature an importance value by quantifying its marginal contribution to the prediction across all possible feature combinations. Higher SHAP values indicate features that play a more decisive role in distinguishing between classes.
Forests 16 01328 g014
Figure 15. Ray cell recognition under low image quality. The image (specimen PACw 4673) shows ray cells recognition results after segmentation and binarization under low image quality conditions.
Figure 15. Ray cell recognition under low image quality. The image (specimen PACw 4673) shows ray cells recognition results after segmentation and binarization under low image quality conditions.
Forests 16 01328 g015
Table 2. Wood species information in the dataset. The information includes Family, Species, WOCI (IAWA code 192: Wood of Commercial Importance) status, CITES listing, and Geographic Origin.
Table 2. Wood species information in the dataset. The information includes Family, Species, WOCI (IAWA code 192: Wood of Commercial Importance) status, CITES listing, and Geographic Origin.
IDFamilySpeciesWOCICITES ListedGeographic Origin
1MELIACEAESwietenia macrophylla(183, 184)
2ANACARDIACEAEPistacia chinensis (164, 167, 171, 173)
3MELIACEAEEntandrophragma angolense (178, 179)
4EBENACEAECyrilla racemiflora (182)
5EBENACEAEDiospyros crassiflora(178, 179)
6EBENACEAEDiospyros virginiana (184)
7FAGACEAEQuercus alba (184)
8JUGLANDACEAEJuglans nigra (182)
9LEGUMINOSAE CAESALPINIOIDEAEGleditsia triacanthos (182)
10LEGUMINOSAE CAESALPINIOIDEAEAlbizia ferruginea (178, 179)
11LEGUMINOSAE CAESALPINIOIDEAEPiptadeniastrum africanum (178, 179)
12LEGUMINOSAE DETARIOIDEAEIntsia bijuga (168, 169)
13LEGUMINOSAE PAPILIONOIDEAEPericopsis elata (178, 179)
14LEGUMINOSAE PAPILIONOIDEAEPterocarpus soyauxii(178, 179)
15LEGUMINOSAE PAPILIONOIDEAERobinia pseudoacacia (182)
16MALVACEAE HELICTEROIDEAETriplochiton scleroxylon (178, 179)
17MELIACEAEGuarea cedrata (178, 179)
18MELIACEAEGuarea laurentii (178, 179)
19MELIACEAELovoa trichilioides (178, 179)
20MELIACEAEMelia azedarach (164, 166)
21MORACEAEMilicia excelsa (178, 179)
22RUBIACEAENauclea diderrichii (178, 179)
23SAPINDACEAEAcer rubrum (182)
24FABACEAEDalbergia cochinchinensis(171, 172)
25FABACEAEDalbergia stevensonii (184)
26FABACEAEDalbergia cultrata(171, 172)
27FABACEAEPterocarpus erinaceus(178, 179)
28MELIACEAEEntandrophragma cylindricum (178, 179)
29FABACEAEPterocarpus macrocarpus (171, 172)
30FABACEAEDalbergia oliveri (171, 172)
31FABACEAEPterocarpus angolensis(178, 179)
32FABACEAEDalbergia frutescens(183)
Geographic origin(IAWA Code): 164 Europe and temperate Asia; 166 Mediterranean including Northern Africa and Middle East; 167 Temperate Asia (China), Japan, Russia; 168 Central South Asia; 169 India, Pakistan, Sri Lanka; 171 Southeast Asia and Pacific; 172 Indochina (Thailand, Laos, Vietnam, Cambodia); 173 Indomalesia (Indonesia, Philippines, Malaysia, Brunei, Papua New Guinea, Solomon Islands); 178 Tropical mainland Africa and adjacent islands; 179 Tropical Africa; 181 Southern Africa (south of the Tropic of Capricorn); 182 North America, north of Mexico; 183 Neotropics and temperate Brazil; 184 Mexico and Central America; 186 Tropical South America. Only populations of Diospyros crassiflora from Madagascar are listed under CITES II, while those from continental Africa are not listed.
Table 3. Selected IAWA feature codes and the descriptions.
Table 3. Selected IAWA feature codes and the descriptions.
IAWA CodeDescription
Vessel arrangement and groupings
6Vessels in longitudinal bands
9Vessels exclusively solitary (90% or more)
10Vessels in radial multiples of 4 or more common
11Vessel clusters common
Tangential diameter of vessel lumina
40<=50 μm
4150–100 μm
42100–200 μm
43>=200 μm
Vessels per square millimetre
46<=5 vessels per square millimetre
475–20 vessels per square millimetre
4820–40 vessels per square millimetre
4940–100 vessels per square millimetre
50>=100 vessels per square millimetre
Paratracheal axial parenchyma
79Axial parenchyma vasicentric
80Axial parenchyma aliform
82Axial parenchyma winged-aliform
83Axial parenchyma confluent
85Axial parenchyma bands more than three cells wide
Ray width
96Rays exclusively uniseriate
97Ray width 1 to 3 cells
98Larger rays commonly 4- to 10 seriate
99Larger rays commonly > 10-seriate
Rays per millimetre
114<=4/mm
1154–12/mm
116>=12/mm
Table 4. Wood species classification results. The table presents the precision, recall, and F1-score for each species, along with the overall accuracy and weighted average metrics across all species. The weighted average accounts for the number of samples per species, giving more influence to species with more samples.
Table 4. Wood species classification results. The table presents the precision, recall, and F1-score for each species, along with the overall accuracy and weighted average metrics across all species. The weighted average accounts for the number of samples per species, giving more influence to species with more samples.
SpeciesPrecisionRecallF1-Score
Swietenia macrophylla0.9601.0000.980
Pistacia chinensis0.9351.0000.967
Entandrophragma angolense1.0000.9620.980
Cyrilla racemiflora0.8951.0000.944
Diospyros crassiflora0.9800.9700.980
Diospyros virginiana0.9750.9850.980
Quercus alba0.9441.0000.971
Juglans nigra0.9800.9900.985
Gleditsia triacanthos0.9090.8330.870
Albizia ferruginea0.6920.8180.750
Piptadeniastrum africanum0.9570.9170.936
Intsia bijuga0.8670.9290.897
Pericopsis elata0.6880.6470.667
Pterocarpus soyauxii1.0000.9500.974
Robinia pseudoacacia0.8851.0000.939
Triplochiton scleroxylon0.9051.0000.950
Guarea cedrata0.9410.8890.914
Guarea laurentii0.9700.9800.975
Lovoa trichilioides0.8950.7730.829
Melia azedarach0.7370.7780.757
Milicia excelsa0.8241.0000.903
Nauclea diderrichii0.9380.8330.882
Acer rubrum1.0000.9000.947
Dalbergia cochinchinensis1.0000.9500.974
Dalbergia stevensonii0.9331.0000.966
Dalbergia cultrata0.9601.0000.980
Pterocarpus erinaceus1.0000.8570.923
Entandrophragma cylindricum0.9650.9750.970
Pterocarpus macrocarpus1.0000.9130.955
Dalbergia oliveri0.9700.9850.977
Pterocarpus angolensis0.9750.9800.978
Dalbergia frutescens0.9600.9700.965
Accuracy 0.941
Weighted Average0.9260.9330.927
Table 5. Hyperparameter settings used for YOLO-seg model training.
Table 5. Hyperparameter settings used for YOLO-seg model training.
HyperparameterValueDescription
epochs250Number of training epochs
batch16Batch size per iteration
imgsz640Input image size
learning_rate0.01Initial learning rate
optimizerSGDType of optimizer
weight_decay0.0005Weight decay for regularization
momentum0.937Momentum factor for SGD
Table 6. Detection performance comparison of different models for anatomical structures. Metrics reported include Average Precision (AP) at IoU thresholds 0.50 (AP50) and 0.75 (AP75), and the mean Average Precision (mAP) across different IoU thresholds evaluated for vessel, ray, and parenchyma.
Table 6. Detection performance comparison of different models for anatomical structures. Metrics reported include Average Precision (AP) at IoU thresholds 0.50 (AP50) and 0.75 (AP75), and the mean Average Precision (mAP) across different IoU thresholds evaluated for vessel, ray, and parenchyma.
ModelVesselRayParenchyma
AP50 AP75 mAP AP50 AP75 mAP AP50 AP75 mAP
Mask-RCNN0.7700.7470.6460.948 0.8110.6740.7600.5820.494
PointRend0.7410.7050.5960.9300.8060.6740.5310.5010.457
CondInst0.7640.6600.5440.9030.6130.5510.5330.4960.471
YOLOv8-seg0.9120.8640.7150.9350.8340.7160.7300.6230.572
m A P : average of A P s , mean of ( A P 50 , A P 55 , , A P 95 ). The Bolded values represent the best scores.
Table 7. Recognition accuracy of five classification models evaluated on the train set and test set.
Table 7. Recognition accuracy of five classification models evaluated on the train set and test set.
ModelAccuracy of Test SetAccuracy of Train Set
SVM0.9390.946
Naive Bayes0.941 0.945
Decision Tree0.9000.973
Random Forest0.9370.942
KNN0.9210.940
The Bolded values represent the best scores.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, L.; Qiu, J.; Cao, Y.; Li, Q.; Qian, S.; Sun, Y. An Interpretable Approach to Wood Species Identification Based on Anatomical Features in Microscopic Images. Forests 2025, 16, 1328. https://doi.org/10.3390/f16081328

AMA Style

Liu L, Qiu J, Cao Y, Li Q, Qian S, Sun Y. An Interpretable Approach to Wood Species Identification Based on Anatomical Features in Microscopic Images. Forests. 2025; 16(8):1328. https://doi.org/10.3390/f16081328

Chicago/Turabian Style

Liu, Lei, Jian Qiu, Yong Cao, Qiying Li, Songping Qian, and Yongke Sun. 2025. "An Interpretable Approach to Wood Species Identification Based on Anatomical Features in Microscopic Images" Forests 16, no. 8: 1328. https://doi.org/10.3390/f16081328

APA Style

Liu, L., Qiu, J., Cao, Y., Li, Q., Qian, S., & Sun, Y. (2025). An Interpretable Approach to Wood Species Identification Based on Anatomical Features in Microscopic Images. Forests, 16(8), 1328. https://doi.org/10.3390/f16081328

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop