Next Article in Journal
Asynchronous Patterns Between Vegetation Structural Expansion and Photosynthetic Functional Enhancement on China’s Loess Plateau
Previous Article in Journal
Optimized and Reliable Protoplast Isolation for Transient Gene Expression Studies in the Gymnosperm Tree Species Pinus densiflora
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Recognition of Commercial Tree Species from the Amazon Flora Using Bark Images and Transfer Learning

by
Natally Celestino Gama
1,*,
Luiz Eduardo Soares Oliveira
2,
Samuel de Pádua Chaves e Carvalho
3,
Alexandre Behling
1,
Pedro Luiz de Paula Filho
4,
Márcia Orie de Sousa Hamada
5,
Eduardo da Silva Leal
6 and
Deivison Venicio Souza
5,*
1
Department of Forest Science, Federal University of Parana, Curitiba 80060-000, Brazil
2
Department of Informatics, Federal University of Parana, Curitiba 81531-980, Brazil
3
Forestry Institute, Federal Rural University of Rio de Janeiro, Seropédica 23897-000, Brazil
4
Department of Informatics, Federal Technological University of Parana, Medianeira 85884-000, Brazil
5
Faculty of Forestry Engineering, Federal University of Pará, Altamira 68372-000, Brazil
6
Department of Forest Sciences, Federal Rural University of Amazônia, Capanema 68700-000, Brazil
*
Authors to whom correspondence should be addressed.
Forests 2025, 16(9), 1374; https://doi.org/10.3390/f16091374
Submission received: 13 June 2025 / Revised: 7 July 2025 / Accepted: 18 August 2025 / Published: 27 August 2025
(This article belongs to the Section Forest Inventory, Modeling and Remote Sensing)

Abstract

The application of artificial intelligence (AI) techniques has improved the accuracy of forest species identification, particularly in timber inventories conducted under Sustainable Forest Management (SFM). This study developed and evaluated machine learning models to recognize 16 Amazonian timber species using digital images of tree bark. Data were collected from three SFM units located in Nova Maringá, Feliz Natal, and Cotriguaçu, in the state of Mato Grosso, Brazil. High-resolution images were processed into sub-images (256 × 256 pixels), and two feature extraction methods were tested: Local Binary Patterns (LBP) and pre-trained Convolutional Neural Networks (ResNet50, VGG16, InceptionV3, MobileNetV2). Four classifiers—Support Vector Machine (SVM), Artificial Neural Networks (ANN), Random Forest (RF), and Linear Discriminant Analysis (LDA)—were used. The best result (95% accuracy) was achieved using ResNet50 with SVM, confirming the effectiveness of transfer learning for species recognition based on bark texture. These findings highlight the potential of AI-based tools to enhance accuracy in forest inventories and support decision-making in tropical forest management.

1. Introduction

The Amazonian flora comprises approximately 30% of all arboreal individuals in South America and represents one of the world’s leading sources of tropical timber [1,2]. As timber market activity intensifies in the region, it becomes increasingly necessary to regulate resource extraction through Sustainable Forest Management (SFM) practices [3,4]. SFM is defined as the management of forests to generate economic, social, and environmental benefits while preserving essential ecosystem functions [5]. One of its fundamental requirements is the implementation of a 100% forest inventory, which entails the collection of both quantitative and qualitative data, including the botanical identification of tree species [6,7].
However, accurate botanical identification in the Amazon remains a significant challenge due to the morphological similarity among species. Traditionally, it relies on the collection of reproductive material—leaves, flowers, fruits, and seeds—which must be compared against herbarium specimens [8,9]. In timber-oriented SFM, identification is often conducted by local experts, known as “mateiros”, who rely on vernacular names and macroscopic traits such as bark color, texture, scent, and the presence or absence of exudates [10,11]. It is important to highlight that this step is legally mandated as part of the forest census process, in accordance with the Brazilian regulation Instrução Normativa IBAMA No. 5, 11 dated December 2006. This regulation establishes the technical criteria for the preparation, submission, implementation, and monitoring of Sustainable Forest Management Plans (PMFS) in native forests. Accurate species identification is essential to ensure traceability, legal compliance in timber harvesting, and the conservation of biodiversity.
While vernacular identification is valuable, it often leads to errors such as misidentification of endangered species, confusion in trade nomenclature, and inconsistencies when converting common names to scientific ones based on outdated official species lists [12,13,14]. These issues are well documented in studies that report spelling mistakes, inconsistent nomenclature, and incomplete taxonomic identification in Sustainable Forest Management Plans (SFMPs), largely due to dialectal variation and limited oversight [15,16].
To address these challenges, recent studies have explored computer vision and artificial intelligence techniques to support botanical identification [11,17,18]. These methods involve extracting features from digital images of plant organs [19,20,21] and classifying them using machine learning algorithms such as Artificial Neural Networks (ANN) and Support Vector Machines (SVM) [22,23,24]. Although most models rely on leaf or wood images [25,26,27,28], bark images present a more practical alternative in Amazonian forest conditions, where canopy height and phenology often limit the collection of other structures [18,29,30].
Despite the progress in computational models for plant recognition [31,32,33,34,35] there is currently no automated system tailored for identifying the main commercial species of the Amazon during forest inventories. This study aims to fill that gap by developing and evaluating machine learning models based on bark images for the identification of 16 commercially important species. The models incorporate different feature extraction techniques and classification algorithms, contributing to the improvement of species identification processes in timber-focused SFM.

2. Materials and Methods

2.1. Areas of Study

This study was conducted in three Sustainable Forest Management (SFM) Demonstration Units (DUs) aimed at commercial timber production, located in the state of Mato Grosso, Brazil: DU1—Farm Pérola (Nova Maringá), DU2—Farm Boa Esperança (Feliz Natal), and DU3—Farm São Nicolau (Cotriguaçu) (Figure 1).
These DUs are situated within regions of the Amazon biome, which occupies approximately 53.5% of Mato Grosso’s territory and overlaps with elements of the Cerrado and Pantanal biomes [36]. The predominant forest types in the study areas include Dense and Open Ombrophilous Forests and Seasonal Forests, characterized by high biodiversity and commercial interest for timber. The regional climate is tropical, with average temperatures ranging from 24.3 °C to 25.7 °C and annual rainfall between 2000 and 2500 mm [37,38].
These areas were selected for their ecological representativeness and economic relevance within legal SFM frameworks in the Amazon. They provide a realistic field context for testing artificial intelligence tools aimed at improving species identification in operational forest inventories.

2.2. Image Acquisition

The species were selected based on the list of forest species of most significant commercial interest in the state of Mato Grosso, provided by the Center of Wood Producing and Exporting Industries of Mato Grosso (CWPEIMG), and taking into account the occurrence and abundance of individuals in the collection areas.
In the field, images were taken of the bark (rhytidome) of 10 trees per species, using two devices: a Canon camera and an iOS iPhone 11 device, obtaining images with resolutions of 3024 × 4032 and 4000 × 5328, respectively (Figure 2). On the trees, the region in which the images were captured ranged in height from 30 cm to 1.40 m from the ground, and a distance of between 20 cm and 40 cm from the device to the trunk, depending on the conditions found in the field.
Many factors made it difficult to acquire images, such as low light (or even too much due to direct sunlight on the tree trunk) and the presence of mosses (or lichens) and termites in the bark. Therefore, in areas of the forest with low light levels, a reflector (YN600L II—Pro Led Video Light) was used to improve the light conditions for capturing the image (Figure 3). In addition, in trees with a high incidence of mosses, lichens, or termites, we looked for areas of the trunk free of these living beings to highlight the characteristics of the tree bark in the captured image.
The acquired images were carefully evaluated to identify quality problems or undesirable elements that could affect the extraction of features and subsequent learning of the models. The image base consisted of 2803 bark images belonging to 16 species, 16 genera, and 9 families (Table 1). Figure 4 shows samples of bark images from the 16 tree species.

2.3. Identifying Species and Collection Botanical Material

In the field, species identification is a complex task, but one that is fundamental to ensuring the correct labeling of images and constructing a reliable species recognition system. Therefore, species identification was initially carried out by a Parabotanist, who climbed at least five trees per species to collect botanical material, sampling 83 specimens from 16 species.
All the botanical material was pressed, and then exsiccates were mounted to be sent to the Felisberto Camargo Herbarium (HFC) at the Federal Rural University of Amazonia to confirm the identification (Figure 5).

2.4. Extraction of Sub-Images (Patches)

To improve feature extraction and optimize computational efficiency, the original high-resolution bark images (3024 × 4032 and 4000 × 5328 pixels) were segmented into smaller sub-images. This patch-based approach enhances the detection of local texture patterns and increases dataset variability, which is particularly beneficial for training convolutional neural networks (CNNs).
Initially, all images were resized to 20% of their original resolution to reduce preprocessing time while preserving relevant visual features. From each resized image, 50 random sub-images (patches) of 256 × 256 pixels were extracted, resulting in a total of 140,150 patches from 2803 original images. The patch size was chosen to ensure compatibility with standard CNN input dimensions (e.g., ResNet50: 224 × 224; InceptionV3: 299 × 299).
The number of patches per image was defined considering the trade-off between dataset augmentation and computational cost. Previous studies used between 9 and 80 patches per image depending on class diversity and resource availability [29,39]. In this study, 50 patches were considered sufficient to balance performance and processing time across 16 species classes.
Final image classification was determined by aggregating predictions from the 50 patches using majority voting, a commonly applied ensemble strategy in image classification tasks [29,39,40].

2.5. Extracting Features from Original Images and Sub-Images

The feature extraction process is an essential step in the successful modeling of automatic classifiers [11]. Thus, two strategies for extracting features from digital images of tree bark were evaluated: (i) Local Binary Patterns (LBP) and (ii) pre-trained Convolutional Neural Networks (CNN) (ResNet50, VGG16, InceptionV3 and MobileNetV2). These features were extracted from original high-resolution images (3024 × 4032 and 4000 × 5328 pixels) captured by two devices (CANON and iOS iPhone 11) and from 256 × 256 pixel sub-images randomly extracted from the original images.

2.5.1. Local Binary Patterns (LBP)

Local Binary Patterns (LBP) is a robust and efficient texture descriptor, invariant to grayscale variations, originally proposed by [41]. It combines statistical and structural texture analysis by encoding the local neighborhood of each pixel into a binary code that reflects primitive texture patterns such as edges, spots, and corners.
Due to its simplicity, low computational cost, and effectiveness in capturing fine-grained textures, LBP has been widely used in image classification tasks, including bark recognition.
Mathematically, an LBP code for a given pixel is calculated as follows:
L B P P , R = P = 0 P 1 s x p x c 2 p s x = 1 ,   x   0 0 ,   x   < 0
where P is the number of circular neighbors, R is the radius, xc is the intensity of the central pixel, and xp are the neighboring pixel values.
To enhance rotation invariance and compactness, the rotation-invariant uniform LBP ( L B P P , R r i u 2 ) proposed by [42] was also used. Uniform patterns are defined as those with at most two transitions between 0 and 1 in the circular binary sequence, resulting in fixed-length feature vectors for each configuration: 10 bins for L B P 8 , 1 r i u 2 , 18 bins for L B P 16 , 2 r i u 2 , and 26 bins for L B P 24 , 3 r i u 2 . Non-invariant uniform versions ( L B P P , R u 2 ) have P (P − 1) + 2 bins (e.g., 59, 243, and 555 bins for P = 8, 16, 24, respectively), capturing directional texture information more explicitly. Figure 6 illustrates the different circular neighborhood configurations for three typical values of P and R used in LBP feature extraction.
In this study, RGB images were first converted to grayscale, and six LBP configurations were applied to both the original and patch images:
  • Uniform and rotation-invariant: L B P 8 , 1 r i u 2 , L B P 16 , 2 r i u 2 , L B P 24 , 3 r i u 2 ;
  • Uniform and non-invariant: L B P 8 , 1 u 2 , L B P 16 , 2 u 2 , L B P 24 , 3 u 2 .
From each configuration, histograms of LBP code occurrences were extracted. To standardize the feature distributions and improve model convergence, Z-score normalization was applied:
z = x i μ σ

2.5.2. Transfer Learning

Convolutional Neural Networks (CNNs) are a class of deep learning models characterized by feedforward architectures incorporating convolutional operations and depth [43]. However, training CNNs typically requires large datasets and substantial computational resources. Transfer Learning offers an alternative by adapting pre-trained networks to new but related tasks, thereby leveraging previously acquired knowledge and reducing training demands [23,44].
This study employed four pre-trained CNN architectures—ResNet50, VGG16, InceptionV3, and MobileNetV2—selected based on their proven performance in similar plant classification applications [45,46,47]. ResNet50, a 50-layer residual network, was trained on over one million images from the ImageNet dataset, capable of classifying up to 1000 categories [33]. VGG16 comprises 16 trainable layers and was developed by the Visual Geometry Group at the University of Oxford, widely applied in computer vision [48]. InceptionV3 includes approximately 48 layers, utilizing multiple kernel sizes and pooling methods to optimize computational efficiency [49]. MobileNetV2, with around 53 layers, balances computational cost and accuracy, making it suitable for mobile devices [50].
All networks were initialized with ImageNet weights, and their final classification layers were replaced by a Global Average Pooling (GAP) layer to serve as feature extractors. GAP effectively reduces feature map dimensionality and the number of parameters, minimizing overfitting risk without substantial information loss. Images were pre-processed according to each network’s standards, and the resulting feature vectors were input to subsequent classification algorithms.

2.6. Algorithms, Cross-Validation and Performance Metrics

2.6.1. Classification Algorithms

Four machine learning algorithms were employed: Support Vector Machine (SVM), Artificial Neural Networks (ANN), Random Forest (RF), and Linear Discriminant Analysis (LDA). SVM is a supervised learning method for classification and regression that maps features into a new space and defines a decision boundary separating classes; multi-class problems are addressed via multiple binary classifiers [51,52,53].
ANNs mimic human brain information processing, providing a framework for regression and classification through adjustable connections and specific architectures [54]. RF is an ensemble technique combining outputs of multiple decision trees for classification and regression [55,56]. LDA is a supervised algorithm used for classification and dimensionality reduction, projecting data into lower-dimensional space by maximizing between-class variance relative to within-class variance [57,58].

2.6.2. Image Division and Cross-Validation

The bark image dataset (n = 2803) was initially split into training (80%; n = 2237) and test (20%; n = 566) sets with stratification by class to maintain balanced proportions. The split was performed using the splitfolders library (v 0.5.1) in Python. Subsequently, sub-images of 256 × 256 pixels were generated, resulting in training (80%; n = 111,850) and test (20%; n = 28,300) sets, approximately 50 times larger. Hyperparameter tuning for each classifier employed Random Search with a predefined grid, optimized via fivefold cross-validation on the training data using RandomizedSearchCV from scikit-learn (Refer to the Appendix ATable A1 for additional details).
StratifiedGroupKFold from the scikit-learn library was employed for cross-validation to generate stratified folds preserving class proportions while ensuring non-overlapping groups. This approach allowed (i) maintaining species class balance across validation folds and (ii) preventing data leakage by excluding images from the same sample (tree) in both training (k − 1 folds) and validation (k fold) sets.
Owing to the high computational cost, hyperparameters optimized via random search on the original image resolution were used to train models based on sub-image features, applying 10-fold cross-validation. The entire workflow, including feature extraction, model training, and performance evaluation, was implemented in Python using Google Colaboratory resources and is publicly available at https://github.com/NatallyCelestino/Bark_Recogition.git (accessed on 16 August 2025).

2.6.3. Performance Metrics

The evaluation of the generalization capacity of the classification models was carried out using the following performance metrics obtained in the cross-validation and test set: accuracy (Equation (3)), recall (Equation (4)), and f1-score (Equation (5)). In addition, the confusion matrix was examined to identify the main classification errors.
  • Accuracy: represents the number of correct predictions made by the model:
    A c c u r a c y = T P + T N T P + F P + T N + F N
    where TP—true positive; TN—true negative; FP—false positive; and FN—false negative.
  • Recall or Sensitivity: Metric recommended when there is class imbalance. It represents the classification model’s ability to predict the positive class:
    R e c a l l = T P T P + F N
    where TP—true positive; and FN—false negative.
  • F1—score is the harmonic mean between Recall and Precision (which represents the number of observations classified correctly). It is considered a suitable metric for problems with unbalanced classes:
    F 1 s c o r e = 2 × P R E C I S I O N × R E C A L L ( P R E C I S I O N + R E C A L L )

3. Results

Performance of the Classifiers

A total of 160 trees were sampled from 16 forest species of commercial timber value, belonging to 16 genera and nine families. A total of nine classifiers of external bark images were adjusted, using the LBP operator as a feature extractor, individually and combined, for each machine learning algorithm, using original images (3024 × 4032 pixels and 4000 × 5328 pixels) and sub-images (256 × 256 pixels).
In the approach to extracting features from images with original sizes, in most of the experiments, the classifiers built using uniform and non-invariant LBP feature vectors ( L B P P , R u 2 ) performed better, indicating a better ability to extract discriminative texture patterns from bark images. In addition, the combination of feature vectors extracted by the LBP operators contributed, in some cases, to achieving more accurate classifiers (Table 2).
The C8 and C9 classifiers showed similar accuracy in cross-validation and test set. However, the C8 classifier seems to be more practical in terms of extracting texture features, as it only depends on the combination of the three configurations of the uniform and non-invariant LBP operator ( L B P 8 , 1 u 2 + L B P 16 , 2 u 2 + L B P 24 , 3 u 2 ), with a lower dimensionality of feature vectors. The best classifier in this approach (C8) was developed using multilayer neural networks, whose optimal tuning hyperparameters were: (hidden layer sizes = (250), activation = identity, solver = lbfgs, alpha = 0.0826, learning rate = constant), with an accuracy of 60.30% in cross-validation and 72% in the test set.
The approach of extracting features from sub-images (patches), combined with the majority voting technique, based on the sum of the maximum class probabilities, allowed for constructing more accurate classifiers in most experiments.
In this approach, the C6 classifier, built with sub-image feature vectors extracted by the operator L B P 16 , 2 u 2 and the multilayer neural network algorithm, showed superior performance in cross-validation (accuracy = 81.72% and F1 = 81.55%) and test set (accuracy = 79% and F1 = 79%) (Table 3). The optimal fitting hyperparameters were: ANN (hidden layer sizes = (500), activation = relu, solver = lbfgs, alpha = 0.0859, learning rate = invscaling). Thus, the sub-image approach showed a 21 percent increase in cross-validation accuracy compared to the best classifier (C8) built using original size images.
Using pre-trained convolutional neural networks to extract representative features from bark images contributed to achieving more accurate classifiers than using LBP operators. When using images with original sizes, the classifiers built using ResNet50 as a feature extractor, and ANN (hidden layer sizes = (450,), activation = relu, solver = adam, alpha = 0.0695, learning rate = constant) or SVM (C = 6.9535, kernel = linear) as classification algorithms, showed similar performance in cross-validation and test set (Table 4).
The classifier built using ResNet50 and SVM obtained an increase of 7 percentage points in accuracy in cross-validation, and approximately 10 percentage points in the test set, compared to the C8 classifier ( L B P P , R u 2 and ANN), using images with original sizes.
In the approach using sub-images, an increase of approximately 14 percentage points was observed in the accuracy of cross-validation using the classifier built from the ResNet50 network and SVM, compared to the best results obtained by the LBP operators, with the C6 model ( L B P 16 , 2 u 2 and ANN). This classifier achieved accuracy and F1-score values of 95% with a standard deviation of less than 2%, showing high capacity and stability in predicting the class of new samples (Table 5).
Figure 7a,b present the confusion matrices for the best-performing classifiers: C6 ( L B P 16 , 2 u 2 and ANN) and ResNet50 with SVM, respectively. In the C6 model, Hymenolobium petraeum, Goupia glabra, and Dipteryx odorata showed higher misclassification rates. H. petraeum was mainly confused with Astronium lecointei and Bowdichia nítida; G. glabra with D. odorata, H. petraeum, Protium acrense, and Vatairea sericea; and D. odorata with Mezilaurus itauba. The ResNet50 + SVM classifier exhibited fewer errors, with notable confusion between D. odorata and Apuleia leiocarpa, and Parkia pendula and A. lecointei.
Figure 8 compares species-level recall between classifiers. The C6 model showed generally lower recall, especially for H. petraeum and G. glabra. In contrast, ResNet50 with SVM achieved higher recall across all species, reaching 100% accuracy for six of them.

4. Discussion

4.1. Images Sets: Characteristics, Sources of Variation and Dificulties

The “Amazon Bark” image set includes species exhibiting substantial intra- and interspecific variability in outer bark morphology (Figure 9). Field observations indicate that bark structure is primarily influenced by tree size, with additional variation linked to humidity and diameter, both closely associated with tree age.
Image collection in tropical forests is hindered by operational challenges, notably inconsistent lighting due to understory shading and overexposure in open areas. These factors, coupled with morphological variability, complicate automated species recognition. Therefore, standardized protocols for image acquisition—regarding lighting, camera angle, and distance—are critical for effective field application of recognition systems.
These challenges also affect the reliability of forest inventories under Sustainable Forest Management (SFM), where accurate species identification is required for legal compliance and biodiversity monitoring. In Brazil, field inventories must comply with official protocols (e.g., IN n 5/2006, IBAMA), and misidentification can result in illegal harvesting or misclassification of protected species.

4.2. Local Binary Standards

Although rotation-invariant uniform ( L B P P , R r i u 2 ) is generally robust for texture classification, the non-invariant variant ( L B P P , R u 2 ) outperformed it in this study, suggesting that bark texture orientation has limited influence on species discrimination, consistent with previous wood image classification research [11,17,59]. Early fusion of multiple LBP descriptors further enhanced feature representation and classification accuracy [60]; for example, L B P 24 , 3 u 2 alone achieved 66% accuracy, while combining L B P 8 , 1 u 2 , L B P 16 , 2 u 2 and L B P 24 , 3 u 2 improved accuracy by 6%.
The practical relevance of the superior performance of non-rotation-invariant uniform LBP lies in its computational efficiency and suitability for field applications. LBP-based models, particularly those applied to sub-image patches, demonstrated reasonably high accuracy and potential for rapid, low-cost species recognition in forest inventories. These models are especially advantageous in remote settings with limited computational resources, where deploying deep learning may be impractical.
However, their susceptibility to misclassification due to similar bark textures and environmental noise highlights the need for standardized imaging and proper lighting—challenges common in dense Amazonian forests. While promising for preliminary species filtering or flagging ambiguous cases, these models may require integration with morphological features or expert validation in regulatory frameworks.
From a policy perspective, LBP models could facilitate swift assessments during timber harvesting inspections, enhancing traceability and enforcement under Brazil’s Forest Code. By identifying likely misclassifications early, such tools can help forest agents and concession managers prioritize verification efforts, alleviating pressure on limited human resources in environmental monitoring.

4.3. Transfer Learning

Numerous studies have confirmed the efficacy of pre-trained convolutional neural networks (CNNs) for plant classification across various structures, including leaves [47,61], wood [34,46,62,63], bark [18,64], and charcoal [39,65]. However, deep learning applications using bark images remain scarce [66], likely due to task complexity and limited access to large, high-quality datasets.
As established in the literature, larger datasets enhance model generalization and mitigate overfitting [67,68]. In this study, extracting 256 × 256 pixel sub-images from high-resolution bark images significantly improved classification accuracy across all CNN architectures, likely due to increased data volume and improved learning from fine-scale features.
This sub-image strategy, combined with majority voting, has also proven effective in related domains. [39] attained 95.7% accuracy with charcoal images from 44 forest species, and [69] reported 93% with wood images from 14 European species using similar methods. In the present study, ResNet50 coupled with either Artificial Neural Networks (ANN) or Support Vector Machines (SVM) yielded the best results, corroborating prior plant imaging research [70,71,72].
While dense neural networks are frequently applied to bark classification with strong performance [18,66,73,74], SVM slightly outperformed them in this study. Notably, Linear Discriminant Analysis (LDA) offered competitive accuracy (92%) with significantly reduced training time—two hours compared to over six for SVM—highlighting an efficient alternative [75].
Within the context of Sustainable Forest Management (SFM) in the Brazilian Amazon, this work marks a significant advancement toward a robust, generalizable bark-based species recognition system. The proposed model demonstrated high accuracy and holds potential to reduce identification errors during forest inventories, particularly for the 16 commercially important species assessed.

4.4. Implications for Sustainable Forest Management and Future Perspectives

Although the models achieved high classification performance, some misclassifications were noted, particularly in those using LBP descriptors. Visual similarities in bark traits—such as reticulated rhytidome and grey to brown color variation [76]—likely contributed to these errors, as did the presence of lichens. Comparable misclassifications occurred in ResNet50 and SVM models (Figure 10a,b).
These findings highlight how morphological similarities, shaped by microenvironmental factors, can challenge species differentiation by automated classifiers. Environmental and structural factors—such as edaphoclimatic conditions, forest typology, and the sociological position of trees—are known to influence bark morphology.
These sources of variability highlight the need to expand both the taxonomic and morphological diversity of training datasets. Constructing a more generalizable recognition system requires not only the inclusion of additional species, but also a broader sampling effort that captures individual variability within species. Sampling strategies based on the diameter distribution of each species could improve representativeness, though this remains a complex task in Amazonian forests.
From a practical perspective, the integration of automated species recognition tools into Sustainable Forest Management (SFM) operations offers substantial advantages. By reducing taxonomic errors during forest inventories, such systems enhance the accuracy of management plans, strengthen compliance with legal frameworks—such as Brazil’s Forest Code—and mitigate the risks associated with unauthorized logging. Moreover, these technologies hold significant potential to support biodiversity monitoring and long-term ecological assessments by enabling large-scale, standardized species identification. It is also important to emphasize that field image acquisition would not entail additional operational costs, as data collection activities during forest inventories are legally mandated and already incorporated into the standard procedures of forest management operations.
To ensure operational viability, future work should prioritize:
  • Expansion of datasets to include regional and structural bark variability;
  • Validation of models using mobile devices under field conditions;
  • Integration of other plant organs (e.g., leaves, fruits) for multimodal classification;
  • Development of lightweight architectures for deployment in remote forest environments.
Considering that field validation represents a fundamental step for the practical application of the proposed system, future studies may explore established strategies to adapt deep learning models for deployment on mobile and edge devices. Among these strategies, techniques such as model compression and pruning are noteworthy, including the structured removal of redundant filters or neurons, with the aim of reducing computational complexity and inference time while maintaining satisfactory levels of accuracy. Additional approaches include compression methods based on weight clustering or low-rank factorization, as well as post-training quantization, which converts floating-point weights into lower-precision integer representations, thereby reducing memory usage and computational demands.
Moreover, lightweight convolutional neural network architectures specifically designed for resource-constrained environments—such as MobileNetV3, EfficientNet-Lite, and SqueezeNet—may be considered, given their favorable trade-off between accuracy and computational efficiency. Finally, knowledge distillation is highlighted as a complementary approach, wherein compact “student” models are trained to replicate the behavior of larger, more robust “teacher” models, allowing for the preservation of high performance while significantly reducing computational requirements.
This study also contributes to the operationalization of automated tools for SFM. Reliable species recognition supports legal compliance, inventory planning, and biodiversity monitoring. While current limitations include a restricted number of species and high computational demand, the system shows clear potential for practical use in forest governance.

5. Conclusions

This study represents a step forward in automated species identification within the framework of Sustainable Forest Management (SFM) in the Brazilian Amazon, where accurate botanical identification is critical yet often hindered by operational and taxonomic limitations. By using bark sub-image analysis and deep learning models—particularly ResNet50 combined with classifiers such as SVM and ANN—the system achieved up to 95% accuracy in identifying 16 commercial timber species, demonstrating both technical feasibility and field scalability.
Despite these promising results, practical deployment requires expanding species coverage, validating models in real-world settings, and developing lightweight versions suitable for mobile devices. Integration with official forest inventory platforms and updated taxonomic databases could enhance traceability within the timber supply chain, reduce identification errors, and support compliance with environmental regulations and international commitments. Thus, the proposed models show strong potential as complementary tools for data-driven tropical forest management.

Author Contributions

Conceptualization, N.C.G., L.E.S.O., S.d.P.C.e.C. and D.V.S.; methodology, N.C.G., L.E.S.O., S.d.P.C.e.C. and D.V.S.; software, N.C.G., L.E.S.O. and D.V.S.; validation, N.C.G., L.E.S.O., S.d.P.C.e.C., A.B., P.L.d.P.F., E.d.S.L. and D.V.S.; formal analysis, N.C.G., L.E.S.O., S.d.P.C.e.C., A.B., P.L.d.P.F., E.d.S.L. and D.V.S.; investigation, N.C.G., S.d.P.C.e.C., M.O.d.S.H., E.d.S.L. and D.V.S.; resources, S.d.P.C.e.C., M.O.d.S.H. and E.d.S.L.; data curation, N.C.G. and D.V.S.; writing—original draft preparation, N.C.G.; writing—review and editing, N.C.G., L.E.S.O., S.d.P.C.e.C., A.B., P.L.d.P.F. and D.V.S.; visualization, N.C.G., L.E.S.O. and D.V.S.; supervision, L.E.S.O., S.d.P.C.e.C., A.B. and D.V.S.; project administration, S.d.P.C.e.C. and D.V.S.; funding acquisition, S.d.P.C.e.C. and D.V.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Center of Wood Producing and Exporting Industries of the State of Mato Grosso (CWPEISMG), with administrative and financial support from the Foundation for Research and Development Support (FADESP), and by FAPEMAT, grant number PRO.000091/2023.

Data Availability Statement

The dataset used in this study is not yet publicly available, as it represents the first stage of the “Deep Flora” project. In future stages, the dataset is expected to be expanded to include a greater diversity of flora species and subsequently made publicly accessible. However, the data may be provided by the corresponding author upon reasonable request. All source codes developed during the research are openly available at: https://github.com/NatallyCelestino/Bark_Recogition.git (accessed on 18 August 2025).

Acknowledgments

The authors would like to thank the Center of Wood Producing and Exporting Industries of the State of Mato Grosso (CWPEISMG) for their support. We also acknowledge the Mato Grosso State Research Support Foundation (FAPEMAT) for providing financial support under grant number PRO.000091/2023.

Conflicts of Interest

The authors declare they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Appendix A

Table A1. Candidate hyperparameters for each machine learning algorithm.
Table A1. Candidate hyperparameters for each machine learning algorithm.
AlgorithmCandidate Hyperparameters
Artificial neural networkshidden_layer_sizes =
(50), (100), (150), (200), (250), (300), (350), (400), (450) e (500)
activation = relu e identity
solver = adam e lbfgs
alpha = uniform(loc = 0.0001, scale = 0.09).rvs(size = 20, random_state = 10)
learning_rate = constant, adaptive e invscaling
Support vector machineC = uniform(loc = 0.1, scale = 10).rvs(size = 20, random_state = 10)
kernel = linear, rbf, poly e sigmoid
degree = 2, 3 e 4
gamma = scale e auto + list(np.logspace(−9, 3, 13)
Random forestn_estimators = np.arange(40, 320, 20)
max_depth = list(np.arange(10, 100, step = 10)) + [None]
max_features = list(np.arange(30, 60, 5)) + [‘sqrt’, “log2”]
criterion = gini e entropy
min_samples_leaf = np.arange(10, 110, 10)
min_samples_split = np.arange(2, 10, 2)
bootstrap = True e False
Linear discriminant analysissolver = lsqr e eigen
tol = 0.0001, 0.0002 e 0.0003
Where hidden_layer_sizes—number of neurons in each hidden layer; activation—activation function; solver—optimization algorithm used to update the network weights during training; alpha—penalty or regularization applied to the network weights; learning_rate—learning rate; C—controls the regularization of the model; kernel—type of function to be used in the algorithm to map the data to a higher-dimensional feature space; degree—used only for the polynomial kernel defining the degree of the polynomial; gamma—used for all kernels except linear; n_estimators—number of trees in the forest; max_depth—maximum depth of each tree; max_features—maximum number of features; criterion—defines the function to measure the quality of the split in a decision tree; min_samples_leaf—minimum number of samples needed to be a leaf in a tree; min_samples_split—minimum number of samples needed to split a node; bootstrap—indicates whether the samples are sampled with replacement (True) or not (False); tol—tolerance parameter.

References

  1. Loyola, R.; Machado, N.; Nova, D.V.; Martins, E.; Martinelli, G. Áreas Prioritárias Para Conservação e Uso Sustentável Da Flora Brasileira Ameaçada de Extinção; Instituto de Pesquisas Jardim Botânico do Rio de Janeiro (JBRJ): Rio de Janeiro, Brazil, 2014; ISBN 978-85-88742-67-3. Available online: https://dspace.jbrj.gov.br/jspui/handle/doc/28 (accessed on 14 February 2024).
  2. The Brazil Flora Group. Flora Do Brasil 2020. Jard. Botânico do Rio Jan. 2021. [Google Scholar] [CrossRef]
  3. de Melo, R.R.; Rocha, M.J.; Rodolfo Junior, F.; Stangerlin, D.M. Análise Da Influência Do Diâmetro No Rendimento Em Madeira Serrada de Cambará (Qualea sp.). Pesqui. Florest. Bras. 2017, 36, 393. [Google Scholar] [CrossRef]
  4. Stragliotto, M.C.; Pereira, B.L.C.; Oliveira, A.C. Indústrias Madeireiras E Rendimento Em Madeira Serrada Na Amazônia Brasileira. In Engenharia Florestal: Desafios, Limites e Potencialidade; Editora Científica Digital Ltda: São Paulo, Brazil, 2020; pp. 499–518. [Google Scholar] [CrossRef]
  5. CONAMA Brasil, Ministério Do Meio Ambiente. Resolução CONAMA N° 406 de 02 de Fevereiro de 2009; CONAMA Brasil, Ministério do Meio Ambiente: Brasília, Brazil, 2009; pp. 1–5.
  6. Hadlich, H.L.; Durgante, F.M.; dos Santos, J.; Higuchi, N.; Chambers, J.Q.; Vicentini, A. Recognizing Amazonian Tree Species in the Field Using Bark Tissues Spectra. For. Ecol. Manag. 2018, 427, 296–304. [Google Scholar] [CrossRef]
  7. Encinas, J.I.; Antônio, M.; Ferreira, C.; Riesco, G.; Alboreca, A.R. Introduction Forest Management in the Sustained Yield Regime Considers That in the Management. Rev. Mex. Cienc. For. 2022, 13, 30–54. [Google Scholar] [CrossRef]
  8. Engel, J.; Brousseau, L.; Baraloto, C. GuiaTreeKey, a Multi-Access Electronic Key to Identify Tree Genera in French Guiana. PhytoKeys 2016, 68, 27–44. [Google Scholar] [CrossRef]
  9. Lang, C.; Almeida, D.R.A.; Costa, F.R.C. Discrimination of Taxonomic Identity at Species, Genus and Family Levels Using Fourier Transformed Near-Infrared Spectroscopy (FT-NIR). For. Ecol. Manag. 2017, 406, 219–227. [Google Scholar] [CrossRef]
  10. Procópio, L.C.; Secco, R.D.S. A Importância Da Identificação Botânica Nos Inventários Florestais: O Exemplo Do “Tauari” (Couratari spp. e Cariniana spp.—Lecythidaceae) Em Duas Áreas Manejadas No Estado Do Pará. Acta Amaz. 2008, 38, 31–44. [Google Scholar] [CrossRef]
  11. Venicio, D.; Joielan, S.; Santos, X.; Cristina, H.; Lorena, T.; Silvana, N.; Luiz, N.; Oliveira, E.S. An Automatic Recognition System of Brazilian Flora Species Based on Textural Features of Macroscopic Images of Wood. Wood Sci. Technol. 2020, 54, 1065–1090. [Google Scholar] [CrossRef]
  12. Menezes, M.; Bicudo, C.E.M.; Moura, C.W.N.; Alves, A.M.; Santos, A.A.; Pedrini, A.d.G.; Araújo, A.; Tucci, A.; Fajar, A.; Malone, C.; et al. Update of the Brazilian Floristic List of Algae and Cyanobacteria. Rodriguésia 2015, 66, 1047–1062. [Google Scholar] [CrossRef]
  13. Novaes1, T.V.; Ramalho, F.M.G.; da Silva Araujo, E.; Lima, M.D.R.; da Silva, M.G.; Ferreira, G.C.; Hein, P.R.G. Discrimination of Amazonian Forest Species by NIR Spectroscopy: Wood Surface Effects. Eur. J. Wood Wood Prod. 2023, 81, 159–172. [Google Scholar] [CrossRef]
  14. de Lacerda, A.E.B.; Nimmo, E.R. Can We Really Manage Tropical Forests without Knowing the Species within? Getting Back to the Basics of Forest Management through Taxonomy. For. Ecol. Manag. 2010, 259, 995–1002. [Google Scholar] [CrossRef]
  15. Cysneiros, V.C.; Mendonça Júnior, J.O.; Lanza, T.R.; Moraes, J.C.R.; Samor, O.J.M. Espécies Madeireiras Da Amazônia: Riqueza, Nomes Populares e Suas Peculiaridades. Pesqui. Florest. Bras. 2018, 38, 1–14. [Google Scholar] [CrossRef]
  16. Ferreira, R.L.A.; Cerqueira, R.M.; Cardoso Junior, R.C. Análise Da Identificação Botânica Em Inventários Florestais de Planos de Manejo Sustentáveis No Oeste Paraense. Nat. Conserv. 2020, 13, 136–145. [Google Scholar] [CrossRef]
  17. Filho, P.L.P.; Oliveira, L.S.; Nisgoski, S.; Britto, A.S. Forest Species Recognition Using Macroscopic Images. Mach. Vis. Appl. 2014, 25, 1019–1031. [Google Scholar] [CrossRef]
  18. Carpentier, M.; Giguere, P.; Gaudreault, J. Tree Species Identification from Bark Images Using Convolutional Neural Networks. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1075–1081. [Google Scholar]
  19. Gogul, I.; Kumar, V.S. Flower Species Recognition System Using Convolution Neural Networks and Transfer Learning. In Proceedings of the 2017 Fourth International Conference on Signal Processing, Communication and Networking (ICSCN), Chennai, India, 16–18 March 2017; pp. 1–6. [Google Scholar]
  20. Ghosh, A.; Roy, P. An Automated Model for Leaf Image-Based Plant Recognition: An Optimal Feature-Based Machine Learning Approach. Innov. Syst. Softw. Eng. 2022, 20, 705–718. [Google Scholar] [CrossRef]
  21. Ran, J.; Shi, Y.; Yu, J.; Li, D. A Multi-Feature Convolution Neural Network for Automatic Flower Recognition. J. Circuits Syst. Comput. 2021, 30, 1–15. [Google Scholar] [CrossRef]
  22. Anubha Pearline, S.; Sathiesh Kumar, V.; Harini, S. A Study on Plant Recognition Using Conventional Image Processing and Deep Learning Approaches. J. Intell. Fuzzy Syst. 2019, 36, 1997–2004. [Google Scholar] [CrossRef]
  23. Kaya, A.; Keceli, A.S.; Catal, C.; Yalic, H.Y.; Temucin, H.; Tekinerdogan, B. Analysis of Transfer Learning for Deep Neural Network Based Plant Classification Models. Comput. Electron. Agric. 2019, 158, 20–29. [Google Scholar] [CrossRef]
  24. Kumar, M. Plant Species Recognition Using Morphological Features and Adaptive Boosting Methodology. IEEE Access 2019, 7, 163912–163918. [Google Scholar] [CrossRef]
  25. Turkoglu, M.; Hanbay, D. Leaf-Based Plant Species Recognition Based on Improved Local Binary Pattern and Extreme Learning Machine. Phys. Stat. Mech. Its Appl. 2019, 527, 121297. [Google Scholar] [CrossRef]
  26. Keivani, M.; Mazloum, J.; Sedaghatfar, E.; Tavakoli, M.B. Automated Analysis of Leaf Shape, Texture, and Color Features for Plant Classification. Trait. Signal 2020, 37, 17–28. [Google Scholar] [CrossRef]
  27. Bisen, D. Deep Convolutional Neural Network Based Plant Species Recognition through Features of Leaf. Multimed. Tools Appl. 2021, 80, 6443–6456. [Google Scholar] [CrossRef]
  28. Wang, B.; Li, H.; You, J.; Chen, X.; Yuan, X.; Feng, X. Fusing Deep Learning Features of Triplet Leaf Image Patterns to Boost Soybean Cultivar Identification. Comput. Electron. Agric. 2022, 197, 106914. [Google Scholar] [CrossRef]
  29. Misra, D.; Crispim-Junior, C.; Tougne, L. Patch-Based CNN Evaluation for Bark Classification. In Proceedings of the ECCV 2020: European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 197–212, ISBN 9783030654146. [Google Scholar]
  30. Remeš, V.; Haindl, M. Bark Recognition Using Novel Rotationally Invariant Multispectral Textural Features. Pattern Recognit. Lett. 2019, 125, 612–617. [Google Scholar] [CrossRef]
  31. Bertrand, S.; Cerutti, G.; Tougne, L. Bark Recognition to Improve Leaf-Based Classification in Didactic Tree Species Identification. In Proceedings of the 12th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, Porto, Portugal, 27 February–1 March 2017; SCITEPRESS—Science and Technology Publications: Setubal, Portugal, 2017; Volume 4, pp. 435–442. [Google Scholar]
  32. de Geus, A.R.; Backes, A.R.; Gontijo, A.B.; Albuquerque, G.H.Q.; Souza, J.R. Amazon Wood Species Classification: A Comparison between Deep Learning and Pre-Designed Features. Wood Sci. Technol. 2021, 55, 857–872. [Google Scholar] [CrossRef]
  33. Taslim, A.; Saon, S.; Mahamad, A.K.; Muladi, M.; Hidayat, W.N. Plant Leaf Identification System Using Convolutional Neural Network. Bull. Electr. Eng. Inform. 2021, 10, 3341–3352. [Google Scholar] [CrossRef]
  34. He, J.; Sun, Y.; Yu, C.; Cao, Y.; Zhao, Y.; Du, G. An Improved Wood Recognition Method Based on the One-Class Algorithm. Forests 2022, 13, 1350. [Google Scholar] [CrossRef]
  35. Thanikkal, J.G.; Dubey, A.K.; Thomas, M.T. An Efficient Mobile Application for Identification of Immunity Boosting Medicinal Plants Using Shape Descriptor Algorithm. Wirel. Pers. Commun. 2023, 131, 1189–1205. [Google Scholar] [CrossRef]
  36. da Silveira, A.B.; Santos, J.d.P.d.; Rebellato, L. Guia de Boas Práticas: Restauração de Áreas de Preservação Permanente Degradadas (APPDs), Experiência Da Fazenda São Nicolau; Editora PETRA: Cuiabá, Brazil, 2017; ISBN 978-85-94211-00-2. [Google Scholar]
  37. Borges, H.B.N.; Silveira, E.A.; Vendramin, L.N. Flora Arbórea de Mato Grosso—Tipologia Vegetais e Suas Éspecies; Editora Entrelinhas: Cuiabá, Brazil, 2014; ISBN 978-85-7992-072-1. [Google Scholar]
  38. Benini, R.; Santana, P.; Borgo, M.; Girão, V.; Campos, M.; Klein, F.; Kummer, O.P.; Netto, D.S.d.A.; Rodrigues, R.R.; Nave, A.G.; et al. Manual De Restauração Da Vegetação Nativa, Alto Teles Pires, Mt Expediente; Editora The Nature Conservancy (TNC): Brasília, Brazil, 2016; 136p. [Google Scholar]
  39. Maruyama, T.M.; Oliveira, L.S.; Britto, A.S.; Nisgoski, S. Automatic Classification of Native Wood Charcoal. Ecol. Inform. 2018, 46, 1–7. [Google Scholar] [CrossRef]
  40. Robert, M.; Dallaire, P.; Giguere, P. Tree Bark Re-Identification Using a Deep-Learning Feature Descriptor. In Proceedings of the 2020 17th Conference on Computer and Robot Vision (CRV), Ottawa, ON, Canada, 13–15 May 2020; pp. 25–32. [Google Scholar] [CrossRef]
  41. Ojala, T.; Pietikhenl, M.; Harwood, D. Performance Evaluation of Texture Measures with Classification Based on Kullback Discrimination of Distributions. In Proceedings of the 12th International Conference on Pattern Recognition, Jerusalem, Israel, 9–13 October 1994; pp. 582–585. [Google Scholar]
  42. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary Patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  43. Li, J.; Sun, S.; Jiang, H.; Tian, Y.; Xu, X. Image Recognition and Empirical Application of Desert Plant Species Based on Convolutional Neural Network. J. Arid Land 2022, 14, 1440–1455. [Google Scholar] [CrossRef]
  44. Huang, Z.; He, C.; Wang, Z.-N.; Xi, J.; Wang, H.; Hou, L. Cinnamomum Camphora Classification Based on Leaf Image Using Transfer Learning. In Proceedings of the 2019 IEEE 4th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chengdu, China, 20–22 December 2019; pp. 1426–1429. [Google Scholar]
  45. Cai, X.; Huo, Y.; Chen, Y.; Xi, M.; Tu, Y.; Sun, C.; Sun, H. Real-Time Leaf Recognition Method Based on Image Segmentation and Feature Extraction. Int. J. Pattern Recognit. Artif. Intell. 2022, 36, 1–24. [Google Scholar] [CrossRef]
  46. Figueroa-Mata, G.; Mata-Montero, E.; Valverde-Otárola, J.C.; Arias-Aguilar, D.; Zamora-Villalobos, N. Using Deep Learning to Identify Costa Rican Native Tree Species from Wood Cut Images. Front. Plant Sci. 2022, 13, 789227. [Google Scholar] [CrossRef] [PubMed]
  47. Zhang, R.; Zhu, Y.; Ge, Z.; Mu, H.; Qi, D.; Ni, H. Transfer Learning for Leaf Small Dataset Using Improved ResNet50 Network with Mixed Activation Functions. Forests 2022, 13, 2072. [Google Scholar] [CrossRef]
  48. Soni, P.; Dhavale, S.; Yenishetti, S.; Panat, L.; Karajkhede, G. Medicinal Plant Species Identification Using AI. In Proceedings of the 2023 IEEE 11th Region 10 Humanitarian Technology Conference (R10-HTC), Rajkot, India, 16–18 October 2023; pp. 662–668. [Google Scholar]
  49. Kini, A.S.; Prema, K.V.; Pai, S.N. Early Stage Black Pepper Leaf Disease Prediction Based on Transfer Learning Using ConvNets. Sci. Rep. 2024, 14, 1404. [Google Scholar] [CrossRef]
  50. Faisal, S.; Javed, K.; Ali, S.; Alasiry, A.; Marzougui, M.; Khan, M.A.; Cha, J.H. Deep Transfer Learning Based Detection and Classification of Citrus Plant Diseases. Comput. Mater. Contin. 2023, 76, 895–914. [Google Scholar] [CrossRef]
  51. Thirumala, K.; Pal, S.; Jain, T.; Umarikar, A.C. Neurocomputing A Classification Method for Multiple Power Quality Disturbances Using EWT Based Adaptive Filtering and Multiclass SVM. Neurocomputing 2019, 334, 265–274. [Google Scholar] [CrossRef]
  52. Toğaçar, M.; Ergen, B.; Cömert, Z. Classification of Flower Species by Using Features Extracted from the Intersection of Feature Selection Methods in Convolutional Neural Network Models. Measurement 2020, 158, 107703. [Google Scholar] [CrossRef]
  53. Dourado Filho, L.A.; Calumby, R.T. Data Augmentation Policies and Heuristics Effects over Dataset Imbalance for Developing Plant Identification Systems Based on Deep Learning: A Case Study. Rev. Bras. Comput. Apl. 2022, 14, 85–94. [Google Scholar] [CrossRef]
  54. Nunes, M.H.; Görgens, E.B. Artificial Intelligence Procedures for Tree Taper Estimation within a Complex Vegetation Mosaic in Brazil. PLoS ONE 2016, 11, e0154738. [Google Scholar] [CrossRef]
  55. Fan, G.; Yu, M.; Dong, S.; Yeh, Y.; Hong, W. Forecasting Short-Term Electricity Load Using Hybrid Support Vector Regression with Grey Catastrophe and Random Forest Modeling. Util. Policy 2021, 73, 101294. [Google Scholar] [CrossRef]
  56. Sahu, S.K.; Pandey, M. An optimal hybrid multiclass SVM for plant leaf disease detection using spatial fuzzy C-means model. Expert Syst. Appl. 2023, 214, 118989. [Google Scholar] [CrossRef]
  57. da Rosa, A.C.F.; Galdamez, E.V.C.; de Souza, R.C.T.; Melo, M.d.G.M.; Villarinho, A.L.C.F.; Leal, G.C.L. Uso de Técnicas de Aprendizado de Máquina Para Classificação de Fatores Que Influenciam a Ocorrência de Dermatites Ocupacionais. Rev. Bras. Saúde Ocup. 2023, 48, 1–10. [Google Scholar] [CrossRef]
  58. Wu, Q.; Luo, J.; Fang, H.; He, D.; Liang, T. Spectral Classification Analysis of Recycling Plastics of Small Household Appliances Based on Infrared Spectroscopy. Vib. Spectrosc. 2024, 130, 103636. [Google Scholar] [CrossRef]
  59. Rahiddin, R.N.N.; Hashim, U.R.; Salahuddin, L.; Kanchymalay, K.; Wibawa, A.P.; Chun, T.H. Local Texture Representation for Timber Defect Recognition Based on Variation of LBP. Int. J. Adv. Comput. Sci. Appl. 2022, 13, 443–448. [Google Scholar] [CrossRef]
  60. Tortorici, C.; Werghi, N. Early Features Fusion over 3D Face for Face Recognition. In Communications in Computer and Information Science; Ben Amor, B., Chaieb, F., Ghorbel, F., Eds.; Springer International Publishing: Cham, Switzerland, 2017; Volume 684, pp. 56–64. ISBN 978-3-319-60653-8. [Google Scholar]
  61. Vizcarra, G.; Bermejo, D.; Mauricio, A.; Zarate Gomez, R.; Dianderas, E. The Peruvian Amazon Forestry Dataset: A Leaf Image Classification Corpus. Ecol. Inform. 2021, 62, 101268. [Google Scholar] [CrossRef]
  62. Ravindran, P.; Owens, F.C.; Wade, A.C.; Shmulsky, R.; Wiedenhoeft, A.C. Towards Sustainable North American Wood Product Value Chains, Part I: Computer Vision Identification of Diffuse Porous Hardwoods. Front. Plant Sci. 2022, 12, 758455. [Google Scholar] [CrossRef]
  63. Yang, Y.; Wang, H.; Jiang, D.; Hu, Z. Surface Detection of Solid Wood Defects Based on SSD Improved with ResNet. Forests 2021, 12, 1419. [Google Scholar] [CrossRef]
  64. Kim, T.K.; Hong, J.; Ryu, D.; Kim, S.; Byeon, S.Y.; Huh, W.; Kim, K.; Baek, G.H.; Kim, H.S. Identifying and Extracting Bark Key Features of 42 Tree Species Using Convolutional Neural Networks and Class Activation Mapping. Sci. Rep. 2022, 12, 4772. [Google Scholar] [CrossRef]
  65. Menon, L.T.; Laurensi, I.A.; Penna, M.C.; Oliveira, L.E.S.; Britto, A.S. Data Augmentation and Transfer Learning Applied to Charcoal Image Classification. In Proceedings of the 2019 International Conference on Systems, Signals and Image Processing (IWSSIP), Osijek, Croatia, 5–7 June 2019; pp. 69–74. [Google Scholar]
  66. Cui, L.; Chen, S.; Mu, Y.; Xu, X.; Zhang, B.; Zhao, X. Tree Species Classification over Cloudy Mountainous Regions by Spatiotemporal Fusion and Ensemble Classifier. Forests 2023, 14, 107. [Google Scholar] [CrossRef]
  67. Khosla, C.; Saini, B.S. Enhancing Performance of Deep Learning Models with Different Data Augmentation Techniques: A Survey. In Proceedings of the 2020 International Conference on Intelligent Engineering and Management (ICIEM), London, UK, 17–19 June 2020. [Google Scholar] [CrossRef]
  68. Nanni, L.; Paci, M.; Brahnam, S.; Lumini, A. Comparison of Different Image Data Augmentation Approaches. J. Imaging 2021, 7, 254. [Google Scholar] [CrossRef]
  69. Fabijańska, A.; Danek, M.; Barniak, J. Wood Species Automatic Identification from Wood Core Images with a Residual Convolutional Neural Network. Comput. Electron. Agric. 2021, 181, 105941. [Google Scholar] [CrossRef]
  70. Durairajah, V.; Gobee, S.; Muneer, A. Automatic Vision Based Classification System Using DNN and SVM Classifiers. In Proceedings of the 2018 3rd International Conference on Control, Robotics and Cybernetics (CRC), Penang, Malaysia, 26–28 September 2018; pp. 6–14. [Google Scholar]
  71. Zhao, Y.; Gao, X.; Hu, J.; Chen, Z. Tree Species Identification Based on the Fusion of Bark and Leaves. Math. Biosci. Eng. 2020, 17, 4018–4033. [Google Scholar] [CrossRef] [PubMed]
  72. Diwedi, H.K.; Misra, A.; Tiwari, A.K. CNN-based Medicinal Plant Identification and Classification Using Optimized SVM. Multimed. Tools Appl. 2024, 83, 33823–33853. [Google Scholar] [CrossRef]
  73. Ido, J.; Saitoh, T. CNN-Based Tree Species Identification from Bark Image. In Proceedings of the Tenth International Conference on Graphics and Image Processing (ICGIP 2018), Chengdu, China, 12–14 December 2018; Yu, H., Pu, Y., Li, C., Pan, Z., Eds.; SPIE: Philadelphia, PA, USA, 6 May 2019; p. 63. [Google Scholar]
  74. Faizal, S. Automated Identification of Tree Species by Bark Texture Classification Using Convolutional Neural Networks. Int. J. Res. Appl. Sci. Eng. Technol. 2022, 10, 1384–1392. [Google Scholar] [CrossRef]
  75. Sachar, S.; Kumar, A. Survey of Feature Extraction and Classification Techniques to Identify Plant through Leaves. Expert Syst. Appl. 2021, 167, 114181. [Google Scholar] [CrossRef]
  76. Ferreira, C.A.; Silva, H.D. da Eucalyptus Para a Região Amazônica, Estados de Rondônia e Acre. Embrapa Florestas 2004, 1, 1–4. [Google Scholar]
Figure 1. Location of the three Demonstrative Units (DUs) of Forest Management in the state of Mato Grosso, Brazil—Farm Pérola (Nova Maringá), DU2—Farm Boa Esperança (Feliz Natal), and DU3—Farm São Nicolau (Cotriguaçu).
Figure 1. Location of the three Demonstrative Units (DUs) of Forest Management in the state of Mato Grosso, Brazil—Farm Pérola (Nova Maringá), DU2—Farm Boa Esperança (Feliz Natal), and DU3—Farm São Nicolau (Cotriguaçu).
Forests 16 01374 g001
Figure 2. Representation of the collection procedures for the construction of the set of bark images of commercial trees of Amazonian forest species. (a) Ten individuals were sampled for each species; (b) For each individual, bark images were acquired using two different devices (a Canon camera and an iPhone 11, iOS) from multiple circumferential positions around the trunk, in order to build the bark image dataset.
Figure 2. Representation of the collection procedures for the construction of the set of bark images of commercial trees of Amazonian forest species. (a) Ten individuals were sampled for each species; (b) For each individual, bark images were acquired using two different devices (a Canon camera and an iPhone 11, iOS) from multiple circumferential positions around the trunk, in order to build the bark image dataset.
Forests 16 01374 g002
Figure 3. (A) Low-light area; (B) use of the reflector.
Figure 3. (A) Low-light area; (B) use of the reflector.
Forests 16 01374 g003
Figure 4. Sample images of the bark of the 16 Amazonian forest species collected.
Figure 4. Sample images of the bark of the 16 Amazonian forest species collected.
Forests 16 01374 g004
Figure 5. (A) Climbing and collecting botanical material; (B) assembling the exsiccates.
Figure 5. (A) Climbing and collecting botanical material; (B) assembling the exsiccates.
Forests 16 01374 g005
Figure 6. Variations of the LBP operator with different circular neighborhood configurations. The number of neighbors (P) and the radius (R) determine the sampling pattern around the central pixel (xc). Each configuration defines the spatial resolution and scale at which local texture patterns are captured. Neighboring pixels (xp) are sampled along a circular path centered on xc, with interpolation applied when necessary.
Figure 6. Variations of the LBP operator with different circular neighborhood configurations. The number of neighbors (P) and the radius (R) determine the sampling pattern around the central pixel (xc). Each configuration defines the spatial resolution and scale at which local texture patterns are captured. Neighboring pixels (xp) are sampled along a circular path centered on xc, with interpolation applied when necessary.
Forests 16 01374 g006
Figure 7. (a) Confusion matrix in the C6 classifier test set, constructed using feature vectors of sub-images extracted by the operator L B P 16 , 2 u 2 and the multilayer neural network algorithm. (b) Confusion matrix in the test set for the classifier built using ResNet50 as the sub-image feature extractor and SVM as the classification algorithm.
Figure 7. (a) Confusion matrix in the C6 classifier test set, constructed using feature vectors of sub-images extracted by the operator L B P 16 , 2 u 2 and the multilayer neural network algorithm. (b) Confusion matrix in the test set for the classifier built using ResNet50 as the sub-image feature extractor and SVM as the classification algorithm.
Forests 16 01374 g007
Figure 8. Recognition rate per species (recall), on the test set, for the classifiers ( L B P 16 , 2 u 2 and ANN—classifier C6) and ResNet50 with SVM, built from sub-images.
Figure 8. Recognition rate per species (recall), on the test set, for the classifiers ( L B P 16 , 2 u 2 and ANN—classifier C6) and ResNet50 with SVM, built from sub-images.
Forests 16 01374 g008
Figure 9. Images of different trees of Trattinnickia burserifolia (AC) and Parkia pendula (DF).
Figure 9. Images of different trees of Trattinnickia burserifolia (AC) and Parkia pendula (DF).
Forests 16 01374 g009
Figure 10. (a) Images of the species Hymenolobium petraeum that were misclassified as Bowdichia nítida and Astronium lecointei by the C6 classifier ( L B P 16 ,   2 u 2 e ANN); (b) Images of the species Hymenolobium petraeum and Dipteryx odorata that were misclassified as Apuleia leiocarpa by the classifier built using ResNet50 and SVM.
Figure 10. (a) Images of the species Hymenolobium petraeum that were misclassified as Bowdichia nítida and Astronium lecointei by the C6 classifier ( L B P 16 ,   2 u 2 e ANN); (b) Images of the species Hymenolobium petraeum and Dipteryx odorata that were misclassified as Apuleia leiocarpa by the classifier built using ResNet50 and SVM.
Forests 16 01374 g010
Table 1. Family, scientific name, vernacular name, number of samples, and images (n) per species.
Table 1. Family, scientific name, vernacular name, number of samples, and images (n) per species.
ID 1FamilyScientific NameVernacular NameSamplesN 2
1FabaceaeApuleia leiocarpa (Vogel) J. F. Macbr.Garapeira10155
2AnacardiaceaeAstronium lecointei DuckeMuiracatiara10182
3MoraceaeBagassa guianensis Aubl.Tatajuba10185
4FabaceaeBowdichia nitida Spruce ex Benth.Sucupira10194
5MeliaceaeCedrela odorata L.Cedro-Rosa10198
6FabaceaeDipteryx odorata (Aubl.) Forsyth f.Cumaru10149
7VochysiaceaeErisma uncinatum Warm.Cedrinho10158
8GoupiaceaeGoupia glabra Aubl.Cupiúba10162
9FabaceaeHymenelobium petraeum DuckeAngelim-Pedra10179
10LauraceaeMezilaurus itauba (Meisn.) Taub. ex MezItauba10164
11FabaceaeParkia pendula (Willd.) Benth. ex Walp.Angelim-Saia10183
12BurseraceaeProtium acrense DalyAmescla-Aroeira10188
13VochysiaceaeQualea paraensis DuckeCambara10168
14SimaroubaceaeSimarouba amara Aubl.Marupá10197
15BurseraceaeTrattinnickia burserifolia Mart.Amescla10176
16FabaceaeVatairea sericea (Ducke) DuckeAngelim-amargoso10165
1 Species identification number; 2 Number of images per species.
Table 2. Performance of the classifiers built using features extracted from original images and the invariant uniform LBP operators ( L B P P , R r i u 2 ) and non-invariant ( L B P P , R u 2 ).
Table 2. Performance of the classifiers built using features extracted from original images and the invariant uniform LBP operators ( L B P P , R r i u 2 ) and non-invariant ( L B P P , R u 2 ).
Training Set (cv = 5, n = 2237)
ClassifierVector sizeStatisticsSVMANNRFLDA
Accuracy (%)F1 (%)Accuracy (%)F1 (%)Accuracy (%)F1 (%)Accuracy (%)F1 (%)
C1 ( L B P 8.1 r i u 2 )10Average26.1425.6831.5431.1122.2621.5217.5816.17
sd1.241.401.511.621.391.122.101.84
C2 ( L B P 16.2 r i u 2 )18Average32.3031.1638.9538.2025.4724.4022.7921.54
sd2.062.432.262.200.910.771.651.21
C3 ( L B P 24.3 r i u 2 )26Average34.5734.0341.5841.2126.1925.1426.5925.41
sd2.602.792.092.042.041.832.011.62
C4 ( L B P P . R r i u 2 ) All54Average43.8943.5750.2349.9628.5527.0035.1534.48
sd2.973.061.932.052.702.642.352.35
C5 ( L B P 8.1 u 2 )59Average38.3137.9642.9642.6729.2128.6831.4130.81
sd2.021.852.192.061.241.401.791.90
C6 ( L B P 16.2 u 2 )243Average48.7548.2752.3652.0432.4331.8140.8140.48
sd1.241.552.582.542.322.352.512.86
C7 ( L B P 24.3 u 2 )555Average51.6950.9755.1754.7633.3232.3548.5848.04
sd1.301.422.402.533.193.422.913.42
C8 ( L B P P . R u 2 ) All857Average54.7354.0060.3059.9035.6334.4955.4054.95
sd1.571.782.622.822.963.323.574.07
C9
( L B P P . R r i u 2 + L B P P . R u 2 )
All
911Average55.4054.7460.6560.2635.5934.5056.2555.84
sd2.122.261.591.842.963.024.114.61
Test Set (n = 566)
ClassifierVector sizeSVMANNRFLDA
Accuracy (%)F1 (%)Accuracy (%)F1 (%)Accuracy (%)F1 (%)Accuracy (%)F1 (%)
C1 ( L B P 8.1 r i u 2 )1038.6838.4841.0040.8233.0032.6220.0018.82
C2 ( L B P 16.2 r i u 2 )1841.7140.9850.0049.8134.0033.6329.0027.57
C3 ( L B P 24.3 r i u 2 )2646.8847.1252.0051.4432.0031.9434.0032.46
C4 ( L B P P . R r i u 2 ) All5454.5554.6360.0060.6039.0037.6742.0041.57
C5 ( L B P 8.1 u 2 )5947.0647.1860.0060.0341.0041.3837.0037.14
C6 ( L B P 16.2 u 2 )24362.9263.0963.0063.2445.0044.2551.0051.16
C7 ( L B P 24.3 u 2 )55565.6065.6666.0066.1344.0043.1456.0055.92
C8 ( L B P P . R u 2 ) All85767.9168.1072.0072.4248.0048.2364.0063.79
C9
( L B P P . R r i u 2 + L B P P . R u 2 )
All
91168.6368.7972.0071.8945.0045.1765.0064.91
Legend. Where P—number of points; R—radius; sd—standard deviation; SVM—Support Vector Machine; ANN—Artificial Neural Networks; RF—Random Forest; LDA—Linear Discriminant Analysis. The value in bold indicates the classifier with the best performance.
Table 3. Performance of the classifiers in cross-validation and testing using the features obtained through the uniform invariant LBP operators ( L B P P , R r i u 2 ) and non-invariant ( L B P P , R u 2 ) and bark image patches.
Table 3. Performance of the classifiers in cross-validation and testing using the features obtained through the uniform invariant LBP operators ( L B P P , R r i u 2 ) and non-invariant ( L B P P , R u 2 ) and bark image patches.
Training Set (cv = 5, n = 2237)
ClassifierVector sizeStatisticsSVMANNRFLDA
Accuracy (%)F1 (%)Accuracy (%)F1 (%)Accuracy (%)F1 (%)Accuracy (%)F1 (%)
C1 ( L B P 8.1 r i u 2 )10Average50.0349.5847.8447.2244.0443.3731.1229.27
sd2.482.664.134.293.603.682.362.19
C2 ( L B P 16.2 r i u 2 )18Average36.1734.2652.4051.3941.9440.6035.1833.29
sd3.133.472.882.883.864.164.004.03
C3 ( L B P 24.3 r i u 2 )26Average36.8836.8056.6055.8548.8247.7538.8137.31
sd1.521.522.222.233.013.163.093.05
C4 ( L B P P . R r i u 2 )
All
54Average57.8457.5472.5672.3654.7254.0955.4354.94
sd2.512.632.712.803.803.973.083.19
C5 ( L B P 8.1 u 2 )59Average66.7966.5174.1673.9056.9556.1153.9652.87
sd1.771.772.662.802.352.432.672.73
C6 ( L B P 16.2 u 2 )243Average67.8167.4781.7281.5559.2858.5661.7461.08
sd0.630.643.123.191.922.113.893.97
C7 ( L B P 24.3 u 2 )555Average67.9067.8978.2378.0753.8252.3463.2162.45
sd2.492.502.722.762.672.653.263.31
C8 ( L B P P . R u 2 ) All857Average77.6177.2977.5277.4160.2259.3873.7273.46
sd2.852.813.163.172.192.293.633.71
C9
( L B P P . R r i u 2 + L B P P . R u 2 ) All
911Average77.2176.977.5677.4860.9360.0773.9473.66
sd2.922.93.283.272.212.273.683.75
Test set (n = 566)
ClassifierVector sizeSVMANNRFLDA
Accuracy (%)F1 (%)Accuracy (%)F1 (%)Accuracy (%)F1 (%)Accuracy (%)F1 (%)
C1 ( L B P 8.1 r i u 2 )1047.0046.0049.0048.0042.0042.0031.0029.00
C2 ( L B P 16.2 r i u 2 )1837.0035.0053.0053.0041.0040.0036.0034.00
C3 ( L B P 24.3 r i u 2 )2630.0030.0055.0055.0049.0048.0039.0037.00
C4 ( L B P P . R r i u 2 )
All
5456.0056.0073.0073.0054.0053.0053.0052.00
C5 ( L B P 8.1 u 2 )5969.0069.0075.0075.0056.0055.0054.0053.00
C6 ( L B P 16.2 u 2 )24366.0066.0079.0079.0057.0055.0061.0060.00
C7 ( L B P 24.3 u 2 )55567.0067.0075.0075.0051.0048.0061.0060.00
C8 ( L B P P . R u 2 ) All85775.0074.0075.0074.0060.0059.0073.0072.00
C9
( L B P P . R r i u 2 + L B P P . R u 2 ) All
91175.0075.0075.0074.0060.0059.0073.0072.00
Legend. Where P—number of points; R—radius; sd—standard deviation; SVM—Support Vector Machine; ANN—Artificial Neural Networks; RF—Random Forest; LDA—Linear Discriminant Analysis. The value in bold indicates the classifier with the best performance.
Table 4. Performance of the classifiers built using features extracted from images with original sizes and pre-trained convolutional neural networks.
Table 4. Performance of the classifiers built using features extracted from images with original sizes and pre-trained convolutional neural networks.
CNNResNet50VGG16Inception_V3MobileNet_V2
Vector size204851220481280
Statistics (%)AveragesdAveragesdAveragesdAveragesd
Training set
(cv = 10, n = 2237)
SVMAccuracy67.362.5356.942.3251.941.4860.301.69
F166.792.7356.352.3651.361.2459.601.39
ANNAccuracy69.330.4460.391.5752.582.4361.281.68
F168.730.7859.941.4451.832.4760.601.47
RFAccuracy57.652.0953.472.5645.922.2848.952.25
F155.741.9350.712.5343.382.3946.292.58
LDAAccuracy63.871.7953.011.4152.223.0059.002.10
F163.541.9853.081.3052.282.5458.791.41
Test set
(n = 566)
SVMAccuracy82.6973.6763.4373.32
F182.6373.7863.2572.87
ANNAccuracy81.9874.0363.2576.50
F182.0873.9863.4476.24
RFAccuracy69.9659.1950.5357.42
F169.0557.5348.7556.16
LDAAccuracy77.7460.4263.7871.55
F177.8360.7864.1471.62
Legend. Where sd—standard deviation; SVM—Support Vector Machine; ANN—Artificial Neural Networks; RF—Random Forest; LDA—Linear Discriminant Analysis. Values in bold indicate the best performing classifiers.
Table 5. Performance of the classifiers built using features extracted from sub-images (patches) and pre-trained convolutional neural networks.
Table 5. Performance of the classifiers built using features extracted from sub-images (patches) and pre-trained convolutional neural networks.
CNNResNet50VGG16Inception_V3MobileNet_V2
Vector size204851220481280
Statistics (%)AveragesdAveragesdAveragesdAveragesd
Training set
cv = 10, n = 2.237)
SVMAccuracy95.571.0791.421.2893.211.9494.551.53
F195.571.0791.381.2793.21.9794.531.57
ANNAccuracy95.351.6391.282.5391.241.4592.361.99
F195.341.6491.212.6191.181.4592.312.01
RFAccuracy84.762.8380.823.1271.713.4376.093.04
F184.482.9480.313.4070.483.8275.153.28
LDAAccuracy90.932.2380.332.4484.942.2185.972.22
F191.012.2480.462.5684.872.2885.902.22
Test set
(n = 566)
SVMAccuracy95.0091.0092.0094.00
F195.0091.0092.0094.00
ANNAccuracy94.0089.0090.0091.00
F194.0089.0090.0091.00
RFAccuracy83.0081.0067.0074.00
F183.0080.0066.0073.00
LDAAccuracy92.0081.0083.0086.00
F192.0081.0083.0086.00
Legend. Where sd—standard deviation; SVM—Support Vector Machine; ANN—Artificial Neural Networks; RF—Random Forest; LDA—Linear Discriminant Analysis. Values in bold indicate the best performing classifiers.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gama, N.C.; Oliveira, L.E.S.; Carvalho, S.d.P.C.e.; Behling, A.; de Paula Filho, P.L.; Hamada, M.O.d.S.; Leal, E.d.S.; Souza, D.V. Automatic Recognition of Commercial Tree Species from the Amazon Flora Using Bark Images and Transfer Learning. Forests 2025, 16, 1374. https://doi.org/10.3390/f16091374

AMA Style

Gama NC, Oliveira LES, Carvalho SdPCe, Behling A, de Paula Filho PL, Hamada MOdS, Leal EdS, Souza DV. Automatic Recognition of Commercial Tree Species from the Amazon Flora Using Bark Images and Transfer Learning. Forests. 2025; 16(9):1374. https://doi.org/10.3390/f16091374

Chicago/Turabian Style

Gama, Natally Celestino, Luiz Eduardo Soares Oliveira, Samuel de Pádua Chaves e Carvalho, Alexandre Behling, Pedro Luiz de Paula Filho, Márcia Orie de Sousa Hamada, Eduardo da Silva Leal, and Deivison Venicio Souza. 2025. "Automatic Recognition of Commercial Tree Species from the Amazon Flora Using Bark Images and Transfer Learning" Forests 16, no. 9: 1374. https://doi.org/10.3390/f16091374

APA Style

Gama, N. C., Oliveira, L. E. S., Carvalho, S. d. P. C. e., Behling, A., de Paula Filho, P. L., Hamada, M. O. d. S., Leal, E. d. S., & Souza, D. V. (2025). Automatic Recognition of Commercial Tree Species from the Amazon Flora Using Bark Images and Transfer Learning. Forests, 16(9), 1374. https://doi.org/10.3390/f16091374

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop