Next Article in Journal
Modeling Natural Forest Fire Regimes Based on Drought Characteristics at Various Spatial and Temporal Scales in P. R. China
Next Article in Special Issue
GIS-Based Multi-Criteria Analysis for Urban Afforestation Planning in Semi-Arid Cities
Previous Article in Journal
Influences of the China–Russia Crude Oil Pipelines on the Characteristics of Soil Bacterial and Fungal Communities in Permafrost Regions of the Da Xing’anling Mountains, Northeast China
Previous Article in Special Issue
GIS-Based Spatial Modeling of Soil Erosion and Wildfire Susceptibility Using VIIRS and Sentinel-2 Data: A Case Study of Šar Mountains National Park, Serbia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Tree Species in Poland Using CNNs Tabular-to-Pseudo Image Approach Based on Sentinel-2 Annual Seasonality Data

Department of Forest Resources Management, Faculty of Forestry, University of Agriculture in Krakow, 31-120 Cracow, Poland
*
Author to whom correspondence should be addressed.
Forests 2025, 16(7), 1039; https://doi.org/10.3390/f16071039
Submission received: 17 May 2025 / Revised: 11 June 2025 / Accepted: 16 June 2025 / Published: 20 June 2025

Abstract

Tree species classification provides invaluable information across various sectors, from forest management to conservation. This task is most commonly performed using remote sensing; however, this method is prone to classification errors, which modern computational approaches aim to minimize. Convolutional neural networks (CNNs) used to model tabular data have recently gained popularity as a highly efficient classification tool. In the present study, a variation of this method is used to classify satellite multispectral data from the Sentinel-2 mission to distinguish between 18 common Polish tree species. The novel model is trained and tested on data from species-homogeneous forest stands. The data form a multi-seasonal time series and cover five years of observations. The model achieved an overall accuracy of 80% and Cohen Kappa of 0.80 of the raw output and increased to 93% with post-processing procedures. Considering the large number of species classified, this is a promising and encouraging result. The presented results indicate the importance of early vegetation season reflectance data in model training. The spectral bands representing the infrared, red-edge and green wavelengths had the greatest impact on the model.

1. Introduction

Remote-sensing-based tree species classification provides valuable information for a variety of sectors ranging from forest management to conservation [1,2,3,4]. The steadily growing availability of airborne and satellite sensor data coupled with the increasing processing power and availability of cloud-based computational services [5,6] greatly facilitates the classification process. Recently, the process is most often entrusted to machine learning algorithms of different architectures. The choice of the classification algorithm is often determined by the character of the dataset—whether it is multispectral or point cloud, tabular or images. Among the popular approaches to tackle the classification problem, random forest (RF) [7,8,9,10] and support vector machine (SVM) [11,12,13] are the most popular when the data is in tabular format. Imaging data are often processed using a variety of convolutional neural networks (CNNs) adaptations [14,15,16].
With respect to Earth observation datasets, species classification tasks are now often based on Sentinel-2 mission imagery due to its suitable resolution, frequent revisits and relative ease of access [8,17,18,19,20,21,22]. Due to spatial and time resolution limitations, the use of imagery from the Landsat mission is gradually becoming less popular [11,19,23,24], but new applications are still being proposed [25] with a large spatial scale analyses in mind. There are also successful studies using high-resolution WorldView-2 [7,16,26], WorldView-3 [12] and Formosat-2 [27] satellite imagery. A separate group of studies are classifiers based on hyperspectral data from airborne vehicles (drones, aircraft) [13,28] or commercially available datasets [29].
While standard machine learning algorithms, such as random forest and SVM, have been widely utilized in forest classification due to their robustness and speed [10,11], recently, deep learning approaches, particularly CNNs, have emerged as powerful tools, demonstrating superior performance in various remote sensing tasks, including tree species classification [30,31,32,33,34]. This enhanced capability is largely attributed to their capacity for automated feature learning, enabling CNNs to iteratively learn relevant data transformations and hierarchical features directly from raw input, thereby making traditional, heuristic feature engineering obsolete. Furthermore, CNNs are uniquely adept at exploiting spatial patterns through successive convolutions and pooling operations, providing a crucial advantage for analyzing complex remote sensing imagery and significantly enhancing their effectiveness, especially with very high spatial resolution data [35]. Recent innovative CNN (convolutional neural network) -based methods have progressed beyond discrete classification to quantify tree species proportions at the pixel level, allowing for mapping across all forest compositions from pure to highly mixed stands. This approach provides a more realistic representation of heterogeneous forest environments, optimizing the use of available forest inventory data that may not offer pixel-level labels [30].
The approach proposed in this study—transforming tabular data into pseudo-images—offers a promising pathway for leveraging the powerful capabilities of (CNN), which are inherently designed for spatial data and typically cannot be directly applied to tabular structures due to the lack of inherent locality and spatial relationships. In the case of analyzing medium-resolution satellite imagery such as Sentinel-2, the standard approach is to treat them as tabular data rather than as image data in the sense used by CNN approaches. To the best of our knowledge, our study represents the first example of transferring an approach—converting tabular data into two-dimensional CNN—that has already been tested in other fields [36,37] to the context of remote sensing.
In the current paper, a classifier capable of discriminating 18 common tree species (Table 1) based on spectral data from the Sentinel-2 mission [38] is presented. The classifier was trained and evaluated on data from homogeneous (species-wise) forest stands scattered over the entire territory of Poland. The data were pixel-based in tabular format, but the choice of the modelling algorithm was influenced by the recently popular formula of using CNN on tabular data transformed into pseudo-images [39,40]. The architecture of the proposed two-dimensional CNN is inspired by one-dimensional CNN, which achieved high accuracy in Kaggle competition in classification of tabular data [41]. Additional inspiration was an approach using a one-dimensional CNN on time series data proposed by Xi [42].
This paper presents a modern approach to tree species classification in the hope of promoting state-of-the-art modelling algorithms offering better classification accuracy than commonly used ensemble learning models. The presented study introduces a comprehensive classification framework that significantly advances the scope of tree species identification by enabling the accurate classification of a broad range of species. This contrasts with prior approaches, which have predominantly concentrated on limited, species-specific datasets [8,19,42]. By addressing a wider taxonomic spectrum, this methodology represents a substantial improvement in automated classification, with implications for enhanced scalability and applicability across diverse ecological and forest management contexts.

2. Materials and Methods

2.1. Data and Preprocessing

The study area spanned an extensive region across Poland, encompassing a vast area from 14°07″ to 24°09″ east longitude and from 49°00″ to 54°50″ north latitude. The data was obtained from the area belonging to the State Forests, which were covered with forest. The reference data was derived from the Forest Data Bank (FDB) database. The FDB online service (https://www.bdl.lasy.gov.pl/portal accessed on 15 June 2025) provides free access and the ability to download the data set with a multitude of forest stand variables, including species composition, age, average height of species, average diameter at breast height (DBH), and so forth. The data are created during preparation of forest management plans and are annually updated. The data extracted for this study included information on the geometry of the forest stands, the species composition, the age of the species, and the site type. A total of eighteen tree species were selected for analysis. The list of species and the number of stands chosen for analysis is presented in Table 1. Only single-species stands with an area exceeding 0.4 hectares were selected for analysis. The age of the stands ranged from 5 to 160 years. For species where the number of available stands in the database was less than 500, all stands were used for creating a reference dataset. For species where the number of available stands was between 500 and 1000, 500 stands were randomly selected. For species with more than 1000 available stands in the database, 600 polygons per species were randomly drawn using stratification. The stands were grouped by age class (eight classes of 20-year intervals) and site type (21 classes). Then, for each group within a species, the same number of stands were drawn in the stratification group to give a final count of 600 polygons for that species. Initially selected polygons for all species were subsequently verified manually on the basis of an up-to-date orthophoto map with a resolution of 0.25 m, which was available on the national geoportal (geoportal.gov.pl). During the on-screen verification, the boundaries of the stands’ polygons were edited if another species were identified, for example, close to a stand boundary. If more than one species was identified within the polygon boundaries and it was not possible to clearly define the range of a particular species, such polygons were removed and not considered as reference.
After manual verification a total of 5258 reference stands were selected for further analysis. The number of chosen reference stands by species is summarized in Table 1. The location of the reference forest stands used in the analysis is shown in Figure 1. Based on the data set, training and test sets were prepared.
The data from the Sentinel-2 satellite for the period from 2018 to 2022 were acquired through Google Earth Engine [5]. For each year, four mosaics were generated for a two-month interval covering the months of April to May, June to July, August to September and October to November, which we refer to as “seasons”. Cloud masking was conducted via the s2cloudless algorithm (Sentinel-Hub) [43]. Mosaics were generated by computing the median value of cells from overlapping Sentinel-2 scenes. Mosaics for each season were prepared from 10 spectral bands, including B2-B8a and B11-B12. A raster layer containing 40 bands (4 seasons * 10 spectral bands) was created for each year.
To ensure uniformity across seasons, the layers were standardized using the calculated mean and standard deviation, according to the following formula: z = (x − μ)/σ [44].
Next, the values of raster cells (observations) were extracted from the prepared spatial layers for the 5258 reference fields. To reduce the influence of outliers, observations below the 5th percentile and greater than the 95th percentile were excluded from the data set. Similarly, spatial layer cells marked as missing data were also excluded from further analysis.
This sampling procedure resulted in 4,468,452 valid pixel observations (896,596 for the year 2018, 896,651 for the year 2019, 882,366 for the year 2020, 896,183 for the year 2021 and 896,656 for the year 2022). Table 1 shows the number of observations for each tree species.
Next, the input dataset was randomly divided into training and testing sets in a 75%/25% ratio. When dividing the input data set, care was taken to ensure that all data from a reference polygon, but from different years, were stored either within the training set or the test set, and were never mixed.
Each observation in the datasets represents a time series with location coordinate information appended. The time series covers the reflectance variation during the growing season in one year. It provides insight into the annual seasonality. The five years studied allow a broader variability to be captured.
Figure 2 shows the mean values of standardized reflectance by tree species and season. Each line represents the average reflectance magnitude in the different Sentinel-2 bands used in analysis.

2.2. Method

The data were organized in a tabular format. A convolutional neural network (CNN) was employed to analyze them. CNNs are primarily utilized for image analysis. The novel approach of applying such networks to tabular data within the context of image classification is used in this study. The workflow of the conducted analysis is presented in Figure 3a and is described in the following sections of the paper.

2.2.1. CNN Topology

The CNN architecture employed in this research is inspired by one-dimensional CNNs that have achieved high levels of accuracy in classifying tabular data [41]. However, a two-dimensional CNN architecture demonstrated superior performance on this particular dataset and was ultimately utilized in the study. The model was implemented in the Tensorflow [45]/Keras [46] environment. The topology of the network used is illustrated in Figure 3b.
The 40-dimensional input data vector was fed into a 100-dimensional fully connected layer, which employed a rectified linear unit activation function to expand the input. The layer was subsequently transformed into a 2D pseudo-image representation comprising 10 × 10 pixels. A set of two convolutional layers were subsequently implemented; the first comprised 32 filters (kernel size 2 × 2, stride 1, and 0 padding, a Rectified Linear Unit (RELU) activation function), while the second incorporated 64 filters (kernel size 3 × 3, stride 1, and 0 padding, RELU activation function). The output was then flattened and subjected to regularization through the introduction of a 0.5 dropout. A fully connected layer (32 nodes, RELU activation function) and a fully connected output layer (18 nodes, softmax activation function) were subsequently added. The number of layers and associated hyperparameters were initially tuned using the KerasTuner optimization framework [47] and further adjusted experimentally based on model performance. The Adam optimization algorithm was employed to learn the model. In an attempt to mitigate model overfitting, a relatively modest learning rate (0.00001) and an early-stopping mechanism were implemented. The learning curves of the CNN designed to classify tree species from tabular data are presented in Figure 4.
The effect of input variables on the model output was examined using SHAP index [48]. The magnitude of importance is a measure of how strongly a variable influences the model outcome [49].

2.2.2. Post-Processing

Two approaches were tested to improve tree species recognition. In the first approach, a species was selected based on network responses above 0.8. In the second approach, results were combined due to belonging to a single reference field and a score was assigned when all network responses indicated the same species. The first approach favored quality over quantity. It allowed for a significant increase in accuracy. However, at the same time, this approach reduced the number of classified pixels. The second approach is limited to single-species forest stands.

2.2.3. Model Quality Assessment

The following measures were used to assess model quality: Overall Classification Accuracy (CA), Cohen Kappa Index (CK), Precision, Recall, F1 Score, Receiver Operating Characteristics curve (ROC), Area under the ROC curve (AUC) and Confusion Matrix.

3. Results

The final CNN model achieved high performance with overall classification accuracy of 0.803 and Cohen’s Kappa of 0.804. All performance metrics are presented in Table 2. The ROC curves (with AUC values) and the confusion matrix are shown in Figure 5 and Figure 6, respectively.
The values of the resulting SHAP index for each input variable are shown in Figure 7. The figure provides only the 20 variables that have the most significant impact on changing the CNN’s response. The variables are ordered in descending order of strength of influence. The most influential variables were bands representing shortwave infrared, green and vegetation red edge wavelengths for the early vegetation season (April–May).
In the current study, modeling was particularly successful in recognizing four conifer species: dwarf mountain pine, Scots pine, silver fir, and European larch. Similarly well classified were selected deciduous species: robinia pseudoacacia, European beech, European black alder and silver birch.
However, there were tree species that were more difficult to classify (see Figure 5 and Figure 6 and Table 2 for details). The poorest performance was observed for sycamore maple, for which the classifier confused with silver birch and European black alder. The second worst species was small-leaved linden, which was most often confused with European black alder, oak, European ash, and silver birch. This was followed by European hornbeam, which was misclassified with a variety of deciduous species. The other problematic species were Douglas fir. This species was misclassified as either silver fir or Scots pine. Black pine was mistaken for Scots pine.
Despite these classification issues, the overall accuracy and Cohen’s Kappa are over 0.8. That shows the potential of the approach based on tabular data analysis.
The classifier successfully distinguished between deciduous and coniferous trees in 96% of cases. It is worth noting that there is no significant relationship between the sample size of the test set (see Table 2 column: Support) and the classification score of the species (Spearman’s Rho = 0.22); thus, the performance of the classifier is most likely due to the characteristics of the spectral response of the species encoded internally by CNN as an artificial image.
The map of classified tree species for the whole country is presented in Figure 8.
Post-processing procedures have made it possible to improve the accuracy of the model output. The overall accuracy was increased in two proposed scenarios. The results are as follows: after selecting the output above the maximum probability threshold of 0.8, the accuracy reached 0.932; after polygon grouping, the accuracy of the classifier reached 0.835.

4. Discussion

The tree species used in the study cover the most common species in Poland. There was no particular key for species selection, and the 18 species selected were the best represented in the available dataset (FDB), which ensured proper model training.
Seasonal changes in canopy spectral response are a strong discriminator of tree species [50,51,52,53] and were included in the current study as a sequential set of repeated measurements of spectral variables averaged over four two-month intervals from April to November. The length of the two-month intervals was determined by the quality and acquisition frequency of the satellite imagery. In this way, the composites provided nearly cloud-free mosaic coverage of the study area. The observations were a seasonal time series. The spectral response is also expected to vary between years due to various meteorological and biological factors [54,55]. Capturing this variability by including satellite images from five consecutive years was intended to introduce a climate influence into the model.
During data pre-processing, the training and test datasets were prepared to eliminate spatial autocorrelation. Data from a given training polygon were assigned exclusively to either the training or test dataset, regardless of the date of acquisition. In addition, the test dataset was normalized based on the statistics of the training dataset, both to promote the model’s ability to classify upcoming, unseen data.
The training polygons were deliberately chosen to represent the stands with homogeneous tree species. Such an approach allows for an efficient model training process. However, there were strong outliers in the raw spectral data set. Two approaches were tested to resolve them. The first approach excluded polygons with a variance of more than 2 σ in any of the variables in a given year, while the second approach used Winsorization and showed a positive effect on model performance. Therefore, the second procedure was selected as the outlier treatment for the study.
CNNs are typically designed to extract information from images where spatially meaningful patterns are expected [56]. They are most commonly used for computer vision and object classification tasks. Recently, CNNs have been shown to outperform regular deep learning models when trained on tabular data preprocessed into a pseudo-image format [39,40]. The choice of CNN dimensionality in the case of tabular data is determined less by the original data format and more by performance metrics. In the current study, a two-dimensional CNN model was used after preliminary testing. The basic architecture of the model was based on the well-performing competition architecture [41]. The model architecture was adapted to the input dataset and roughly hyperparameterized using the automatic hyperparameter optimization tool [47]. The implementation of the MaxPooling layer was also tested due to its reported positive effect on classification results [57]. It was ultimately removed due to a lack of positive impact on model performance. The model used in this study appears complex compared to the relatively small dataset and could be prone to overfitting, but thanks to strong regularization with a dropout layer, a low learning rate, and an early stopping mechanism, it was possible to benefit from the model’s good feature extraction ability while maintaining its ability to generalize.
Complicated machine learning models are difficult to interpret [58]. The use of the SHAP values method [48,59] revealed factors influencing the presented model. The variables that most discriminated species were those representing the shortwave infrared (bands B11 and B12), red edge (bands B5 and B8a), and green (band B3) wavelength spectra (Figure 7). This result is expected and reassuring that the models are working correctly. This is because the red edge is postulated to be specifically sensitive to the chlorophyll content of vegetation (see [60] for references). Short-wave infrared has been reported to be sensitive to leaf water content [61], vegetation density [62], and nitrogen content [63]. In addition, the “green” spectral region has been reported to aid in the detection of chlorophyll content where “red” bands fail [64]. The early growing season (April–May) clearly stands out as the most influential on the model (Figure 2 and Figure 7), most likely due to the influence it can have on the separation between deciduous and coniferous species [8].
In order to extract meaningful information from the model output, certain post-processing steps are usually required. In this research, two basic solutions are presented. Both solutions depend on different practical scenarios of model application. For the classification of mixed-species or unmanaged forests, the most useful approach seems to be to set the threshold of the output maximum probability. A supervised classifier assumes that the class with the highest probability value is assigned as the most likely output. However, if the data does not match the classes learned by the model, the highest probability value may still be relatively low. This indicates the inability of the model to classify with confidence. In such a case, setting a probability threshold (e.g., 0.8) and considering only outputs above this threshold helps ensure that the assigned class is not a near-random guess. However, this approach reduces the number of output datasets. In the case of spatial analysis, this will result in leaving empty (No Data) spatial cells. However, this approach increased the accuracy from 0.804 (raw output) to 0.932 (post-processing probability threshold) in the study presented. Another potential post-processing solution used in this study was to calculate the mode of the outputs associated with a coherent single-species area. This approach seems to be useful only in scenarios where monocultures are studied, and has the advantage of eliminating data noise. This approach, where the output mode is calculated for test polygons regardless of the year of data collection, improved the classification results, increasing the accuracy from 0.804 (raw output) to 0.835 (area mode post-processing).
By converting tabular data into image formats, various approaches aim to incorporate two-dimensional spatial information that CNNs can process efficiently, potentially leading to improved model performance [36,37,65]. A significant advantage observed in several studies is the enhanced classification accuracy when using the transformation to tabular data to pseudo-image. For instance, the Low Mixed-Image Generator for Tabular Data method, which transforms tabular data into 2D grayscale images, combined with CNNs, demonstrated superior area under the ROC curve values for cardiovascular disease prediction compared to traditional machine learning (ML) models like least absolute shrinkage and selection operator (LASSO) and Tabular Prior Fitted Network (TabPFN), especially when transfer learning was applied [36]. Similarly, the Novel Algorithm for Convolving Tabular Data (NCTD) approach, which uses mathematical transformations to create synthetic images, consistently surpassed conventional ML algorithms in classification accuracy across seven out of ten diverse datasets [65]. Furthermore, the Feature Clustering-Visualization (FC-Viz) method, by structuring tabular data as color-encoded images with preserved feature relationships, achieved the highest average test accuracy compared to SVM, RF, Decision Tree, and eXtreme Gradient Boosting across ten benchmark datasets [66].
While some studies reported that CNNs on pseudo-images could outperform machine learning algorithms, it is also noted that the improvements might not always be statistically significant or consistently superior to the best-optimized machine learning models [37]. This suggests that while conversion adds spatial information, it might not always sufficiently equip the CNN to grasp underlying patterns with significantly greater accuracy than highly optimized typical machine learning models based on tabular models.
Beyond accuracy, the pseudo-image approach often offers significant advantages in parameter efficiency and model size, leading to faster training and inference times [66]. However, it is also acknowledged that some pseudo-image methods might generate large image representations, which could increase computational demands for storage and processing, potentially limiting scalability in resource-constrained environments [65].
The general trend from studies conducted in other fields indicates that a well-designed tabular-to-image conversion can yield models that are not only competitive in predictive performance but also more efficient in terms of computational resources and model footprint. In our study, we transferred experience from the analysis of medical images to remote sensing applications. Although this was not the primary aim of our study, future research should focus on comparing the proposed approach with standard machine learning methods based on tabular data, which are commonly used for tree species classification based on medium-resolution satellite images.

5. Conclusions

This study presents a modern approach to tree species classification using a CNN architecture designed for tabular data—a method that is gaining popularity in the field. The proposed CNN model demonstrated strong performance in classifying 18 tree species in Poland, utilizing publicly available satellite imagery data from the Sentinel-2 mission. To enhance the model’s accuracy, specific post-processing steps were recommended, resulting in an accuracy of 0.93 in this study. Regarding data collection, images captured during the early vegetation season proved to be the most valuable for this classification task. Additionally, infrared, red-edge, and green wavelengths were identified as the most influential in enabling the model to effectively distinguish between tree species.

Author Contributions

Conceptualization, Ł.M., P.H., P.N. and J.S.; methodology, Ł.M., P.H., P.N., J.T. and J.S.; software, Ł.M. and P.N.; validation, Ł.M., P.H. and P.N.; formal analysis, Ł.M.; investigation, Ł.M., P.H., P.N., J.T., N.Z. and J.S.; resources, Ł.M., P.H., P.N., N.Z. and J.S.; data curation, Ł.M., P.H., P.N. and N.Z; writing—original draft preparation, Ł.M.; writing—review and editing, Ł.M., P.H., P.N., J.T., N.Z. and J.S.; visualization, Ł.M. and J.T.; supervision, J.S.; project administration, J.S.; funding acquisition, J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science Center, Poland, and carried within the project “Assessment of the impact of weather conditions on forest health status and forest disturbances at regional and national scale based on the integration of ground and space-based remote sensing datasets” No. 2021/41/B/ST10/04113.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AUCArea Under the Curve
CAClassification Accuracy
CNNConvolutional Neural Network
CKCohen Kappa Index
DBHDiameter at Breast Height
FC-VizFeature Clustering-Visualization
FDBForest Data Bank
LASSOLeast Absolute Shrinkage and Selection Operator
NCTDNovel Algorithm for Convolving Tabular Data
RELURectified Linear Unit
RFRandom Forest
ROCReceiver Operating Characteristics
SCASpecies Classification Accuracy
SVMSupport Vector Machine
TabPFNTabular Prior Fitted Network

References

  1. Lindenmayer, D.B.; Margules, C.R.; Botkin, D.B. Indicators of Biodiversity for Ecologically Sustainable Forest Management. Conserv. Biol. 2000, 14, 941–950. [Google Scholar] [CrossRef]
  2. Wulder, M.A.; Hall, R.J.; Coops, N.C.; Franklin, S.E. High Spatial Resolution Remotely Sensed Data for Ecosystem Characterization. Bioscience 2004, 54, 511–521. [Google Scholar] [CrossRef]
  3. McDermid, G.J.; Hall, R.J.; Sanchez-Azofeifa, G.A.; Franklin, S.E.; Stenhouse, G.B.; Kobliuk, T.; LeDrew, E.F. Remote Sensing and Forest Inventory for Wildlife Habitat Assessment. For. Ecol. Manag. 2009, 257, 2262–2269. [Google Scholar] [CrossRef]
  4. Fassnacht, F.E.; Latifi, H.; Stereńczak, K.; Modzelewska, A.; Lefsky, M.; Waser, L.T.; Straub, C.; Ghosh, A. Review of Studies on Tree Species Classification from Remotely Sensed Data. Remote Sens. Environ. 2016, 186, 64–87. [Google Scholar] [CrossRef]
  5. Gorelick, N.; Hancher, M.; Dixon, M.; Ilyushchenko, S.; Thau, D.; Moore, R. Google Earth Engine: Planetary-Scale Geospatial Analysis for Everyone. Remote Sens. Environ. 2017, 202, 18–27. [Google Scholar] [CrossRef]
  6. Bisong, E. Building Machine Learning and Deep Learning Models on Google Cloud Platform; Apress: Berkeley, CA, USA, 2019; ISBN 978-1-4842-4469-2. [Google Scholar]
  7. Immitzer, M.; Atzberger, C.; Koukal, T. Tree Species Classification with Random Forest Using Very High Spatial Resolution 8-Band WorldView-2 Satellite Data. Remote Sens. 2012, 4, 2661–2693. [Google Scholar] [CrossRef]
  8. Persson, M.; Lindberg, E.; Reese, H. Tree Species Classification with Multi-Temporal Sentinel-2 Data. Remote Sens. 2018, 10, 1794. [Google Scholar] [CrossRef]
  9. Loiola, T.M.; Fantinel, R.A.; Dos Santos, F.D.; De Bastos, F.; Schuh, M.S.; Fernandes, P.; Simões, B.A.; Pereira, R.S. Use of Machine Learning Algorithms in the Classification of Forest Species. Anuário Inst. Geociências 2023, 46, 50490. [Google Scholar] [CrossRef]
  10. Grabska-Szwagrzyk, E.; Tiede, D.; Sudmanns, M.; Kozak, J. Map of Forest Tree Species for Poland Based on Sentinel-2 Data. Earth Syst. Sci. Data Discuss. 2024, 16, 2877–2891. [Google Scholar] [CrossRef]
  11. Zhang, Z.; Liu, X. Support Vector Machines for Tree Species Identification Using LiDAR-Derived Structure and Intensity Variables. Geocarto Int. 2013, 28, 364–378. [Google Scholar] [CrossRef]
  12. Deur, M.; Gašparović, M.; Balenović, I. Tree Species Classification in Mixed Deciduous Forests Using Very High Spatial Resolution Satellite Imagery and Machine Learning Methods. Remote Sens. 2020, 12, 3926. [Google Scholar] [CrossRef]
  13. Ballanti, L.; Blesius, L.; Hines, E.; Kruse, B. Tree Species Classification Using Hyperspectral Imagery: A Comparison of Two Classifiers. Remote Sens. 2016, 8, 445. [Google Scholar] [CrossRef]
  14. Mubin, N.A.; Nadarajoo, E.; Shafri, H.Z.M.; Hamedianfar, A. Young and Mature Oil Palm Tree Detection and Counting Using Convolutional Neural Network Deep Learning Method. Int. J. Remote Sens. 2019, 40, 7500–7515. [Google Scholar] [CrossRef]
  15. Yuan, Q.; Shen, H.; Li, T.; Li, Z.; Li, S.; Jiang, Y.; Xu, H.; Tan, W.; Yang, Q.; Wang, J.; et al. Deep Learning in Environmental Remote Sensing: Achievements and Challenges. Remote Sens. Environ. 2020, 241, 111716. [Google Scholar] [CrossRef]
  16. Li, H.; Hu, B.; Li, Q.; Jing, L. CNN-Based Individual Tree Species Classification Using High-Resolution Satellite Imagery and Airborne LiDAR Data. Forests 2021, 12, 1697. [Google Scholar] [CrossRef]
  17. Immitzer, M.; Vuolo, F.; Atzberger, C. First Experience with Sentinel-2 Data for Crop and Tree Species Classifications in Central Europe. Remote Sens. 2016, 8, 166. [Google Scholar] [CrossRef]
  18. Immitzer, M.; Neuwirth, M.; Böck, S.; Brenner, H.; Vuolo, F.; Atzberger, C. Optimal Input Features for Tree Species Classification in Central Europe Based on Multi-Temporal Sentinel-2 Data. Remote Sens. 2019, 11, 2599. [Google Scholar] [CrossRef]
  19. Soleimannejad, L.; Ullah, S.; Abedi, R.; Dees, M.; Koch, B. Evaluating the Potential of Sentinel-2, Landsat-8, and Irs Satellite Images in Tree Species Classification of Hyrcanian Forest of Iran Using Random Forest. J. Sustain. For. 2019, 38, 615–628. [Google Scholar] [CrossRef]
  20. Illarionova, S.; Trekin, A.; Ignatiev, V.; Oseledets, I. Tree Species Mapping on Sentinel-2 Satellite Imagery with Weakly Supervised Classification and Object-Wise Sampling. Forests 2021, 12, 1413. [Google Scholar] [CrossRef]
  21. Polyakova, A.; Mukharamova, S.; Yermolaev, O.; Shaykhutdinova, G. Automated Recognition of Tree Species Composition of Forest Communities Using Sentinel-2 Satellite Data. Remote Sens. 2023, 15, 329. [Google Scholar] [CrossRef]
  22. Ahlswede, S.; Schulz, C.; Gava, C.; Helber, P.; Bischke, B.; Förster, M.; Arias, F.; Hees, J.; Demir, B.; Kleinschmit, B. TreeSatAI Benchmark Archive: A Multi-Sensor, Multi-Label Dataset for Tree Species Classification in Remote Sensing. Earth Syst. Sci. Data Discuss. 2023, 15, 681–695. [Google Scholar] [CrossRef]
  23. Chiang, S.H.; Valdez, M.; Chen, C.-F. Forest tree species distribution mapping using landsat satellite imagery and topographic variables with the maximum entropy method in mongolia. ISPRS—Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 41, 593–596. [Google Scholar] [CrossRef]
  24. Grigorieva, O.; Brovkina, O.; Saidov, A. An Original Method for Tree Species Classification Using Multitemporal Multispectral and Hyperspectral Satellite Data. Silva Fenn. 2020, 54, 10143. [Google Scholar] [CrossRef]
  25. Konrad Turlej, C.; Ozdogan, M.; Radeloff, V.C. Mapping Forest Types over Large Areas with Landsat Imagery Partially Affected by Clouds and SLC Gaps. Int. J. Appl. Earth Obs. Geoinf. 2022, 107, 102689. [Google Scholar] [CrossRef]
  26. Pu, R.; Landry, S. A Comparative Analysis of High Spatial Resolution IKONOS and WorldView-2 Imagery for Mapping Urban Tree Species. Remote Sens. Environ. 2012, 124, 516–533. [Google Scholar] [CrossRef]
  27. Sheeren, D.; Fauvel, M.; Josipović, V.; Lopes, M.; Planque, C.; Willm, J.; Dejoux, J.-F. Tree Species Classification in Temperate Forests Using Formosat-2 Satellite Image Time Series. Remote Sens. 2016, 8, 734. [Google Scholar] [CrossRef]
  28. Dalponte, M.; Ørka, H.O.; Ene, L.T.; Gobakken, T.; Næsset, E. Tree Crown Delineation and Tree Species Classification in Boreal Forests Using Hyperspectral and ALS Data. Remote Sens. Environ. 2014, 140, 306–317. [Google Scholar] [CrossRef]
  29. Gašparović, M.; Dobrinić, D.; Pilaš, I. Mapping of Allergenic Tree Species in Highly Urbanized Area Using Planetscope Imagery—A Case Study of Zagreb, Croatia. Forest 2023, 14, 1193. [Google Scholar] [CrossRef]
  30. Bolyn, C.; Lejeune, P.; Michez, A.; Latte, N. Mapping Tree Species Proportions from Satellite Imagery Using Spectral–Spatial Deep Learning. Remote Sens. Environ. 2022, 280, 113205. [Google Scholar] [CrossRef]
  31. Illarionova, S.; Trekin, A.; Ignatiev, V.; Oseledets, I. Neural-Based Hierarchical Approach for Detailed Dominant Forest Species Classification by Multispectral Satellite Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1810–1820. [Google Scholar] [CrossRef]
  32. Mouret, F.; Morin, D.; Planells, M.; Vincent-Barbaroux, C. Tree Species Classification at the Pixel Level Using Deep Learning and Multispectral Time Series in an Imbalanced Context. Remote Sens. 2025, 17, 1190. [Google Scholar] [CrossRef]
  33. Pei, H.; Owari, T.; Tsuyuki, S.; Hiroshima, T.; Hong, D. Combined Gated Graph Convolution Neural Networks with Multi-Modal Geospatial Data for Forest Type Classification. Int. J. Appl. Earth Obs. Geoinf. 2025, 136, 104372. [Google Scholar] [CrossRef]
  34. Ma, Y.; Zhen, Z.; Li, F.; Feng, F.; Zhao, Y. An Innovative Lightweight 1D-CNN Model for Efficient Monitoring of Large-Scale Forest Composition: A Case Study of Heilongjiang Province, China. GIScience Remote Sens. 2023, 60, 2271246. [Google Scholar] [CrossRef]
  35. Kattenborn, T.; Leitloff, J.; Schiefer, F.; Hinz, S. Review on Convolutional Neural Networks (CNN) in Vegetation Remote Sensing. ISPRS J. Photogramm. Remote Sens. 2021, 173, 24–49. [Google Scholar] [CrossRef]
  36. Lara-Abelenda, F.J.; Chushig-Muzo, D.; Peiro-Corbacho, P.; Gómez-Martínez, V.; Wägner, A.M.; Granja, C.; Soguero-Ruiz, C. Transfer Learning for a Tabular-to-Image Approach: A Case Study for Cardiovascular Disease Prediction. J. Biomed. Inform. 2025, 165, 104821. [Google Scholar] [CrossRef]
  37. Neto, L.M.; da Silva Neto, S.R.; Endo, P.T. A Comparative Analysis of Converters of Tabular Data Into Image for the Classification of Arboviruses Using Convolutional Neural Networks. PLoS ONE 2023, 18, e0295598. [Google Scholar] [CrossRef]
  38. Copernicus Sentinel Data, 2018–2021. Retrieved from Google Earth Engine, Processed by ESA. Available online: https://browser.dataspace.copernicus.eu (accessed on 23 February 2023).
  39. Sharma, A.; Vans, E.; Shigemizu, D.; Boroevich, K.A.; Tsunoda, T. DeepInsight: A Methodology to Transform a Non-Image Data to an Image for Convolution Neural Network Architecture. Sci. Rep. 2019, 9, 11399. [Google Scholar] [CrossRef]
  40. Zhu, Y.; Brettin, T.; Xia, F.; Partin, A.; Shukla, M.; Yoo, H.; Evrard, Y.A.; Doroshow, J.H.; Stevens, R.L. Converting Tabular Data into Images for Deep Learning with Convolutional Neural Networks. Sci. Rep. 2021, 11, 11325. [Google Scholar] [CrossRef]
  41. TMP. 2nd Place Solution—With 1D-CNN (Private LB: 0.01601), Kaggle. 2021. Available online: https://Www.Kaggle.Com/c/Lish-Moa/Discussion/202256 (accessed on 23 February 2023).
  42. Xi, Y.; Ren, C.; Tian, Q.; Ren, Y.; Dong, X.; Zhang, Z. Exploitation of Time Series Sentinel-2 Data and Different Machine Learning Algorithms for Detailed Tree Species Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7589–7603. [Google Scholar] [CrossRef]
  43. Sentinel-Hub. Sentinel2-Cloud-Detector. Available online: https://Github.Com/Sentinel-Hub/Sentinel2-Cloud-Detector (accessed on 21 February 2023).
  44. Abramowitz, M.; Stegun, I. Handbook of Mathematical Functions: With Formulas, Graphs, and Mathematical Tables; Dover Publications: Mineola, NY, USA, 1965. [Google Scholar]
  45. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  46. Chollet, F. Keras. Available online: https://Github.com/Fchollet/Keras (accessed on 21 February 2023).
  47. O’Malley, T.; Bursztein, E.; Long, J.; Chollet, F.; Jin, H.; Invernizzi, L. Keras Tuner. 2019. Available online: https://Github.Com/Keras-Team/Keras-Tuner (accessed on 3 March 2023).
  48. Lundberg, S.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. arXiv 2017, arXiv:1705.07874. [Google Scholar]
  49. Štrumbelj, E.; Kononenko, I. Explaining Prediction Models and Individual Predictions with Feature Contributions. Knowl. Inf. Syst. 2014, 41, 647–665. [Google Scholar] [CrossRef]
  50. Papes, M.; Tupayachi, R.; Martinez, P.; Peterson, A.T.; Asner, G.P.; Powell, G.V.N. Seasonal Variation in Spectral Signatures of Five Genera of Rainforest Trees. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 6, 339–350. [Google Scholar] [CrossRef]
  51. Blackburn, G.A.; Milton, E.J. Seasonal Variations in the Spectral Reflectance of Deciduous Tree Canopies. Int. J. Remote Sens. 1995, 16, 709–720. [Google Scholar] [CrossRef]
  52. Kodani, E.; Awaya, Y.; Tanaka, K.; Matsumura, N. Seasonal Patterns of Canopy Structure, Biochemistry and Spectral Reflectance in a Broad-Leaved Deciduous Fagus Crenata Canopy. For. Ecol. Manag. 2002, 167, 233–249. [Google Scholar] [CrossRef]
  53. Wu, Q.; Song, C.; Song, J.; Wang, J.; Chen, S.; Yu, B. Impacts of Leaf Age on Canopy Spectral Signature Variation in Evergreen Chinese Fir Forests. Remote Sens. 2018, 10, 262. [Google Scholar] [CrossRef]
  54. Clark, D.A.; Clark, D.B. Climate-Induced Annual Variation in Canopy Tree Growth in a Costa Rican Tropical Rain Forest. J. Ecol. 1994, 82, 865. [Google Scholar] [CrossRef]
  55. Mazza, G.; Sarris, D. Identifying the Full Spectrum of Climatic Signals Controlling a Tree Species’ Growth and Adaptation to Climate Change. Ecol. Indic. 2021, 130, 108109. [Google Scholar] [CrossRef]
  56. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  57. Scherer, D.; Müller, A.; Behnke, S. Evaluation of Pooling Operations in Convolutional Architectures for Object Recognition; Springer: Berlin/Heidelberg, Germany, 2010; Volume 6354, pp. 92–101. [Google Scholar] [CrossRef]
  58. Rudin, C. Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef]
  59. Lundberg, S.M.; Erion, G.G.; Lee, S.-I. Consistent Individualized Feature Attribution for Tree Ensembles. arXiv 2018, arXiv:1802.03888. [Google Scholar]
  60. Schuster, C.; Förster, M.; Kleinschmit, B. Testing the Red Edge Channel for Improving Land-Use Classifications Based on High-Resolution Multi-Spectral Satellite Data. Int. J. Remote Sens. 2012, 33, 5583–5599. [Google Scholar] [CrossRef]
  61. Tucker, C.J. Remote Sensing of Leaf Water Content in the near Infrared. Remote Sens. Environ. 1980, 10, 23–32. [Google Scholar] [CrossRef]
  62. Bing, W.W.; Kun, J.J.; ShunLin, L.L. Assessment of Sentinel-2 MSI Spectral Band Reflectances for Estimating Fractional Vegetation Cover. Remote Sens. 2018, 10, 1927. [Google Scholar] [CrossRef]
  63. Herrmann, I.; Karnieli, A.; Bonfil, D.J.; Cohen, Y.; Alchanatis, V. SWIR-Based Spectral Indices for Assessing Nitrogen Content in Potato Fields. Int. J. Remote. Sens. 2010, 31, 5127–5143. [Google Scholar] [CrossRef]
  64. Gitelson, A.A.; Kaufman, Y.J.; Merzlyak, M.N. Use of a Green Channel in Remote Sensing of Global Vegetation from EOS-MODIS. Remote Sens. Environ. 1996, 58, 289–298. [Google Scholar] [CrossRef]
  65. Alenizy, H.A.; Berri, J. Transforming Tabular Data into Images via Enhanced Spatial Relationships for CNN Processing. Sci. Rep. 2025, 15, 17004. [Google Scholar] [CrossRef]
  66. Damri, A.; Last, M.; Cohen, N. Towards Efficient Image-Based Representation of Tabular Data. Neural Comput. Appl. 2024, 36, 1023–1043. [Google Scholar] [CrossRef]
Figure 1. Map showing the location of the reference forest stands used in the analysis. The dots represent the centroids of the reference fields, and the color of the dots indicates the tree species of each field.
Figure 1. Map showing the location of the reference forest stands used in the analysis. The dots represent the centroids of the reference fields, and the color of the dots indicates the tree species of each field.
Forests 16 01039 g001
Figure 2. Graphs showing the evolution of the average reflectance for the selected bands recorded by the Sentinel-2 satellite. The four plots show data for four seasons. Line colors represent tree species.
Figure 2. Graphs showing the evolution of the average reflectance for the selected bands recorded by the Sentinel-2 satellite. The four plots show data for four seasons. Line colors represent tree species.
Forests 16 01039 g002
Figure 3. The workflow of the conducted analysis (a) and the topology of the applied convolutional neural network (b).
Figure 3. The workflow of the conducted analysis (a) and the topology of the applied convolutional neural network (b).
Forests 16 01039 g003
Figure 4. This figure illustrates the learning curves of convolutional neural networks (CNNs) for tree species recognition: (a) the change in the loss function (loss) for each learning epoch is displayed; (b) the variation of species classification accuracy (sca) for successive learning epochs is shown. The color indicates the curves for the training set and the test set.
Figure 4. This figure illustrates the learning curves of convolutional neural networks (CNNs) for tree species recognition: (a) the change in the loss function (loss) for each learning epoch is displayed; (b) the variation of species classification accuracy (sca) for successive learning epochs is shown. The color indicates the curves for the training set and the test set.
Forests 16 01039 g004
Figure 5. Receiver Operating Characteristic (ROC) curves for all classes (solid lines) and averaged curves (dotted lines) with respective Area Under the Curve (AUC) values.
Figure 5. Receiver Operating Characteristic (ROC) curves for all classes (solid lines) and averaged curves (dotted lines) with respective Area Under the Curve (AUC) values.
Forests 16 01039 g005
Figure 6. Confusion matrix of CNN model output (values on the plot are the fractions of a ‘true label’ sum).
Figure 6. Confusion matrix of CNN model output (values on the plot are the fractions of a ‘true label’ sum).
Forests 16 01039 g006
Figure 7. Influence of variables on the response of the classifier model (CNN network) expressed by SHAP index.
Figure 7. Influence of variables on the response of the classifier model (CNN network) expressed by SHAP index.
Forests 16 01039 g007
Figure 8. The final high-resolution map of eighteen classified tree species for Poland.
Figure 8. The final high-resolution map of eighteen classified tree species for Poland.
Forests 16 01039 g008
Table 1. List of tree species selected to analyze. For each species, the table lists the abbreviation, name, number of 10 × 10 m cells, and number of forest stands with Sentinel-2 satellite information.
Table 1. List of tree species selected to analyze. For each species, the table lists the abbreviation, name, number of 10 × 10 m cells, and number of forest stands with Sentinel-2 satellite information.
AbbreviationLatin NamePopular NameCells SampledNumber of Stands
ASHFraxinus excelsior L.European ash196,470336
BEEFagus sylvatica L.European beech458,448315
BLCRobinia pseudoacacia L.Black locust167,099351
BPNPinus nigra Arn. Black pine65,240161
DMPPinus mugo TurraDwarf mountain pine117,172127
EBAAlnus glutinosa (L.) Gaertn.European black alder506,234520
FIRPseudotsuga menziesii (Mirb.) FrancoDouglas fir71,295178
HBMCarpinus betulus L.European Hornbeam86,110176
LARLarix decidua Mill.European larch331,760300
OAKQuercus undefinedOak undefined398,095360
PINPinus sylvestris L.Scots pine436,300355
ROAQuercus rubra L.northern red oak162,934278
SBRBetula pendula RothSilver birch332,170355
SFRAbies alba Mill.Silver Fir506,935303
SLLTilia cordata Mill.Small-leaved linden57,385127
SPRPicea abies (L.) H.KarstEuropean spruce231,753269
SYCAcer pseudoplatanus L.Sycamore maple50,063163
WPPPopulus alba L.White poplar292,989584
Table 2. Model performance metrics calculated for species.
Table 2. Model performance metrics calculated for species.
SpeciesAccuracyPrecisionRecallF1 ScoreSupport
ASH0.560.540.560.5537,356
BEE0.860.760.870.8180,108
BLC0.850.800.860.8332,165
BPN0.550.770.560.6518,165
DMP0.960.960.970.9635,749
EBA0.850.870.860.86169,852
FIR0.550.740.560.6417,040
HBM0.550.670.550.6119,620
LAR0.880.840.890.8657,875
OAK0.730.770.730.75114,673
PIN0.920.850.930.89112,260
ROA0.690.690.690.6935,600
SBR0.810.800.820.8182,715
SFR0.910.860.910.88121,970
SLL0.400.480.400.447830
SPR0.750.830.750.7957,179
SYC0.340.460.350.399918
WPP0.690.770.690.7384,900
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mikołajczyk, Ł.; Hawryło, P.; Netzel, P.; Talaga, J.; Zdunek, N.; Socha, J. Classification of Tree Species in Poland Using CNNs Tabular-to-Pseudo Image Approach Based on Sentinel-2 Annual Seasonality Data. Forests 2025, 16, 1039. https://doi.org/10.3390/f16071039

AMA Style

Mikołajczyk Ł, Hawryło P, Netzel P, Talaga J, Zdunek N, Socha J. Classification of Tree Species in Poland Using CNNs Tabular-to-Pseudo Image Approach Based on Sentinel-2 Annual Seasonality Data. Forests. 2025; 16(7):1039. https://doi.org/10.3390/f16071039

Chicago/Turabian Style

Mikołajczyk, Łukasz, Paweł Hawryło, Paweł Netzel, Jakub Talaga, Nikodem Zdunek, and Jarosław Socha. 2025. "Classification of Tree Species in Poland Using CNNs Tabular-to-Pseudo Image Approach Based on Sentinel-2 Annual Seasonality Data" Forests 16, no. 7: 1039. https://doi.org/10.3390/f16071039

APA Style

Mikołajczyk, Ł., Hawryło, P., Netzel, P., Talaga, J., Zdunek, N., & Socha, J. (2025). Classification of Tree Species in Poland Using CNNs Tabular-to-Pseudo Image Approach Based on Sentinel-2 Annual Seasonality Data. Forests, 16(7), 1039. https://doi.org/10.3390/f16071039

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop