1. Introduction
The accurate and systematic monitoring of croplands at the regional and global scale is crucial for many agricultural applications, ranging from the monitoring of crop productivity and biomass estimation, to managing agricultural subsidies. Due to the dynamic structure of crops, timely updates of the crop maps are needed for crop yield forecasting and the planning of agricultural policies [
1,
2]. Due to its temporal and broad-coverage monitoring capabilities, remote sensing offers practical solutions to these challenges. To oversee the spatial and temporal changes of crops, as well as for accurate crop identification, the classification of multi-temporal images is a common approach, providing a cost-effective solutions as compared to traditional techniques, such as land surveying or cadastral surveying for mapping [
3,
4,
5]. The unavailability of cloud-free images from optical sensors can restrict the temporal monitoring of agricultural lands in some particular regions (e.g., tropical regions). Cloud coverage and haze could be a problem for large-scale agricultural applications, especially during the growing (or vegetation) season of crops. In such cases, Synthetic Aperture Radar (SAR) images are essential, which provide high temporal coverage in time-critical agricultural applications [
1,
4,
6]. The requirements for multi-temporal and multi-polarization SAR data for successful crop classification have been clearly stated and verified in many previous research papers, due to crops overall, as well as within various phenological stages, demonstrating different scattering mechanisms in the temporal domain [
1,
3,
4].
SAR signals are sensitive to crop geometric structure and dielectric properties, and hence they could provide complementary information to the optical sensors for crop analysis and monitoring. The radar backscatter, after calibration, is highly dependent upon the relationship between the target parameters (plant density, surface roughness, geometric structure, dielectric properties, etc.) and sensor configuration (incidence angle, polarization of wave, wavelength, etc.) [
7,
8,
9]. When compared to single or dual-polarization modes, full polarimetric SAR (PolSAR) sensors record the complete scattering characteristics of the targets (i.e., vegetation or crops) and enable the analysis of the independent elements of physical scattering [
7,
10]. These complete scattering mechanisms can be analyzed and simplified using polarimetric target decomposition techniques. Polarimetric target decompositions are a powerful approach for the interpretation and analysis of complex scattering mechanisms, since they separate the polarimetric measurements (covariance/coherency matrices) into several independent secondary elements, and then provide significant information regarding various scattering mechanisms. Polarimetric target decompositions can be grouped into two categories: (i) coherent decompositions and (ii) incoherent decompositions [
7,
11,
12]. Coherent decompositions deal with the single look scattering matrix
S, and can characterize the coherent or deterministic scatterer (i.e., man-made targets). In practice, targets that are observed by the SAR sensors have different scattering characteristics, known as distributed or incoherent scatterers (i.e., natural targets). A distributed scatterer can only be characterized or modelled by using second-order statistics that can be obtained by either coherency
T or covariance
C matrices. Incoherent decompositions, which utilize
T or
C matrices, are applied in order to analyze the complex scattering mechanisms of incoherent scatterers, and they are divided into two categories: eigenvector based decompositions (e.g., Cloude–Pottier and Van Zyl) and model based decompositions (e.g., Freeman–Durden and Yamaguchi) [
11,
12,
13,
14].
The ability of polarimetric target decompositions to classify land use/cover has been investigated in several studies in the past [
15,
16,
17]. However, only few studies have investigated radar-only crop-dominated classification by exploiting polarimetric target decompositions from multi-temporal C-band polarimetric RADARSAT-2 imagery. For this purpose, a wealth of classification algorithms has been utilized, such as support vector machines (SVM), maximum likelihood classification (MLC), random forest (RF), and nearest neighbor (NN). Ma et al. [
18] used the polarimetric parameters of Pauli decomposition for crop classification to assess the potential of multi-temporal RADARSAT-2 PolSAR data. They achieved an overall classification accuracy of 87.8% while using MLC. Liu et al. [
19] tested the multi-temporal RADARSAT-2 PolSAR images for the monitoring of crop growth. In order to analyze the scattering characteristics of crops, they exploited Pauli decomposition and used MLC for crop mapping. They reached up to 85% classification accuracy. Jiao et al. [
4] investigated the Cloude–Pottier and Freeman–Durden decomposition methods for crop mapping, using object-oriented classification and multi-temporal RADARSAT-2 PolSAR data. Their experimental results demonstrated that the Cloude–Pottier decomposition parameters achieved higher classification accuracy than the Freeman–Durden decompositions and linear backscatter polarizations. Larrañaga and Álvarez-Mozos [
2] evaluated the Pauli and Cloude–Pottier decompositions, as well as other several polarimetric features for crop mapping, with object based RF classification. They pointed out that C-band dual pol data could be sufficient if the dates of the images represent the key phenological changes of the crops. Shuai et al. [
20] proposed a novel method for maize mapping using a parcel-based SAR classification with SVM. In their experiment, they extracted the polarimetric features from Pauli and Cloude–Pottier decompositions. Their experimental research concluded that their proposed method obtained a classification accuracy of 94.7%, and proved the added benefit of multi-temporal PolSAR data for maize mapping.
These aforementioned studies explored the potential of coherent and incoherent polarimetric target decompositions in crop classification while using multi-temporal C-band PolSAR data. However, the inter-comparison of incoherent decompositions (eigenvector based decompositions versus model based decompositions) in discriminating crop types from multi-temporal C-band PolSAR data has yet to be fully explored in the literature. Jiao et al. [
4] used only one method (Freeman–Durden) from the model-based decompositions in their experiment. Furthermore, none of these above studies compared the capabilities of the
C or
T matrices alongside polarimetric target decompositions in crop classification using multi-temporal PolSAR data.
Over the last two decades, an ensemble of classifiers (multiple classifiers) have been widely exploited in the field of pattern recognition and machine learning for classification/regression problems, as ensemble classifiers are believed to attain higher accuracy when compared to single classifiers [
21,
22]. RF is the most widely used and recognized ensemble learning method in remote sensing due to its high accuracy, short computation time, and notable efficiency [
22]. Recently, new ensemble learning algorithms, such as canonical correlation forest (CCF) (2015), extreme gradient boosting (XgBoost) (2016), and Light Gradient Boosting Machine (LightGBM) (2017), have been introduced to the machine learning community [
23,
24,
25]. A very limited number of papers have been published regarding CCF [
26,
27] and XgBoost [
28,
29,
30] for classification purposes in remote sensing; however, no study has been published yet using the recently launched LightGBM, which is a highly efficient gradient boosting decision tree that was developed by Microsoft Research in the field of remote sensing for classification purposes. Only in one paper by Liu, Ji, and Buchroithner [
31] has LightGBM been tested, in this case for soil property retrieval by combining partial least squares. To the best of our knowledge, this experimental study is the first implementation of LightGBM for crop classification while using PolSAR data.
The main contributions of this research can be summarized, as follows:
- (1)
The performances of incoherent polarimetric target decompositions (eigenvector-based versus model-based) for discriminating crops using multi-temporal C-band PolSAR data were compared.
- (2)
For the first time, LightGBM was used for crop classification using the polarimetric features of multi-temporal C-band PolSAR data.
- (3)
The performance of polarimetric target decomposed parameters for crop classification was compared with the original polarimetric features (linear backscatter coefficients, and T and C matrices).
2. Study Site and Dataset
In this section, the test site will be introduced and the details regarding the dataset will be presented. The test site contains about 42 km
2 of agricultural land, and it is situated approximately 65 km north of Konya city center (see
Figure 1), extending from 38°43′ to 38°53′ North latitudes and 32°43′ to 32°53′ East longitudes. The terrain of the test site is flat, making the area convenient for precision farming. Five crop types are grown in the test site: maize, potato, wheat, sunflower, and alfalfa.
The multi-temporal RADARSAT-2 quad-polarimetric dataset was acquired from the test site using the following four acquisition dates: 13 June, 7 July, 31 July, and 24 August in 2016 (see
Figure 2). All of the images were acquired as single look complex data and in fine quad-polarization acquisition mode. The specifications of the RADARSAT-2 images are presented in
Table 1.
The agricultural fields were visited on the acquisition dates of the images and the ground truth data with the phenological stages for each crop type were recorded.
Table 2 presents the details for the training and testing dataset (number of pixels).
The
Biologische
Bundesanstalt, Bundessortenamt and
CHemical industry (BBCH) scale [
32] described the crop growth development, as leaf development, stem elongation, heading, and flowering in general (
Table 3).
The phenological changes for each crop type in the multi-temporal dataset leads to better discrimination of the crops, as the backscattered signal for each phenological stage provides different information regarding crop structures and growth. There is no exact planting time for alfalfa crops.
4. Experimental Results and Discussion
In this subsection, the classification results (and class based accuracies) are presented, and the impacts of polarimetric target decomposition are then discussed. The inter-comparison of the incoherent polarimetric target decomposition methods, along with the original PolSAR features for crop classification, was performed in this study by using the LightGBM ensemble learning classifier. The accuracies of the classified images (and class predictions) were derived from the error matrix (confusion matrix).
Figure 4 displays the overall classification accuracies. The highest classification accuracy was obtained by the Yamaguchi decomposition method (92.07%), followed by the Van Zyl (91.76%) and Freeman–Durden (91.60%) decompositions. All of the polarimetric target decomposition methods, with the exception of Cloude–Pottier (78.83%), outperformed the backscatter coefficients (87.65%) in terms of overall classification accuracy. The model based decompositions (Yamaguchi and Freeman-Durden) outperformed the eigenvector based decompositions, with the exception of the Van Zyl method. Even though the Cloude–Pottier resulted in a lower classification accuracy than the backscatter coefficient and other polarimetric target decompositions, it obtained a higher classification accuracy than the
T matrix (76.47%) and
C matrix (75.99%). Furthermore, when comparing the classification performances of the
T and
C matrices, the
T matrix outperformed the
C matrix. The possible reason for this is because the
T matrix includes information regarding the physical scattering mechanism, however the
C matrix includes only the second order formalism of the scattering matrix and allows for the direct analysis of polarization channels [
39].
Alfalfa had the lowest F1-score value from among the crop types, ranging from 0.23 (as obtained by the
C matrix) to 0.37 (as obtained by the backscatter coefficient). Even though the polarimetric target decomposition methods, with the exception of the Cloude–Pottier, outperformed the backscatter coefficient in terms of overall classification accuracy, the backscatter coefficient had the highest F1-score value for alfalfa. The second highest value for this class was obtained by Yamaguchi (0.36), followed by Van Zyl (0.34). No feature type (original features or polarimetric target decompositions) could reach the F1-score value of 0.50 (see
Figure 5) for alfalfa, as this class was mostly misclassified as maize, as it can be deduced from the normalized confusion matrices in
Table 6. The row and column in the confusion matrix represent the true (actual) and predicted class, respectively. Additionally, the diagonal elements (in bold characters) in the matrix show the degree of correctly predicted classes. The major (and possible) reason for this failure is that alfalfa is the minor class in our experiment, and the classifiers mostly fail to classify the minor classes. Furthermore, the backscatter values of alfalfa and maize are close or equal to each other in the temporal domain (see
Figure 6) on some polarizations (i.e., VV or VH channels). The highest value of F1-score for the maize class was obtained by Yamaguchi, Van Zyl, and Freeman–Durden (0.92), followed by the backscatter coefficient (0.86) and Cloude–Pottier (0.76) method. With the exception of the backscatter coefficient, the original feature, had the lowest F1-score value, at 0.74 for the maize class. Among the polarimetric target decompositions, only the Cloude–Pottier obtained a lower F1-score value than the backscatter coefficient for maize. The sunflower class is the most accurate classified class type, as Yamaguchi, Van Zyl, and Freeman–Durden obtained the maximum value of F1-score (1.00). Polarimetric target decompositions had higher F1-score values than the original features for sunflowers.
The Cloude–Pottier decomposition outperformed the backscatter coefficient by 0.04 in terms of the F1-score value for sunflowers, whereas the backscatter coefficient obtained higher accuracy than Cloude–Pottier by 8.82% for overall accuracy. This inconsistency clearly demonstrated that each crop type could have various scattering characteristics and impact on each polarimetric feature (both original features and polarimetric target decompositions) in the classification process. The F1-score values for sunflowers were less than 0.90 from the T (0.80) and C (0.79) matrices.
With the exception of the Cloude-Pottier decomposition, all of the decomposition methods obtained the highest F1-score values for potatoes (0.97), followed by the backscatter coefficient (0.94). The Cloude-Pottier decomposition had the lowest F1-score value (0.71) for potatoes, which is an exception for this class. For the other class types, the
C matrix obtained the lowest F1-score values. Potatoes were partially confused with maize for the classification of the
C matrix and Cloude–Pottier decomposition (
Table 6).
For the all of the features (original or polarimetric target decomposition), wheat was classified with a higher value than 0.9 in terms of the F1-score (see
Figure 5). This was a predictable result, since the backscatter response of wheat in the temporal domain for all polarization channels is clearly distinguishable from others (see
Figure 6). The obtained results for wheat can also be considered to be an exception in our experiment, because the F1-score values that were obtained for the other classes were lower than 0.9 from the
T or
C matrix. The classified images for all incoherent polarimetric target decompositions and original features are depicted in
Figure 7.
5. Conclusions
This paper tested the capabilities of polarimetric target decompositions for crop classification using LightGBM, and evaluated the inter-comparison of polarimetric target decomposition in terms of overall and class based accuracies. Our experimental results demonstrated that all polarimetric target decomposition methods, with the exception of the Cloude–Pottier decomposition, were superior to the original features in terms of overall classification accuracy. The highest classification accuracy (92.07%) was achieved by Yamaguchi, whereas the lowest (75.99%) was obtained from the C matrix. Additionally, model based decompositions can be concluded to be more powerful than eigenvector based decomposition in terms of class based accuracies (based on F1-score values) from our study. When class-based accuracies were evaluated (F1-score values), alfalfa and sunflowers were identified as the least and most correctly predicted class types, respectively. Alfalfa was mostly confused with maize and it had the lowest F1-score values. This result was predictable, since classification algorithms often fail to classify minor classes and the backscatter values of alfalfa were close or equal to maize in the temporal domain. Our results clearly prove the added benefits of model based decompositions for crop type classification from PolSAR data.
The main limitation of the data used in this research is that RADARSAT-2 is the commercial earth observation satellite data and it is not freely available to everyone. In most cases, obtaining the multi-temporal PolSAR data can be costly. In this case, there was not any freely available alternative to full polarimetric RADARSAT-2 in C band frequency for the investigation of polarimetric target decomposition. Additionally, we observed that the main drawback of the classification model (LightGBM) was the need for tuning of the parameters.
In addition, this study investigated, for the first time, the capabilities of LightGBM in the classification of crops from multi-temporal PolSAR data. Our future research will focus on testing the different types of polarimetric target decompositions (e.g., Touzi decomposition, Van Zyl/Arii Decompositions, etc.) for crop classification, using the CatBoost (gradient boosting with categorical features support), as well as the comparison with LightGBM.