Next Article in Journal
Satellite Monitoring of Mass Changes and Ground Subsidence in Sudan’s Oil Fields Using GRACE and Sentinel-1 Data
Next Article in Special Issue
Agreement Index for Burned Area Mapping: Integration of Multiple Spectral Indices Using Sentinel-2 Satellite Images
Previous Article in Journal
An Assessment of Global Forest Change Datasets for National Forest Monitoring and Reporting
Previous Article in Special Issue
An Interaction Methodology to Collect and Assess User-Driven Requirements to Define Potential Opportunities of Future Hyperspectral Imaging Sentinel Mission
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Similarity Metrics Enforcement in Seasonal Agriculture Areas Classification

1
Computing and Informatics School, Mackenzie Presbyterian University—Rua da Conslação, 930 Consolação, São Paulo SP 01302-90, Brazil
2
Embrapa Informatica Agropecuária, Environmental Modeling Lab—Av. Andre Tosello, 209 Unicamp Campus, Campinas SP 13083-886, Brazil
3
Getúlio Vargas Foundation/São Paulo School of Economics—Rua Itapeva, 474 Bela Vista, São Paulo SP 01332-000, Brazil
*
Author to whom correspondence should be addressed.
Remote Sens. 2020, 12(11), 1791; https://doi.org/10.3390/rs12111791
Submission received: 2 April 2020 / Revised: 26 May 2020 / Accepted: 27 May 2020 / Published: 2 June 2020

Abstract

:
Accurate identification of agriculture areas is a key piece in the building blocks strategy of environment and economics resources management. The challenge requires one to deal with landscape complexity, sensors and data acquisition limitations through a proper computational approach to timely deliver accurate information. In this paper, a Machine Learning (ML) based method to enhance the classification process of areas dedicated to seasonal crops (row crops) is proposed. To this objective, a broad exploration of data from Moderate Resolution Imaging Spectro-radiometer sensors (MODIS) was made using pixel time-series combined with time-series similarity metrics. The experiment was performed in Brazil, covered 61% of the total agriculture areas, five different states specifically selected to demonstrate biome differences and the country’s diversity. The validation was made against independent data from EMBRAPA (Brazilian Agriculture Research Corporation), RapidEye Sensor Scene Maps. For the eight tested algorithms, the results were enhanced and demonstrate that the method can rate the classification accuracy up to 98.5%, average value for the tested algorithms. The process can be used to timely monitor large areas dedicated to row crops and enables the application of state of art classification techniques, two levels classification process, to identify crops according to each specific need within the areas.

Graphical Abstract

1. Introduction

The identification of agriculture areas is valuable information for the scientific community, government agencies, farmers, and other members of the society. Agriculture main commodity areas are monitored on a global scale to predict production, yield, demand, prices, climate risks, and so forth [1,2,3,4,5]. Related statistics are produced and published to reduce the economic externalities impact and to balance market information asymmetry [6,7]. Ongoing, accurate, and timely crop information remains a challenge, as data becomes available long after the harvesting time, and the machine learning classification is dependent on reliable data sources and processing needs enhancements, despite the available capacity [8,9,10]. The accuracy of the latest results in distiguishing crops of agriculture areas range from 84% to 95% [11,12,13,14]. The state of the art publications in agriculture remote sensing explore the importance of machine learning applied in combined data sources and types in order to enhance the process and accuracy for multiple crops as classes.
When it comes to process improvements and machine learning, Convolutional Neural Networks (CNN) was applied to achieved significant results in image recognition tasks by automatically learning a hierarchical feature representation from raw data, and through combinations of time series (temporal dimension) with 2D texture images (spatial dimension) to enhance features that could not be gained in a single dimension. The unified frame work, times series, and CNN, demonstrated competitive accuracy when compared with the existing deep architectures and the state-of-the art time series algorithms [15]. The exploration of synergies between different sources of data to improve classification of high-level spatial features produced by hierarchical learning (i.e., scene labeling) to contrast with low-level features such as spectral information (morphological properties) has also been made. Temporal and angular features played more important roles in classification performance, especially abundant vegetation growth information. Multispectral and hyper-spectral fusion successfully discriminated natural vegetation types diversity [16]. Also using fusion methods techniques, in this case to mitigate spatiotemporal limitations of multi and hyper-spectral data from multiple sensors, a compatibility between the hyper-spectral data and Sentinel-2 (multi spectral) data has been validated. The method opened new possibilities for classifying complex and heterogeneous land covers in multiple environments with the combination multiple data sources [17]. On the other side, when accuracy enhancement is the major focus, an automated mapping process for soybean and corn using crop phenology characteristics with time series and topographic features from multiple sources combined was proposed by Reference [18]. The classification achievements range from 87% to 95%, which is an increase of 2.86% on previous published works [19]. For the case of row-crop areas in Brazil, as is the case of this work, international publications that addresses this need were observed. A experiment that covered 3 crops (corn, soybeans and cotton) in Mato Grosso, used time series correlation coefficient and successive classifications has been performed to detect agricultural areas. The approach was capable to achieve 95% accuracy and kappa index of 0.98 [14]. In another study, a rigorous multiyear evaluation of the applicability of time-series for crop classification in Mato Grosso was performed. The conclusion showed progress in refined crop-specific classification and appointed the need for grouping of crops as classes. The results were consistently near or above 80% accuracy and Kappa values were above 0.60. The authors also highlight the need for additional research to evaluate agricultural intensification and extensification in this region of the world [20]. The combination of different channels (red and near-infrared) were explored with five algorithms (Maximum Likelihood, Support Vector Machines, Random Forest, Decision Tree, and Neural Networks). The methods accuracy ranged from 85% to 95%, and demonstrated that 250 m imagery is efficient to map fields down to 20 ha. Results also suggested that cropland diversity could be addressed using regional and specific landscapes training sets [13]. As a reference, the accuracy assessment of a supervised classification on Landsat 8 satellite images was performed. The results indicated that the object classification was better than the classifications by pixel and the best thematic map was generated by the SEGCLASS classifier. The accuracy achieved was 74% and kappa index 0.57 [21].
In common is that References [11,12,13,18,21,22,23,24] share the best results in the crop classification with their methodologies within the studied area, and present the challenges related to the limited amount of training data to scale up the process. As each of the studies was performed within unique field information and conditions, including different crop varieties contrasted, the results cannot be compared properly.
In this research the aim is to propose a process to enhance classification accuracy in agriculture areas dedicated to row crops and put them in evidence to support the land monitoring process and generate a first knowledge layer to support specifics needs and fine-tuning classification. Although time series has been extensively explored to classify agriculture and it is a well-stablished process [25,26], the combination of time series and similarity metrics to explore the classification as proposed is a novel, explore the crop growing season cycle of a variety of temporary crops with multiple algorithms.

1.1. Remote Sensing

The use of the pixel as a sensor in remote sensing to monitor agriculture cycles requires quality images and fine specs to be granularly classified accordingly to each unique purpose, so unexpected dynamics can be detected and properly managed [8,9], (e.g., crop phenology stage, chemicals administration, mechanization, and irrigation management, among others). The critical components for remote sensing are: accurate and current information for training the classifiers; an affordable source of data that qualifies for the specific objective; a storage and processing capacity [10,27,28]. By using MODerate-resolution Imaging Spectrometer (MODIS) products that include the Enhanced Vegetation Index (EVI) [29], among others, we have access to a complete record of data from each of the Terra and Aqua MODIS sensors, at varying spatial (250 m, 0.05 degree) and temporal (8-day) resolutions validated with accuracies depicted by a pixel reliability flag and with globally averaged uncertainties of 0.015 units. Further, the MODIS/EVI combination is a robust set for exploring seasonal crops (soybeans) [22]. The spatial resolution adopted are consistent in expressing accurate cropland information in fields that are larger than 20 ha [13], and appropriate to this task [30]. This experiment relies on the achievements and specifications above to explore the spectral dynamics of extensive areas, in accordance with the specifics purpose stablished for this research.

1.2. Brazilian Agriculture in Numbers

In Brazil, agriculture represents almost a quarter of the country GDP’s, 24.1% in 2017 [31]. According to the Brazilian Institute of Geography and Statistics (IBGE) rural census [32], the harvested area during the 2016/2017 season represented 7.9% of the total area with 73,797,057 hectare (ha) dedicated for temporary and semi-temporary crops and another 5,184,813 ha for permanent crops, with the final results published in 2019. For the purpose of this study, detailed statistics of the studied areas are summarized on the Table 1. The column “Total” shows the total area (country/State). The column “Agriculture” presents the harvested area (country/State). It is important to highlight that, due to geographical and environmental conditions, which include climate and/or technologies (e.g., irrigation), it is common to have multiple crops per year in tropical areas. Therefore, the conclusion is that the harvests are larger than the total area dedicated to agriculture, a total of 67,547,537 ha according to EMBRAPA [33,34] and MAPA [35].
The main crops cultivated locally are presented below, in Table 2, and the data is organized as a percentage of total area occupied to provided a land use perspective view, and put in evidence of a larger area per crop cultivated.

2. Methods

2.1. Studied Area

The study was performed over Brazil at the state level, see Figure 1. The locations were selected to bring the diversity (biome, crop, climate, etc.) and its complexity into the analysis. The states studied were: Goiás (GO), Minas Gerais (MG), Mato Grosso (MT), Parana (PR) and São Paulo (SP), Table 1. In this study the growing season (phonological) cycle is the key aspect taken in consideration, crop recognition requires specifics and may be performed in accordance to them, combining scenes and the growing season maps.

2.2. IBGE Census

As a reference in this study, IBGE Agriculture census data was used to contrast results, that is, quantify the total area and row crop areas for each studied state. The original numbers were consolidated and published by municipalities and by state, consequently. For this research objective the productive clusters are represented in maps and at the state level, without municipal geopolitical boundaries. The preliminary census results were published in 2018 and final results late in 2019.

2.3. Ontology

The use of expert knowledge management based on ontology process to support the collective understanding of a single event (Data driven approach) was applied to proper address crop growing season specifics to expand the comprehension possibilities in a multidisciplinary context [36]. The same author defined ontology scope as being the “specialist of specifics”, meaning the understanding of a domain group. In this environment the essential is the potential for information and knowledge sharing [37]. The proposed Ontology is presented in the concept map below, Figure 2, where the target class are highligted and the profiles are demonstrated. The data was segmented in accordance to the Global Food Security (GFS) [38] and MODIS ontology into two large classes, Agriculture and Other. Agriculture, as the target class, included: Irrigated and dry seasonal crops areas potentially featuring corn, soybean, cotton, barley, potato, alfalfa, sorghum, rye, canola, peanut, manioc, and beet. The Others class included: perennial and semi-perennial/semi-temporary crops, natural vegetation (forest, cerrado, amazon forest, tropical forest), pasture, urban areas, and water surface, among others. The agriculture class areas are those where we expected to find a high level of EVI values dynamics, which is the purpose of this study, and is also compatible with the objective of the sensor selected. As a reference, the profile images that are representing the target group are from SatVeg [39]. Only the crops that can be found within the region and are described as part of sensor ontology are here in evidence.

2.4. Datasets Processing

The criteria for the selection of the sensor was based on the most recent publications and goal, as cited above. The .tiff (Tagged Image Format) dataset files were downloaded from the MODIS repository as .hdf (Hierarchical Data Format) files according to provided guidelines, and then framed in accordance to each studied state and a country view. The experiment was performed using EVI from MOD13Q1 with 250 m of spatial resolution.
Accord to the methodology outlined as a flowchart in Figure 3, a large amount of random Pixels Time Series (PTS) were extracted from each studied area to achieve 99 % 1 + 1 of confidence level, with 17,000 points for each dataset, and 5 datasets per state. PTS values were extracted as vectors and stored as .csv (Comma Separated Values) files using a expert system created for this purpose. Each vector contained the pixel position, EVI raw data as time series values, and calculated computational distances, with similarity metrics proposed. The data were then pre-classified into two large groups, Agriculture and Others. Agriculture was the target group, which expected areas with high spectral dynamics level, and the Others group contained lower dynamics level areas. Each experiment used 20% of data set for training.
Given the inherent complexity of vegetation and environments which are all reflected within the region selected, for each state, a group of similarity metrics were selected and tested as a way to put in evidence the expected characteristics and work as a hybrid index in the classification process, as in Reference [40]. The computational distances used in this experiment are: Manhattan distance, Minkowski distance, Sum, Mean, Median, Standard deviation, Coefficient of Variation, Variance, and Difference [41,42]. All the similarity metrics that used the distances were calculated for each pixel time-series. In particular, the Minkowski distance used has been explored by References [43,44] to cluster time series that have different temporal resolutions, as the case for the Dynamic Time Warp (DTW), and in this case it was used to measure the length of the time-series. The possible values for the c are useful to accommodate the time difference into the time series. The equation is demonstrated below, Equation (1):
D ( p , q ) = ( i = 1 n | p i q i | c ) 1 / c .
where: the distance D, p and q are the data points ( x i + 1 and x i , respectively) and used c = 2 .

2.5. Time-Frame

For this research, 36 raster layers from the MODIS sensor were used, with an 18 month time frame coverage, from July 2016 to December of 2017 to match IBGE census data also used as a reference in this study. The time frame selected meets the objective of measuring annual dynamics for temporary crops, thus excluding semi-temporary crops (e.g., sugarcane) from the analysis.

2.6. Validation

Tree levels of validation were considered—the classification accuracy, the concordance between algorithms, and the computational metrics relevance. EMBRAPA scene maps were used as an external reference data source to contrast and validate accuracy. The spatial resolution of the the validation maps was 5 m to provide quality data for the training set used, to preserve the characteristics that are relevant to the experiment [8], and to match the field size.
The classification accuracy validation was made according to the proposed ontology, a confusion matrix used to contrast results as True-Positive (TP), False-Negative (FN), False-Positive (FP) and True-Negative (TN). Accuracy stands for all that were correctly classified, as in the Equation (2), and recall stands for positives that were correctly classified, as in the Equation (3), below:
a c c u r a c y = T P + T N T P + F P + T N + F N ,
r e c a l l = T P T P + F N .
The concordance between algorithms assessment were made using the Cohen’s Kappa coefficient or simply (k), which has been largely used to evaluated classification algorithms performance as described in Reference [45]. Kappa values ranges from 1 to 1, where 1 represents “complete disagreement”, 0 is a “random classification”, and 1 is a “perfect agreement”.
The computational metrics used are the sensitivity and specificity, Equations (4) and (5) below, where we expect to identify the impact of the selected similarity metrics composition on the classification process of the target group.
s e n s i t i v i t y = t r u e p o s i t i v e p o s i t i v e
s p e c i f i c i t y = t r u e n e g a t i v e n e g a t i v e ,
where “true positive” is the number of correctly predicted areas, “positive” is the number of agriculture, “true negative” is the number of correctly predicted as others, and “negative” is the number of others shown in the classification.
The impact of each similarity metric as an attribute over the target class was evaluated using Shannon entropy index to reveal in a [0:1] scale the cluster behavior [46,47]. The calculation of the distances was made according to the Equation below, Equation (6):
E i = i = 1 n p i log ( p i ) ,
where E is the information strengh of the attibute n over the class i, and p is the probability of the class in n.

2.7. Computation Tolls

All data manipulation and tests were performed using R ® language, Matlab ® , RapidMiner ® , and QGIS ® . Using these tools to evaluate the effectiveness of the proposal, we explored the data with eight Machine Learning algorithms: Naive Bayes, Logistic Regression, Decision Tree, Gradient Boosted Tree, Generalized Linear Model, Deep Learning, Random Forest. This collection represents most of the algorithms used in the cited publications.
The data used were min-max normalized to comply with the classification tolls specs. The values were linearly reduced to a scale between [0:1], where 0 and 1 are the minimum and maximum values, respectively. The z is the normalized value accord to Equation (7) below:
z i = x i m i n ( x ) m a x ( x ) m i n ( x ) .

3. Results

By performing the study as proposed, we observed that the addition improved the specificity and sensitivity of the algorithms. The sensitivity of the algorithms can be verified by contrasting ROC comparison curve (detection probability in machine learning). The Figure 4 shows the classification process for data without similarity metrics and Figure 5 shows the sme process for data with similarity metrics. Specificity raised from 98.6% to 99.2% with a direct effect on accuracy, sensitivity enhancement from 94.62% to 97.4% which improved processing performance despite the data dimensional increase. As we can verify, data with similarity metrics increased sensitivity for all tested algorithms.
The enhancement for each method is demonstrated below, Table 3, where the results for each state are presented. The caption MT, GO, MG, SP and PR are the abbreviations of the states mentioned on Section 2.1. The columns RD and DSM represent the accuracy with the different type od data, Raw Data (RD) and Data with Similarity Metrics (DSM). It is important to point the difference in results between the algorithms, these differences reflect the way each of them internally organizes the data toward the best solution.
The concordance of the algorithms index (Kappa) and the enhancement range achieved for each method with Raw Data and Data with Similarity Metrics (RD-DSM) is demonstrated in the Table 4. The highest impact was achieved with Deep Learning, increased the accuracy from 83.9% to 98.6% with Kappa index of 0.95. However, the highest accuracy was achieved with Gradient Boosted Tree, results were enhanced from 95.2% to 99.6% and with Kappa index of 0.96.
The hierarchal clustering analysis below, Figure 6, put in evidence the hierarchal importance of each metric for the process. The relevance of each attribute is calculated accord to Shannon information and the normalized score is represented hierarchically. The Minkowski distance played a key part in the classification process and demonstrated to be very representative of the target class, highest score. Canberra distance, Standard deviation, Mean, Coefficient of Variation, and Variance were also very relevant in the high score process, but not as decisive. The removal of the Manhattan distance, Sum, Median, and Difference had no impact on the process accuracy.
The correlation between a group of indexes found in the bibliography were also evaluated as alternatives to the EVI. Results are presented in Table 5, where the correlations are contrasted. From this perspective, none of the indexes were found as strong enough to be used as alternative or to be replaced within the same method.
Finally, from the process it was possible to generate a scene map layer identifying row crops areas, as presented on Figure 7, below. The figure shows where the row crops clusters are located.

4. Discussion

As presented in the Results section, the proposed approach enhanced the classification accuracy for all tested methods. The similarity metrics in combination with time series improved the average accuracy from 88.7% to 98.5%, as presented on Table 4. Therefore, the proper selection of similarity metrics showed potential to enhance the classification efficiency, increase classification accuracy without extra processing time.
The results achieved are higher than all previous work that addressed classification of agriculture row crop areas in Brazil and, in particular, the results achieved by References [13,14,20,21,22,24] as presented in Table 6. In regard to Reference [14], which achieved the highest result prior to this work, despite the large area, they only study one state (Mato Grosso) and tree crops. The second highest result achieved [13], compared results for 6 algorithms and results ranged from 84% to 96%, covered 5 countries and noticed the lowest accuracy for Brazil (84%). The same author appointed as a key finding that “the site effect dominates the method effect”.
It is important to highlight that in Brazil we have summer crop season which is in general rainy and cloudy, more than 90% of the production is rainfed. According to References [49,50], less than 40% of temporal images are useful for analysis. In this context, in this research, we bring the need for the combination of high temporal resolution with high spatial resolution for validation, MODIS and RapidEye, respectively.
According to Reference [48] the combination of temporal and spectral information can improve classification accuracy than only using spectral information, 10–15% higher. In this work, we demonstrated that the proper similarity metric can enhance time series classification accuracy, 3–14%, 10% on average for the tested algorithms. In fact, the combination created condition to increase specificity (accuracy) and sensibility (convergence). Both aspects are demonstrated on Figure 4 and Figure 5 above. The sensibility, in this case the positive impact on the classification process, for each of the algorithms were different. Below we present the sensibility for each one of the algorithms:
  • Low: Suport Vector Machine, Deep Learning, Gradient Boosted Tree
  • Moderate: Naive Bayes
  • High: Logistic Regression, Decision Tree, Gradient Boosted Tree, Random Forest
Two level classification strategy, as proposed by Reference [18], or multi-level as used by [14] are a possibility from the achieved results. The layer identifying temporary crop clusters, as presented on Figure 7, meets the demand presented by References [12,20,22,48], among others.
Another important aspect is the potential of the process to remotely identify agricultural areas as a dynamic census process. Timely spatial information about productive areas with a high confidence level ( 99 % 1 + 1 ) is relevant for policy makers and for the private sector. This information can support IBGE statics and contribute to their monitoring process. In regard to private sector, precision agriculture services can be provided in large scale when geospatial data organized and available.
The process also demonstrated some limits that are related to the objective of each search and complexity of the environment. An accurate time-frame collection of maps is required to enhance the uniqueness of the target. It is also important to remember that on top of spectral resolution, temporal and spatial resolutions are key to put in evidence specifics of cultivars. Therefore, sugarcane and cultivated pasture have specifics in growing season, and requires specifics in spectral and temporal resolution to enhance the classification process [51].

5. Conclusions

This study explored the classification of agriculture areas dedicated to temporary crops using well-known ML algorithms to examine time-series from a collection of vegetative indexes in combination with similarity metrics. As a conclusion we have that:
  • As the primary objective, the results demonstrated that the approach enhanced the classification accuracy for all tested algorithms. This is the highest accuracy for the classification of agriculture areas in Brazil, 99.6% with kappa index of 0.96 using Gradient Boosted Tree algoriothm;
  • The similarity metrics worked to increase the accuracy within the context proposed, EVI data reflecting the growing season dynamics of temporary crops. The similarity metric added 3–14% points to the accuracy;
  • The process increased accuracy and without extra computational cost.
  • The results are robust to support policy maker and precision farming, 99 % 1 + 1 of confidence level.
The field knowledge (scene map generated) allows crop level classification improvements, that is, crop differentiation using the available techniques accordingly to each specific need. This information is useful for further research and also to support the private sector and public sectors on monitoring and spatial planning of annual crops in Brazil.
To enhance results and explore the approach, modifications in the Minkowski equation, c value, should be tested to address specifics in phenology time frame. The size of the data set for training should be quantified, it is expected that the a smaller data set size would maintain accuracy. Moreover, DTW can be used to explore the differences in growing season for the two levels crop classification process, among others. Specifics in computational costs and cloud computing benefits should also be explored and demonstrated.

Author Contributions

M.A.S.S.: Conceptualization of this study, Methodology, Software coding and manipulation, Validation Methodology, Writing—Original draft preparation; E.D.A.: Review of the methodology and validation process; A.C.G.: Writing review; N.O.: Supervisor. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

We thank Mackenzie Presbyterian Univesity, Embrapa Agroinformatica, Fundação Getúlio Vargas, Capes, for the support with this research. We also thank Vanessa Pugliero and Eduardo Pavão from Embrapa for the support with organization of the information used during the validation process.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Johansson, R.; Luebehusen, E.; Morris, B.; Shannon, H.; Meyer, S. Monitoring the impacts of weather and climate extremes on global agricultural production. Weather Clim. Extrem. 2015, 10, 65–71. [Google Scholar] [CrossRef] [Green Version]
  2. Shi, Z.H.; Li, L.; Yin, W.; Ai, L.; Fang, N.F.; Song, Y.T. Use of multi-temporal Landsat images for analyzing forest transition in relation to socioeconomic factors and the environment. Int. J. Appl. Earth Obs. Geoinf. 2011, 13, 468–476. [Google Scholar] [CrossRef]
  3. Kaliraj, S.; Chandrasekar, N.; Magesh, N.S. Evaluation of multiple environmental factors for site-specific groundwater recharge structures in the Vaigai River upper basin, Tamil Nadu, India, using GIS-based weighted overlay analysis. Environ. Earth Sci. 2015, 74, 4355–4380. [Google Scholar] [CrossRef]
  4. Herman, E.; Haesen, D.; Rem, F.; Urbano, F.; Tote, C.; Bydekerke, L. Monitoring the impacts of weather and climate extremes on global agricultural production. ISPRS J. Photogramm. Remote Sens. 2015, 53, 154–162. [Google Scholar] [CrossRef] [Green Version]
  5. Lahousse, T.; Chang, K.T.; Lin, Y.H. Landslide mapping with multi-scale object-based image analysis-a case study in the Baichi watershed, Taiwan. Nat. Hazards Earth Syst. Sci. 2011, 11, 2715–2726. [Google Scholar] [CrossRef]
  6. Coase, R.H. The Nature of the Firm 1937. Economica 1937, 4, 386–405. [Google Scholar] [CrossRef]
  7. Barzel, Y.; Kochin, L.A. Ronald Coase on the Nature of Social Cost as a Key to the Problem of the Firm. Scand. J. Econ. 1992, 94, 19–31. [Google Scholar] [CrossRef]
  8. Guijarro, M.; Pajares, G.; Riomoros, I.; Herrera, P.J.; Burgos-Artizzu, X.P.; Ribeiro, A. Automatic segmentation of relevant textures in agricultural images. Comput. Electron. Agric. 2011, 75, 75–83. [Google Scholar] [CrossRef] [Green Version]
  9. Mulyono, S.; Pianto, T.A.; Fanany, M.; Basaruddin, T. An ensemble incremental approach of Extreme Learning Machine (ELM) For paddy growth stages classification using MODIS remote sensing images. Adv. Comput. Sci. Inf. Syst. 2013, 309–314. [Google Scholar] [CrossRef]
  10. Jones, H.G.; Vaughan, R.A. Remote Sensing of Vegetation: Principles, Techniques and Applications; Oxford: New York, NY, USA, 2010; p. 353. [Google Scholar] [CrossRef]
  11. Cariou, C.; Chehdi, K. Unsupervised nearest neighbors clustering with application to hyperspectral Images. IEEE J. Sel. Top. Signal Process. 2015, 9, 1105–1116. [Google Scholar] [CrossRef] [Green Version]
  12. Kussul, N.; Lavreniuk, M.; Skakun, S.; Shelestov, A. Deep Learning Classification of Land Cover and Crop Types Using Remote Sensing Data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 778–782. [Google Scholar] [CrossRef]
  13. Waldner, F.; De Abelleyra, D.; Verón, S.R.; Zhang, M.; Wu, B.; Plotnikov, D.; Bartalev, S.; Lavreniuk, M.; Skakun, S.; Kussul, N.; et al. Towards a set of agrosystem-specific cropland mapping methods to address the global cropland diversity. Int. J. Remote Sens. 2016, 37. [Google Scholar] [CrossRef] [Green Version]
  14. Arvor, D.; Jonathan, M.; Meirelles, M.S.P.; Dubreuil, V.; Durieux, L. Classification of MODIS EVI time series for crop mapping in the state of Mato Grosso, Brazil. Int. J. Remote Sens. 2011, 32, 7847–7871. [Google Scholar] [CrossRef]
  15. Hatami, N.; Gavet, Y.; Debayle, J. Classification of Time-Series Images Using Deep Convolutional Neural Networks. In Proceedings of the 10th International Conference Machine Vision, Vienna, Austria, 13–15 November 2017; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  16. Rezaee, M.; Mahdianpari, M.; Zhang, Y.; Salehi, B. Deep Convolutional Neural Network for Complex Wetland Classification Using Optical Remote Sensing Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018. [Google Scholar] [CrossRef]
  17. Transon, J.; D’Andrimont, R.; Maugnard, A.; Defourny, P. Survey of hyperspectral Earth Observation applications from space in the Sentinel-2 context. Remote Sens. 2018, 10, 157. [Google Scholar] [CrossRef] [Green Version]
  18. Chen, B.; Huang, B.; Xu, B. Multi-source remotely sensed data fusion for improving land cover classification. ISPRS J. Photogramm. Remote Sens. 2017, 124, 27–39. [Google Scholar] [CrossRef]
  19. Huang, B.; Wei, D.K.; Li, H.X.; Zhuang, Y.L. Using a rough set model to extract rules in dominance-based interval-valued intuitionistic fuzzy information systems. Inf. Sci. 2013, 221, 215–229. [Google Scholar] [CrossRef]
  20. Brown, J.C.; Kastens, J.H.; Coutinho, A.C.; de Castro Victoria, C.; Bishop, C.R. Classifying multiyear agricultural land use data from Mato Grosso using time-series MODIS vegetation index data. Remote Sens. Environ. 2013, 130, 39–50. [Google Scholar] [CrossRef] [Green Version]
  21. Mastella, A.F.; Vieira, C.A. Acurácia temática para classificação de imagens utilizando abordagens por pixel e por objetos. Rev. Bras. Cartogr. 2018, 70, 1618–1643. [Google Scholar] [CrossRef]
  22. Silva Junior, C.; Frank, T.; Rodrigues, T. Discriminação de áreas de soja por meio de imagens EVI/MODIS e análise baseada em geo-objeto. Rev. Bras. Eng. Agrícola Ambient. 2014, 18, 44–53. [Google Scholar] [CrossRef]
  23. Eerens, H.; Haesen, D.; Rembold, F.; Urbano, F.; Tote, C.; Bydekerke, L. Image time series processing for agriculture monitoring. Environ. Model. Softw. 2014, 53, 154–162. [Google Scholar] [CrossRef]
  24. Furtado, L.F.D.A.; Francisco, C.N.; de Almeida, C.M. Análise de Imagem Baseada em Objeto Para Classificação das Fisionomias da Vegetação em Imagens de Alta Resolução Espacial; Unesp Geociências; UNESP, Ed.; UNESP: São Paulo, Brazil, 2013; Volume 32, pp. 441–451. [Google Scholar]
  25. Abade, N.A.; De Carvalho, O.A.; Guimarães, R.F.; De Oliveira, S.N. Comparative analysis of MODIS time-series classification using support vector machines and methods based upon distance and similarity measures in the brazilian cerrado-caatinga boundary. Remote Sens. 2015, 7, 12160–12191. [Google Scholar] [CrossRef] [Green Version]
  26. Tu, B.; Kuang, W.; Zhao, G.; Fei, H. Hyperspectral Image Classification via Superpixel Spectral Metrics Representation. IEEE Signal Process. Lett. 2018, 25, 1520–1524. [Google Scholar] [CrossRef]
  27. Bailey, J.T.; Boryan, C.G. Remote Sensing Applications in Agriculture at the USDA National Agricultural Statistics Service; Technical Report; USDA: Washington, DC, USA, 2010.
  28. Hatfield, J.L.; Gitelson, A.A.; Schepers, J.S.; Walthall, C.L. Application of spectral remote sensing for agronomic decisions. Agron. J. 2008, 100, 117–131. [Google Scholar] [CrossRef] [Green Version]
  29. NASA. Modis Data Products. Available online: https://modis.gsfc.nasa.gov/data/dataprod/ (accessed on 27 July 2019).
  30. Angel, C.; Asha, S. A Survey on Ambient Intelligence in Agricultural Technology. Int. J. Biol. Biomol. Agric. Food Biotechnol. Eng. 2015, 9, 210–213. [Google Scholar] [CrossRef]
  31. CEPEA. Center for Applied Economics Studies. Available online: https://www.cepea.esalq.usp.br/ (accessed on 20 June 2019).
  32. IBGE. Brazilian Institute of Geography and Statistics. Rural Census. Available online: https://censos.ibge.gov.br/agro/2017/templates/censo_agro/resultadosagro/index.html (accessed on 20 November 2019).
  33. EMBRAPA. Brazilian Agricultural Research Corporation. Available online: www.embrapa.br/ (accessed on 26 July 2019).
  34. EMBRAPA Informática. Embrapa Agricultural Informatics. Available online: https://www.embrapa.br/en/informatica-agropecuaria (accessed on 23 July 2019).
  35. MAPA. Ministry of Agriculture, Livestock and Food Supply. Available online: http://www.agricultura.gov.br (accessed on 26 July 2019).
  36. Gruber, T.R. A translation approach to portable ontology specifications. Knowl. Acquis. 1993, 5, 199–220. [Google Scholar] [CrossRef]
  37. Oliveira, A.B.F.; Werneck, V.M.B. Construindo ontologias a partir de recursos existentes: Uma prova de conceito no domínio da educação. IME USP 2003, 15, 226. [Google Scholar] [CrossRef]
  38. GFS. Global Food Security. Available online: www.cropland.org (accessed on 27 July 2019).
  39. EMBRAPA. Vegetation Temporal Analysis System. Available online: https://www.satveg.cnptia.embrapa.br/satveg/login.html (accessed on 26 July 2019).
  40. Kraska, T.; Beutel, A.; Chi, E.H.; Dean, J.; Polyzotis, N. The Case for Learned Index Structures. CoRR 2017. [Google Scholar] [CrossRef] [Green Version]
  41. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–541. [Google Scholar] [CrossRef]
  42. Zhou, J. Enhancing Time Series Clustering by Incorporating Multiple Distance Measures with Semi-Supervised Learning. J. Comput. Sci. Technol. 2015, 30, 859–873. [Google Scholar] [CrossRef]
  43. Montero, P.; Vilar, J. TSclust: An R Package for Time Series Clustering. JSS J. Stat. Softw. 2014, 62, 1–43. [Google Scholar] [CrossRef] [Green Version]
  44. Lee, K.S.; Jin, D.; Yeom, J.M.; Seo, M.; Choi, S.; Kim, J.J.; Han, K.S. New Approach for Snow Cover Detection through Spectral Pattern Recognition with MODIS Data. J. Sens. 2017, 2017. [Google Scholar] [CrossRef]
  45. Pereira, M.E.; Ferreira, F.D.O.; Martins, A.H.; Cupertino, C.M. Introdução ao processamento de imagem de sensoriamento remoto. Estudos de Psicologia 2002, 7, 389–397. [Google Scholar] [CrossRef] [Green Version]
  46. Rocchini, D.; Foody, G.M.; Nagendra, H.; Ricotta, C.; Anand, M.; He, K.S.; Schmidtlein, S.; Feilhauer, H.; Amici, V.; Kleinschmit, B.; et al. Uncertainty in ecosystem mapping by remote sensing. Comput. Geosci. 2013, 128–135. [Google Scholar] [CrossRef]
  47. Stanton, J. An Introduction to Data Science. Syracuse Univ. 2012, 1–157. [Google Scholar] [CrossRef]
  48. Cai, Y.; Guan, K.; Peng, J.; Wang, S.; Seifert, C.; Wardlow, B.; Li, Z. A high-performance and in-season classi fi cation system of fi eld-level crop types using time-series Landsat data and a machine learning approach. Remote Sens. Environ. 2018, 210, 35–47. [Google Scholar] [CrossRef]
  49. Luiz, A.J.B.; Eberhardt, I.D.R.; Schultz, B.; Formaggio, A.R. Visualização de dados de imagens de sensoriamento remoto. Rev. da Estatística UFOP 2014, III, 260–265. [Google Scholar]
  50. Rezende, C.L.; Scarano, F.R.; Assad, E.D.; Joly, C.A.; Metzger, J.P.; Strassburg, B.B.N.; Mittermeier, R.A. From hotspot to hopespot: An opportunity for the Brazilian Atlantic Forest. Perspect. Ecol. Conserv. 2018, 16, 208–214. [Google Scholar] [CrossRef]
  51. Zhou, Z.; Huang, J.; Wang, J.; Zhang, K.; Kuang, Z.; Zhong, S.; Song, X. Object-oriented classification of sugarcane using time-series middle-resolution remote sensing data based on AdaBoost. PLoS ONE 2015, 10, e0142069. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The experiment was performed over GO, MG, MT, PR, and SP, and a country-wide view.
Figure 1. The experiment was performed over GO, MG, MT, PR, and SP, and a country-wide view.
Remotesensing 12 01791 g001
Figure 2. A concept map of the ontology used—Target groups highlighted above share seasonal characteristics with each other and within the time frame selected from EVI collected maps. Sources: EMBRAPA, MODIS, IBGE.
Figure 2. A concept map of the ontology used—Target groups highlighted above share seasonal characteristics with each other and within the time frame selected from EVI collected maps. Sources: EMBRAPA, MODIS, IBGE.
Remotesensing 12 01791 g002
Figure 3. Process Flow—Target groups selected—Agriculture & Other. Elaborated by the authors.
Figure 3. Process Flow—Target groups selected—Agriculture & Other. Elaborated by the authors.
Remotesensing 12 01791 g003
Figure 4. ROC—Classification process effectiveness—Data witout similarity metrics on the data set. Comparison—Naive Bayes, Logistic Regression, Decision Tree, Gradient Boosted Tree, Generalized Linear Model, Deep Learning, Random Forest.
Figure 4. ROC—Classification process effectiveness—Data witout similarity metrics on the data set. Comparison—Naive Bayes, Logistic Regression, Decision Tree, Gradient Boosted Tree, Generalized Linear Model, Deep Learning, Random Forest.
Remotesensing 12 01791 g004
Figure 5. ROC—Classification process effectiveness—Data with similarity metrics as part of the data set. Comparison—Naive Bayes, Logistic Regression, Decision Tree, Gradient Boosted Tree, Generalized Linear Model, Deep Learning, Random Forest.
Figure 5. ROC—Classification process effectiveness—Data with similarity metrics as part of the data set. Comparison—Naive Bayes, Logistic Regression, Decision Tree, Gradient Boosted Tree, Generalized Linear Model, Deep Learning, Random Forest.
Remotesensing 12 01791 g005
Figure 6. Shannon information with Euclidian distance for the key main decision levels. In this case: Minkowski Distance (Minkowski); Coefficient of Variation (CV); Standard Deviation (SD); Canberra distance (Canberra); Mean value (Mean).
Figure 6. Shannon information with Euclidian distance for the key main decision levels. In this case: Minkowski Distance (Minkowski); Coefficient of Variation (CV); Standard Deviation (SD); Canberra distance (Canberra); Mean value (Mean).
Remotesensing 12 01791 g006
Figure 7. Country view without geopolitical boundaries.
Figure 7. Country view without geopolitical boundaries.
Remotesensing 12 01791 g007
Table 1. Studied areas (ha).
Table 1. Studied areas (ha).
Country/StateTotalAgriculture(%)
Brazil (BR)851,605,39467,547,5377.9
Goias (GO)34,011,1786,106,27917.9
Minas Gerais (MG)58,652,2124,814,4388.2
Mato grosso (MT)90,336,61914,872,04516.5
Paraná (PR)19,930,7928,776,87144.0
São Paulo (SP)24,822,3626,835,74127.5
State Subtotal227,753,03841,405,37461.3
Table 2. Temporary crops.
Table 2. Temporary crops.
Crop TypeArea (ha)Fraction of Total (%)
Soybean30,622,46045.3%
Corn17,985,76426.6%
Sugar cane9,153,70913.6%
Beans3,069,6222.1%
Wheat1,796,0652.6%
Rice1,778,1902.6%
Manioc943,3231.4%
Cotton910,0571.3%
Total area64,660,18393.7%
Table 3. Accuracy classification results.
Table 3. Accuracy classification results.
AlgorithmsGO MG MT PR SP
RDDSMRDDSMRDDSMRDDSMRDDSM
Naive Bayes88.494.584.293.885.893.189.294.180.090.5
Generalized Linear Model93.498.496.999.394.999.391.198.489.797.8
Logistic Regression97.497.497.798.697.498.991.098.796.397.8
Deep Learning83.997.986.999.288.999.285.198.283.997.6
Decision Tree85.298.692.399.492.399.185.198.786.597.5
Randon Forest85.898.292.399.395.899.686.998.976.297.6
Gradient Boosted Trees96.098.497.199.497.499.696.098.895.297.8
Support Vector Machine97.998.398.199.597.899.197.498.997.092.2
RD: Raw data and SM: Data with Similarity metrics.
Table 4. Concordance.
Table 4. Concordance.
Tested Models(RD-DSM)Kappa
Naive Bayes80.0–93.8%0.89
Generalized Linear Model89.7–99.3%0.97
Logistic Regression91.0–99.3%0.96
Deep Learning83.9–98.6%0.95
Decision Tree86.2–99.4%0.94
Randon Forest86.6–99.5%0.97
Gradient Boosted tree95.2–99.6%0.96
Support Vector Machine97.0–99.5%0.95
Average88.798.6%0.95
Table 5. Correlation.
Table 5. Correlation.
CorrelaçãoEVINDVINIRMIR
EVI1
NDVI0.19641
NIR0.65100.10881
MIR0.32270.32830.24171
Table 6. Results contrasted.
Table 6. Results contrasted.
AuthorArea (ha)AccuracyKappaComments
Publications that share at least one common aspect: Brazil as studied area
[14]5,617,25095.0%0.98- 3 tree seasonal crops in Mato grosso/
[13]507,72884.0% - used the same algorithms, achieved 84% accuracy for Brazil/
[22]724,29384.0%0.56- differentiated Soybean and non-soybean/
[24]160083.0%0.78- used high spatial resolution images/
[21]965874.6%0.57- seasonal crops group as part of the results/
International Publications that share at least one similarity
[48]258,50097.0% - seasonal crops group as part of the results/
[12]2,800,00094.6% - seasonal crops classification/

Share and Cite

MDPI and ACS Style

Santos, M.A.S.; Assad, E.D.; Gurgel, A.C.; Omar, N. Similarity Metrics Enforcement in Seasonal Agriculture Areas Classification. Remote Sens. 2020, 12, 1791. https://doi.org/10.3390/rs12111791

AMA Style

Santos MAS, Assad ED, Gurgel AC, Omar N. Similarity Metrics Enforcement in Seasonal Agriculture Areas Classification. Remote Sensing. 2020; 12(11):1791. https://doi.org/10.3390/rs12111791

Chicago/Turabian Style

Santos, Marcio A. S., Eduardo D. Assad, Angelo C. Gurgel, and Nizam Omar. 2020. "Similarity Metrics Enforcement in Seasonal Agriculture Areas Classification" Remote Sensing 12, no. 11: 1791. https://doi.org/10.3390/rs12111791

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop