Next Article in Journal
Satellite-Based Analysis of Lake Okeechobee’s Surface Water: Exploring Machine Learning Classification for Change Detection
Previous Article in Journal
An Integrated Modeling Framework to Estimate Time Series of Evapotranspiration on a Regional Scale Using MODIS Data and a Two-Source Energy Balance Model
 
 
Please note that, as of 4 December 2024, Environmental Sciences Proceedings has been renamed to Environmental and Earth Sciences Proceedings and is now published here.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Enhancing Corn Yield Prediction in Iowa: A Concatenate-Based 2D-CNN-BILSTM Model with Integration of Sentinel-1/2 and SoilGRIDs Data †

1
School of Surveying and Geospatial Engineering, College of Engineering, University of Tehran, Tehran 14399-57131, Iran
2
Ludwig-Franzius-Institute for Hydraulic, Estuarine and Coastal Engineering, Leibniz University Hannover, Nienburger Str. 4, 30167 Hannover, Germany
*
Author to whom correspondence should be addressed.
Presented at the 5th International Electronic Conference on Remote Sensing, 7–21 November 2023; Available online: https://ecrs2023.sciforum.net/.
Environ. Sci. Proc. 2024, 29(1), 2; https://doi.org/10.3390/ECRS2023-15852
Published: 6 November 2023
(This article belongs to the Proceedings of ECRS 2023)

Abstract

:
Ensuring food security in precision agriculture demands early prediction of corn yield in the USA at international, regional, and local levels. Accurate corn yield estimation can play a crucial role in averting famine by offering insights into food availability during the growing season. To address this, we propose a Concatenate-based 2D-CNN-BILSTM model that integrates Sentinel-1, Sentinel-2, and Soil GRIDS (global gridded soil information) data for corn yield estimation in Iowa State from 2018 to 2021. This approach utilizes Sentinel-2 features, including spectral bands (Blue, Green, Red, Red Edge 1/2/3, NIR, n-NIR, and SWIR 1/2), and vegetation indices (NDVI, LSWI, DVI, RVI, WDRVI, SAVI, VARIGREEN, and GNDVI), alongside Sentinel 1 features (VV, VH, difference VV, and VH, and RVI), and soil data (Silt, Clay, Sand, CEC, and pH) as initial inputs. To extract high-level features from this data each month, a dedicated 2D-CNN was designed. This 2D-CNN concatenates high-level features from the previous month with low-level features of the subsequent month, serving as input features for the model. Additionally, to incorporate single-time soil data features, another 2D-CNN was implemented. Finally, high-level features from soil, Sentinel-1, and Sentinel-2 data were concatenated and fed into a BILSTM layer for accurate corn yield prediction. Comparative analysis against random forest (RF), Concatenate-based 2D-CNN, and 2D-CNN models, using metrics like RMSE, MAE, MAPE, and the Index of Agreement, revealed the superiority of our model. It achieved an Index of Agreement of 84.67% with an RMSE of 0.698 t/ha. The Concatenate-based 2D-CNN model also performed well with an RMSE of 0.799 t/ha and an Index of Agreement of 72.71%. The 2D-CNN model followed closely with an RMSE of 0.834 t/ha and an Index of Agreement of 69.90%. In contrast, the RF model lagged with an RMSE of 1.073 t/ha and an Index of Agreement of 69.60%. Integration of Sentinel 1–2 and Soil-GRIDs data with the Concatenate-based 2D-CNN-BILSTM model significantly improved accuracy. Combining soil data with Sentinel 1–2 features reduced the RMSE by 16 kg and increased the Index of Agreement by 2.59%. This study highlighted the potential of advanced machine learning (ML)/deep learning (DL) models in achieving precise and reliable predictions, which could support sustainable agricultural practices and food-security initiatives.

1. Introduction

Corn is a highly significant crop in the United States (U.S.) due to its abundance of protein, oil, and its high water consumption [1,2]. As the largest corn producer globally, the U.S. recorded a corn production of 15.1 billion bushels in 2021 (https://www.nass.usda.gov/Newsroom/2022/01-12-2022.php, accessed on 12 January 2022).
With the rapid increase in population, the use of Remote Sensing (RS) technology in agriculture has become of paramount economic importance. By using multisensor RS images and soil and weather data, researchers can accurately predict crop yield. Optical RS data, obtained from satellites such as Sentinel-2 and Landsat-8, provide high/moderate-resolution imagery in visible and near-infrared bands. This enables precise assessment of vegetation health through indices like NDVI, while also offering detailed information on land cover for accurate identification of crop types, including corn [3]. Additionally, optical RS data allow the monitoring of phenological changes, facilitating the tracking of crop growth stages and overall health [4]. On the other hand, SAR (Synthetic Aperture Radar) images facilitate the structural analysis of vegetation by using different polarizations and microwave frequencies. Additionally, SAR signals penetrates cloud cover, providing valuable insights into the physical structure of crops [5]. The applications of RS in agriculture are diverse and encompass product and irrigation management, predicting crop performance, disease and fertilizer management, as well as crop classification, among other factors [3]. However, the effectiveness of these applications hinges on various factors, including temperature, rainfall, growth indicators, soil type, genotype structure, management practices, and nutrient elements [6]. Additionally, radiometric distortions have the potential to adversely affect the spectral bands of optical RS images [7]. To mitigate these challenges and enhance the accuracy of yield predictions, a multi-faceted approach is recommended. This involves integrating RS data with advanced machine learning (ML) models and employing data fusion techniques.
For example, Ma et al. suggested using the Bayesian neural network to estimate corn yield using MODIS images, GLDAS dataset, PRISM dataset, and SSURGO at the county level in the United States between 2005 and 2019 [8]. Desloires et al. introduced a stack of machine learning techniques, namely RF, SVR, XG-Boost, and MLP, to predict corn yield based on Sentinel-2 images captured at field scale in Iowa and Nebraska from 2017 to 2021 [9]. Khaki et al. proposed the Deep-Corn network for enhancing crop yield at the field scale by counting corn kernels, which used a shortened VGG-16 for feature extraction at different scales [10]. Shah-Hosseini et al. developed the Stacked LASSO method for predicting corn yield in Illinois, Indiana, and Iowa between 2000 and 2018 using observed corn yield, management data, plant population, planting date, and environmental features (weather and soil) [11]. Shah-Hosseini et al. also proposed a new CNN-DNN method for estimating corn using historical management, environmental, and yield data in the United States from 1980 to 2019 [12]. San et al. suggested using the CNN-RNN method for predicting corn yield using MODIS images, weather data, and soil features to extract multi-level spatiotemporal features at the county level from 2013 to 2016 [13]. Dhaliwal et al. proposed the Random Forest model for predicting corn yield using crop management data, weather data, and field-level data in the United States between 1992 and 2018 [14]. Shah-Hosseini et al. suggested combining the APSIM model with machine learning methods using plant population, planting date, and weather data for estimating corn yield in the United States between 1984 and 2018 [15].
Recent studies have demonstrated satisfactory outcomes in estimating crop yield [11,16,17,18,19]. However, they have given less consideration to the combination of radar and optical images, along with soil data for corn yield prediction. Moreover, most of the studies used CNN-LSTM for feature extraction, and they have not fully explored the benefits of combining high-level features from the previous month with low-level features from the subsequent month to improve corn yield [20]. Also, they have mostly employed Long Short-Term Memory (LSTM) networks for yield prediction, overlooking the potential of Bidirectional-LSTM (Bi-LSTM) networks which can integrate both past and future information to enhance corn yield forecasting [21].
In response to these limitations, we introduce a novel Concatenate-based 2D-CNN-BiLSTM model for corn yield estimation at the county level in Iowa. Leveraging Sentinel-1 and Sentinel-2 images along with Soil GRIDS, which provide global gridded soil information, our model aims to enhance performance during the growing season. This model offers an innovative approach to feature integration, effectively capturing short-term fluctuations and long-term trends in corn growth patterns. Additionally, the incorporation of Soil GRIDS data provides crucial insights into soil characteristics, augmenting the model’s capacity to account for diverse soil conditions.

2. Materials and Methods

2.1. Study Area

The study area was located in the state of Iowa in the U.S. (see Figure 1). The research was conducted on corn during the years 2018 to 2021. Corn is planted in Iowa when the soil is warm enough for the seeds to grow, but not too early in order to avoid frost damage. The timing varies depending on the location, with southern counties planting as early as April and northern counties waiting until several weeks later. Farmers in Iowa typically begin harvesting corn in mid-September, with the majority of the harvest taking place in October. However, in cooler years, the harvest may not take place until November (https://www.iowacorn.org/education/faqs, accessed on 15 January 2022).

2.2. Dataset

A variety of datasets were used to forecast corn yield from 2018 to 2021, including Sentinel-1 SAR (COPERNICUS/S1_GRD), Sentinel-2 SR (S2_SR_HARMONIZED), Soil-Grids (https://www.isric.org/explore/soilgrids, accessed on 4 May 2020), USDA Yield (https://quickstats.nass.usda.gov/, accessed on 15 January 2021), Crop Land Data Layer (CDL), and County Boundaries data [22,23]. Sentinel-1 and -2 were downloaded from the Google Earth Engine (GEE) cloud computing platform [24]. Table 1 displays the statistical characteristics of yield observations for both the training and test datasets.

2.3. Methodology

The aim of the proposed method is to improve the corn yields prediction accuracy at the county level in Iowa during the growing season prior to the harvest during the month of August. As displayed in Figure 2, our proposed method includes two main steps: 1—extracting features derived from Sentinel-1, Sentinel-2, and Soil GRIDS in the GEE system, and 2—using the proposed Concatenate-based 2D-CNN-BiLSTM model to predict corn yield. The details of each step have been briefly explained in the following subsections.

2.3.1. Feature Selection

As discussed before, various features derived from Sentinel1\2 and Soil GRIDS were used to predict corn yield in our study area. Informative spectral bands of Sentinel-2 images (i.e., Blue, Green, Red, Red Edge 1/2/3, NIR, n-NIR, and SWIR 1/2 bands), along with its Vis (i.e., NDVI, LSWI, DVI, RVI, WDRVI, SAVI, VARI-GREEN, and GNDVI), were used as input optical features [25,26,27,28,29,30,31,32]. Additionally, VV, VH, the difference between VV and VH, and the Radar Vegetation Index (RVI) [33] were extracted from Sentinel-1 SAR images [34]. Soil data including silt, clay, sand, cec, and pH were also collected at various depths ranging from 0 cm to 200 cm to construct our feature set (https://www.isric.org/explore/soilgrids, accessed on 4 May 2020).

2.3.2. Corn Yields Prediction Using the Concatenate-Based 2D-CNN-BiLSTM Model

As prediction of corn yield is so challenging, improvement of the advanced and novel deep learning model for accurately predicting corn yield is important. In this way we proposed the Concatenate-based 2D-CNN-BiLSTM model which have two main parts (see Figure 3) including feature extraction using a 2D-CNN network, and corn yield prediction using a Bi-LSTM network. The 2D-CNN network extracts high-level spatial features from input data and concatenates them with low-level features from subsequent months [35]. Additionally, a separate 2D-CNN network was created to incorporate single-time soil data features. Finally, the high-level features from soil, Sentinel-1, and Sentinel-2 data were concatenated and fed into a Bi-LSTM layer to accurately predict corn yield. The Bi-LSTM layer is able to overcome significant time lags between inputs across any time period and enhance its ability to represent temporal patterns at different frequencies using backward and forward information [21]. This makes it particularly advantageous for analyzing crop growth cycles of varying durations. Monthly Block consists of Conv2D-1 > Linear activation function > Concatenate layer > Conv2D-2 > Linear activation function. Monthly composites (XMonthly) pass through the Monthly Block, and Monthly features (FMonthly) are extracted. In addition, Soil Block consists of Conv2D-1 > Linear activation function. Soil features pass through the Soil Block, and soil features (S) are then extracted. FMonthly and S are then concatenated together and fed into Bi-LSTM layer with a ReLU activation function to predict corn yield. Finally, the output of the Bi-LSTM layer passes through a dense layer with a linear activation function to obtain yield values.
Overall, our Concatenate-based 2D-CNN-BiLSTM model is a promising approach for accurately predicting corn yield by incorporating various data sources and effectively capturing temporal patterns.

3. Results and Discussion

For this study, a total of 250, 27, and 83 samples were selected for training, validation, and testing of the Concatenate-based 2D-CNN-BiLSTM model, respectively. The Conv2D-1, and Conv2D-2 layers were set to have 16 and 22 filters, respectively, with a kernel size of 1 × 1. The Bi-LSTM layer had 16 filters. The model was trained using the Adam optimizer for 30 epochs with a batch size of 10. The best weight was obtained based on the minimum Validation Loss. The performance of the proposed model was compared with Concatenate-based 2D-CNN, 2D-CNN, and RF in two scenarios: (1) using Sentinel-1 and -2 data, and (2) using both Sentinel-1 and -2 data along with Soil Grids. Table 2 displays the performance of the proposed models and the compared models, measured in terms of RMSE, MAPE, MAE, RRMSE, and Index of Agreement (D).
Table 2 reveals that the Concatenate-Based 2D-CNN-BiLSTM model outperforms the Concatenate-based 2D-CNN, 2D-CNN, and RF methods significantly. The best performance of the Concatenate-based 2D-CNN-BiLSTM model is achieved when combining Sentinel-1 and -2 and Soil GRIDS, with an RMSE of 0.698 (t/ha), MAPE of 4.47%, MAE of 0.556 (t/ha), RRMSE of 5.55%, and D of 84.67%. Our proposed model improves D by 14.77% compared to the 2D-CNN.
Figure 4 depicts the scatter plots of predicted yield versus observed yield between our proposed method and compared methods in 2021. The scatter plots demonstrate that the fit line is close to the diagonal line in the Concatenate-based 2D-CNN model and far away from it in the RF model.
Additionally, Figure 5 illustrates that the proposed model outperforms the compared models, resulting in a reduction in Error maps and generation of a brighter error map. This confirms the efficacy of utilizing Soil Grids data for yield estimation.
A visual representation of the distribution of the corn yield value is presented in Figure 6, which compares the USDA yield with the predicted yield obtained using our proposed method. The results displayed in Figure 6 indicate a significant level of agreement between the observed and predicted corn yield, thereby reinforcing the reliability and accuracy of our proposed method’s predictions.

4. Conclusions

Forecasting corn yield is a crucial aspect of agriculture management in IOWA. Recent studies have demonstrated that remote sensing, soil data, and deep learning methods are effective when it comes to estimating corn yield. In order to accurately predict corn yield, it is important to consider both temporal and spatial features. To achieve this, we propose a novel Concatenate-Based 2D-CNN-BiLSTM model that extracts both spatial and temporal features. The CNNs extract spatial features while Bi-LSTM extracts temporal features. The inputs for our model include remote sensing data (Sentinel-1 and -2) and Soil GRIDS data. We conducted experiments with the proposed model on Iowa corn from 2018 to 2021 at the county level. Our results demonstrate the effectiveness and advantages of our approach compared to other methods. By considering both spatial and temporal features, our model is able to accurately forecast corn yield, which can aid in making informed decisions for agriculture management in Iowa.

Author Contributions

Conceptualization, M.F., R.S.-H. and A.M.; Methodology, M.F., R.S.-H. and A.M.; Project administration, R.S.-H.; Resources, M.F., R.S.-H. and A.M.; Validation, M.F., R.S.-H. and A.M.; Supervision, R.S.-H.; Writing—original draft, M.F., R.S.-H. and A.M.; Writing—review and editing, M.F., R.S.-H. and A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data that support the findings of this study are available on reasonable request from the corresponding author.

Acknowledgments

We would like to express our sincere gratitude to Google for providing access to Earth Engine and Colab. These pivotal tools greatly facilitated the execution and analysis of this research. Additionally, we extend our heartfelt thanks to the United States Department of Agriculture (USDA) for generously providing the essential yield values that contributed significantly to the success of this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fathi, M.; Shah-Hosseini, R.; Moghimi, A. Comparison of Some Deep Neural Networks for Corn and Soybean Mapping in Iowa State using Landsat imagery. Earth Obs. Geomat. Eng. 2022, 6, 57–66. [Google Scholar] [CrossRef]
  2. Fathi, M.; Shah-Hosseini, R. Automatic Corn and Soybean Mapping Based on Deep Learning Methods (case Study: Hamilton, Hardin, Boone, Story, Dallas, Polk, and Jusper Counties in Lowa State). ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2023, 10, 187–193. [Google Scholar] [CrossRef]
  3. Ali, A.M.; Abouelghar, M.; Belal, A.; Saleh, N.; Yones, M.; Selim, A.I.; Amin, M.E.; Elwesemy, A.; Kucher, D.E.; Maginan, S. Crop yield prediction using multi sensors remote sensing. Egypt. J. Remote Sens. Space Sci. 2022, 25, 711–716. [Google Scholar]
  4. Amani, M.; Kakooei, M.; Moghimi, A.; Ghorbanian, A.; Ranjgar, B.; Mahdavi, S.; Davidson, A.; Fisette, T.; Rollin, P.; Brisco, B. Application of google earth engine cloud computing platform, sentinel imagery, and neural networks for crop mapping in Canada. Remote Sens. 2020, 12, 3561. [Google Scholar] [CrossRef]
  5. Moghimi, A.; Mohammadzadeh, A.; Khazai, S. Integrating thresholding with level set method for unsupervised change detection in multitemporal SAR images. Can. J. Remote Sens. 2017, 43, 412–431. [Google Scholar] [CrossRef]
  6. Dharani, M.; Thamilselvan, R.; Natesan, P.; Kalaivaani, P.; Santhoshkumar, S. Review on crop prediction using deep learning techniques. In Proceedings of the Journal of Physics: Conference Series, International E-Conference on Data Analytics, Intelligent Systems and Information Security & ICDIIS 2020, Pollachi, India, 11–12 December 2020; p. 012026. [Google Scholar]
  7. Moghimi, A.; Celik, T.; Mohammadzadeh, A. Tensor-based keypoint detection and switching regression model for relative radiometric normalization of bitemporal multispectral images. Int. J. Remote Sens. 2022, 43, 3927–3956. [Google Scholar] [CrossRef]
  8. Ma, Y.; Zhang, Z.; Kang, Y.; Özdoğan, M. Corn yield prediction and uncertainty analysis based on remotely sensed variables using a Bayesian neural network approach. Remote Sens. Environ. 2021, 259, 112408. [Google Scholar] [CrossRef]
  9. Desloires, J.; Ienco, D.; Botrel, A. Out-of-year corn yield prediction at field-scale using Sentinel-2 satellite imagery and machine learning methods. Comput. Electron. Agric. 2023, 209, 107807. [Google Scholar] [CrossRef]
  10. Khaki, S.; Pham, H.; Han, Y.; Kuhl, A.; Kent, W.; Wang, L. Deepcorn: A semi-supervised deep learning method for high-throughput image-based corn kernel counting and yield estimation. Knowl.-Based Syst. 2021, 218, 106874. [Google Scholar] [CrossRef]
  11. Shahhosseini, M.; Hu, G.; Archontoulis, S.V. Forecasting corn yield with machine learning ensembles. Front. Plant Sci. 2020, 11, 1120. [Google Scholar] [CrossRef]
  12. Shahhosseini, M.; Hu, G.; Khaki, S.; Archontoulis, S.V. Corn yield prediction with ensemble CNN-DNN. Front. Plant Sci. 2021, 12, 709008. [Google Scholar] [CrossRef] [PubMed]
  13. Sun, J.; Lai, Z.; Di, L.; Sun, Z.; Tao, J.; Shen, Y. Multilevel deep learning network for county-level corn yield estimation in the us corn belt. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5048–5060. [Google Scholar] [CrossRef]
  14. Dhaliwal, D.S.; Williams, M.M. Sweet corn yield prediction using machine learning models and field-level data. Precis. Agric. 2023, 25, 51–64. [Google Scholar] [CrossRef]
  15. Shahhosseini, M.; Hu, G.; Huber, I.; Archontoulis, S.V. Coupling machine learning and crop modeling improves crop yield prediction in the US Corn Belt. Sci. Rep. 2021, 11, 1606. [Google Scholar] [CrossRef]
  16. Rao, C.; Liu, Y. Three-dimensional convolutional neural network (3D-CNN) for heterogeneous material homogenization. Comput. Mater. Sci. 2020, 184, 109850. [Google Scholar] [CrossRef]
  17. Mohite, J.; Sawant, S.; Pandit, A.; Agrawal, R.; Pappula, S. Soybean Crop Yield Prediction by Integration of Remote Sensing and Weather Observations. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2023, 48, 197–202. [Google Scholar] [CrossRef]
  18. Schwalbert, R.A.; Amado, T.; Corassa, G.; Pott, L.P.; Prasad, P.V.; Ciampitti, I.A. Satellite-based soybean yield forecast: Integrating machine learning and weather data for improving crop yield prediction in southern Brazil. Agric. For. Meteorol. 2020, 284, 107886. [Google Scholar] [CrossRef]
  19. You, J.; Li, X.; Low, M.; Lobell, D.; Ermon, S. Deep gaussian process for crop yield prediction based on remote sensing data. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  20. Sun, J.; Di, L.; Sun, Z.; Shen, Y.; Lai, Z. County-level soybean yield prediction using deep CNN-LSTM model. Sensors 2019, 19, 4363. [Google Scholar] [CrossRef]
  21. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. The performance of LSTM and BiLSTM in forecasting time series. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; pp. 3285–3292. [Google Scholar]
  22. Malenovský, Z.; Rott, H.; Cihlar, J.; Schaepman, M.E.; García-Santos, G.; Fernandes, R.; Berger, M. Sentinels for science: Potential of Sentinel-1,-2, and -3 missions for scientific observations of ocean, cryosphere, and land. Remote Sens. Environ. 2012, 120, 91–101. [Google Scholar] [CrossRef]
  23. Boryan, C.; Yang, Z.; Mueller, R.; Craig, M. Monitoring US agriculture: The US department of agriculture, national agricultural statistics service, cropland data layer program. Geocarto Int. 2011, 26, 341–358. [Google Scholar] [CrossRef]
  24. Amani, M.; Ghorbanian, A.; Ahmadi, S.A.; Kakooei, M.; Moghimi, A.; Mirmazloumi, S.M.; Moghaddam, S.H.A.; Mahdavi, S.; Ghahremanloo, M.; Parsian, S. Google earth engine cloud computing platform for remote sensing big data applications: A comprehensive review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5326–5350. [Google Scholar] [CrossRef]
  25. Sonobe, R.; Yamaya, Y.; Tani, H.; Wang, X.; Kobayashi, N.; Mochizuki, K.-i. Crop classification from Sentinel-2-derived vegetation indices using ensemble learning. J. Appl. Remote Sens. 2018, 12, 026019. [Google Scholar] [CrossRef]
  26. Gitelson, A.A.; Kaufman, Y.J.; Merzlyak, M.N. Use of a green channel in remote sensing of global vegetation from EOS-MODIS. Remote Sens. Environ. 1996, 58, 289–298. [Google Scholar] [CrossRef]
  27. Wang, C.; Wu, Y.; Hu, Q.; Hu, J.; Chen, Y.; Lin, S.; Xie, Q. Comparison of Vegetation Phenology Derived from Solar-Induced Chlorophyll Fluorescence and Enhanced Vegetation Index, and Their Relationship with Climatic Limitations. Remote Sens. 2022, 14, 3018. [Google Scholar] [CrossRef]
  28. Richardson, A.J.; Everitt, J.H. Using spectral vegetation indices to estimate rangeland productivity. Geocarto Int. 1992, 7, 63–69. [Google Scholar] [CrossRef]
  29. Christian, J.I.; Basara, J.B.; Lowman, L.E.; Xiao, X.; Mesheske, D.; Zhou, Y. Flash drought identification from satellite-based land surface water index. Remote Sens. Appl. Soc. Environ. 2022, 26, 100770. [Google Scholar] [CrossRef]
  30. Broge, N.H.; Leblanc, E. Comparing prediction power and stability of broadband and hyperspectral vegetation indices for estimation of green leaf area index and canopy chlorophyll density. Remote Sens. Environ. 2001, 76, 156–172. [Google Scholar] [CrossRef]
  31. Eng, L.S.; Ismail, R.; Hashim, W.; Baharum, A. The use of VARI, GLI, and VIgreen formulas in detecting vegetation in aerial images. Int. J. Technol. 2019, 10, 1385–1394. [Google Scholar] [CrossRef]
  32. Huete, A.R. A soil-adjusted vegetation index (SAVI). Remote Sens. Environ. 1988, 25, 295–309. [Google Scholar] [CrossRef]
  33. Arii, M.; van Zyl, J.J.; Kim, Y. A general characterization for polarimetric scattering from vegetation canopies. IEEE Trans. Geosci. Remote Sens. 2010, 48, 3349–3357. [Google Scholar] [CrossRef]
  34. Holtgrave, A.-K.; Röder, N.; Ackermann, A.; Erasmi, S.; Kleinschmit, B. Comparing Sentinel-1 and-2 data and indices for agricultural land use monitoring. Remote Sens. 2020, 12, 2919. [Google Scholar] [CrossRef]
  35. Yu, C.; Han, R.; Song, M.; Liu, C.; Chang, C.-I. A simplified 2D-3D CNN architecture for hyperspectral image classification based on spatial–spectral fusion. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2485–2501. [Google Scholar] [CrossRef]
Figure 1. Study area.
Figure 1. Study area.
Environsciproc 29 00002 g001
Figure 2. Flowchart of the proposed method.
Figure 2. Flowchart of the proposed method.
Environsciproc 29 00002 g002
Figure 3. Flowchart of the proposed Concatenate-based 2D-CNN-BILSTM model for corn yield prediction.
Figure 3. Flowchart of the proposed Concatenate-based 2D-CNN-BILSTM model for corn yield prediction.
Environsciproc 29 00002 g003
Figure 4. The scatter plots of predicted yield versus observed yield between our proposed method and compared methods in 2021.
Figure 4. The scatter plots of predicted yield versus observed yield between our proposed method and compared methods in 2021.
Environsciproc 29 00002 g004
Figure 5. Error maps generated by the proposed and compared models in 2021.
Figure 5. Error maps generated by the proposed and compared models in 2021.
Environsciproc 29 00002 g005
Figure 6. Map of USDA corn yield and predicted corn yield in 2021.
Figure 6. Map of USDA corn yield and predicted corn yield in 2021.
Environsciproc 29 00002 g006
Table 1. Sample plot yield statistics for year in study area.
Table 1. Sample plot yield statistics for year in study area.
TypeYearNumber of SamplesMin
(Ton Ha−1)
Max
(Ton Ha−1)
Mean
(Ton Ha−1)
Std
(Ton Ha−1)
Train2018939.3814.1812.131.25
Train2019889.5014.7312.191.06
Train2020955.5412.9910.921.17
Test2021849.5714.4912.571.05
Table 2. Comparison of performance of proposed Concatenate-based 2D-CNN-BILSTM Model versus other considered methods for corn yield prediction.
Table 2. Comparison of performance of proposed Concatenate-based 2D-CNN-BILSTM Model versus other considered methods for corn yield prediction.
ModelSentinel 1 and 2Sentinel 1 and 2 and Soil Grids
RMSERRMSEMAEMAPEDRMSERRMSEMAEMAPED
Proposed Model0.7145.680.5614.5582.080.6985.550.5564.4784.67
Concatenate-Based 2D-CNN0.8496.750.6865.6067.580.7996.350.6205.0272.71
2D-CNN0.8486.740.6945.6464.800.8346.630.6775.5169.90
RF1.0898.660.9357.9569.041.0738.540.9187.7869.60
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fathi, M.; Shah-Hosseini, R.; Moghimi, A. Enhancing Corn Yield Prediction in Iowa: A Concatenate-Based 2D-CNN-BILSTM Model with Integration of Sentinel-1/2 and SoilGRIDs Data. Environ. Sci. Proc. 2024, 29, 2. https://doi.org/10.3390/ECRS2023-15852

AMA Style

Fathi M, Shah-Hosseini R, Moghimi A. Enhancing Corn Yield Prediction in Iowa: A Concatenate-Based 2D-CNN-BILSTM Model with Integration of Sentinel-1/2 and SoilGRIDs Data. Environmental Sciences Proceedings. 2024; 29(1):2. https://doi.org/10.3390/ECRS2023-15852

Chicago/Turabian Style

Fathi, Mahdiyeh, Reza Shah-Hosseini, and Armin Moghimi. 2024. "Enhancing Corn Yield Prediction in Iowa: A Concatenate-Based 2D-CNN-BILSTM Model with Integration of Sentinel-1/2 and SoilGRIDs Data" Environmental Sciences Proceedings 29, no. 1: 2. https://doi.org/10.3390/ECRS2023-15852

APA Style

Fathi, M., Shah-Hosseini, R., & Moghimi, A. (2024). Enhancing Corn Yield Prediction in Iowa: A Concatenate-Based 2D-CNN-BILSTM Model with Integration of Sentinel-1/2 and SoilGRIDs Data. Environmental Sciences Proceedings, 29(1), 2. https://doi.org/10.3390/ECRS2023-15852

Article Metrics

Back to TopTop