Next Article in Journal
Muscle Synergies in Parkinson’s Disease
Next Article in Special Issue
Low-Cost Automated Vectors and Modular Environmental Sensors for Plant Phenotyping
Previous Article in Journal
Timed Up and Go and Six-Minute Walking Tests with Wearable Inertial Sensor: One Step Further for the Prediction of the Risk of Fall in Elderly Nursing Home People
Previous Article in Special Issue
Performances Evaluation of a Low-Cost Platform for High-Resolution Plant Phenotyping
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Precise Estimation of NDVI with a Simple NIR Sensitive RGB Camera and Machine Learning Methods for Corn Plants

Department of Agricultural and Biological Engineering, Purdue University, 225 S. University St., West Lafayette, IN 47907, USA
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(11), 3208; https://doi.org/10.3390/s20113208
Submission received: 1 May 2020 / Revised: 2 June 2020 / Accepted: 3 June 2020 / Published: 5 June 2020
(This article belongs to the Special Issue Low-Cost Sensors and Vectors for Plant Phenotyping)

Abstract

:
The normalized difference vegetation index (NDVI) is widely used in remote sensing to monitor plant growth and chlorophyll levels. Usually, a multispectral camera (MSC) or hyperspectral camera (HSC) is required to obtain the near-infrared (NIR) and red bands for calculating NDVI. However, these cameras are expensive, heavy, difficult to geo-reference, and require professional training in imaging and data processing. On the other hand, the RGBN camera (NIR sensitive RGB camera, simply modified from standard RGB cameras by removing the NIR rejection filter) have also been explored to measure NDVI, but the results did not exactly match the NDVI from the MSC or HSC solutions. This study demonstrates an improved NDVI estimation method with an RGBN camera-based imaging system (Ncam) and machine learning algorithms. The Ncam consisted of an RGBN camera, a filter, and a microcontroller with a total cost of only $70 ~ 85. This new NDVI estimation solution was compared with a high-end hyperspectral camera in an experiment with corn plants under different nitrogen and water treatments. The results showed that the Ncam with two-band-pass filter achieved high performance (R2 = 0.96, RMSE = 0.0079) at estimating NDVI with the machine learning model. Additional tests showed that besides NDVI, this low-cost Ncam was also capable of predicting corn plant nitrogen contents precisely. Thus, Ncam is a potential option for MSC and HSC in plant phenotyping projects.

1. Introduction

Plant chlorophyll absorbs red light and mostly reflects near-infrared (NIR) light, thus distinguishing plants from other materials such as soil and water [1]. Based on this characteristic, the normalized difference vegetation index (NDVI) is calculated as the ratio between red and NIR reflectance to monitor plant health. Since the 1970s, NDVI has been widely used in remote sensing to discriminate vegetation and other materials in satellite images [2,3,4,5]. It has also been proved that NDVI has significant correlations with green biomass and crop yield [3]. Usually, NDVI is obtained from images captured by a multispectral camera (MSC) or hyperspectral camera (HSC), which captures the accurate red and NIR reflectance of plants [6,7,8,9,10,11,12]. However, both MSC and HSC are heavy, expensive, and require professional operation for imaging and data processing [13,14]. For push-broom hyperspectral cameras in remote sensing, the geo-referencing has also been challenging, whereas the popular multi-lenses multispectral cameras require long imaging distances so they cannot be used in greenhouses and in-door imaging stations. All these issues of MSC and HSC limit their applications in plant phenotyping for the broader agricultural community.
Compared with MSC or HSC, RGB cameras have lower cost, lighter weight, and are easy to operate during imaging. Therefore, RGB cameras are widely used in remote sensing by being mounted to unmanned aerial vehicles (UAV) or ground-based vehicles [15,16]. The standard RGB camera has been explored for calculating a visible NDVI (vNDVI) to estimate NDVI [17,18]. Costa et al. proposed a new vNDVI formula using a genetic algorithm with a standard RGB camera [18]. The vNDVI works as an estimate of NDVI but is obtained with the regular RGB band values instead of the red and near-infrared bands values.
Another approach for achieving highly precise NDVI is to use an RGBN (NIR-sensitive RGB) camera. A standard RGB camera usually contains a NIR rejection filter, which makes the RGB images close to human vision. After simply removing the filter, an RGBN camera is obtained for capturing the red and NIR reflectance needed for NDVI calculation. Thus, many studies have explored the feasibility of using an RGBN camera to measure NDVI. A two-camera system containing a standard RGB camera and an RGBN camera is reported to calculate NDVI [19]. The standard RGB camera captures the red reflectance, whereas the RGBN camera captures the NIR reflectance. However, the two-camera system creates new challenges, such as image matching and cross-calibration. An RGBN camera with a filter-switching unit is reported to calculate NDVI, getting rid of the limitations of the two-camera system [20]. With the filter-switching unit, the visible image and NIR image are captured by switching the NIR rejection filter, respectively. Subsequently, the NDVI is calculated from both images. However, the position switching of the filter-switching unit is time consuming, which creates the challenge of image matching when the camera is mounted on a moving platform, such as a UAV. An RGBN camera with a new filter to help obtain pure NIR is studied in recent years, which avoids the image matching issue in two-camera and filter-switching systems [1,21]. The new filter is usually either a long-pass filter or a two-band-pass filter. The long-pass filter cuts off the visible light from a particular wavelength, whereas the two-band-pass filter allows two narrow bands of light to pass through. Therefore, the RGBN camera records visible and NIR light in a specified channel(s) depending on the filter applied. Most previous studies calculated NDVI still with the formula (NIR - Red)/(NIR + Red) by using the corresponding channel values obtained from RGBN camera as Red and NIR, respectively [1,16,22,23,24,25,26,27]. However, these results were somehow different from the NDVI by HSC or MSC. Since NIR reflectance is mixed with the visible reflectance in each channel, the pure R and NIR cannot be obtained precisely at the same time with the RGBN camera. Therefore, the NDVI calculated from these earlier reported methods could not be used to replace the NDVI from MSC and HSC completely.
Targeting this problem, this study proposed an improved NDVI measuring solution based on machine learning and a low-cost imaging system, Ncam. The Ncam was comprised of a widely available RGBN camera, a filter, and a microcontroller. This new NDVI estimation solution was compared with a high-end HSC in an experiment with corn plants under different nitrogen and water treatments in one of the phenotyping greenhouses at Purdue University. Furthermore, additional tests were also conducted on the Ncam images for their potential of assessing other plant physiological features such as nitrogen contents.

2. Materials and Methods

2.1. Ncam System Development

Two Ncam systems, Ncam lp and Ncam tp, with different filters, were built for evaluation and comparison. Each of the Ncam imaging systems was composed of one NIR sensitive RGB imaging sensor, a filter, and a microcontroller. The Raspberry Pi Noir camera module V2 (Pi camera, Raspberry Pi foundation, Cambridge, UK) could sense visible and NIR light and was widely available at a low price. Therefore, the Pi camera was selected as the imaging sensor in this study. The value of RGB channels from Pi camera is the product of plant reflectance and each channel sensitivity to different light. It is described with Equation (1).
[ R C G C B C ] = [ a 1 b 1 c 1 d 1 a 2 b 2 c 2 d 2 a 3 b 3 c 3 d 3 ] [ R G B N   ] = [ a b c d   ] [ R G B N   ]
where RC, GC, and BC are the values of RGB channels, respectively; R, G, B, and N are the plant reflectance of red, green, blue, and NIR, respectively; a, b, c, and d are the sensitivities to red, green, blue and NIR light for RGB channels of Pi camera, respectively. As Figure 1a shows, the RGB channels mostly record the corresponding band reflectance and NIR reflectance; thus, b1, c1, a2, c2, a3, and b3 have very small values.
To calculate NDVI, red and NIR reflectance are needed. However, from Equation (1), there are four variables, R, G, B, and N, but only three responses RC, GC, and BC are obtained. Red and NIR reflectance cannot be estimated accurately with the original Pi camera. In this study, the filter blocking blue and green light was applied on Pi camera in both Ncam systems to decrease spectral range and increase spectral resolution. In the system of Ncam lp, the blue and green rejection filter (long-pass filter 66043, Edmund Optics, Barrington, New Jersey, USA) cutting off the light before 580 nm was placed in front of the lens of Pi camera. As shown in Figure 1b, each channel of Ncam lp senses red and NIR reflectance from 580 to 1100 nm with different sensitivities. The value of RGB channels for Ncam lp is described with Equation (2).
[ R C G C B C ] = [ j 1 k 1 j 2 k 2 j 3 k 3 ] [ R l p N l p ] = [ j k   ] [ R l p N l p ]
where Rlp is the red reflectance from 580 nm to 700 nm; Nlp is the NIR reflectance from 700 nm to 1100 nm; j and k are the sensitivities to Rlp and Nlp light for RGB channels of Ncam lp; j1 >> j2 > j3 (Figure 1b).
In the system of Ncam tp, a two-band-pass filter (FU-650850LGP, Shenzhen Fuzhe Technology Co., Ltd., Shenzhen, Guangdong, China) transmitting 640 nm (red) and 850 nm (NIR) bands light with full width at half maximum of 40 nm was chosen. As shown in Figure 1c, each channel of Ncam tp senses the narrow red and NIR reflectance with different sensitivities. The value of RGB channels for Ncam tp is described with Equation (3).
[ R C G C B C ] = [ l 1 m 1 l 2 m 2 l 3 m 3 ] [ R t p N t p ] = [ l m   ] [ R t p N t p ]
where Rtp is the red reflectance from 620 nm to 690 nm, Ntp is the NIR reflectance from 810 nm to 890 nm; l and m are the sensitivities to Rtp and Ntp light for RGB channels of Ncam tp; l1l2l3, and m1m2m3 (Figure 1c).
One of the most popular, low-cost single-board computers, Raspberry Pi 3B (Raspberry Pi foundation, Cambridge, UK), was used as the microcontroller for both systems [28]. Each of them captured raw data and formed an array with a size of 3280 × 2464 and a depth of 16-bit. Then the demosaicing operation was applied in the controller to convert the raw data to an RGB image with a size of 3280 × 2464 × 3 and a depth of 16-bit. The RGB image was finally saved as a Tiff format image to the local SD card.

2.2. Plant Samples

A total of 56 corn plants (genotype: Hybrid B73 × Mo17) were grown in the Purdue Lilly Greenhouse (40°25′19.7″ N, 86°55′7.8″ W) for this study. The plants were grown in mixed soil of Greens GradeTM (Profile Products LLC, Buffalo Grove, Illinois, USA) and Metro-Mix® 510 (Sun Gro Horticulture, Agawam, Massachusetts, USA). The temperature in the greenhouse was 23–29 °C, and supplemental lighting was on 12 h from 8 AM to 8 PM a day. To minimize the effects of the microclimate in the greenhouse, all pots (one corn plant per pot) were placed on a closed-loop conveyor system. The system was programmed to run for only 30 s with every 30 min break phase to avoid shaking disturbance to plant growth [29]. Two different water regimes (well-watered and drought-stressed) and two nitrogen levels (low and high) were applied to the plants. The high nitrogen and low nitrogen treatments were achieved with 200 ppm and 25 ppm nitrogen fertilizers, respectively. All plants were irrigated 600 mL water by automatic water station on the conveyor every morning. However, the drought-stressed plants were stopped irrigating four days before imaging. The experiment, consisting of four treatments (two water regimes × two nitrogen levels), was replicated 14 times in a full factorial design.

2.3. Image Acquisition

As the top view image of plants contained more information to monitor plant health than the side view image [30], these two Ncams were mounted on the top gantry in an imaging tower reported by Ma et al. [29]. An HSC (MSV 500, Middleton Spectral Vision, Middleton, WI, USA) was also mounted on the gantry to capture hyperspectral images of plants for NDVI “ground truth” collection (Figure 2). The HSC sensed top view plant reflectance from 380 nm to 1000 nm with a spectral resolution of 1.3 nm. There were eight halogen bulbs above plants inside the tower as the imaging light sources.
Plants were imaged at the V6 stage. The HSC and the two Ncams were initialized before imaging with the parameters shown in Table 1. The corn plants were imaged one by one, and were brought by the conveyor into the imaging tower. To make sure leaves were still while imaging, each plant was imaged 20 s after the conveyor had stopped. The HSC and two Ncams captured images simultaneously to get three images per plant, including two NIR sensitive RGB images and a hyperspectral image. After all plants were imaged, a polyvinyl chloride whiteboard was imaged by these three cameras as white references, respectively. Finally, a total of 171 images (56 plants × three images per plant + three whiteboard images) were duplicated on a computer for further image processing.

2.4. Image Processing

The images from Ncams and HSC were processed in Python language (version 3.6.8). The image processing procedure is shown in Figure 3. Firstly, all images of each plant were calibrated when divided by the corresponding whiteboard images, respectively. Secondly, the background was removed by segmentation. For the RGB image from Ncams, (BC – RC) / (BC + RC) was used to segment the image with Equation (4). To segment hyperspectral images of plants, the convolution methodology was applied on the red edge to enlarge the difference between plant and background pixels [30]. Thirdly, all pixel values of the segmented image were averaged, giving the averaged RGB values from Ncams and averaged spectra from HSC for each plant. These averaged RGB values and averaged spectra were written into a table for further data analysis.
i x , y   = { p x , y , n x , y t 0 ,   n x , y < t
where ix,y is the pixel value of the segmented RGB image, px,y is the pixel value of the RGB image after calibration, nx,y is the pixel value calculated by (BC – RC) / (BC + RC), and (x,y) is the index of the pixel in the image.
The true-color image was generated from the calibrated and segmented hyperspectral image to show the actual color of the plants. The RGB values in the true-color image were calculated with a convolution of the pixel value in the hyperspectral image and the sensitivity of a typical RGB camera. The Pi camera with the NIR rejection filter was a typical RGB camera capturing the true-color image, and its sensitivity reported by Pagnutti et al. was used in this study [31]. Then the RGB values of the true-color images were compared with those of Ncams images for channel components analysis.

2.5. Data Analysis

2.5.1. Dataset Analysis

Several statistics of the raw dataset, including mean, minimum value, maximum value, skewness, variance, and coefficient of variation, were obtained to analyze the variability of the dataset. The skewness was calculated with Equation (5).
S k e n e s s = i = 1 n ( x i μ ) 3 n s 3
where x i is the value of the ith sample; μ is the mean of the samples; n is the sample size; s is the standard deviation.

2.5.2. Data Preprocessing

Due to the light source distribution and plants’ 3D structure, the effect of ununiform lighting could not be removed by the calibration based on the flat whiteboard. To reduce the impact, the value of each channel of Ncams was scaled by dividing the sum of its corresponding RGB channels’ values. For the averaged spectra from HSC, log transformation, multiplicative scatter correction, and mean center operations were used to extract the useful signals. As the preprocessed RGB values of Ncams were summed to 1, we did not include preprocessed GC values to avoid data redundancy in the following modeling.

2.5.3. NDVI Estimation

NDVI from Ncams images was estimated with two types of methods: machine learning algorithms with the values of all channels and the typical formula method with the values of blue and red channels. The results from these two methods were discussed and compared with that from HSC.
The machine learning algorithms, including multiple linear regression (MLR), Gaussian process (GP), and support vector regression (SVR), were applied to the preprocessed RC and BC values from Ncams to estimate NDVI. MLR, GP, and SVR are typical regression methods based on different principles and are suitable for different kinds of data. MLR is a widely used linear regression method that is easy to compute and interpret. In contrast, GP and SVR are robust for non-linear data due to the kernel tricks to the models [32,33]. Thus, these three methods were selected for NDVI prediction modeling. When training the model, the MLR models without interaction were fitted, and the kernels with the highest accuracy of GP and SVR models were selected for analysis. The leave-one-out cross-validation was applied to avoid overfitting. Additionally, the NDVI “ground truth” (NDVIHSC) for the model development was extracted from the hyperspectral images. The NDVIHSC was calculated with Equation (6).
N D V I H S C = b 850 b 680 b 850 + b 680
where b850 and b680 are the intensity at 850 nm and 680 nm for each hyperspectral image, respectively.
To evaluate the performance of each model and camera, the R2, RMSE, mean absolute error (MAE) and mean percentage error (MPE) was used as the evaluation metrics. MAE is the mean of the absolute difference between the NDVI from the HSC and the predicted NDVI from Ncams using different models; MPE is MAE divided by the NDVI from the HSC. Besides, the Bland-Altman diagram was used to compare the NDVI measurements from Ncams and the HSC.
The typical formula method for NDVI estimation is used widely with the RGBN camera. It obtains NDVI by replacing the red and NIR reflectance in the calculation formula with the corresponding channel values [21]. In the Ncam system, the red and NIR reflectance were recorded mostly in the red and blue channels, respectively. Therefore, the NDVI from the typical formula method was also calculated with Equation (7) for each plant. Then the calculated NDVI was regressed to NDVIHSC. The R2 and RMSE of the regression were obtained for evaluation.
N D V I = B c R c B c + R c
where BC and RC are the averaged values of blue and red channels from each Ncam image, respectively.

2.5.4. Other Applications

To explore the additional potential of Ncams on plant health monitoring, the abilities to predict nitrogen contents were discussed briefly in this paper. To train the prediction model of nitrogen contents, the ground truth nitrogen contents of all collared leaves of each plant was measured by a FlashEA 1112 Nitrogen and Carbon Analyzer (Thermo Fisher Scientific, Waltham, MA USA) after imaging.
MLR, which was one of the most common methods for regression, was selected for the prediction of nitrogen contents for Ncams. The performance was evaluated by comparing it with HSC. The partial least square regression was applied to predict the nitrogen contents of plants for HSC. The leave-one-out cross-validation was introduced when training models to avoid overfitting. The R2, RMSE, MAE, and MPE were computed for evaluation. All models in this study were developed with Statistics and Machine Learning, and PLS toolbox in MATLAB (2019a).

3. Results and Discussion

3.1. Dataset Analysis

The NDVI from HSC ranged from 0.587 to 0.749, with a mean of 0.686, a variance of 0.002, and a coefficient of variation of 0.058. The distribution of NDVI was slightly left-skewed with the skewness of −0.334 but still was approximately normal.
The distributions of RGB values of Ncams were all right-skewed, with skewness ranging from 0.420 to 0.533 (Figure 4), which was consistent with the distribution of the plants’ height. As the light source of the imaging tower was above the plant, higher plants tended to be brighter than the shorter plants in the images. Calibration with a flat whiteboard did not remove the light intensity impact on different height plants. Thus, the preprocessing for the dataset by dividing the sum of the corresponding RGB values was necessary. The distributions of the RGB values after preprocessing became more symmetric (Figure 5d).

3.2. Channel Components of Ncams Images

The true-color image, which was used to show the actual RGB color of plants, was obtained from the corresponding hyperspectral image. As shown in Figure 5a–c, the true-color image was green due to the plant characteristic reflecting more green light than red and blue. In comparison, Ncams’ images were bluer than the true-color images since the special sensitivities of Ncams. According to Equations (2)–(3), the calibrated value of a channel in Ncams, Cc, could be described with Equation (8).
C c = α N p + β R p α N w + β R w = α λ N w + β θ R w α N w + β R w = α β λ N w + θ R w α β N w + R w = λ + R w ( θ λ ) α β N w + R w
where Np and Rp are NIR and red band reflectance from the plant, respectively, whereas Nw and Rw are NIR and red band reflectance from the whiteboard; α and β are the sensitivity of the channel to NIR and red light, respectively; λ and θ are the reflectance ratio of the plant to the whiteboard in NIR and red bands, respectively; and α/β is the sensitivity ratio of NIR to red (NIR-red ratio).
The NDVI value of plants is up to 0.8 [34]. According to Equation (5), plants reflect up to ten times more NIR than red light, which suggests that λ is much higher than θ. Therefore, the larger NIR-red ratio would give a lager Cc (Equation (8)). In this study, Ncams had the largest NIR-red ratio in the blue channel but the smallest NIR-red ratio in the red channel (Figure 1), thus Bc was the largest whereas Rc was the smallest (Figure 5d). Consequently, the image of Ncams was bluer than the true-color image. Besides, for Ncam, which only sensed red and NIR light in each channel, the NIR-red ratio indicated the purity of red and NIR recorded in the channel. When the NIR-red ratio approximated to zero, the channel would only record red reflectance. Oppositely, when the NIR-red ratio tended to infinite, the channel would only record NIR reflectance. As Cc and NIR-red ratio was positively correlated, it was easy to derive that smaller Cc gave purer red reflectance, and larger Cc gave purer NIR reflectance for the same plant. Bc of Ncam tp was larger than that of Ncam lp, and Rc of Ncam tp was smaller than that of Ncam lp (Figure 5d). Thus, Ncam tp obtained purer red and NIR reflectance than Ncam lp. This finding, while preliminary, suggested that Ncam tp may achieve more accurate NDVI than Ncam lp.

3.3. NDVI Estimation

Ncam tp achieved high performance with both of the models (Table 2). The model of MLR performed a little bit better, with R2 of 0.96, RMSE of 0.0079, MAE of 0.0064, and MPE of 0. 93%, than the other two models. Figure 6a presents the Bland-Altman diagram of NDVI measurements from HSC and Ncam tp using the MLR model. There was no bias between the measurements from Ncam tp and HSC as the points were distributed around zero evenly. The absolute 95% agreement limit was less than 0.02, which indicated a good agreement between these two measurements.
For Ncam lp, the model of MLR, GP, and SVR achieved close performance. The model of MLR performed a little bit better, with R2 of 0.82, RMSE of 0.0169, MAE of 0.0143, and MPE of 2.09%, than the other two models with respect to MAE and MPE (Table 2). The Bland-Altman diagram of measurements from Ncam lp and HSC in Figure 6b showed that the NDVI for the well-watered plants was underestimated, whereas the NDVI for most drought-stressed plants was overestimated. Regarding the whole dataset, the absolute 95% agreement limit was less than 0.04 and there was no bias, still suggesting a good agreement between these two measurements.
Ncam tp outperformed Ncam lp with either of the models. One of the possible explanations might be that Ncam tp obtained purer red and NIR reflectance in corresponding channels than Ncam lp, as mentioned in Section 3.2.
The MLR model was easy to interpret and achieved high performance for both Ncam lp and Ncam tp. Thus, the NDVI estimated by MLR models was compared with that calculated with the typical formula method (Figure 7). The results showed that the MLR model achieved higher accuracy than the typical formula did on NDVI estimation for either of Ncams. The R2 of the MLR model was higher than that of the typical formula by 6% and 7% for Ncam lp and Ncam tp, respectively. The higher performance of MLR models than the typical formula method could be explained from two aspects. First, the NDVI was defined with red and NIR reflectance, which were recorded in both R, G, and B channels in Ncams. The MLR model used all useful information for NDVI estimation by inputting RGB channels values. In contrast, the typical formula method only used blue and red channels’ values. Second, the typical formula method ignored the confounding issue of red and NIR reflectance in the blue and red channels. In the blue channel, Ncams had a large NIR-red ratio, giving an approximate NIR reflectance. However, the red channel record red reflectance as well as NIR reflectance, which was not ignorable. Therefore, the MLR model was more accurate to estimate NDVI compared with the typical formula.

3.4. Other Application

The ability of Ncam to predict nitrogen contents was evaluated by comparing it with the HSC. The MLR model with preprocessed Rc, Bc, and their interaction was fitted for both Ncam lp and Ncam tp. The partial least square regression model with three latent variables was built for the hyperspectral camera. The results showed Ncam lp achieved a close performance (R2 = 0.94, RMSE = 0.2227, MAE = 0.1737,and MPE = 7.35%) with HSC (R2 = 0.95, RMSE = 0.2006, MAE = 0.1433, MPE = 6.24%), whereas Ncam tp obtained a lower accuracy with R2 = 0.84, RMSE = 0.3508, MAE = 0.2891 and MPE = 11.82% (Figure 8). The reason could be Ncam lp obtained adequate spectrum information for nitrogen prediction, while Ncam tp did not. Researchers comment that nitrogen contents are highly correlated with the visible and NIR reflectance, especially the red edge ranging from 700 nm to 800 nm [35,36]. However, according to Figure 1, Ncam tp did not sense the reflectance at the red edge, which resulted in a low accuracy for nitrogen contents prediction. Therefore, Ncam with the long-pass filter could predict nitrogen contents with high accuracy.

3.5. Ncam Advantages

In the greenhouse, Ncam tp achieved high accuracy in estimating NDVI values of plants, and Ncam lp achieved high accuracy on nitrogen contents prediction for corn plants. Compared to the high-end HSC applied in this study, Ncam had more advantages, such as quicker imaging speed, more compact structure, and easier operation (Table 3). Researchers also proved these advantages of RGB camera compared to HSC [14,37,38]. Moreover, the cost of Ncam is much lower than the HSC. In the Ncam imaging system, the Pi camera costed $24.58, the Raspberry Pi 3 B costed $34.49, the long pass filter costed $25.50, and the two-band-pass filter costed about $10.00. In total, the gross materials cost of Ncam tp and Ncam lp was $69 and $85, respectively.

3.6. Ncam Limitations

The Ncam system was specifically designed for these plant phenotyping researchers who are looking for high-resolution but low-cost sensors for plant NDVI measurement, especially for the indoor phenotyping studies with a light-controlled environment. Many studies have demonstrated that the value of NDVI is easily impacted by the light changes and weather conditions, especially in the field [10,39]. It is unknown how will the weather, sunlight, and the reflection from non-vegetations such as soil and water impact the NDVI measurement with Ncam in the field. Thus, it is necessary to adjust the parameters of Ncam and the machine learning models with field tests to adapt it to the field environment. For nitrogen contents prediction, more nitrogen level treatments and crop species are required to evaluate the performance of Ncams.

3.7. Opportunities for Future Work

Compared to the approach getting the NDVI from a standard RGB camera, the NIR sensitive RGB camera is expected to get a higher NDVI estimation accuracy, as NIR reflection is obtained and stored in the RGB bands [1,20]. However, with the development of the application of advanced algorithms, such as genetic algorithm, on the plant phenotyping, a standard RGB camera is proved to be showing a high ability to estimate NDVI values [18]. Thus, it will be valuable to compare the performance to estimate NDVI between a standard RGB camera and a NIR sensitive RGB camera. Based on this, a standard RGB camera accompanying Ncam will be applied in field experiments to predict the NDVI values of plants. The performance of the standard RGB camera and Ncam will be compared to analyze the impact of filters on the estimation of NDVI. Furthermore, we also expected to construct a generalized formula as the substitute of the typical NDVI formula for RGBN cameras based on the model from machine learning.

4. Conclusion

A low-cost solution, Ncam, to estimate NDVI with a widely used RGBN camera was proposed based on machine learning. Two systems with different filters, Ncam lp with long-pass filter and Ncam tp with two-band-pass filter, were developed and evaluated. The Ncams were tested and compared with an HSC in an experiment with corn plants under different nitrogen and water treatments in the greenhouse. The machine learning algorithms, including MLR, GP, and SVR, were applied to estimate NDVI with the ground truth calculated from the hyperspectral image. The results showed: 1) Ncam tp with the machine learning models achieved high performance on NDVI estimation with the R2 of 0.96 and RMSE of 0.0079. 2) The NDVI estimates with machine learning achieved higher accuracy than that calculated from the typical formula. 3) Ncam was also capable of predicting nitrogen contents for corn plants. Therefore, Ncam is an affordable and promising approach to measure NDVI as an alternative to MSC and HSC in the light-controlled environment. In the next experiment, we will test Ncam in the field and evaluate its performance under various weather conditions.

Author Contributions

All authors made significant contributions to this research. Conceptualization, L.W. and J.J.; methodology, L.W. and Y.D.; software, L.W.; validation, L.W.; formal analysis, Y.D.; investigation, Y.D. and L.W.; resources, T.U.R., D.M. and J.J.; data curation, L.W. and Y.D.; writing—original draft preparation, L.W.; writing—review and editing, Y.D., L.Z. and J.J.; visualization, Y.D. and L.W.; supervision, J.J.; project administration, J.J.; funding acquisition, J.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding

Acknowledgments

The authors would like to thank Meng-Yang Lin in the Department of Agronomy at Purdue University for his great work in cultivating the corn plants.

Conflicts of Interest

The authors declare no conflict of interest

References

  1. Rabatel, G.; Gorretta, N.; Labbé, S. Getting simultaneous red and near-infrared band data from a single digital camera for plant monitoring applications: Theoretical and practical study. Biosyst. Eng. 2014, 117, 2–14. [Google Scholar] [CrossRef] [Green Version]
  2. Kumar, A.; Bhandari, A.K.; Padhy, P. Improved normalised difference vegetation index method based on discrete cosine transform and singular value decomposition for satellite image processing. IET Signal Process. 2012, 6, 617–625. [Google Scholar] [CrossRef]
  3. Genovese, G.; Vignolles, C.; Nègre, T.; Passera, G. A methodology for a combined use of normalised difference vegetation index and CORINE land cover data for crop yield monitoring and forecasting. A case study on Spain. Agronomie 2001, 21, 91–111. [Google Scholar] [CrossRef] [Green Version]
  4. Sannier, C.A.D.; Taylor, J.C.; Du Plessis, W. Real-time monitoring of vegetation biomass with NOAA-AVHRR in Etosha National Park, Namibia, for fire risk assessment. Int. J. Remote Sens. 2002, 23, 71–89. [Google Scholar] [CrossRef]
  5. Kriegler, F.J.; Malila, W.A.; Nalepka, R.F.; Richardson, W. Preprocessing transformations and their effects on multispectral recognition. In Proceedings of the 6th International Symposium on Remote Sensing of Environment, Ann Arbor, MI, USA, 13–16 October 1969; pp. 97–131. [Google Scholar]
  6. Deng, L.; Mao, Z.; Li, X.; Hu, Z.; Duan, F.; Yan, Y. UAV-based multispectral remote sensing for precision agriculture: A comparison between different cameras. ISPRS J. Photogramm. Remote Sens. 2018, 146, 124–136. [Google Scholar] [CrossRef]
  7. Duan, T.; Chapman, S.C.; Guo, Y.; Zheng, B. Dynamic monitoring of NDVI in wheat agronomy and breeding trials using an unmanned aerial vehicle. Field Crop. Res. 2017, 210, 71–80. [Google Scholar] [CrossRef]
  8. Cho, M.A.; Skidmore, A.; Corsi, F.; van Wieren, S.E.; Sobhan, I. Estimation of green grass/herb biomass from airborne hyperspectral imagery using spectral indices and partial least squares regression. Int. J. Appl. Earth Obs. Geoinf. 2007, 9, 414–424. [Google Scholar] [CrossRef]
  9. Thenkabail, P.S.; Smith, R.B.; De Pauw, E. Hyperspectral vegetation indices and their relationships with agricultural crop characteristics. Remote Sens. Environ. 2000, 71, 158–182. [Google Scholar] [CrossRef]
  10. Galvão, L.S.; Ponzoni, F.J.; Epiphanio, J.C.N.; Rudorff, B.F.T.; Formaggio, A.R. Sun and view angle effects on NDVI determination of land cover types in the Brazilian Amazon region with hyperspectral data. Int. J. Remote Sens. 2004, 25, 1861–1879. [Google Scholar] [CrossRef]
  11. Wang, L.; Jin, J.; Song, Z.; Wang, J.; Zhang, L.; Rehman, T.U.; Ma, D.; Carpenter, N.R.; Tuinstra, M.R. LeafSpec: An accurate and portable hyperspectral corn leaf imager. Comput. Electron. Agric. 2020, 169. [Google Scholar] [CrossRef]
  12. Wang, L.; Duan, Y.; Zhang, L.; Wang, J.; Li, Y.; Jin, J. LeafScope: A Portable High-Resolution Multispectral Imager for In Vivo Imaging Soybean Leaf. Sensors 2020, 20, 2194. [Google Scholar] [CrossRef] [PubMed]
  13. Li, Q.; He, X.; Wang, Y.; Liu, H.; Xu, D.; Guo, F. Review of spectral imaging technology in biomedical engineering: Achievements and challenges. J. Biomed. Opt. 2013, 18, 100901. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Li, L.; Zhang, Q.; Huang, D. A review of imaging techniques for plant phenotyping. Sensors 2014, 14, 20078–20111. [Google Scholar] [CrossRef] [PubMed]
  15. Li, Y.; Chen, D.; Walker, C.N.; Angus, J.F. Estimating the nitrogen status of crops using a digital camera. F. Crop. Res. 2010, 118, 221–227. [Google Scholar] [CrossRef]
  16. Lebourgeois, V.; Bégué, A.; Labbé, S.; Houlès, M.; Martiné, J.F. A light-weight multi-spectral aerial imaging system for nitrogen crop monitoring. Precis. Agric. 2012, 13, 525–541. [Google Scholar] [CrossRef]
  17. Qiu, C.; Liao, G.; Tang, H.; Liu, F.; Liao, X.; Zhang, R.; Zhao, Z. Derivative parameters of hyperspectral NDVI and its application in the inversion of rapeseed leaf area index. Appl. Sci. 2018, 8, 1300. [Google Scholar] [CrossRef] [Green Version]
  18. Costa, L.; Nunes, L.; Ampatzidis, Y. A new visible band index (vNDVI) for estimating NDVI values on RGB images utilizing genetic algorithms. Comput. Electron. Agric. 2020, 172, 105334. [Google Scholar] [CrossRef]
  19. Sakamoto, T.; Gitelson, A.A.; Nguy-Robertson, A.L.; Arkebauer, T.J.; Wardlow, B.D.; Suyker, A.E.; Verma, S.B.; Shibayama, M. An alternative method using digital cameras for continuous monitoring of crop status. Agric. For. Meteorol. 2012, 154–155, 113. [Google Scholar] [CrossRef] [Green Version]
  20. Petach, A.R.; Toomey, M.; Aubrecht, D.M.; Richardson, A.D. Monitoring vegetation phenology using an infrared-enabled security camera. Agric. For. Meteorol. 2014, 195–196, 143–151. [Google Scholar] [CrossRef]
  21. Nijland, W.; de Jong, R.; de Jong, S.M.; Wulder, M.A.; Bater, C.W.; Coops, N.C. Monitoring plant condition and phenology using infrared sensitive consumer grade digital cameras. Agric. For. Meteorol. 2014, 184, 98–106. [Google Scholar] [CrossRef] [Green Version]
  22. Montes de Oca, A.; Arreola, L.; Flores, A.; Sanchez, J.; Flores, G. Low-cost multispectral imaging system for crop monitoring. In Proceedings of the 2018 International Conference on Unmanned Aircraft Systems (ICUAS), Dallas, TX, USA, 12–15 June 2018; pp. 443–451. [Google Scholar] [CrossRef]
  23. Sakamoto, T.; Shibayama, M.; Kimura, A.; Takada, E. Assessment of digital camera-derived vegetation indices in quantitative monitoring of seasonal rice growth. ISPRS J. Photogramm. Remote Sens. 2011, 66, 872–882. [Google Scholar] [CrossRef]
  24. Netto, A.F.A.; Martins, R.N.; De Souza, G.S.A.; Dos Santos, F.F.L.; Rosas, J.T.F. Evaluation of a low-cost camera for agricultural applications. J. Exp. Agric. Int. 2019, 1–9. [Google Scholar] [CrossRef]
  25. Zheng, H.; Cheng, T.; Li, D.; Zhou, X.; Yao, X.; Tian, Y.; Cao, W.; Zhu, Y. Evaluation of RGB, color-infrared and multispectral images acquired from unmanned aerial systems for the estimation of nitrogen accumulation in rice. Remote Sens. 2018, 10, 824. [Google Scholar] [CrossRef] [Green Version]
  26. Stow, D.; Hope, A.; Chen, D.; Garrison, C.; Service, D.; Richardson, D. Potential of colour-infrared digital camera imagery for inventory and mapping of alien plant invasions in South African shrublands. Int. J. Remote Sens. 2000, 21, 2965–2970. [Google Scholar] [CrossRef]
  27. Dworak, V.; Selbeck, J.; Dammer, K.H.; Hoffmann, M.; Zarezadeh, A.A.; Bobda, C. Strategy for the development of a smart NDVI camera system for outdoor plant detection and agricultural embedded systems. Sensors 2013, 13, 1523–1538. [Google Scholar] [CrossRef] [PubMed]
  28. Raza, A.; Ikram, A.A.; Amin, A.; Ikram, A.J. A review of low cost and power efficient development boards for IoT applications. In Proceedings of the 2016 Future Technologies Conference (FTC), San Francisco, CA, USA, 6–7 December 2017; pp. 786–790. [Google Scholar] [CrossRef]
  29. Ma, D.; Carpenter, N.; Amatya, S.; Maki, H.; Wang, L.; Zhang, L.; Neeno, S.; Tuinstra, M.R.; Jin, J. Removal of greenhouse microclimate heterogeneity with conveyor system for indoor phenotyping. Comput. Electron. Agric. 2019, 166, 104979. [Google Scholar] [CrossRef]
  30. Zhang, L.; Maki, H.; Ma, D.; Sánchez-Gallego, J.A.; Mickelbart, M.V.; Wang, L.; Rehman, T.U.; Jin, J. Optimized angles of the swing hyperspectral imaging system for single corn plant. Comput. Electron. Agric. 2019, 156, 349–359. [Google Scholar] [CrossRef]
  31. Pagnutti, M.; Ryan, R.E.; Cazenavette, G.; Gold, M.; Harlan, R.; Leggett, E.; Pagnutti, J. Laying the foundation to use Raspberry Pi 3 V2 camera module imagery for scientific and engineering purposes. J. Electron. Imaging 2017, 26, 013014. [Google Scholar] [CrossRef] [Green Version]
  32. Quiñonero-Candela, J.; Rasmussen, C.E. A unifying view of sparse approximate Gaussian process regression. J. Mach. Learn. Res. 2005, 6, 1939–1959. [Google Scholar]
  33. Drucker, H.; Surges, C.J.C.; Kaufman, L.; Smola, A.; Vapnik, V. Support vector regression machines. In Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA, 1–6 December 1997; pp. 155–161. [Google Scholar]
  34. Masialeti, I.; Egbert, S.; Wardlow, B. A comparative analysis of phenological curves for major crops in Kansas. GISci. Remote Sens. 2010, 47, 241–259. [Google Scholar] [CrossRef]
  35. Smith, M.-L.; Ollinger, S.V.; Martin, M.E.; Aber, J.D.; Hallett, R.A.; Goodale, C.L. Direct estimation of aboveground forest productivity through hyperspectral remote sensing of canopy nitrogen. Ecol. Appl. 2002, 12, 1286–1302. [Google Scholar] [CrossRef]
  36. Clevers, J.G.P.W.; Kooistra, L. Using hyperspectral remote sensing data for retrieving canopy chlorophyll and nitrogen content. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 574–583. [Google Scholar] [CrossRef]
  37. Taghizadeh, M.; Gowen, A.A.; O’donnell, C.P. Comparison of hyperspectral imaging with conventional RGB imaging for quality evaluation of Agaricus bisporus mushrooms. Biosyst. Eng. 2011, 108, 191–194. [Google Scholar] [CrossRef]
  38. Prey, L.; von Bloh, M.; Schmidhalter, U. Evaluating RGB imaging and multispectral active and hyperspectral passive sensing for assessing early plant vigor in winter wheat. Sensors 2018, 18, 2931. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Wang, J.; Rich, P.M.; Price, K.P. Temporal responses of NDVI to precipitation and temperature in the central Great Plains, USA. Int. J. Remote Sens. 2003, 24, 2345–2364. [Google Scholar] [CrossRef]
Figure 1. Channel sensitivities of (a) Pi camera and long-pass filter, two-band-pass filter, (b) Ncam lp, and (c) Ncam tp.
Figure 1. Channel sensitivities of (a) Pi camera and long-pass filter, two-band-pass filter, (b) Ncam lp, and (c) Ncam tp.
Sensors 20 03208 g001
Figure 2. Imaging systems of Ncams and hyperspectral camera.
Figure 2. Imaging systems of Ncams and hyperspectral camera.
Sensors 20 03208 g002
Figure 3. Flowchart of image processing. HSI, hyperspectral image.
Figure 3. Flowchart of image processing. HSI, hyperspectral image.
Sensors 20 03208 g003
Figure 4. The boxplot of the corn plants raw dataset from Ncams and HSC after calibration with the whiteboard.
Figure 4. The boxplot of the corn plants raw dataset from Ncams and HSC after calibration with the whiteboard.
Sensors 20 03208 g004
Figure 5. Images of (a) true-color RGB, (b) Ncam lp, and (c) Ncam tp; and (d) Preprocessed RGB values of true-color RGB and Ncams images.
Figure 5. Images of (a) true-color RGB, (b) Ncam lp, and (c) Ncam tp; and (d) Preprocessed RGB values of true-color RGB and Ncams images.
Sensors 20 03208 g005
Figure 6. The Bland-Altman diagram of NDVI measurements from (a) HSC and Ncam tp, and (b) HSC and Ncam lp using the model of MLR. LNLW, low nitrogen and drought-stressed. HNLW, high nitrogen and drought-stressed. LNHW, low nitrogen and well-watered. HNHW, high nitrogen and well-watered.
Figure 6. The Bland-Altman diagram of NDVI measurements from (a) HSC and Ncam tp, and (b) HSC and Ncam lp using the model of MLR. LNLW, low nitrogen and drought-stressed. HNLW, high nitrogen and drought-stressed. LNHW, low nitrogen and well-watered. HNHW, high nitrogen and well-watered.
Sensors 20 03208 g006
Figure 7. The performances of Ncams on estimating NDVI with (a) machine learning methods and (b) typical formula. lp, Ncam lp; tp, Ncam tp.
Figure 7. The performances of Ncams on estimating NDVI with (a) machine learning methods and (b) typical formula. lp, Ncam lp; tp, Ncam tp.
Sensors 20 03208 g007
Figure 8. The performances of Ncams on estimating nitrogen (N) contents with (a) Ncams and (b) HSC. lp, Ncam lp; tp, Ncam tp.
Figure 8. The performances of Ncams on estimating nitrogen (N) contents with (a) Ncams and (b) HSC. lp, Ncam lp; tp, Ncam tp.
Sensors 20 03208 g008
Table 1. Initialization parameters of HSC, Ncam lp, and Ncam tp for the performance testing experiment.
Table 1. Initialization parameters of HSC, Ncam lp, and Ncam tp for the performance testing experiment.
ParametersHSCNcam lpNcam tp
Frame rate (FPS)6022
Exposure time (ms)61016
Resolution1760 × 16203280 × 24643280 × 2464
Note: FPS, frame per second.
Table 2. The performances of Ncams on NDVI estimation with different models.
Table 2. The performances of Ncams on NDVI estimation with different models.
Ncam tpNcam lp
ModelR2RMSEMAEMPER2RMSEMAEMPE
MLR0.960.00790.00640.93%0.820.01690.01432.09%
GP Matern0.960.00790.00670.97%0.820.01680.01442.10%
SVR quadratic0.960.00810.00650.94%0.810.01730.01492.18%
Note: R2, RMSE, MAE, and MPE are from the cross-validated values.
Table 3. Comparison of Ncams and HSC.
Table 3. Comparison of Ncams and HSC.
ParametersHSCNcam lpNcam tp
ConfigurationSpatial resolution1760 × 16203280 × 24643280 × 2464
Effective bands51333
Spectral range (nm)400 to 1000580 to 1000620 to 690 and 810 to 890
Imaging methodImaging typePush broom scanOne-shotOne-shot
Imaging speed1 minLess than 1 sLess than 1 s
Size and weightDimension (mm)320 × 90 × 96 25 × 24 × 925 × 24 × 9
Weight (kg)5About 0.1About 0.1
Operation--ProfessionalEasyEasy
Price--Over $10,000$85 $69

Share and Cite

MDPI and ACS Style

Wang, L.; Duan, Y.; Zhang, L.; Rehman, T.U.; Ma, D.; Jin, J. Precise Estimation of NDVI with a Simple NIR Sensitive RGB Camera and Machine Learning Methods for Corn Plants. Sensors 2020, 20, 3208. https://doi.org/10.3390/s20113208

AMA Style

Wang L, Duan Y, Zhang L, Rehman TU, Ma D, Jin J. Precise Estimation of NDVI with a Simple NIR Sensitive RGB Camera and Machine Learning Methods for Corn Plants. Sensors. 2020; 20(11):3208. https://doi.org/10.3390/s20113208

Chicago/Turabian Style

Wang, Liangju, Yunhong Duan, Libo Zhang, Tanzeel U. Rehman, Dongdong Ma, and Jian Jin. 2020. "Precise Estimation of NDVI with a Simple NIR Sensitive RGB Camera and Machine Learning Methods for Corn Plants" Sensors 20, no. 11: 3208. https://doi.org/10.3390/s20113208

APA Style

Wang, L., Duan, Y., Zhang, L., Rehman, T. U., Ma, D., & Jin, J. (2020). Precise Estimation of NDVI with a Simple NIR Sensitive RGB Camera and Machine Learning Methods for Corn Plants. Sensors, 20(11), 3208. https://doi.org/10.3390/s20113208

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop