Next Article in Journal
Drought Stress Drives Sex-Specific Physiological and Biochemical Differences in Female and Male Litsea cubeba
Previous Article in Journal
Microbial-Enhanced Abiotic Stress Tolerance in Grapevines: Molecular Mechanisms and Synergistic Effects of Arbuscular Mycorrhizal Fungi, Plant Growth-Promoting Rhizobacteria, and Endophytes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Study on the Detection of Chlorophyll Content in Tomato Leaves Based on RGB Images

by
Xuehui Zhang
1,2,
Huijiao Yu
1,2,
Jun Yan
1,2 and
Xianyong Meng
1,2,*
1
College of Information Science and Engineering, Shandong Agricultural University, Taian 271000, China
2
The Key Laboratory of Smart Agriculture Technology in Huang-Huai-Hai Region, Ministry of Agriculture and Rural Affairs, Tai’an 271018, China
*
Author to whom correspondence should be addressed.
Horticulturae 2025, 11(6), 593; https://doi.org/10.3390/horticulturae11060593
Submission received: 17 April 2025 / Revised: 20 May 2025 / Accepted: 22 May 2025 / Published: 26 May 2025
(This article belongs to the Section Vegetable Production Systems)

Abstract

:
Chlorophyll is a key substance in plant photosynthesis, and its content detection methods are of great significance in the field of agricultural AI. These methods provide important technical support for crop growth monitoring, pest and disease identification, and yield prediction, playing a crucial role in improving agricultural productivity and the level of intelligence in farming. This paper aims to explore an efficient and low-cost non-destructive method for detecting chlorophyll content (SPAD) and investigate the feasibility of smartphone image analysis technology in predicting chlorophyll content in greenhouse tomatoes. This study uses greenhouse tomato leaves as the experimental object and analyzes the correlation between chlorophyll content and image color features. First, leaf images are captured using a smartphone, and 42 color features based on the red, green, and blue (R, G, B) color channels are constructed to assess their correlation with chlorophyll content. The experiment selects eight color features most sensitive to chlorophyll content, including B, (2G − R − B)/(2G + R + B), GLA, RGBVI, g, g − b, ExG, and CIVE. Based on this, this study constructs and evaluates the predictive performance of multiple models, including multiple linear regression (MLR), ridge regression (RR), support vector regression (SVR), random forest (RF), and the Stacking ensemble learning model. The experimental results indicate that the Stacking ensemble learning model performs the best in terms of prediction accuracy and stability (R2 = 0.8359, RMSE = 0.8748). The study confirms the feasibility of using smartphone image analysis for estimating chlorophyll content, providing a convenient, cost-effective, and efficient technological approach for crop health monitoring and precision agriculture management. This method helps agricultural workers to monitor crop growth in real-time and optimize management decisions.

1. Introduction

Tomato (Solanum lycopersicum L.) [1] is an annual herbaceous plant belonging to the Solanaceae family and the Solanum genus, with plants reaching up to 2 m in height. As one of the most widely cultivated fruit vegetables worldwide, tomatoes play a significant role in major producing countries such as the United States, Italy, and China. Tomatoes are not only rich in nutrients, containing carotene, vitamin C, and B vitamins [2], but also have a unique flavor, making them suitable for both raw consumption and various cooking methods, such as boiling and stir-frying. In addition, tomatoes are widely used in the food processing industry, where they are made into products like beverages, dried fruits, and seasonings. They hold a crucial position in both global agriculture and the food industry [3].
Chlorophyll is a key substance for photosynthesis in green plants and an important indicator for evaluating plant nutritional status [4]. Obtaining real-time and accurate chlorophyll content data is crucial for scientific fertilization, pest and disease control, and crop management [5], facilitating healthy crop growth. In the process of crop growth, key indicators of plant health include chlorophyll fluorescence, chlorophyll content, and nutritional and water status. Among these, the chlorophyll level in the leaves shows a significant correlation with the plant’s growth condition and overall health [6]. Therefore, as an important economic crop, studying the chlorophyll levels in tomato leaves is crucial for ensuring the healthy growth of tomato plants.
Traditionally, chlorophyll content is measured through chemical analysis, typically using specialized instruments in a laboratory setting [7]. In contrast, the portable handheld chlorophyll meter (SPAD meter) is based on the principle of Beer’s Law and uses two characteristic wavelengths, 650 nm and 940 nm, for simultaneous detection. It enables rapid and non-destructive measurement of chlorophyll content in plant leaves. This instrument is low-cost and easy to carry and operate, and obtains SPAD values that are highly correlated with the chlorophyll content of crops [8]. Uddling et al. [9] studied the relationship between SPAD values and chlorophyll concentrations in birch, wheat, and potato leaves, demonstrating significant correlations. Similarly, Markwell et al. [10] measured the chlorophyll concentration in soybean and corn leaves using organic extraction and spectrophotometry, and compared the results with the output values from the SPAD-502 chlorophyll meter. The study showed a significant correlation between the two methods. However, SPAD meters have limitations, such as their small measurement area during a single test, making it difficult to meet the demand for large-scale rapid detection. This constrains their application and widespread adoption in practical production scenarios [11]. Li et al. [12] used a SPAD meter to measure the chlorophyll content of Hami melon leaves and verified the relationship between RGB image features and the measured chlorophyll content. The results indicated that RGB images can effectively predict chlorophyll content.
With the advancement of sensor technology, chlorophyll detection methods based on spectrometers or spectral imagers have been extensively studied. By acquiring near-infrared spectra [13], chlorophyll fluorescence spectra [14], multispectral images [15], and hyperspectral images [16,17] of crop leaves or canopies, and extracting spectral features, chlorophyll content can be predicted. Hyperspectral imaging technology, which combines imaging and spectroscopy, enables rapid and non-destructive detection of chlorophyll content, addressing the limitations of traditional manual measurement methods. For instance, Qiao Lang et al. [18] used drone remote sensing technology to collect aerial images of field corn. By constructing color features and building a maize canopy chlorophyll content detection model based on a BP neural network, their model achieved an R2 of 0.72. However, spectrometers are expensive, and their data processing and analysis procedures are complex and time-consuming, limiting their widespread application [19]. The aforementioned methods all rely on specialized and costly instruments, resulting in high measurement costs. Additionally, traditional methods not only require destructive sampling of the leaves, but also involve complex procedures and significant delays [7], making large-scale application difficult. Therefore, it is crucial to find a simple, cost-effective, and easily scalable method for estimating chlorophyll content in plant leaves.
The combination of digital image processing technology and visible light devices enables the fast and non-destructive segmentation of leaf images. By utilizing the red, green, and blue (RGB) channels, different color models are used to quantify the color information in digital images [20], allowing for the rapid, non-destructive detection of chlorophyll content in crops. Related studies have shown that changes in chlorophyll content and other pigments within plants lead to alterations in the color of crop leaves. Therefore, by combining digital image processing technology to extract color features highly correlated with chlorophyll content, rapid and non-destructive detection of chlorophyll content in crops can be achieved [21].
In the past, researchers primarily used the RGB color model to monitor plant health. Most studies employed digital cameras to capture leaf images and analyzed the relationship between the R, G, and B values and the chlorophyll and nitrogen content [22]. The color characteristics of plant leaves are primarily influenced by their internal spectral absorption and reflection properties. Chlorophyll a and b have the strongest absorption capabilities in the blue light (430–450 nm) and red light (640–660 nm) bands, while their absorption of green light (500–600 nm) is weaker. This results in leaves primarily reflecting green light, giving them their green appearance [23]. This absorption characteristic allows for the use of R, G, B values and their combined color indices to estimate chlorophyll content and plant health. For instance, Riccardi et al. [24] found that, based on the best-fit regression model for individual color component variables, in amaranth, the R value shows a higher correlation with chlorophyll content, while in quinoa, the G value has a higher correlation with chlorophyll content. Hu et al. [25] used a SPAD meter to measure leaf chlorophyll content and found that RGB color features such as R, G, R + G + B, R−B, R + B, and R + G were significantly correlated with chlorophyll content. Sulistyo [26] used a standard digital camera to directly capture wheat field images, extracted 12 statistical features from the RGB color space, and established a neural network model for predicting wheat leaf nitrogen content. Amaral et al. [27] extracted color features of tropical tree seedling species during the nursery stage using RGB images. The study found that R, 2R(G − B)/(G + B), and 2G(G − B)/(G + B) were most beneficial for analyzing the physiological status and quality of tropical seedlings. Guo et al. [28] extracted color features from maize RGB images and analyzed their correlation with chlorophyll content. Based on this, they developed chlorophyll content estimation models using a backpropagation neural network (BP), support vector machine regression (SVM), and random forest (RF), with the highest coefficient of determination (R2) reaching 0.85.
Research on extracting color features from RGB images has also made progress in plants such as cotton [29], microalgae [30], and citrus [31]. Additionally, studies on crops such as spinach [32] and willow [33] have confirmed that color features extracted from RGB images can achieve rapid detection of leaf chlorophyll content.
With the widespread popularity of smartphones, phone cameras have become a common feature. The RGB images captured by these cameras are not only easy to obtain but also relatively low in cost [34]. By providing spectral information of color images such as red, green, blue (RGB), and hue, saturation, intensity (HSI), constructing different color indices based on RGB and HSI helps researchers estimate the chlorophyll concentration in leaves. Therefore, the use of RGB image processing technology to obtain plant chlorophyll content has gradually gained widespread application due to its non-destructive, real-time, simple, and accurate characteristics. Mohan et al. [35] obtained images of rice leaves and their chlorophyll content using a smartphone and a SPAD meter, and employed digital image analysis methods to detect chlorophyll content. The study found that the RGB model and the DGCI-rgb model both performed well in predicting chlorophyll content, with the latter achieving a correlation coefficient of 0.63. Similarly, Li et al. [36] proposed a low-cost method for acquiring plant leaf RGB images using a smartphone combined with an auxiliary photographing device, and for instantly detecting chlorophyll content. Through color difference correction and optimization with various machine learning models, they constructed a prediction model. The study showed that the SVR model had higher accuracy and better performance, with a correlation coefficient reaching 0.73. Li et al. [12] used a smartphone to measure the chlorophyll content in Hami melon leaves. They selected (B − G − R)/(B + G) and (G − B)B/(R + G) as feature combinations. Through model prediction verification, the results showed that the random forest regression model performed the best.
The aforementioned studies provide references for extracting image features from mobile devices and modeling chlorophyll detection. At the same time, color correction [36] technology corrects and adjusts the color differences in images through mathematical models and algorithms, ensuring that the image displays consistent colors under different devices and lighting conditions. Due to differences in sensors, lenses, and color processing algorithms used by different models of smartphones and cameras, images of the same object captured on different devices may exhibit color deviations. Therefore, color calibration technology plays a key role in the image acquisition and processing process, providing compatibility for different smartphone models when capturing images. It prevents color deviation issues caused by device differences, thereby ensuring the consistency and accuracy of image data.
However, most current research focuses on estimating chlorophyll content in field crops such as wheat and corn, typically conducted in controlled environments, which limits these studies’ applicability and flexibility. In comparison, research on non-destructive, close-range estimation of chlorophyll content in tomato leaves in natural environments is relatively scarce. Different crops exhibit distinct mathematical fitting behaviors during chlorophyll content estimation, and significant differences exist between species in this regard [37]. Therefore, research methods need to be adjusted according to the specific plant species and its growth stage.
Greenhouse cultivation, as an efficient and sustainable agricultural production method, has gradually gained widespread application due to its ability to effectively control environmental factors and improve crop yield and quality. However, the spatial limitations of greenhouse environments make it challenging for traditional drone spectrometers to efficiently capture crop leaf images, which poses a challenge for precise monitoring of crop growth. In contrast, the complex and variable background of crop leaves presents additional difficulties for the application of digital image technology. Although previous studies [12,35,36] have explored digital image-based chlorophyll measurement methods, most existing approaches rely on controlled environments or expensive equipment, lacking a universal solution suitable for greenhouse environments and complex backgrounds. To address this research gap, this paper proposes an innovative, smartphone-based, rapid, non-destructive, and environmentally adaptable chlorophyll content recognition method. By using black cardboard to block complex backgrounds, non-destructive RGB images of tomato leaves are captured, and the correlation between image features and chlorophyll content is analyzed. Machine learning techniques are then employed to construct prediction models. This method does not rely on controlled environments and can be widely applied to various practical scenarios. It not only significantly improves the convenience of chlorophyll content recognition but also ensures the integrity of the plants, making it easier to conduct further research on other parameters at multiple time points. The results of this study can be widely applied in smart agriculture practices, enabling real-time monitoring of tomato plant growth, optimizing fertilization management, and improving agricultural resource utilization efficiency.

2. Materials and Methods

2.1. Leaf Data Collection

The experiment was conducted in 2024 at the cucumber science and technology yard in Juye, Shandong Province. The tested tomato variety was “Shuang fei 8.” Thid study focuses on tomato leaves at the flowering stage (leaf age 22–28 days), a critical phase in tomato growth and development. During this period, the physiological state of the leaves and the measurement conditions are optimal, making it an ideal time for chlorophyll content determination. The flowering stage is typically associated with high photosynthetic efficiency, and previous studies [38] have shown a significant correlation between leaf physiological status and fruit set rate. In contrast, younger leaves (around 15 days old) may not be fully developed, while older leaves (over 40 days old) often exhibit varying degrees of chlorophyll degradation. In the experimental design, a random sampling method was employed: from 114 tomato plants at the flowering stage, one healthy leaf with uniform color and no visible pests or diseases was selected from each plant as a sample. All samples were collected using non-destructive methods to ensure that the normal growth and development of the plants were not affected. To ensure the reliability of the experimental materials, the selected leaves should have a consistent color distribution across the entire area, with minimal color variation. There should be no spots, shadows, yellowing areas, or color gradients present. Data collection included measuring the chlorophyll content (SPAD value) of tomato leaves and capturing images of the leaves. The images were taken immediately after measuring the SPAD value to ensure each leaf image corresponded precisely to its SPAD value.
Images were captured using an iPhone 13 Pro smartphone (Foxconn Zhengzhou Factory, Zhengzhou, China). To ensure consistency in RGB values, photographs were taken under stable lighting conditions on clear, windless days during two time windows—9:00–11:00 a.m. and 3:00–5:00 p.m.—reducing the impact of light fluctuations on RGB values. At the same time, to further improve the accuracy and consistency of the image colors, color calibration technology was applied to process the captured images, correcting color deviations caused by differences in device sensors, lenses, and color processing algorithms. During shooting, the camera lens should be kept perpendicular to the leaf surface to avoid color distortion caused by angle deviations, thereby enhancing the reliability and comparability of the image data.
To reduce the interference of complex backgrounds on image feature extraction, a narrow slit was created above the black cardboard. During the image capture, the petiole of the leaf was passed through the slit, and the leaf was laid flat on the surface of the cardboard. This method effectively highlights the leaf’s color features while blocking out interference from soil, weeds, and other factors in the field environment. As a result, RGB images are captured non-destructively, providing a reliable foundation for subsequent color extraction and analysis. The captured leaf image is shown in Figure 1. Additionally, filtering and image smoothing techniques were applied to reduce noise caused by background glare or reflections. All images were stored in JPEG format with a resolution of 3024 × 4032 pixels to ensure high-quality data acquisition. Each image was named using a unique identifier corresponding to the leaf number for subsequent analysis.

2.2. Chlorophyll Content Measurement

After capturing each image, the SPAD-502 chlorophyll meter (Beijing Zhongke Weihe, TYS-4N, Weihe, China) was used to measure the chlorophyll content. To minimize measurement error, measurements were taken by avoiding large veins. For each leaf, six points were sampled as shown in Figure 2, and the average value was calculated as the chlorophyll content for that leaf. A total of 114 samples were measured during the experiment.

2.3. Leaf Color Feature Parameter Extraction

Initially, the resolution of the images was reduced, and the Mask-RCNN algorithm was used to segment the leaf images, removing the black background and completing image preprocessing. Python 3.11, along with the OpenCV library, was then used to further process the segmented tomato leaf images. Gaussian filtering and morphological closing operations were applied to eliminate noise and enhance the images, with the enhanced images shown in Figure 3. Next, image processing techniques were used to extract the region of interest (RoI) from the images. The RGB channels of the RoI images were extracted, and the mean values of R, G, and B were calculated. Subsequently, additional color features were computed through RGB algebraic operations, and the data were converted into other color spaces such as HSI and Lab. A total of 42 color features were obtained. The detailed color features are shown in Table 1. High-correlation color features were then selected for subsequent analysis and modeling. Based on this foundation, an ensemble learning approach was adopted, integrating multiple machine learning models using the Stacking method to further enhance model performance. The experimental workflow is illustrated in Figure 4.

2.4. Construction of Chlorophyll Prediction Models

First, the Pearson correlation between each color feature and chlorophyll content was analyzed to identify color features with high correlations. Then, multiple modeling methods were applied to the dataset, including multiple linear regression (MLR), support vector regression (SVR), ridge regression (RR), random forest (RF), and ensemble learning (Stacking). The performance of these models was compared. During the modeling process, 70% of the data was randomly selected as the training set and the remaining 30% as the test set.
The performance evaluation of the prediction models is critical for assessing their accuracy and stability, which are key factors in determining the quality of the models and the precision of prediction results. Common indicators for evaluating model performance include the coefficient of determination (R2), root mean square error (RMSE), and normalized root mean square error (NRMSE), with their respective formulas shown in Equations (1)–(3). A higher R2 value indicates better predictive modeling performance, while lower RMSE and NRMSE values signify stronger model stability.
R 2 = 1 i = 1 n y i y ^ i 2 i = 1 n y i y ¯ i 2
R M S E = 1 n i = 1 n y i y ^ i 2
N R M S E = 1 n i = 1 n y i y ^ i 2 y ¯ m a x y ¯ m i n
where y i is the actual observed value; y ^ i is the predicted value from the model; y ¯ i is the mean of the actual observed values; n is the total number of samples; y ¯ m a x and y ¯ m i n represent the maximum and minimum actual observed values, respectively.

3. Results and Analysis

3.1. Correlation Analysis Between Color Features and Chlorophyll Content

The Pearson correlation between color features and chlorophyll content was analyzed. The Pearson correlation coefficient ranges from −1 to 1, where a positive value indicates a positive correlation, meaning the two variables change in the same direction, while a negative value indicates a negative correlation, meaning the two variables change in opposite directions. By converting RGB, HSI, and Lab color spaces and applying algebraic operations, 42 color features were calculated. The results of the correlation analysis with SPAD values are shown in Table 2.
The color features with higher correlation coefficients listed in the table were selected for analysis, and Figure 5 provides a more intuitive illustration of the relationship between these features and SPAD values.
Based on the 42 color features mentioned above, it was found that B, (2G − R − B)/(2G + R + B), GLA, RGBVI, g, g − b, ExG, and CIVE values exhibit a strong correlation with SPAD values. Among them, the B value exhibits a correlation of 0.82 with the SPAD parameter. The scatter plots and regression lines between these eight color features and chlorophyll content are shown in Figure 6.
The analysis results indicate that among the RGB primary colors, the correlation between R and G values with chlorophyll content is relatively low, while the correlation between the B value and chlorophyll content is the highest and positive. For normalized color features (r, g, b), the g value had a higher correlation with chlorophyll content and showed a negative correlation. In terms of color ratio parameters, (2G − R − B)/(2G + R + B) had a strong negative correlation with chlorophyll content. For color difference parameters, g−b demonstrated a higher positive correlation. By converting the RGB color space into HSI and Lab color spaces, it was observed that in both spaces the S value had a higher negative correlation with chlorophyll content. Additionally, in vegetation indices, most color features showed high correlation with chlorophyll content, particularly CIVE, RGBVI, ExG, and GLA, which had strong correlation coefficients. Based on the above color features, constructing a model to estimate chlorophyll content in tomato leaves is statistically feasible and provides a reliable foundation for rapid estimation of chlorophyll content in tomatoes.

3.2. Chlorophyll Content Prediction Model

The superior performance of the Stacking model primarily stems from its ability to integrate the predictive capabilities of multiple base learners and leverage a meta-learner to refine their outputs. The random forest (RF) model excels in handling nonlinear relationships by constructing multiple decision trees and aggregating their results through voting or averaging, effectively capturing complex patterns and nonlinear features within the data. Support vector regression (SVR) performs more stably on small sample datasets, optimizing the prediction boundary by maximizing the margin, which helps mitigate overfitting in limited data scenarios. Meanwhile, multiple linear regression (MLR) is more efficient in handling simple linear relationships, providing quick and accurate linear predictions.
Based on the above analysis, this study proposes a chlorophyll content estimation model for leaves based on the Stacking ensemble approach. Multiple linear regression (MLR), support vector regression (SVR), and random forest (RF) models are selected as base learners, with ridge regression (Ridge) as the meta-learner. A brief introduction to each model is provided below.
(1)
Multiple Linear Regression Model (MLR)
MLR is a model that extends simple linear regression to handle the linear relationships between multiple independent variables (features) and a dependent variable (target value). The calculation formula for MLR is shown in Equation (4).
Y = β 0 + β 1 x 1 + β 2 x 2 + + β p x p + ϵ
In the formula, β0 is the intercept (constant term); β1, β2, …, βp are the regression coefficients, representing the extent to which each independent variable influences the dependent variable; x1, x2, …, xp are the independent variables (feature variables); p is the number of independent variables; and ϵ is the error term, representing the random error between the dependent variable and the predicted value from the regression model.
(2)
Support Vector Regression Model (SVR)
SVR is based on the principles of SVM (support vector machine). It works by finding the optimal hyperplane to minimize the distance from sample points to the hyperplane, thereby predicting continuous variables. The core idea is to define a specific loss function to differentiate data points based on their distances and apply corresponding penalties. In this study, the SVR model parameters were selected using grid search, and the best penalty coefficient (C), kernel function (kernel), and kernel function coefficient (gamma) are shown in Table 3.
(3)
Random Forest Model (RF)
RF is an ensemble learning algorithm that generates multiple decision trees by randomly sampling data using the Bootstrap method and combines the predictions of all trees to make a final decision. This algorithm can efficiently handle high-dimensional data and effectively overcome the overfitting issue commonly found in individual decision trees. In this study, the grid search method was used to optimize and select the key parameters of the RF model, including the number of decision trees (n_estimators), maximum tree depth (max_depth), minimum samples required for node splitting (min_samples_split), and minimum samples required for leaf nodes (min_samples_leaf). The specific parameters are shown in Table 3.
(4)
Ridge Regression Model (RR)
Ridge regression is based on ordinary least squares regression, but it applies L2 regularization to the sum of the squares of the regression coefficients. The loss function formula is shown in Equation (5). A larger λ value means stronger regularization, which makes the model simpler but may sacrifice some fitting ability. In ridge regression, the goal is to find a linear model that minimizes the loss function. In this study, the best λ value for the ridge regression model was selected using grid search, as shown in Table 3.
L o s s = i = 1 n y i y ^ i 2 + λ j = 1 p β j 2 = y x β 2 2 + λ β 2 2
In the formula, y i represents the true observed value of the i-th sample; y ^ i is the predicted value for the i-th sample; β j denotes the coefficient for the j-th feature; λ is the regularization parameter that controls the influence of regularization.
The training parameters for each regression model are shown in Table 3.

3.3. Model Performance Comparison

Prediction models for the chlorophyll content of tomato leaves were established using optimal parameters, and the performance of each model was compared, as shown in Figure 7.
Through correlation analysis and the results shown in Figure 5, it can be observed that some color features (such as CIVE, RGBVI, and GLA) have a high correlation with SPAD values. However, the Pearson correlation coefficient does not reach 1, indicating that there may be a complex nonlinear relationship between them. The multivariate linear regression (MLR) model assumes a linear relationship between input variables and the target variable [39], which limits its ability to handle nonlinear features. When there are interactions between color features or higher-order nonlinear dependencies, linear models may fail to capture these characteristics effectively, leading to reduced prediction accuracy. Experimental results also confirm this, as the R2 value of the MLR model is only 0.7359, indicating weaker prediction performance.
In contrast, nonlinear models such as support vector regression (SVR) and random forest (RF) are better at modeling complex feature relationships. SVR maps data to high-dimensional space using kernel functions, which allows it to learn nonlinear features to some extent. However, it may become unstable in cases of high-dimensional data and excessive feature redundancy. On the other hand, random forest (RF) uses an ensemble of decision trees to adaptively learn the nonlinear dependencies between variables, offering strong feature selection and generalization capabilities. However, RF may be affected by factors such as tree depth, data noise, and hyperparameter settings, which can limit the model’s predictive power in certain cases.
Unlike single models, Stacking combines multiple base models (such as MLR, SVR, and RF) to leverage the advantages of each model across different data patterns, allowing the final meta-learner to capture the complex relationships between input features and SPAD values more comprehensively. Specifically, Stacking employs a layered architecture, where the first layer uses multiple base models to extract different levels of feature information, and the second layer meta-learner further learns the prediction patterns of each base model to optimize the final output. This fusion strategy effectively reduces the bias and variance of individual models, enhancing the model’s robustness in handling nonlinearity, feature interactions, and data noise. As a result, the Stacking model achieves a coefficient of determination (R2) of 0.8359, significantly outperforming the single models. It not only demonstrates higher stability and generalization ability but also more effectively integrates the strengths of different algorithms.
While the Stacking model improves prediction performance, it also has some limitations. First, training multiple base models and the meta-learner is time-consuming, especially when dealing with large datasets. Second, the model’s performance is sensitive to the choice of base models: too little diversity among the base models may result in insufficient variation, while too much difference may increase the risk of overfitting. Additionally, hyperparameter tuning is complex, as each model requires independent tuning, which increases the computational cost.
In summary, although the Stacking ensemble method theoretically improves prediction performance through model combination, its applicability should be carefully evaluated in practical applications. Especially when dealing with large datasets, scenarios requiring strict model interpretability, or poor data quality, it is advisable to consider the balance between computational cost, model complexity, and performance improvement, and select an appropriate modeling strategy.

4. Discussion

Although Stacking demonstrates high accuracy in chlorophyll content prediction, several influencing factors require further optimization. First, the shooting angle of the smartphone may affect the extraction of leaf color features, as different angles can lead to variations in light reflection, thereby impacting the acquisition of RGB color features [36]. To minimize the effect of angular deviation, the camera lens should be kept perpendicular to the leaf surface during shooting, and a fixed stand or calibration tool should be used to maintain a consistent shooting angle. Additionally, the choice of shooting time is equally critical, as different time periods (such as morning, noon, and evening) have varying lighting conditions that may cause color feature shifts, affecting model stability [40]. To reduce the impact of lighting variations, images should be captured during stable light periods (e.g., 9:00–11:00 a.m. and 3:00–5:00 p.m.), and color correction techniques should be applied to enhance data consistency.
Although nonlinear models (such as RF and Stacking) have overcome some of the limitations of traditional linear models, the effectiveness of chlorophyll prediction models is still limited by the selection of color features. Currently, color features primarily come from color spaces such as RGB, HSI, and Lab. While these features can partially reflect changes in chlorophyll, their representational ability may not be sufficient to fully capture the complex variations in chlorophyll content. Therefore, introducing deep learning models to extract hidden features, or exploring more complex color spaces (such as HSV or YCbCr), holds promise for further improving the predictive performance of the model. Moreover, environmental factors such as light intensity, humidity, and temperature have not been fully considered in the current model, despite their dynamic variations potentially affecting chlorophyll content measurement accuracy and increasing prediction errors. Therefore, incorporating these environmental variables into the model is crucial for improving prediction accuracy and stability. Future studies could introduce environmental factors as additional input features or conduct correlation analyses during model training to minimize external influences on prediction results. This approach would not only optimize the model’s generalization ability but also enhance its applicability under varying environmental conditions, further strengthening its value in smart agriculture by ensuring precise crop health monitoring and management.
The proposed model provides a rapid and non-destructive chlorophyll sensing method for precision agriculture, enabling farmers to promptly assess the nutritional status of tomato plants and optimize fertilization management, thereby reducing resource waste. Future research can further refine the Stacking model structure to minimize redundant computations and enhance its adaptability to large-scale datasets. Additionally, existing modeling approaches still have room for improvement, such as integrating more color features, incorporating other types of sensor data, or refining algorithms to further enhance prediction accuracy and model applicability. Although this study focused on a single tomato variety at a specific growth stage, variations in chlorophyll content, leaf color characteristics, and spectral reflectance among different tomato varieties or genotypes may impact the model’s predictive performance [41]. Therefore, future research should expand trials to multiple crops, varieties, and growth stages to validate the method’s robustness and applicability. Furthermore, exploring additional image-based texture features and other potential information could improve the model’s predictive capability, enhancing its adaptability and applicability in complex agricultural environments. Future studies will also integrate Internet of Things (IoT) technology to develop a real-time automated chlorophyll sensing system for greenhouse environments. This system will enable real-time data acquisition and visualization at the regional level, providing a solid technical foundation and data support for precision crop management and large-scale deployment.

5. Conclusions

This study investigates a method for predicting chlorophyll content in greenhouse tomato leaves based on RGB color features extracted from smartphone images. The results show that chlorophyll content is significantly correlated with several color features (such as the B value, (2G − R − B)/(2G + R + B), GLA, RGBVI, g, g − b, ExG, and CIVE), demonstrating the feasibility and effectiveness of using smartphone RGB images for rapid and non-destructive chlorophyll estimation. Based on these sensitive features, multiple predictive models were developed, including multiple linear regression, ridge regression, support vector regression, random forest, and a Stacking ensemble model. Among them, the Stacking model performed the best, achieving the highest prediction accuracy (R2 = 0.8359, RMSE = 0.8748), significantly enhancing the estimation capability for chlorophyll content.
Compared to traditional hyperspectral imaging and digital camera technologies, smartphones offer a more practical and scalable solution for chlorophyll detection due to their high portability, low cost, and ease of use. This study not only enables efficient and non-destructive detection of tomato chlorophyll content but also lays a foundation for precise assessment of plant nutritional status and growth dynamics. The proposed method provides strong technical support for AI-driven chlorophyll recognition and precision agriculture management, promoting the advancement of smart farming and showing great potential for application and adoption.

Author Contributions

The authors confirm contribution to the paper as follows: Study conception and design: X.Z., X.M. and J.Y. data collection: X.Z. and H.Y.; Analysis and interpretation of results: X.Z. and X.M.; Draft manuscript preparation: X.Z. and X.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Shandong Province Higher Education Program for the Introduction and Cultivation of Young Innovative Talents (2021).

Data Availability Statement

The data that support the findings of this study are available from the corresponding author, X.M. upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Guo, J.T.; Zhao, Y.; Sun, Y.J. Recent Advances in Research on Flavor Substances in Tomato Fruit and Their Influential Factors. Food Sci. 2023, 44, 169–177. [Google Scholar]
  2. Tieman, D.; Zhu, G.; Resende, M.F., Jr.; Lin, T.; Nguyen, C.; Bies, D.; Rambla, J.L.; Beltran, K.S.O.; Taylor, M.; Zhang, B. A Chemical Genetic Roadmap to Improved Tomato Flavor. Science 2017, 355, 391–394. [Google Scholar] [CrossRef] [PubMed]
  3. Yao, Y.C.; Zhao, Y.; Zhao, S.N.; Zhu, C.H.; Li, H.N. Analysis of Current Situation and Development Trend of Tomato Products. China Fruit Veg. 2021, 41, 144–148. [Google Scholar] [CrossRef]
  4. Xiao, Q.; Tang, W.; Zhang, C.; Zhou, L.; Feng, L.; Shen, J.; Yan, T.; Gao, P.; He, Y.; Wu, N. Spectral Preprocessing Combined with Deep Transfer Learning to Evaluate Chlorophyll Content in Cotton Leaves. Plant Phenomics 2022, 2022, 9813841. [Google Scholar] [CrossRef]
  5. Wang, Z.; Li, Y.; Wu, F.; Zou, X.C. Estimation of Winter Wheat Chlorophyll Relative Content Combing with Canopy Spectrum Red Edge Parameters and Random Forest Machine Learning. Trans. Chin. Soc. Agric. Eng. 2024, 40, 171–182. [Google Scholar] [CrossRef]
  6. Pérez-Patricio, M.; Camas-Anzueto, J.L.; Sanchez-Alegría, A.; Aguilar-González, A.; Gutiérrez-Miceli, F.; Escobar-Gómez, E.; Voisin, Y.; Rios-Rojas, C.; Grajales-Coutiño, R. Optical Method for Estimating the Chlorophyll Contents in Plant Leaves. Sensors 2018, 18, 650. [Google Scholar] [CrossRef]
  7. Zhang, F.; Zhang, Y.K.; Mao, P.J.; Wang, J.; Qiu, Z.M. Status and Development of Measuring Method in Plant Chlorophyll Content. J. Agric. Mech. Res. 2014, 4, 238–241. [Google Scholar] [CrossRef]
  8. Lin, F.F.; Deng, J.S.; Shi, Y.Y.; Chen, L.S.; Wang, K. Investigation of SPAD Meter-Based Indices for Estimating Rice Nitrogen Status. Comput. Electron. Agric. 2010, 71, S60–S65. [Google Scholar] [CrossRef]
  9. Uddling, J.; Gelang-Alfredsson, J.; Piikki, K.; Pleijel, H. Evaluating the Relationship between Leaf Chlorophyll Concentration and SPAD-502 Chlorophyll Meter Readings. Photosynth. Res. 2007, 91, 37–46. [Google Scholar] [CrossRef]
  10. Markwell, J.; Osterman, J.C.; Mitchell, J.L. Calibration of the Minolta SPAD-502 Leaf Chlorophyll Meter. Photosynth. Res. 1995, 46, 467–472. [Google Scholar] [CrossRef]
  11. Tan, L.; Zhou, L.; Zhao, N.; He, Y.; Qiu, Z. Development of a Low-Cost Portable Device for Pixel-Wise Leaf SPAD Estimation and Blade-Level SPAD Distribution Visualization Using Color Sensing. Comput. Electron. Agric. 2021, 190, 106487. [Google Scholar] [CrossRef]
  12. Li, L.J.; Shi, Y.; Liu, Y.C.; Guo, J.X. Research on Predicting Hami Melon Leaf Chlorophyll Based on RGB Image Processing. J. Chin. Agric. Mech. 2024, 45, 149–155. [Google Scholar] [CrossRef]
  13. Zhang, J.; Han, W.; Huang, L.; Zhang, Z.; Ma, Y.; Hu, Y. Leaf Chlorophyll Content Estimation of Winter Wheat Based on Visible and Near-Infrared Sensors. Sensors 2016, 16, 437. [Google Scholar] [CrossRef]
  14. Wang, H.; Zhu, R.; Zhang, J.; Ni, L.; Shen, H.; Xie, P. A Novel and Convenient Method for Early Warning of Algal Cell Density by Chlorophyll Fluorescence Parameters and Its Application in a Highland Lake. Front. Plant Sci. 2018, 9, 869. [Google Scholar] [CrossRef] [PubMed]
  15. Dian, R.; Li, S.; Sun, B.; Guo, A. Recent Advances and New Guidelines on Hyperspectral and Multispectral Image Fusion. Inf. Fusion 2021, 69, 40–51. [Google Scholar] [CrossRef]
  16. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking Hyperspectral Image Classification with Transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5518615. [Google Scholar] [CrossRef]
  17. Lu, B.; Dao, P.D.; Liu, J.; He, Y.; Shang, J. Recent Advances of Hyperspectral Imaging Technology and Applications in Agriculture. Remote Sens. 2020, 12, 2659. [Google Scholar] [CrossRef]
  18. Qiao, L.; Zhang, Z.Y.; Chen, L.S.; Sun, H.; Li, M.Z.; Li, L.; Ma, J.Y. Detection of Chlorophyll Content in Maize Canopy from UAV Imagery. IFAC-Pap. 2019, 52, 330–335. [Google Scholar] [CrossRef]
  19. Yuan, Y.; Wang, X.; Shi, M.; Wang, P. Performance Comparison of RGB and Multispectral Vegetation Indices Based on Machine Learning for Estimating Hopea Hainanensis SPAD Values under Different Shade Conditions. Front. Plant Sci. 2022, 13, 928953. [Google Scholar] [CrossRef]
  20. Liang, Y.; Urano, D.; Liao, K.-L.; Hedrick, T.L.; Gao, Y.; Jones, A.M. A Nondestructive Method to Estimate the Chlorophyll Content of Arabidopsis Seedlings. Plant Methods 2017, 13, 26. [Google Scholar] [CrossRef]
  21. Li, L.; Cheng, L. A Method for Rapid Detection of Chlorophyll Content in Tomato Leaves Based on RGB Model. Mol. Plant 2021, 19, 6906–6909. [Google Scholar] [CrossRef]
  22. Mercado-Luna, A.; Rico-García, E.; Lara-Herrera, A.; Soto-Zarazúa, G.; Ocampo-Velázquez, R.; Guevara-González, R.; Herrera-Ruiz, G.; Torres-Pacheco, I. Nitrogen Determination on Tomato (Lycopersicon esculentum Mill.) Seedlings by Color Image Analysis (RGB). Afr. J. Biotechnol. 2010, 9, 5326–5332. [Google Scholar]
  23. Gitelson, A.A.; Kaufman, Y.J.; Merzlyak, M.N. Use of a Green Channel in Remote Sensing of Global Vegetation from EOS-MODIS. Remote Sens. Environ. 1996, 58, 289–298. [Google Scholar] [CrossRef]
  24. Riccardi, M.; Mele, G.; Pulvento, C.; Lavini, A.; d’Andria, R.; Jacobsen, S.-E. Non-Destructive Evaluation of Chlorophyll Content in Quinoa and Amaranth Leaves by Simple and Multiple Regression Analysis of RGB Image Components. Photosynth. Res. 2014, 120, 263–272. [Google Scholar] [CrossRef]
  25. Hu, H.; Liu, H.Q.; Zhang, H.; Zhu, J.H.; Yao, X.G.; Zhang, X.B.; Zheng, K.F. Assessment of Chlorophyll Content Based on Image Color Analysis, Comparison with SPAD-502. In Proceedings of the 2010 2nd International Conference on Information Engineering and Computer Science, Wuhan, China, 25–26 December 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 1–3. [Google Scholar] [CrossRef]
  26. Sulistyo, S.B.; Woo, W.L.; Dlay, S.S. Regularized Neural Networks Fusion and Genetic Algorithm Based On-Field Nitrogen Status Estimation of Wheat Plants. IEEE Trans. Ind. Inform. 2016, 13, 103–114. [Google Scholar] [CrossRef]
  27. do Amaral, E.S.; Vieira Silva, D.; Dos Anjos, L.; Schilling, A.C.; Dalmolin, Â.C.; Mielke, M.S. Relationships Between Reflectance and Absorbance Chlorophyll Indices with RGB (Red, Green, Blue) Image Components in Seedlings of Tropical Tree Species at Nursery Stage. New For. 2019, 50, 377–388. [Google Scholar] [CrossRef]
  28. Guo, Y.; Yin, G.; Sun, H.; Wang, H.; Chen, S.; Senthilnath, J.; Wang, J.; Fu, Y. Scaling Effects on Chlorophyll Content Estimations with RGB Camera Mounted on a UAV Platform Using Machine-Learning Methods. Sensors 2020, 20, 5130. [Google Scholar] [CrossRef]
  29. Yu, S.; Bu, H.; Hu, X.; Dong, W.; Zhang, L. Establishment and Accuracy Evaluation of Cotton Leaf Chlorophyll Content Prediction Model Combined with Hyperspectral Image and Feature Variable Selection. Agronomy 2023, 13, 2120. [Google Scholar] [CrossRef]
  30. Salgueiro, J.L.; Pérez, L.; Sanchez, Á.; Cancela, Á.; Míguez, C. Microalgal Biomass Quantification from the Non-Invasive Technique of Image Processing through Red–Green–Blue (RGB) Analysis. J. Appl. Phycol. 2022, 34, 871–881. [Google Scholar] [CrossRef]
  31. Naqvi, S.M.Z.A.; Awais, M.; Khan, F.S.; Afzal, U.; Naz, N.; Khan, M.I. Unmanned Air Vehicle Based High Resolution Imagery for Chlorophyll Estimation Using Spectrally Modified Vegetation Indices in Vertical Hierarchy of Citrus Grove. Remote Sens. Appl. Soc. Environ. 2021, 23, 100596. [Google Scholar]
  32. Agarwal, A.; Gupta, S.D. Assessment of Spinach Seedling Health Status and Chlorophyll Content by Multivariate Data Analysis and Multiple Linear Regression of Leaf Image Features. Comput. Electron. Agric. 2018, 152, 281–289. [Google Scholar] [CrossRef]
  33. Zhang, H.C.; Bian, L.M.; Ge, Y.F.; Li, X.P. Estimation and Visualization of Chlorophyll Content in Plant Based on YOLO V5. Trans. Chin. Soc. Agric. Mach. 2021, 53, 313–321. [Google Scholar] [CrossRef]
  34. Montanaro, G.; Petrozza, A.; Rustioni, L.; Cellini, F.; Nuzzo, V. Phenotyping Key Fruit Quality Traits in Olive Using RGB Images and Back Propagation Neural Networks. Plant Phenomics 2023, 5, 0061. [Google Scholar] [CrossRef]
  35. Mohan, P.J.; Gupta, S.D. Intelligent Image Analysis for Retrieval of Leaf Chlorophyll Content of Rice from Digital Images of Smartphone under Natural Light. Photosynthetica 2019, 57, 388–398. [Google Scholar] [CrossRef]
  36. Li, X.H.; Lu, X.J.; Xi, J.Y.; Zhang, Y.H.; Zhang, M.Q. Univeisal Method to Detect the Chlorophyll Content in Plant Leaves with RGB Images Captured by Smart Phones. Trans. Chin. Soc. Agric. Eng. 2021, 37, 145–151. [Google Scholar] [CrossRef]
  37. Jiang, C.; Johkan, M.; Hohjo, M.; Tsukagoshi, S.; Maruo, T. A Correlation Analysis on Chlorophyll Content and SPAD Value in Tomato Leaves. HortResearch 2017, 71, 37–42. [Google Scholar] [CrossRef]
  38. Went, F.W. Plant Growth under Controlled Conditions. II. Thermoperiodicity in Growth and Fruiting of the Tomato. Am. J. Bot. 1944, 31, 135–150. [Google Scholar] [CrossRef]
  39. Abdikan, S.; Sekertekin, A.; Narin, O.G.; Delen, A.; Sanli, F.B. A Comparative Analysis of SLR, MLR, ANN, XGBoost and CNN for Crop Height Estimation of Sunflower Using Sentinel-1 and Sentinel-2. Adv. Space Res. 2023, 71, 3045–3059. [Google Scholar] [CrossRef]
  40. Özreçberoğlu, N.; Kahramanoğlu, İ. Mathematical Models for the Estimation of Leaf Chlorophyll Content Based on RGB Colours of Contact Imaging with Smartphones: A Pomegranate Example. Folia Hortic. 2020, 32, 57–67. [Google Scholar] [CrossRef]
  41. Al-Gaadi, K.A.; Tola, E.; Madugundu, R.; Zeyada, A.M.; Alameen, A.A.; Edrris, M.K.; Edrees, H.F.; Mahjoop, O. Response of Leaf Photosynthesis, Chlorophyll Content and Yield of Hydroponic Tomatoes to Different Water Salinity Levels. PLoS ONE 2024, 19, e0293098. [Google Scholar] [CrossRef]
Figure 1. Captured images of tomato leaves. (a) Original image. (b) Image dataset.
Figure 1. Captured images of tomato leaves. (a) Original image. (b) Image dataset.
Horticulturae 11 00593 g001
Figure 2. Leaf measurement points. The numbers 1–6 in the image represent Sampling Points 1 to 6, respectively.
Figure 2. Leaf measurement points. The numbers 1–6 in the image represent Sampling Points 1 to 6, respectively.
Horticulturae 11 00593 g002
Figure 3. Comparison of image enhancement methods. (a) Color-segmented original image. (b) Image after Gaussian filtering and closing operation.
Figure 3. Comparison of image enhancement methods. (a) Color-segmented original image. (b) Image after Gaussian filtering and closing operation.
Horticulturae 11 00593 g003
Figure 4. Experimental workflow.
Figure 4. Experimental workflow.
Horticulturae 11 00593 g004
Figure 5. Correlation analysis of color features.
Figure 5. Correlation analysis of color features.
Horticulturae 11 00593 g005
Figure 6. Scatter plots between color features and chlorophyll content. The blue solid line in the figure represents the fitted linear regression line, and the blue shaded area indicates the 95% confidence interval of the model. (a) (2G − R − B)/(2G + R + B). (b) GLA. (c) B. (d) RGBVI. (e) g. (f) g − b. (g) ExG. (h) CIVE.
Figure 6. Scatter plots between color features and chlorophyll content. The blue solid line in the figure represents the fitted linear regression line, and the blue shaded area indicates the 95% confidence interval of the model. (a) (2G − R − B)/(2G + R + B). (b) GLA. (c) B. (d) RGBVI. (e) g. (f) g − b. (g) ExG. (h) CIVE.
Horticulturae 11 00593 g006aHorticulturae 11 00593 g006b
Figure 7. Regression scatter plot of predicted vs. measured values. (a) MLR estimation scatter plot. (b) SVR estimation scatter plot. (c) RF estimation scatter plot. (d) RR estimation scatter plot. (e) Stacking estimation scatter plot.
Figure 7. Regression scatter plot of predicted vs. measured values. (a) MLR estimation scatter plot. (b) SVR estimation scatter plot. (c) RF estimation scatter plot. (d) RR estimation scatter plot. (e) Stacking estimation scatter plot.
Horticulturae 11 00593 g007aHorticulturae 11 00593 g007b
Table 1. Selected color features.
Table 1. Selected color features.
Color SpaceNumberColor FeaturesFormula and Meaning
RGB19RR
GG
BB
R/BR/B
G/BG/B
B/RB/R
B/GB/G
R − BR − B
G − BG − B
R − B − GR − B − G
(G + B − R)/(2B)(G + B − R)/(2B)
(G + B − R)/(2G)(G + B − R)/(2G)
(G + B − R)/(2R)(G + B − R)/(2R)
(B − G − R)/(B + G)(B − G − R)/(B + G)
(2G − R − B)/(2G + R + B)(2G − R − B)/(2G + R + B)
(R − G − B)/(R + G)(R − G − B)/(R + G)
(R − G − B)/(G + B)(R − G − B)/(G + B)
(B − G − R)/(R + B)(B − G − R)/(R + B)
(B − G − R)/(G + R)(B − G − R)/(G + R)
Normalized RGB8rR/(R + G + B)
gG/(R + G + B)
bB/(R + G + B)
r/br/b
g/bg/b
r − br − b
g − bg − b
r − gr − g
HSI3HH channel average value,
Represents the type of color (Hue).
SS channel average value,
Represents the purity of the color (Saturation).
II channel average value,
Represents the overall brightness of the color (Intensity).
Lab3LL channel average value,
Represents the brightness (Lightness).
AA channel average value,
Represents the green-red axis (Green-Red).
B_1B channel average value,
Represents the blue-yellow axis (Blue-Yellow).
Others9RGBVI(g2 − r2)/(g2 + r2)
GRVI(g − r)/(g + r)
ExG2g − r − b
GLA(2g − r − b)/(2g + r + b)
ExR1.4r − g
ExGRExG − 1.4r − g
MGRVI(g2 − br)/(g2 + br)
VARI(g − r)/(g + r − b)
CIVE0.441r − 0.881g + 0.3856b + 18.78745
Note: r, g, b: Normalized R, G, and B values, respectively. H, S, I: Hue, saturation, and intensity in the HSI color space. L, A, B_1: Lightness and two opposing color dimensions in the Lab color space. RGBVI: RGB Vegetation Index. GRVI: Green–Red Vegetation Index. ExG: Excess Green Index. GLA: Green Leaf Area Index. ExR: Excess Red Index. ExGR: Excess Green–Red Index. MGRVI: Modified Green–Red Vegetation Index. VARI: Visible Atmospherically Resistant Index. CIVE: Color Index of Vegetation.
Table 2. Correlation analysis between color features and SPAD values.
Table 2. Correlation analysis between color features and SPAD values.
Color
Features
Correlation
Coefficient
Color
Features
Correlation
Coefficient
Color
Features
Correlation
Coefficient
R0.27R − B−0.43CIVE0.75
G0.14G − B−0.51RGBVI−0.75
B0.82R − B − G−0.68GRVI−0.39
r−0.35(B − G − R)/(B + G)0.62ExG−0.75
g−0.75(2G − R − B)/(2G + R + B)−0.75GLA−0.75
b0.66(R − G − B)/(R + G)−0.44ExR0.32
g/b−0.71(R − G − B)/(G + B)−0.35ExGR−0.62
r/b−0.61(B − G − R)/(R + B)0.70MGRVI−0.39
R/B−0.61(B − G − R)/(G + R)0.66VARI−0.18
G/B−0.71(G + B − R)/(2B)−0.69L0.18
B/R0.59(G + B − R)/(2G)0.53A0.50
B/G0.70(G + B − R)/(2R)0.34B_1−0.53
r − g0.49g − b−0.71S−0.70
r − b−0.59H0.18I0.14
Table 3. Model Parameter Table.
Table 3. Model Parameter Table.
ModelParametersMeaningSet Values
CPenalty parameter1
SVRKernel
gamma
kernel function
kernel parameter
linear
scale
RFn_estimators
max_depth
min_samples_split
min_samples_leaf
number of decision trees
maximum depth of the tree
minimum samples required to split a leaf node
minimum samples per leaf
80
3
2
8
RRλcontrol the strength of regularization1
Note: SVR, support vector regression. RF, random forest. RR, ridge regression.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Yu, H.; Yan, J.; Meng, X. Study on the Detection of Chlorophyll Content in Tomato Leaves Based on RGB Images. Horticulturae 2025, 11, 593. https://doi.org/10.3390/horticulturae11060593

AMA Style

Zhang X, Yu H, Yan J, Meng X. Study on the Detection of Chlorophyll Content in Tomato Leaves Based on RGB Images. Horticulturae. 2025; 11(6):593. https://doi.org/10.3390/horticulturae11060593

Chicago/Turabian Style

Zhang, Xuehui, Huijiao Yu, Jun Yan, and Xianyong Meng. 2025. "Study on the Detection of Chlorophyll Content in Tomato Leaves Based on RGB Images" Horticulturae 11, no. 6: 593. https://doi.org/10.3390/horticulturae11060593

APA Style

Zhang, X., Yu, H., Yan, J., & Meng, X. (2025). Study on the Detection of Chlorophyll Content in Tomato Leaves Based on RGB Images. Horticulturae, 11(6), 593. https://doi.org/10.3390/horticulturae11060593

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop