Next Article in Journal
Comparative Life Cycle Assessment of Animal Feed Formulations Containing Conventional and Insect-Based Protein Sources
Previous Article in Journal
Impact of Drying Conditions on Soybean Quality: Mathematical Model Evaluation
Previous Article in Special Issue
Long-Term Blueberry Storage by Ozonation or UV Irradiation Using Excimer Lamp
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spectral-Spatial Fusion for Soybean Quality Evaluation Using Hyperspectral Imaging

1
Agricultural Engineering Technology, School of Agriculture, Tennessee Technological University, Cookeville, TN 38505, USA
2
School of Environmental Studies, Tennessee Technological University, Cookeville, TN 38505, USA
*
Author to whom correspondence should be addressed.
AgriEngineering 2025, 7(9), 274; https://doi.org/10.3390/agriengineering7090274
Submission received: 4 July 2025 / Revised: 3 August 2025 / Accepted: 18 August 2025 / Published: 25 August 2025
(This article belongs to the Special Issue Latest Research on Post-Harvest Technology to Reduce Food Loss)

Abstract

Accurate postharvest quality evaluation of soybeans is essential for preserving product value and meeting industry standards. Traditional inspection methods are often inconsistent, labor-intensive, and unsuitable for high-throughput operations. This study presents a non-destructive soybean classification approach using a simplified reflectance-mode hyperspectral imaging system equipped with a single light source, eliminating the complexity and maintenance demands of dual-light configurations used in prior studies. A spectral–spatial data fusion strategy was developed to classify harvested soybeans into four categories: normal, split, diseased, and foreign materials such as stems and pods. The dataset consisted of 1140 soybean samples distributed across these four categories, with spectral reflectance features and spatial texture attributes extracted from each sample. These features were combined to form a unified feature representation for use in classification. Among multiple machine learning classifiers evaluated, Linear Discriminant Analysis (LDA) achieved the highest performance, with approximately 99% accuracy, 99.05% precision, 99.03% recall and 99.03% F1-score. When evaluated independently, spectral features alone resulted in 98.93% accuracy, while spatial features achieved 78.81%, highlighting the benefit of the fusion strategy. Overall, this study demonstrates that a single-illumination HSI system, combined with spectral–spatial fusion and machine learning, offers a practical and potentially scalable approach for non-destructive soybean quality evaluation, with applicability in automated industrial processing environments.

1. Introduction

Soybean (Glycine max) is a globally important crop, extensively used for food, feed, and industrial purposes due to its high protein and oil content [1]. In 2024, the United States produced approximately 4.37 billion bushels of soybeans [2], highlighting the significance of efficient postharvest quality control to maintain product value and marketability. However, mechanical harvesting can introduce defects such as split, cracked, or diseased seeds and contamination by foreign materials like stems and pods. These impair product quality and demand precise postharvest inspection systems.
Traditional quality evaluation methods, such as manual sorting or chemical analysis, are destructive, labor-intensive, and often inconsistent. They are also inadequate for large-scale industrial operations that require fast, objective, and non-destructive solutions [3,4]. Recent advances in machine vision systems have facilitated automated classification using hyperspectral imaging (HSI) for various agricultural products, including soybean [4], and maize [5], establishing HSI as a powerful technology for non-destructive quality assessment. HSI captures both spectral information, the unique light reflectance profile at different wavelengths, and spatial information, such as shape, surface texture, and structural defects [6]. While spectral data is valuable for detecting internal or chemical changes, spatial features help identify external damage or contamination [7,8]. However, using either modality alone can limit classification accuracy, especially when different objects exhibit overlapping spectral or visual characteristics. To overcome these limitations, spectral-spatial fusion combines both feature types, enabling classifiers to exploit the complementary strengths of each domain. This integrated approach improves class separability, enhances robustness, and reduces errors in distinguishing visually similar categories like normal vs. split soybeans.
Complementing these imaging advances, machine learning techniques have emerged as powerful tools for interpreting high-dimensional HSI data. Algorithms such as Support Vector Machines (SVM), Random Forest (RF), Gradient Boosting, and deep learning models like Convolutional Neural Networks (CNNs) have been used to classify soybean seed defects, estimate nutrient content, and predict yield with high accuracy [9,10,11,12,13]. For example, a hybrid SVM-CNN model achieved 95.89% F1-score in classifying soybean defects using the SOYPR dataset [9], while deep learning frameworks such as SNet, SSDINet, and MobileNetV2-based systems reached accuracies exceeding 96% across multiple defect categories [11,14,15,16]. Gulzar et al. [17,18] applied modified Inception and MobileNetV2 architectures for seed classification using transfer learning and feature augmentation. However, many of these studies are limited in scope, often addressing only specific defect types and struggling to differentiate visually similar classes such as normal and split soybeans. Other approaches relying on morphological or texture-based features may suffer reduced performance in the presence of noise, overlapping objects, or high spectral redundancy [11,12]. Moreover, models trained solely on spectral or spatial features often overfit and fail to generalize across diverse environmental or postharvest conditions [4,19].
Recent studies have also explored integrating spectral and spatial data from hyperspectral images for soybean classification. For example, Tan et al. [20] combined texture features and spectral information in a Takagi-Sugeno fuzzy neural network, achieving 92% accuracy in classifying soybean varieties. Similarly, an ensemble learning model achieved 99.8% accuracy for classifying Proso Millet seed varieties using hyperspectral data [21]. However, many of these studies are limited to controlled experimental conditions or a narrow range of seed types, which can affect model stability and generalizability when applied to broader field conditions.
In previous work, Momin et al. [22] developed a proof-of-concept machine vision system using basic image processing techniques to evaluate soybean quality by detecting materials other than grain (MOGs). The system utilized both front-lit and back-lit imaging and achieved high detection rates, 98% for defective beans and stem/pods, 96% for split beans, and 75% for contaminated beans. However, the reliance on a glass plate for backlighting introduced mechanical challenges. Accumulation of dust and moisture on the glass surface interfered with light transmission, hindering accurate classification of split or broken soybeans. These maintenance and reliability issues limited the system’s scalability and practicality for field deployment or continuous operation.
To address these challenges, particularly the need for simplified lighting in postharvest image acquisition, the present study proposes a non-destructive soybean classification framework using a reflectance-mode HSI system with a single light source. This simplification was made possible by leveraging the capabilities of a hyperspectral camera, which captures rich spectral information across the visible and near-infrared regions using only one illumination source. Each soybean category exhibits distinct spectral signatures, enabling reliable classification without the need for multiple lighting angles. Moreover, spectral reflectance data are directly correlated with the chemical and physical properties of biological materials. Since both soybean and non-soybean objects in our dataset are plant-based, meaningful spectral differences were observed, supporting robust classification. In contrast to earlier approaches that required dual lighting for conventional image processing, the current study integrates spectral and spatial information to classify harvested soybeans using machine learning. This eliminates the need for complex lighting hardware and minimizes mechanical maintenance.
The objectives of this study are to (1) acquire and analyze spectral reflectance data of harvested soybeans across visible and near-infrared wavelengths to characterize different object categories, (2) extract spatial texture features using the Gray Level Co-occurrence Matrix (GLCM) to capture surface patterns associated with different soybean attributes, and (3) implement a spectral–spatial fusion strategy and evaluate its classification performance using various machine learning models, comparing it against spatial-only and spectral-only classifiers. Overall, this study proposes a simplified and potentially scalable imaging solution that reduces mechanical complexity while enabling accurate, non-destructive soybean quality evaluation in postharvest processing environments. The main contributions of this study are as follows:
  • A reflectance-mode HSI system with a single light source is introduced, eliminating the need for complex dual-light configurations typically used in earlier soybean classification systems.
  • A spectral–spatial data fusion framework is proposed that improves classification robustness and accuracy compared to models using only spectral or spatial features.
  • A broad evaluation of eight machine learning classifiers shows that the proposed approach achieves near state-of-the-art classification performance across four soybean categories using a simplified and potentially scalable setup suitable for integration into automated industrial environments.

2. Materials and Methods

2.1. Dataset Preparation of Soybean

Soybean samples were collected during the 2024 harvesting season from an experimental farm located in Cookeville, Tennessee, USA. The soybean cultivar used in this study was Asgrow AG45XF0, a widely cultivated variety in the southeastern United States. Harvesting was conducted using a John Deere S690 combine (John Deere, Moline, IL, USA). Due to the mechanical nature of the harvesting process, the postharvest soybeans often exhibited physical damage and were susceptible to contamination by foreign materials. After collection, all samples were stored at room temperature (20–25 °C) for subsequent analysis.
For classification purposes, the samples were divided into four categories: normal soybeans, split or broken soybeans, diseased soybeans, and foreign materials. Normal soybeans were defined as intact seeds without visible damage, while split or broken soybeans showed cracks, damaged seed coats, or partial crushing. Diseased soybeans were identified based on signs of contamination with soil, fungi, mold, or vegetative matter, indicating physiological degradation. The final category included non-soybean materials, also referred to as foreign materials, consisting of stems, pods, and other plant debris commonly mixed during mechanical harvesting. Representative samples from each category are illustrated in Figure 1. A total of 1140 samples were collected and initially labeled according to the classification criteria defined by Momin et al. [22] The categorization was then verified by an experienced local soybean grower with expertise in postharvest quality traits. Following verification, the samples were sealed in numbered airtight bags for laboratory analysis. The complete dataset was divided into training and testing subsets for model development and validation, as summarized in Table 1. Minor variations in the train-test ratio across object types (e.g., 230/53 vs. 214/53) resulted from practical constraints during sample acquisition rather than deliberate imbalance. Despite these slight variations, the dataset maintains a well-balanced distribution across categories, ensuring robust model training and fair evaluation without introducing significant bias.

2.2. Hyperspectral Image Acquisition

Hyperspectral images were acquired using a push-broom, reflectance-based visible–NIR hyperspectral imaging system (PIKA L-675, Resonon, Bozeman, MT, USA), as illustrated in Figure 2. The system comprised a C-mount objective lens (CITRINE 1.4/23 mm, Schneider Optics, Bad Kreuznach, Germany), a hyperspectral camera, and a motorized translation stage for sample movement. Similar camera systems have been widely used in recent years in non-destructive evaluation of biological and agricultural samples [23,24,25]. Spectral data were captured across a wavelength range of 379.16 to 1020.87 nm, with a spectral interval of 2.02 nm, resulting in 300 bands covering both visible and near-infrared regions. A halogen fiber line light, equipped with a polarizing film, was used for illumination to minimize surface glare. The light source was preheated for 20 min to ensure stable output before image acquisition.
The entire system was enclosed to eliminate ambient light interference, and image acquisition was controlled using Spectronon software (version 3.5.5). An exposure time of 121.129 ms was used to optimize illumination without overexposure. Each sample was scanned at a speed of 3.45 mm/s, producing a hyperspectral data cube with dimensions of 700 lines × 900 spatial pixels per line × 300 spectral bands. The spectral axis recorded wavelength-dependent reflectance values, while the two spatial axes captured the physical structure and arrangement of the soybean samples.

2.3. Proposed Method

The proposed classification framework integrates both spectral and spatial features extracted from hyperspectral imagery, as shown in Figure 3. The process begins with data acquisition and continues with preprocessing steps, including single-band extraction, Gaussian smoothing, intensity thresholding, and morphological erosion. These steps were applied to remove background noise and segment individual soybean objects from the scene. Once the objects were isolated, features were extracted from both spectral and spatial domains. Spectral features consisted of the mean reflectance values across 300 wavelengths within the hyperspectral cube, which were preprocessed (as described in Section 2.4.2) to reduce spectral noise and enhance signal consistency. In parallel, spatial features were computed using the Gray Level Co-occurrence Matrix (GLCM) method, which captures texture characteristics from grayscale images reconstructed from the hyperspectral data. After feature extraction, a feature-level fusion step was performed to combine spectral and spatial features for classification. The resulting fusion vector Ffusion was constructed as follows:
Ffusion = [Fspectral ∥ Fspatial]
where Fspectral is the 300-dimensional preprocessed spectral feature vector, Fspatial is the 14-dimensional spatial texture feature vector, and ∥ denotes vector concatenation. This fusion strategy was not a simple raw concatenation but involved a deliberate integration of noise-reduced spectral data with structured spatial texture features to ensure complementary feature interactions and improve classification robustness. This enriched dataset was used to train classification models. To evaluate model performance, a k-fold cross-validation approach was employed, and metrics such as accuracy, precision, and F1-score were calculated to assess classification reliability and generalizability.

2.4. Data Processing

2.4.1. Spatial Data Processing

Spatial segmentation began with the extraction of a single NIR band at 1000.44 nm, which provided optimal contrast between soybean objects and the background, as shown in Figure 4a,b. This wavelength was selected based on its high reflectance in the dataset, which provided strong object–background contrast and enabled accurate and efficient segmentation. The processing pipeline included Gaussian blurring to reduce high-frequency noise while preserving edge detail, as shown in Figure 4c. Binary thresholding was then applied to generate a mask distinguishing foreground objects from the background, as visualized in Figure 4d. To improve object coherence, morphological erosion was subsequently used to refine object boundaries, followed by an area-based filtering step to remove noise-like elements smaller than 500 pixels, as illustrated in Figure 4e as the final object mask. To support spatial texture analysis, RGB images were reconstructed from selected spectral bands corresponding to red (640.2 nm), green (549.6 nm), and blue (460.6 nm) wavelengths. These images were then converted to grayscale for feature extraction using texture-based methods. The complete preprocessing sequence facilitated reliable object segmentation, enabling robust extraction of both spatial and spectral features.

2.4.2. Spectral Data Processing

Hyperspectral data, comprising up to 300 wavelengths, often contain noise introduced by environmental factors and instrumentation errors [26,27]. This noise can obscure true spectral characteristics and reduce the effectiveness of machine learning models. To mitigate these effects and enhance data quality, spectral preprocessing was performed prior to classification. Three commonly used techniques were evaluated in this study: Multiplicative Scatter Correction (MSC) [28], Standard Normal Variate (SNV) [29], and Savitzky–Golay (SG) smoothing [30]. While MSC and SNV are widely applied to correct scattering effects caused by variations in sample geometry, surface texture, and illumination conditions [31,32], SG smoothing demonstrated the most consistent improvements in classification performance and was therefore selected for further use. All spectral reflectance values were normalized to a range of 0 to 1 to standardize comparisons across samples. SG smoothing was then applied using a window length of 7 and a third-order polynomial, with a derivative order of zero to ensure smoothing without differentiation. This approach effectively reduced high-frequency noise while preserving key spectral features such as peak intensity and curve shape.

2.5. Feature Extraction

2.5.1. Spatial Feature Extraction

GLCM is a statistical method commonly used in image processing to quantify textural patterns by analyzing the spatial relationships between pixel intensities within an image. In this study, RGB images reconstructed from selected hyperspectral bands were converted to grayscale, and GLCM features were extracted to describe the texture characteristics of segmented soybean objects. A total of 14 spatial features were computed based on the method proposed by Haralick et al. [33], including angular second moment, contrast, correlation, entropy, inverse difference moment, difference entropy, difference variance, sum average, sum entropy, sum variance, sum of squares: variance, maximal correlation coefficient, and two measures of information correlation. These features capture various aspects of texture, such as uniformity, intensity variation, randomness, and spatial dependence.
Table 2 presents the GLCM feature values for representative samples from each soybean category. The results show that normal and split soybeans exhibited similar values for features such as angular second moment, contrast, and inverse difference moment, reflecting their visual similarity under standard illumination. In contrast, diseased soybeans and foreign materials showed higher entropy and difference entropy values, indicating more irregular and unpredictable texture patterns. Correlation, which measures the linear relationship between neighboring pixels, was slightly higher for normal and split soybeans compared to diseased and foreign materials. The information measures of correlation also demonstrated strong texture dependency across all categories. Notably, split soybeans showed higher difference variance due to the inclusion of samples in multiple orientations, which introduced variability in local intensity differences. Features such as sum entropy, sum of squares: variance, and sum variance were lower for split soybeans, likely due to reduced surface randomness resulting from compromised seed coats. Foreign materials, which typically have less reflective surfaces, exhibited lower values in brightness-related features such as sum average and maximal correlation coefficient.

2.5.2. Spectral Feature Extraction

To enable effective classification based on spectral features, it is essential to identify distinct spectral signatures and characteristic reflectance patterns across different soybean categories. Figure 5 illustrates representative mean spectral reflectance curves for normal soybeans, diseased soybeans, split soybeans, and foreign materials, across the 379.16 to 1020.87 nm wavelength range. Since these categories exhibit different reflectance behaviors at various wavelengths, all 300 spectral bands were retained to preserve the complete spectral profile. Reducing the number of bands could risk omitting subtle but class-relevant variations, particularly for visually or chemically similar categories.
The spectra revealed that normal and split soybeans shared similar overall curve shapes, characterized by reflectance features in the visible region and increased reflectance in the NIR region. Despite this similarity, noticeable differences were observed in their reflectance intensities. Split soybeans generally exhibited reduced reflectance beyond 700 nm, likely due to compromised seed coats exposing internal structures and altering light scattering and absorption behavior. In contrast, normal soybeans, with intact surfaces, reflected more light in the NIR range, yielding higher intensity values, typically between 0 and 20,000. Split soybeans had different orientation such as, the broken soybean is directly exposed under light or the soybean coat which was compromised is directly under light. However, regardless of orientation, split soybean demonstrated lower intensities, generally ranging from 0 to 15,000. These intensity differences, while preserving the general spectral signature, offer a valuable discriminative feature for classification.
Diseased soybeans displayed significantly altered spectral characteristics. Their mean reflectance curves showed elevated intensities across most wavelengths, particularly in the visible and red-edge regions. These increases suggest structural degradation, chlorophyll loss, and physiological stress due to fungal or microbial infection [34]. The resulting spectral distortion reflects the disruption of normal absorption and scattering mechanisms associated with pathological tissue conditions.
Likewise, foreign materials exhibited spectral signatures that were clearly distinguishable from all soybean categories. Their unique absorption and reflectance profiles stem from differences in surface morphology and internal composition compared to soybean. This distinct spectral behavior provides a reliable foundation for classification models to accurately differentiate between soybean types and effectively exclude foreign materials like contaminants. Collectively, these spectral distinctions reinforce the value of HSI as a non-destructive and efficient tool for postharvest soybean quality assessment.

2.6. Machine Learning Algorithms

To evaluate the classification performance of the proposed spectral–spatial fusion approach, eight widely used machine learning classifiers were employed. These models were selected based on their strong performance in previous soybean-related studies and their ability to capture diverse data characteristics. The selection includes both linear models and ensemble-based methods, chosen to ensure a comprehensive assessment across varying model complexities and assumptions.
The classifiers used in this study include Linear Discriminant Analysis (LDA) [19], Quadratic Discriminant Analysis (QDA) [35], Logistic Regression (LR) [36], K-Nearest Neighbors (KNN) [37], Random Forest (RF), Extra Trees (ET), Gradient Boosting (GB), and Light Gradient Boosting Machine (LightGBM). Among these, ensemble methods such as RF, ET, GB, and LightGBM were included due to their ability to model non-linear relationships and reduce overfitting. For example, RF was utilized by Chen et al. [4] for impurity detection in machine harvested soybean; ET was applied by Dhal et al. [38] to predict water uptake by hydroponically grown soybeans; GB was used by Khalili et al. [39] to classify soybean charcoal rot disease; and LightGBM employed by Li et al. [40] to estimate soybean yield.
LDA was particularly effective in our experiments due to its strength in maximizing class separability in high-dimensional, linearly separable datasets, characteristic of hyperspectral data. KNN and LR were included as baseline models for comparative evaluation. Some classifiers performed better with high-dimensional spectral data (e.g., LDA), while others handled lower-dimensional spatial features more effectively. This diverse selection allowed us to thoroughly assess the impact of spectral, spatial, and fused features across a broad range of learning paradigms.
All models were implemented in Python 3.10 using the PyCaret library. Training and evaluation were conducted on a workstation equipped with an Intel Core i7-14700 processor and 64 GB of RAM. No GPU acceleration was used. Traditional machine learning models used in this study do not follow iterative, epoch-based training processes typical of deep learning frameworks and therefore do not produce training or validation loss curves. To ensure robust evaluation and mitigate overfitting, we employed 10-fold stratified cross-validation and reported standard performance metrics.

2.7. Evaluation Metrics

To comprehensively evaluate the performance of the classification models, several standard metrics were employed, including accuracy, precision, recall, F1-score, the confusion matrix, and the receiver operating characteristic (ROC) curve. These metrics were selected to assess both the overall reliability of the models and their ability to differentiate between high-quality soybeans and defective or foreign materials.
Accuracy reflects the overall proportion of correctly classified instances among all predictions and serves as a broad indicator of model effectiveness. In the context of postharvest quality assessment, it is particularly important for ensuring that normal soybeans are not misclassified or discarded. However, relying solely on accuracy can be misleading in imbalanced datasets. Therefore, class-specific metrics such as precision and recall are also critical. Precision quantifies the proportion of true positive predictions among all samples predicted as positive, offering insight into the model’s ability to avoid false positives, such as misclassifying defective soybeans as normal. Recall, or sensitivity, measures the proportion of actual positive cases correctly identified by the model, which is essential for detecting all instances of defective or diseased soybeans that should be removed from the processing stream.
The F1-score, defined as the harmonic mean of precision and recall, provides a balanced measure that accounts for both types of misclassifications. It is particularly useful when there is a trade-off between false positives and false negatives, as in the case of distinguishing between similar-looking categories such as normal and split soybeans. Additionally, the confusion matrix was used to provide a detailed breakdown of classification results, highlighting true positives and false negatives across all categories. High values along the diagonal indicate strong model performance, while off-diagonal elements reveal areas where misclassifications are occurring.
Finally, the ROC curve was used to evaluate the model’s diagnostic ability by plotting the true positive rate (TPR) against the false positive rate (FPR) across various threshold settings. The area under the curve (AUC) serves as a scalar summary of the model’s class separability, with values approaching 1.0 indicating excellent classification performance. Together, these metrics offer a robust framework for assessing model performance and ensuring its suitability for real-world, non-destructive soybean quality evaluation.

3. Results and Discussion

3.1. Model Performance

This section presents the classification performance of the top five machine learning models introduced in the Section 2.6, evaluated across three types of input features: spatial features derived from GLCM, spectral reflectance features obtained from hyperspectral imaging, and a combination of both through data fusion. Performance metrics including accuracy, precision, recall, and F1-score were reported using 10-fold stratified cross-validation. This approach helped ensure balanced evaluation, address class distribution differences within the dataset and identify the most effective feature representation and modeling approach for soybean quality classification.

3.1.1. Using Spatial Features

The performance of classifiers trained using only spatial texture features derived from GLCM is presented in Table 3. Among all models, the GB classifier achieved the highest overall performance, with an accuracy of 0.8811, precision of 0.8847, recall of 0.8811, and an F1-score of 0.8802. GB outperformed other models such as LightGBM, ET, RF, LDA, and QDA. The superior performance of GB can be attributed to its ability to capture subtle non-linear interactions within the compact set of 14 texture-based spatial features. It effectively boosted weak learners to differentiate among features such as angular second moment, entropy, contrast, and inverse difference moment, which were critical for class separation.

3.1.2. Using Spectral Features

For hyperspectral input, we used the mean reflectance value across 300 wavelengths (379.16 nm–1020.87 nm) for each object. This approach significantly reduced data volume while preserving key spectral characteristics. Various preprocessing techniques were applied to the spectral data, including MSC, SNV, and SG smoothing. As shown in Figure 6, the LDA classifier resulted the highest accuracy (0.9893) when SG smoothing was applied. SG smoothing was most effective at reducing high-frequency noise while preserving critical features necessary for class separation. In contrast, MSC and SNV, although useful for correcting scatter and baseline shifts, sometimes distorted subtle spectral differences between categories.
Based on this result, SG-preprocessed spectral data were used for further analysis. Classifier performance using SG-smoothed features is summarized in Table 4. LDA again outperformed all other models, including LR, QDA, ET, KNN, and LightGBM, likely due to its suitability for high-dimensional data and its ability to project features in a maximally separable space.

3.1.3. Data Fusion

To assess the benefit of combining spatial and spectral features, we fused 14 GLCM-based spatial features with 300 SG-preprocessed spectral features. As shown in Figure 7, LDA achieved the best performance, with all four metrics (accuracy, precision, recall, F1-score) reaching approximately 0.99. Other classifiers, including RF, ET, GB, and LightGBM, yielded strong results in the 0.94–0.96 range but fell short of LDA’s consistency and precision. The effectiveness of LDA under data fusion is attributed to its capacity to leverage complementary information from both spectral and spatial domains, especially in high-dimensional but linearly separable spaces.
Although the improvement in classification accuracy over the spectral-only model is less than 1% and not statistically significant, the fusion-based model offers practical advantages in prediction robustness and generalizability. While the spectral-only model benefits from its use of 300 wavelength-specific features, it lacks spatial and textural context, which may limit its capacity to capture structural object characteristics. In contrast, the spatial-only model demonstrated notably lower performance. The fusion-based approach effectively combines both feature types, resulting in more reliable and interpretable predictions.

3.1.4. Model Performance Comparison

Figure 8 presents confusion matrices comparing the top-performing classifiers across feature types. The GB classifier, trained with only spatial features (Figure 8a), shows noticeable misclassifications, particularly between normal and split soybeans, highlighting the limitations of texture-only data. In contrast, LDA classifiers using only spectral features (Figure 8b) and the combined spectral–spatial feature set (Figure 8c) demonstrated excellent classification accuracy across all categories.
The fusion-based model misclassified two samples, reinforcing that combining spectral reflectance and texture features improves class separability and reduces error rates. In one case, the model likely misinterpreted the hilum region of a normal soybean as a darkened spot, resulting in its misclassification as a diseased soybean. In the other case, the irregular shape and size of a split soybean closely resembled foreign materials, contributing to the incorrect prediction. These instances highlight the inherent challenge of classifying visually ambiguous samples where feature overlap may occur despite spectral–spatial fusion.
Additionally, the ROC curves presented in Figure 9 further validate the advantage of integrating spectral and spatial features for soybean classification. The LDA models (Figure 9b,c) achieved outstanding area under the curve (AUC) scores of 0.99 to 1.00 across all categories, indicating near-perfect sensitivity and specificity. In contrast, the GB model (Figure 9a), which relied solely on spatial features, showed slightly lower AUC values, particularly for normal soybeans (0.94) and split soybeans (0.93), suggesting greater uncertainty in distinguishing these classes. The enhanced performance of the LDA classifiers highlights the effectiveness of combining hyperspectral reflectance data with spatial texture patterns. This superior result is likely attributed to LDA’s strength in maximizing class separability within high-dimensional, linearly separable feature spaces created through data fusion.
To contextualize the performance of our approach, we compared it with previous deep learning-based studies, as summarized in Table 5. Kaler et al. [41] used a CNN to classify two soybean seed categories, achieving 97.72% accuracy. Zheng et al. [42] and Zhang et al. [43] applied ShuffleNet and DenseNet, respectively, to classify four categories with accuracies of 98.36% and 98.48%. Gulzar et al. [18] used an InceptionV3 architecture for five classes, achieving 98.73% accuracy. In contrast, our proposed method employed a data-fusion model that integrates spectral data with spatial attributes to classify four categories of harvested soybean objects, achieving an accuracy of 99.03%. This result is comparable to or better than prior studies, demonstrating the effectiveness of the proposed approach in postharvest soybean quality evaluation.

3.2. Visualization of Model Prediction

To better illustrate the practical implications of the classification models, this section presents visual examples of prediction outcomes on previously unseen hold-out samples. Three scenarios are compared: models trained using only spatial features, only spectral features, and a combination of both through data fusion. These visualizations demonstrate the effectiveness and limitations of each feature set in accurately identifying soybean categories.

3.2.1. Inference Using Spatial Features

Figure 10 shows the prediction results from the GB model trained solely on spatial texture features. The model effectively captured the general structure and spatial positioning of the objects, with the output closely matching the ground truth, except for two misclassified instances. This result highlights the ability of spatial features to represent geometric and textural variations. However, it also reveals their limitations in capturing subtle spectral differences, particularly when objects have similar surface textures but differ in internal composition.

3.2.2. Inference Using Spectral Features

Figure 11 depicts the model output from the LDA classifier trained using only spectral features. The model successfully identified most objects with high fidelity to the ground truth. Spectral features, which represent reflectance patterns across wavelengths, proved effective in differentiating soybean categories. However, one misclassification was observed, suggesting that spectral similarity between certain classes, such as normal and split soybeans, can lead to ambiguity when reflectance characteristics overlap. This inference, performed on a completely independent hold-out set, affirms the robustness and generalizability of the spectral feature-based model.

3.2.3. Inference Using Data Fusion

Figure 12 illustrates the prediction from the LDA classifier trained on the fused dataset comprising both spectral and spatial features. The model produced a highly accurate classification, closely aligned with the ground truth. The fused model effectively captured both spectral reflectance patterns and spatial structural features, leading to more reliable and consistent classification across all object categories. These results confirm the value of data fusion in enhancing classification accuracy and generalization, even when evaluated on unseen samples.

3.3. Limitations

This study has several limitations that warrant consideration. First, classification was performed on isolated soybean objects obtained through controlled image segmentation. While this approach ensured consistency and precise feature extraction, it did not incorporate scenarios involving occlusion, such as overlapping or partially visible beans, conditions common in real-world processing environments. The choice to focus on isolated objects was intentional, allowing for a foundational analysis of spectral–spatial feature fusion without the confounding effects of visual obstruction. Addressing occlusion is a meaningful next step; however, it introduces challenges in segmentation accuracy, object boundary preservation, and feature distortion. Second, lighting variability was minimized by using a halogen fiber line light combined with a polarizing film (Section 2.2), effectively reducing glare and halation. However, performance under variable lighting conditions, typical of industrial settings, remains to be evaluated. Third, we did not perform explicit runtime benchmarking, though hardware specifications were provided (Section 2.6). The system was intentionally designed to reduce computational overhead by avoiding time-intensive operations like pixel-wise segmentation. Lastly, the study focused on a single soybean cultivar (Asgrow AG45XF0), limiting the genetic and phenotypic diversity of the dataset.

4. Conclusions

This study presents a non-destructive approach for postharvest soybean classification using hyperspectral imaging and spectral–spatial data fusion. Analysis across a broad spectral range (379.16–1020.87 nm) revealed distinct reflectance patterns for normal, split, diseased, and foreign materials. The proposed methodology employed a push-broom, reflectance-based hyperspectral imaging system with single-source illumination, which streamlined data acquisition while avoiding the mechanical complexity and maintenance issues of dual-lighting setups. This simplified configuration also minimized the need for extensive image preprocessing, enhancing the system’s practicality for real-world postharvest applications. By integrating 300 spectral features with 14 spatial texture attributes extracted via GLCM, classification performance was significantly enhanced. Among tested classifiers, LDA achieved the highest accuracy, reaching approximately 99%. These findings demonstrate the potential of combining spectral and spatial information to support reliable, scalable, and practical soybean quality assessment in postharvest environments. Future work will focus on extending the system to include realistic field conditions, such as overlapping samples, and exploring advanced segmentation techniques or deep learning-based occlusion-handling methods to enhance model generalizability and robustness. Multiple soybean cultivars will also be incorporated to increase sample diversity and improve applicability across a wider range of postharvest scenarios.

Author Contributions

Conceptualization, A.M. and M.B.R.; methodology, M.B.R., A.T. and A.M.; software, M.B.R. and A.T.; validation, M.B.R., A.T. and A.M.; formal analysis, M.B.R. and A.T.; resources, A.M.; data curation, M.B.R. and A.T.; writing—original draft preparation, M.B.R.; writing—review and editing, A.T. and A.M.; visualization, M.B.R.; supervision and funding acquisition, A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a start-up package provided by Tennessee Technological University, which funded the acquisition of the hyperspectral imaging system used in this study.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data will be made available upon reasonable request.

Acknowledgments

The authors would like to thank the School of Agriculture at Tennessee Technological University for providing laboratory facilities and technical support. Special thanks are extended to Chilcutt Rusty, a local producer, for providing the soybean samples used in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lin, P.; Xiaoli, L.; Li, D.; Jiang, S.; Zou, Z.; Lu, Q.; Chen, Y. Rapidly and Exactly Determining Postharvest Dry Soybean Seed Quality Based on Machine Vision Technology. Sci. Rep. 2019, 9, 17143. [Google Scholar] [CrossRef] [PubMed]
  2. USDA. USDA: 2024 Soybean Production up 5%. Biomass Magazine. 14 January 2025. Available online: https://biomassmagazine.com/articles/usda-2024-soybean-production-up-5 (accessed on 20 June 2025).
  3. Tan, K.; Liu, Q.; Chen, X.; Xia, H.; Yao, S. Estimation of Soybean Internal Quality Based on Improved Support Vector Regression Based on the Sparrow Search Algorithm Applying Hyperspectral Reflectance and Chemometric Calibrations. Agriculture 2024, 14, 410. [Google Scholar] [CrossRef]
  4. Chen, M.; Chang, Z.; Jin, C.; Cheng, G.; Wang, S.; Ni, Y. Classification and Recognition of Soybean Quality Based on Hyperspectral Imaging and Random Forest Methods. Sensors 2025, 25, 1539. [Google Scholar] [CrossRef]
  5. Wang, Z.; Li, J.; Zhang, C.; Fan, S. Development of a General Prediction Model of Moisture Content in Maize Seeds Based on LW-NIR Hyperspectral Imaging. Agriculture 2023, 13, 359. [Google Scholar] [CrossRef]
  6. Ismail, A.; Yim, D.G.; Kim, G.; Jo, C. Hyperspectral Imaging Coupled with Multivariate Analyses for Efficient Prediction of Chemical, Biological and Physical Properties of Seafood Products. Food Eng. Rev. 2023, 15, 41–55. [Google Scholar] [CrossRef]
  7. Tulsi, A.; Momin, A.; Ayres, V. Moisture Prediction in Chicken Litter Using Hyperspectral Data and Machine Learning. Smart Agric. Technol. 2025, 11, 101058. [Google Scholar] [CrossRef]
  8. Aviara, N.A.; Liberty, J.T.; Olatunbosun, O.S.; Shoyombo, H.A.; Oyeniyi, S.K. Potential Application of Hyperspectral Imaging in Food Grain Quality Inspection, Evaluation and Control during Bulk Storage. J. Agric. Food Res. 2022, 8, 100288. [Google Scholar] [CrossRef]
  9. Pereira, G.M.L.; Foleis, J.H.; De Souza Brito, A.; Bertolini, D. A Database for Soybean Seed Classification. In Proceedings of the 2024 37th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Manaus, Brazil, 30 September–3 October 2024. [Google Scholar] [CrossRef]
  10. Huang, Z.; Wang, R.; Cao, Y.; Zheng, S.; Teng, Y.; Wang, F.; Wang, L.; Du, J. Deep Learning Based Soybean Seed Classification. Comput. Electron. Agric. 2022, 202, 107393. [Google Scholar] [CrossRef]
  11. Jin, C.; Liu, S.; Chen, M. Semantic Segmentation-Based Mechanized Harvesting Soybean Quality Detection. Sci. Prog. 2022, 105, 00368504221108518. [Google Scholar] [CrossRef]
  12. Ghimire, A.; Kim, S.H.; Cho, A.; Jang, N.; Ahn, S.; Islam, M.S.; Mansoor, S.; Chung, Y.S.; Kim, Y. Automatic Evaluation of Soybean Seed Traits Using RGB Image Data and a Python Algorithm. Plants 2023, 12, 3078. [Google Scholar] [CrossRef]
  13. Yoosefzadeh-Najafabadi, M.; Earl, H.J.; Tulpan, D.; Sulik, J.; Eskandari, M. Application of Machine Learning Algorithms in Plant Breeding: Predicting Yield From Hyperspectral Reflectance in Soybean. Front. Plant Sci. 2021, 11, 624273. [Google Scholar] [CrossRef]
  14. Zhao, G.; Quan, L.; Li, H.; Feng, H.; Li, S.; Zhang, S.; Liu, R. Real-Time Recognition System of Soybean Seed Full-Surface Defects Based on Deep Learning. Comput. Electron. Agric. 2021, 187, 106230. [Google Scholar] [CrossRef]
  15. Sable, A.; Singh, P.; Kaur, A.; Driss, M.; Boulila, W. Quantifying Soybean Defects: A Computational Approach to Seed Classification Using Deep Learning Techniques. Agronomy 2024, 14, 1098. [Google Scholar] [CrossRef]
  16. Wu, G.; Fei, L.; Deng, L.; Yang, H.; Han, M.; Han, Z.; Zhao, L. Identification of Soybean Mutant Lines Based on Dual-Branch CNN Model Fusion Framework Utilizing Images from Different Organs. Plants 2023, 12, 2315. [Google Scholar] [CrossRef] [PubMed]
  17. Gulzar, Y.; Hamid, Y.; Soomro, A.B.; Alwan, A.A.; Journaux, L. A Convolution Neural Network-Based Seed Classification System. Symmetry 2020, 12, 2018. [Google Scholar] [CrossRef]
  18. Gulzar, Y. Enhancing Soybean Classification with Modified Inception Model: A Transfer Learning Approach. Emir. J. Food Agric. 2024, 36, 1–9. [Google Scholar] [CrossRef]
  19. Wei, Y.; Li, X.; Pan, X.; Li, L. Nondestructive Classification of Soybean Seed Varieties by Hyperspectral Imaging and Ensemble Machine Learning Algorithms. Sensors 2020, 20, 6980. [Google Scholar] [CrossRef]
  20. Tan, K.; Wang, R.; Li, M.; Gong, Z. Discriminating Soybean Seed Varieties Using Hyperspectral Imaging and Machine Learning. J. Comput. Methods Sci. Eng. 2019, 19, 1001–1015. [Google Scholar] [CrossRef]
  21. Ekramirad, N.; Doyle, L.; Loeb, J.; Santra, D.; Adedeji, A.A. Hyperspectral Imaging and Machine Learning as a Nondestructive Method for Proso Millet Seed Detection and Classification. Foods 2024, 13, 1330. [Google Scholar] [CrossRef]
  22. Momin, M.A.; Yamamoto, K.; Miyamoto, M.; Kondo, N.; Grift, T. Machine Vision Based Soybean Quality Evaluation. Comput. Electron. Agric. 2017, 140, 452–460. [Google Scholar] [CrossRef]
  23. Aref, M.H.; Korganbayev, S.; Aboughaleb, I.H.; Hussein, A.A.; Abbass, M.A.; Abdlaty, R.; Sabry, Y.M.; Saccomandi, P.; Youssef, A.B.M. Custom Hyperspectral Imaging System Reveals Unique Spectral Signatures of Heart, Kidney, and Liver Tissues. Spectrochim. Acta A Mol. Biomol. Spectrosc. 2024, 305, 123363. [Google Scholar] [CrossRef]
  24. Zhang, X.; Vinatzer, B.A.; Li, S. Hyperspectral Imaging Analysis for Early Detection of Tomato Bacterial Leaf Spot Disease. Sci. Rep. 2024, 14, 27666. [Google Scholar] [CrossRef] [PubMed]
  25. Sanaeifar, A.; Yang, C.; Min, A.; Jones, C.R.; Michaels, T.E.; Krueger, Q.J.; Barnes, R.; Velte, T.J. Noninvasive Early Detection of Nutrient Deficiencies in Greenhouse-Grown Industrial Hemp Using Hyperspectral Imaging. Remote Sens. 2024, 16, 187. [Google Scholar] [CrossRef]
  26. Wu, H.; Song, Z.; Niu, X.; Liu, J.; Jiang, J.; Li, Y. Classification of Toona Sinensis Young Leaves Using Machine Learning and UAV-Borne Hyperspectral Imagery. Front. Plant Sci. 2022, 13, 940327. [Google Scholar] [CrossRef]
  27. Xing, Z.; Chen, J.; Zhao, X.; Li, Y.; Li, X.; Zhang, Z.; Lao, C.; Wang, H. Quantitative Estimation of Wastewater Quality Parameters by Hyperspectral Band Screening Using GC, VIP and SPA. PeerJ 2019, 7, e8255. [Google Scholar] [CrossRef]
  28. Geladi, P.; MacDougall, D.; Martens, H. Linearization and Scatter-Correction for Near-Infrared Reflectance Spectra of Meat. Appl. Spectrosc. 1985, 39, 491–500. [Google Scholar] [CrossRef]
  29. Barnes, R.J.; Dhanoa, M.S.; Lister, S.J. Standard Normal Variate Transformation and De-Trending of near-Infrared Diffuse Reflectance Spectra. Appl. Spectrosc. 1989, 43, 772–777. [Google Scholar] [CrossRef]
  30. Savitzky, A.; Golay, M.J.E. Smoothing and Differentiation of Data by Simplified Least Squares Procedures. Anal. Chem. 1964, 36, 1627–1639. [Google Scholar] [CrossRef]
  31. Zheng, Y.; Tian, S.; Xie, L. Improving the Identification Accuracy of Sugar Orange Suffering from Granulation through Diameter Correction and Stepwise Variable Selection. Postharvest Biol. Technol. 2023, 200, 112313. [Google Scholar] [CrossRef]
  32. Jiao, Y.; Li, Z.; Chen, X.; Fei, S. Preprocessing Methods for Near-Infrared Spectrum Calibration. J. Chemom. 2020, 34, e3306. [Google Scholar] [CrossRef]
  33. Haralick, R.M.; Dinstein, I.; Shanmugam, K. Textural Features for Image Classification. IEEE Trans. Syst. Man. Cybern. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef]
  34. Gazala, I.F.S.; Sahoo, R.N.; Pandey, R.; Mandal, B.; Gupta, V.K.; Singh, R.; Sinha, P. Spectral Reflectance Pattern in Soybean for Assessing Yellow Mosaic Disease. Indian. J. Virol. 2013, 24, 242–249. [Google Scholar] [CrossRef] [PubMed]
  35. Chelladurai, V.; Karuppiah, K.; Jayas, D.S.; Fields, P.G.; White, N.D.G. Detection of Callosobruchus Maculatus (F.) Infestation in Soybean Using Soft X-Ray and NIR Hyperspectral Imaging Techniques. J. Stored Prod. Res. 2014, 57, 43–48. [Google Scholar] [CrossRef]
  36. Santana, D.C.; de Oliveira, I.C.; de Oliveira, J.L.G.; Baio, F.H.R.; Teodoro, L.P.R.; da Silva Junior, C.A.; Seron, A.C.C.; Ítavo, L.C.V.; Coradi, P.C.; Teodoro, P.E. High-Throughput Phenotyping Using VIS/NIR Spectroscopy in the Classification of Soybean Genotypes for Grain Yield and Industrial Traits. Spectrochim. Acta A Mol. Biomol. Spectrosc. 2024, 310, 123963. [Google Scholar] [CrossRef] [PubMed]
  37. Hossain, E.; Hossain, M.F.; Rahaman, M.A. A Color and Texture Based Approach for the Detection and Classification of Plant Leaf Disease Using KNN Classifier. In Proceedings of the 2019 International Conference on Electrical, Computer and Communication Engineering (ECCE), Cox’sBazar, Bangladesh, 7–9 February 2019. [Google Scholar] [CrossRef]
  38. Dhal, S.B.; Mahanta, S.; Moore, J.M.C.; Kalafatis, S. Machine Learning-Based Analysis of Nutrient and Water Uptake in Hydroponically Grown Soybeans. Sci. Rep. 2024, 14, 24337. [Google Scholar] [CrossRef]
  39. Khalili, E.; Kouchaki, S.; Ramazi, S.; Ghanati, F. Machine Learning Techniques for Soybean Charcoal Rot Disease Prediction. Front. Plant Sci. 2020, 11, 590529. [Google Scholar] [CrossRef]
  40. Li, X.; Chen, M.; He, S.; Xu, X.; He, L.; Wang, L.; Gao, Y.; Tang, F.; Gong, T.; Wang, W.; et al. Estimation of Soybean Yield Based on High-Throughput Phenotyping and Machine Learning. Front. Plant Sci. 2024, 15, 1395760. [Google Scholar] [CrossRef]
  41. Kaler, N.; Bhatia, V.; Mishra, A.K. Deep Learning-Based Robust Analysis of Laser Bio-Speckle Data for Detection of Fungal-Infected Soybean Seeds. IEEE Access 2023, 11, 89331–89348. [Google Scholar] [CrossRef]
  42. Zheng, L.; Zhao, M.; Zhu, J.; Huang, L.; Zhao, J.; Liang, D.; Zhang, D. Fusion of Hyperspectral Imaging (HSI) and RGB for Identification of Soybean Kernel Damages Using ShuffleNet with Convolutional Optimization and Cross Stage Partial Architecture. Front. Plant Sci. 2023, 13, 1098864. [Google Scholar] [CrossRef]
  43. Zhang, N.; Zhang, E.; Li, F.; Ghidoni, S.; Zanuttigh, P.; Zhang, N.; Zhang, E.; Li, F. A Soybean Classification Method Based on Data Balance and Deep Learning. Appl. Sci. 2023, 13, 6425. [Google Scholar] [CrossRef]
Figure 1. Representative samples of harvested soybean classification categories: (a) Normal soybean, (b) Split or broken soybean, (c) Diseased soybean, (d) Foreign materials (e.g., stems, pods, plant debris).
Figure 1. Representative samples of harvested soybean classification categories: (a) Normal soybean, (b) Split or broken soybean, (c) Diseased soybean, (d) Foreign materials (e.g., stems, pods, plant debris).
Agriengineering 07 00274 g001
Figure 2. Hyperspectral imaging system for soybean classification.
Figure 2. Hyperspectral imaging system for soybean classification.
Agriengineering 07 00274 g002
Figure 3. Workflow of the proposed spectral–spatial fusion method for soybean classification.
Figure 3. Workflow of the proposed spectral–spatial fusion method for soybean classification.
Agriengineering 07 00274 g003
Figure 4. Stepwise visualization of spatial data processing: (a) Original RGB image, (b) Single-band image (1000.44 nm), (c) Gaussian smoothing, (d) Binary thresholding, (e) Final object mask.
Figure 4. Stepwise visualization of spatial data processing: (a) Original RGB image, (b) Single-band image (1000.44 nm), (c) Gaussian smoothing, (d) Binary thresholding, (e) Final object mask.
Agriengineering 07 00274 g004
Figure 5. Mean spectral reflectance curves (SG smoothed) of different object categories in the soybean classification dataset. The curves highlight distinct spectral responses across the 379.16–1020.87 nm range, particularly in the NIR and red-edge regions, supporting effective differentiation for classification.
Figure 5. Mean spectral reflectance curves (SG smoothed) of different object categories in the soybean classification dataset. The curves highlight distinct spectral responses across the 379.16–1020.87 nm range, particularly in the NIR and red-edge regions, supporting effective differentiation for classification.
Agriengineering 07 00274 g005
Figure 6. Comparison of spectral preprocessing techniques for soybean classification using LDA.
Figure 6. Comparison of spectral preprocessing techniques for soybean classification using LDA.
Agriengineering 07 00274 g006
Figure 7. Performance metrics for soybean classification using data fusion of spectral and spatial features.
Figure 7. Performance metrics for soybean classification using data fusion of spectral and spatial features.
Agriengineering 07 00274 g007
Figure 8. Confusion matrices for the best-performing classifiers: (a) GB using spatial features, (b) LDA using spectral features, and (c) LDA using data fusion.
Figure 8. Confusion matrices for the best-performing classifiers: (a) GB using spatial features, (b) LDA using spectral features, and (c) LDA using data fusion.
Agriengineering 07 00274 g008
Figure 9. ROC curves of the best classifier for soybean classification: (a) GB using only spatial features, (b) LDA using only spectral features, and (c) LDA using data fusion.
Figure 9. ROC curves of the best classifier for soybean classification: (a) GB using only spatial features, (b) LDA using only spectral features, and (c) LDA using data fusion.
Agriengineering 07 00274 g009
Figure 10. Prediction results using spatial features only.
Figure 10. Prediction results using spatial features only.
Agriengineering 07 00274 g010
Figure 11. Prediction results using spectral features only.
Figure 11. Prediction results using spectral features only.
Agriengineering 07 00274 g011
Figure 12. Prediction results using combined spectral and spatial features (data fusion).
Figure 12. Prediction results using combined spectral and spatial features (data fusion).
Agriengineering 07 00274 g012
Table 1. Distribution of samples across training and testing datasets.
Table 1. Distribution of samples across training and testing datasets.
Object TypeTraining DatasetTest Dataset
Normal soybean23053
Split soybean25050
Diseased soybean24050
Forgien materials21453
Total934206
Table 2. GLCM texture features extracted from representative samples across four soybean classification categories.
Table 2. GLCM texture features extracted from representative samples across four soybean classification categories.
GLCM FeaturesNormal SoybeanSplit SoybeanDiseased SoybeanForeign Materials
Angular second moment0.001166411 0.001468802 0.00060536 0.000800113
Contrast9.9873399 6.4238776441.727726654.2942626
Correlation0.9956900850.9962267270.984167122 0.977210946
Difference entropy2.6903426322.467046877 3.592117992 3.704523883
Difference variance0.000831190.001122403 0.000494607 0.000451044
Entropy10.361285389.94064582411.28466308 11.20165864
Information measure of correlation 1−0.541876642−0.55026881 −0.40437518 −0.37690467
Information measure of correlation 20.999758870.9997142770.9981913150.996879067
Inverse difference moment0.4007466610.417335190.2461505160.262414697
Maximal correlation coefficient3.2534069573.0267766463.9464329394.852678038
Sum average235.7983282229.8465888240.6570153119.0831582
Sum entropy8.0941786127.8471923898.0514742187.892176898
Sum of squares: variance1159.904379852.21207451319.6602861191.164885
Sum variance4629.6301743402.424425236.9134164710.365278
Table 3. Performance evaluation of classifiers using spatial features.
Table 3. Performance evaluation of classifiers using spatial features.
Classifier AccuracyPrecisionRecallF1-Score
GB0.88110.88470.88110.8802
LightGBM0.87470.87860.87470.8739
ET0.87150.87380.87150.8708
RF0.85440.85930.85440.8533
LDA0.83390.84140.83390.8309
Bold values indicate the highest value for each evaluation metric.
Table 4. Performance evaluation of classifiers using spectral reflectance features.
Table 4. Performance evaluation of classifiers using spectral reflectance features.
ClassifierAccuracyPrecisionRecallF1-Score
LDA0.98930.99000.98930.9893
LR0.93140.93780.93140.9311
QDA0.93140.94150.93140.9285
ET0.92290.92630.92290.9227
KNN0.91650.92210.91650.9162
Bold values indicate the highest value for each evaluation metric.
Table 5. Performance comparison between the proposed method and deep learning-based classification methods reported in previous studies.
Table 5. Performance comparison between the proposed method and deep learning-based classification methods reported in previous studies.
WorkModel ArchitectureClasses of ObjectsAccuracy
Proposed MethodData-fusion (Spectral + Spatial)499.03%
Kaler et al. [41]CNN297.72%
Zheng et al. [42]ShuffleNet498.36%
N. Zhang et al. [43]DenseNet498.48%
Gulzar et al. [18]InceptionV3498.73%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rahman, M.B.; Tulsi, A.; Momin, A. Spectral-Spatial Fusion for Soybean Quality Evaluation Using Hyperspectral Imaging. AgriEngineering 2025, 7, 274. https://doi.org/10.3390/agriengineering7090274

AMA Style

Rahman MB, Tulsi A, Momin A. Spectral-Spatial Fusion for Soybean Quality Evaluation Using Hyperspectral Imaging. AgriEngineering. 2025; 7(9):274. https://doi.org/10.3390/agriengineering7090274

Chicago/Turabian Style

Rahman, Md Bayazid, Ahmad Tulsi, and Abdul Momin. 2025. "Spectral-Spatial Fusion for Soybean Quality Evaluation Using Hyperspectral Imaging" AgriEngineering 7, no. 9: 274. https://doi.org/10.3390/agriengineering7090274

APA Style

Rahman, M. B., Tulsi, A., & Momin, A. (2025). Spectral-Spatial Fusion for Soybean Quality Evaluation Using Hyperspectral Imaging. AgriEngineering, 7(9), 274. https://doi.org/10.3390/agriengineering7090274

Article Metrics

Back to TopTop