1. Introduction
Commercial strawberry production has high economic and nutritional value, and demand for fresh berries continues to grow in many countries around the world [
1]. However, strawberry plants are susceptible to a wide range of fungal and bacterial diseases, among which leaf diseases are particularly common [
2,
3]. These diseases significantly reduce yield and product quality and can lead to substantial economic losses for producers if not controlled in a timely manner [
3]. In commercial practice, not only the presence but also the severity of foliar diseases determines pesticide dosage, plant removal decisions and yield forecasting. For this reason, we treat three-level Leaf Spot severity assessment as the central task of this work, rather than a secondary add-on to disease detection. Due to the wide variety of pathogens and the complex dynamics of their development, early diagnosis has traditionally relied on visual inspection of plantings by farmers and agronomists [
4]. However, many leaf diseases of strawberries have similar visual symptoms, which makes it difficult to reliably distinguish between classes and often leads to incorrect identification or underestimation of the severity of the damage [
4].
With the development of computing technology and image processing methods, more and more research is devoted to the intelligent diagnosis of agricultural crop diseases based on computer vision [
4,
5,
6]. Such approaches allow for the automatic extraction of informative features from images of leaves, reduce the time required to examine plantings and labour costs, and provide higher reproducibility and accuracy compared to traditional manual diagnosis [
6,
7]. In particular, deep learning methods and YOLO family object detectors have demonstrated high efficiency in detecting strawberry diseases in field conditions, making them a promising basis for edge-AI monitoring systems and agricultural robots. In addition to architectural modifications, several studies also improve YOLO-based plant disease detectors by introducing specialised activation functions, such as the αSiLU activation proposed for smart agriculture scenarios in [
8].
Recent studies have shown that lightweight YOLO variants can be successfully deployed in smart-agriculture applications; for example, for fruit detection under day- and night-time lighting conditions on mobile devices [
9], for nutritional monitoring of lettuce using point-cloud analysis and YOLO-based object detection [
10], and for livestock monitoring and counting from UAV imagery using Mask-YOLO architectures [
11]. Web-based plant disease classification systems using deep learning have also been proposed [
12], while comprehensive surveys on computer-vision and deep-learning approaches for nutrient-deficiency detection in crops further highlight the importance of non-invasive plant-health monitoring [
13]. Our work follows this line of research but focuses specifically on strawberry leaf diseases and, importantly, on integrating ordinal severity estimation into an edge-AI detection pipeline. In recent years, the YOLO family has evolved from early versions such as YOLOv3 and YOLOv4 to more accurate and lightweight variants including YOLOv5–YOLOv10, many of which are specifically optimised for embedded and edge deployment.
In addition to classifying diseases based on leaf images, more and more work is being carried out on real-time robotic systems for strawberry cultivation. One example is the integration of a YOLO-based detector into the AgriEco self-propelled robot to recognise strawberry bushes when spraying greenhouses. In this setup, an optimised YOLOv3 model running on Jetson TX2 achieves an mAP above 97% with a processing time of approximately 15 ms per frame, demonstrating the feasibility of Edge-AI solutions in greenhouse conditions [
14]. Regarding strawberry leaf diseases, Chen et al. [
15] developed an improved segmentation approach based on YOLOv8 that simultaneously detects strawberry leaves and powdery mildew lesions in natural background conditions, achieving high accuracy and reliability in field scenarios. Taken together, these studies show that YOLO family detectors are well suited for use on mobile agricultural robots, although most existing systems still focus on the localization of plants and fruits rather than on detailed characterisation of leaf condition.
At the same time, lightweight convolutional networks are being developed for recognising strawberry diseases in leaf images. Light-MobileBerryNet is an interpretable model inspired by MobileNetV3-Small that uses inverted residual blocks, depth-wise convolutions, and squeeze-and-excitation modules. The model achieves 96.6% accuracy with < 1 million parameters and a weight volume of about 2 MB, outperforming heavier CNNs (VGG16, EfficientNet, MobileNetV2) in terms of accuracy/complexity ratio [
3]. In addition, the BerryNet-Lite architecture achieves 99.45% accuracy with only 2.87 million parameters, further demonstrating that lightweight convolutional networks can outperform heavier models in strawberry disease identification [
16]. These results demonstrate that lightweight architectures (MobileNet family, EfficientNet-B0, ResNet-18, Swin-Tiny, etc.) are the most promising candidates for deployment on mobile and embedded devices in phytosanitary monitoring tasks.
A separate class of works is devoted not only to the presence of disease, but also to assessing its severity. For example, ref. [
17] investigates eight deep learning models for the simultaneous detection and assessment of the severity of Cercospora leaf spot in chilli peppers under real field conditions. It is shown that segmentation models based on YOLOv8 provide more accurate and faster classification of severity levels compared to Mask R-CNN, achieving 91.4% accuracy for the maximum level of damage with an output time of 27 ms. Reviews emphasise that disease severity assessment tasks require consideration of the ordered nature of classes, special annotation schemes, and often a task formulation that differs from “normal” classification (ordinal classification, assessment of the area of damage, etc.) [
18]. In addition, lightweight frameworks such as PDSE-Lite demonstrate that severity estimation can be implemented efficiently using compact architectures and few-shot learning [
19].
Nevertheless, for strawberries, most studies are limited to binary or multi-class classification according to the “disease type/healthy leaf” scheme without explicit assessment of the degree of damage. Even in modern lightweight models for strawberries, such as Light-MobileBerryNet [
3], the main target metric remains the accuracy of disease type recognition, rather than its severity [
18]. Similarly, work on robotic spraying and harvesting using YOLO architectures focuses on the detection of plants, fruits, or stages of maturity [
14], without offering a quantitative assessment of leaf disease progression at the individual plant level. For strawberry Leaf Spot, there are virtually no public datasets labelled by severity levels, nor are there any studies linking such an assessment to the work of mobile edge-AI platforms. In this study we deliberately focus the severity module on a single disease, Leaf Spot, rather than attempting to model many pathologies at once. Binary or multi-class disease detection (diseased vs. healthy, or different disease types) is already a relatively mature problem in computer vision for agriculture, whereas reliable severity assessment remains a key frontier in precision crop protection. Leaf Spot offers a particularly suitable case study because its visual symptoms develop in a relatively clear and monotonic way, from isolated small lesions to large coalescing necrotic areas, which aligns well with a three-level severity scale.
Unlike previous severity-estimation studies conducted on other crops such as rice and chilli pepper, our work focuses specifically on strawberry leaves, whose morphology and visual symptoms of Leaf Spot differ markedly from those of cereal and vegetable crops. Moreover, we design a two-stage edge-AI pipeline that first performs multiclass detection of seven strawberry leaf diseases using YOLOv10n and then applies an Ordinal MobileViT-S classifier to assign one of three Leaf Spot severity levels (S1–S3). To the best of our knowledge, no prior work has explored such a detection + ordinal severity cascade for strawberry leaves on an embedded platform such as NVIDIA Jetson Nano.
In summary, the main contributions of this work are as follows: (a) we develop StrawberrySeverityNet, a lightweight two-stage edge-AI system that combines YOLOv10n-based multiclass detection of seven strawberry leaf diseases with ordinal three-level severity assessment of Leaf Spot; (b) we construct and describe a custom dataset that includes both disease-level annotations and expert-labelled severity levels linked to the proportion of infected leaf area; (c) we propose an Ordinal MobileViT-S architecture that adds a lightweight cumulative-threshold ordinal head and loss on top of the standard MobileViT-S backbone, explicitly exploiting the ordered nature of the severity levels while keeping the model compact and suitable for edge deployment; and (d) we demonstrate the effectiveness of the proposed framework on an NVIDIA Jetson Nano mobile platform and discuss its potential role within smart-agriculture scenarios for precision irrigation and plant-health monitoring.
From an industrial perspective, StrawberrySeverityNet can also be regarded as a raw-material quality monitoring module within the digital twin of a strawberry-processing enterprise, where leaf disease and three-level Leaf Spot severity estimates serve as upstream quality indicators for subsequent processing stages.
4. Discussion
The results show that the proposed two-stage StrawberrySeverityNet system can serve as a practical solution for monitoring strawberry leaf diseases on peripheral devices. Similar edge-AI architectures for in-field plant disease monitoring have also been reported in recent studies [
23,
24]. The lightweight YOLOv10n detector provides reliable detection of seven disease classes with limited computing resources: on the test sample, it achieves an mAP@0.5 of 0.960, a recall of 0.917, and an F1 score of 0.93. The Ordinal MobileViT-S module allows for additional assessment of Leaf Spot severity on three levels. This design sets the system apart from most existing solutions, which focus only on disease classification or plant and fruit detection without quantitative assessment of the degree of damage.
A comparison of lightweight architectures for the task of three-level severity assessment shows that simply increasing the number of parameters (e.g., using Swin-Tiny) is not always justified in terms of deployment on mobile platforms. Ordinal MobileViT-S achieves the highest accuracy (0.973) with a moderate model size (4.94 million parameters), outperforming both more compact CNNs (MobileNetV3-Small, EfficientNet-B0) and the significantly heavier Swin-Tiny. This confirms that taking into account the ordinal structure of classes and a specialised severity head provide a more significant increase in quality than further complicating the architecture.
Analysis of the error matrix shows that the model practically does not allow gross errors between S1 (mild) and S3 (severe): all severely affected leaves are recognised correctly, and errors occur only at the border between neighbouring levels S1/S2. This is critical for practical precision farming scenarios: the system reliably identifies plants with severe damage that require immediate intervention (removal or targeted treatment), while minor discrepancies between “mild” and “moderate” damage are less critical. This focus on avoiding large misclassifications between mild and severe damage is consistent with general recommendations for designing and interpreting visual disease severity scales in plant pathology [
20].
In practical agronomic terms, confusion between S1 and S2 is usually less critical than confusion involving S3. For example, leaves classified as S1 or S2 may lead to similar recommendations such as closer visual monitoring and, if necessary, localised low-dose fungicide application, whereas S3 plants are candidates for immediate removal or intensive treatment to prevent further spread of the pathogen. By explicitly separating S3 from the two lower levels, the proposed system supports targeted interventions and can help reduce unnecessary blanket spraying.
While a detailed economic analysis is beyond the scope of this study, we expect that severity-aware monitoring could contribute to more cost-effective crop protection. Concentrating fungicide applications and manual labour on S2–S3 plants, and removing severely affected leaves or plants at an early stage, has the potential to lower chemical input, minimise unnecessary treatments on lightly affected plants and reduce yield losses due to late detection.
In a broader systems-engineering context, the edge-deployed StrawberrySeverityNet can be viewed as an IIoT node within the digital twin of a vertically integrated strawberry-processing enterprise, whose field layout can be aligned with CAD-based maps of the production area. Disease type and three-level severity outputs (S1–S3) can be transmitted from the mobile robot to the enterprise information system via standard protocols such as MQTT or OPC UA, visualised on a central dashboard as field- or block-level severity maps and linked to key performance indicators (e.g., expected yield of marketable product, fraction of low-quality batches). In this way, early severity-aware monitoring at the raw-material stage supports lean manufacturing principles by reducing waste associated with poor-quality input and unnecessary treatments.
However, the work has a number of limitations. First, the severity assessment module is trained on a relatively small and manually labelled set of 373 Leaf Spot patches, which may limit the model’s generalizability to other strawberry varieties and to truly uncontrolled open-field illumination (e.g., strong shadows, backlighting and specular reflections). Although using ImageNet-pretrained backbones and data augmentation reduces the risk of overfitting, collecting a larger and more diverse severity dataset remains an important direction for future work. Second, the system evaluates only one disease (Leaf Spot), whereas in real plantings, a combination of several pathologies and physiological stresses is often observed. Finally, the current version does not take into account the temporal dynamics of disease development and relies only on single images of leaves.
Another limitation is that we do not perform a dedicated robustness study with respect to synthetic noise, blur or strong changes in illumination and background. The current evaluation reflects only the natural variability present in the training and test datasets, which already includes different lighting conditions, backgrounds and occlusions. While the use of ImageNet pre-training and mild geometric and photometric augmentations is expected to improve generalisation to moderate appearance changes, we did not carry out an ablation study to isolate their effect, and robustness to severe noise or domain shifts (e.g., different cameras, cultivars or acquisition conditions) remains to be investigated. A systematic robustness analysis of both the detection and severity modules is therefore an important direction for future work.
A further limitation of the present study is that we did not perform a dedicated profiling of inference latency, throughput and energy consumption for the Ordinal MobileViT-S module on the NVIDIA Jetson Nano device. While the small number of parameters and modest input resolution suggest that the severity module is suitable for near real-time operation when combined with YOLOv10n, a systematic latency analysis under different deployment configurations is left as future work. Such a study will be important for further optimising StrawberrySeverityNet for continuous on-board monitoring in commercial greenhouse and open-field environments.
In the future, we plan to expand the Leaf Spot severity dataset to include other common strawberry leaf diseases, as well as conduct full-scale testing on a mobile platform with a closed-loop control system (detection → severity assessment → treatment decision). A separate area of focus is the integration of time series observations and environmental sensor data (humidity, temperature, leaf moisture) to predict disease progression and develop recommendations for targeted fungicide application. In addition, we plan to enlarge the dataset with images from additional strawberry cultivars and growing regions and to explore domain-adaptation techniques to improve cross-site generalisation.
5. Conclusions
This paper presents StrawberrySeverityNet, a lightweight two-stage edge-AI system for monitoring the health of strawberry leaves. In the first stage, the YOLOv10n detector, deployed on an onboard computing module, detects seven major leaf diseases with real-time capability on an embedded module. Experiments have shown that the YOLOv10n detector achieves an mAP@0.5 of 0.960, a recall of 0.917, and an F1 score of 0.93, which corresponds to high-quality localization of affected leaves while maintaining the lightness of the model. In the second stage, patches are formed from the detected Leaf Spot class leaves, for which a three-level severity assessment module based on a modified Ordinal MobileViT-S model is proposed. This combination not only identifies the type of disease but also quantitatively assesses the degree of damage, which is important for making agrotechnological decisions and can be directly linked to actions such as adjusting fungicide dosage, prioritising removal of severely affected plants, and planning site-specific treatments. Overall, the results emphasise that moving from pure disease detection to severity-aware diagnosis can have a direct practical impact on decision making in strawberry production.
A comparison with a number of compact architectures (ResNet-18, EfficientNet-B0, MobileNetV3-Small, Swin-Tiny, basic MobileViT-S) showed that Ordinal MobileViT-S achieves the highest accuracy (0.973) with a moderate number of parameters (4.94 million), providing the best balance between classification quality and computational complexity. Class-wise analysis and error matrix demonstrated high reliability in recognising severely affected leaves and no gross errors between mild and severe degrees of severity.
The practical significance of the system lies in its ability to be deployed on mobile agricultural robots and edge platforms for real-time monitoring of strawberry crops in greenhouses and open fields. The output data in the form of a “disease type—severity level” pair can be directly used to plan spot spraying, remove severely affected plants, and evaluate the effectiveness of the measures taken.
When integrated into the dashboard of a food-enterprise digital twin, these outputs can be used by operators to visualise spatial patterns of disease severity and to adjust harvesting and processing strategies, and the same cascade architecture could be adapted to monitor visible defects on conveyor belts within the processing plant.
In the future, we plan to expand the Leaf Spot severity dataset to include other strawberry leaf diseases, as well as conduct field tests of the system as part of a robotic platform with a closed control loop. Additional areas of development include the use of time series observations to predict disease progression and the integration of multimodal data (images + environmental sensors) for more accurate and robust diagnostics.