Previous Article in Journal
Therapeutic Potential of Natural Products as Innovative and New Frontiers for Combating Parasitic Diseases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MLens: Advancing the Real-Time Detection, Identification, and Counting of Pathogenic Microparasites Through a Web Interface

by
Gustavo Souza Carneiro
1,
Karoliny Caldas Xavier
1,
José Ledamir Sindeaux-Neto
1,
Alanna do Socorro Lima da Silva
2 and
Michele Velasco Oliveira da Silva
1,*
1
Laboratory of Morphological-Molecular Integration and Technologies (LIMT), Federal Rural University of the Amazon (UFRA), Belém 66077-830, PA, Brazil
2
Postgraduate Program in Society, Environment and Quality of Life (PPGSAQ), Federal University of Western Pará (UFOPA), Santarém 68040-255, PA, Brazil
*
Author to whom correspondence should be addressed.
Parasitologia 2025, 5(4), 50; https://doi.org/10.3390/parasitologia5040050
Submission received: 2 November 2024 / Revised: 19 December 2024 / Accepted: 18 September 2025 / Published: 23 September 2025

Abstract

In this study, a diverse collection of images of myxozoans from the genera Henneguya and Myxobolus was created, providing a practical dataset for application in computer vision. Four versions of the YOLOv5 network were tested, achieving an average precision of 97.9%, a recall of 96.7%, and an F1 score of 97%, demonstrating the effectiveness of MLens in the automatic detection of these parasites. These results indicated that machine learning has the potential to make microparasite detection more efficient and less reliant on manual work in parasitology. The beta version of the MLens showed strong performance, and future improvements may include fine-tuning the WebApp hyperparameters, expanding to other myxosporean genera, and refining the model to handle more complex optical microscopy scenarios. This work presented a significant advancement, opening new possibilities for the application of machine learning in parasitology and substantially accelerating parasite detection.

1. Introduction

The class Myxozoa comprises over 70 genera and approximately 2600 species, all of which are obligate endoparasites [1,2,3]. Many of these parasites cause severe diseases known as myxosporidioses. Lom and Arthur [4] described morphological and morphometric variables as indispensable factors in defining myxospores and taxonomic groupings.
Among myxozoans, the genera Henneguya [4] and Myxobolus play significant roles as parasites of ichthyofauna [5]. These polyphyletic genera encompass various members distributed worldwide, are found in both freshwater and marine ichthyofauna, and have also been reported in reptiles and amphibians [1,6]. Their geographic distribution is influenced by environmental factors such as water temperature, water quality, and host availability [7,8,9].
Neural networks aspire to emulate the complexity of biological neuronal functions [10]. Analogous to cellular processes, artificial neurons excel in their ability to receive, process, and transmit information through mathematical operations, thereby enabling them to identify patterns, perform classifications, and predict characteristics [10,11]. This biology–mathematics symbiosis-inspired process opens promising horizons for innovation in the field of AI, where human and artificial minds intertwine in the pursuit of innovative and bold solutions [12].
The application of AI finds fertile ground in parasitology, playing a crucial role in addressing persistent challenges, and convolutional neural networks are tools used in computer vision (CV) and are now essential in veterinary medicine and industry [13]. They are intrinsically associated with CV, which, when combined with image analysis techniques, results in comprehensive specialized systems that offer solutions for detecting and classifying parasitic objects with minimal human intervention and high accuracy [14]. This, in turn, eliminates the errors often associated with human analysis stemming from tasks that are often fatiguing, monotonous, costly, or simply lack expertise [15,16,17].
Moreover, the acquisition of microscopy images establishes a solid foundation for feeding CV systems, revealing new diagnostic and research perspectives in parasitology [18,19].
In this study, a beta version of a WebApp called MLens 1.0.0 beta was developed to perform the automated detection and classification of the genera Myxobolus and Henneguya in light microscopy-acquired images via the YOLOv5 neural network, aiming to increase the efficiency and accuracy of parasitological diagnosis.

Background

Advances in object detection using algorithms such as YOLOv5 emerged as an effective approach for complex identification tasks in optical microscopy images [20]. Networks like these offer fast and accurate solutions, minimizing human intervention and maximizing accuracy even in challenging environments [21]. Previous studies demonstrated the effectiveness of these tools in biology and medicine, including the detection of cellular structures and microparasites [14,22].
As highlighted by Sichman et al. [23], humanity’s relentless pursuit of new heights of progress is undeniably driven by continuous advances in the broad domain of science, as evidenced in the case of artificial intelligence (AI), which played a pivotal role in enabling the development of specialized computational technologies. Therefore, this phenomenon sparked substantial and enduring interest, with its impact continuing to resonate notably in an ever-evolving society [21,24]. However, their application to the diagnosis of myxozoans is still emerging, with gaps related to model generalization in heterogeneous scenarios and the morphological variability of parasites [25,26].
This article was structured into five main sections. In Section 2, we detailed the materials and methods used, including the data collection, preparation, and training process with the YOLOv5 algorithm. Section 3 presented the results obtained and an in-depth discussion of this study’s implications and limitations, highlighting the main findings [27,28]. Section 4 concluded this article and explored future perspectives, At the end of this article, the references that underpin this work [29,30] are listed. This organization aimed to provide a clear and objective view of the MLens system’s contribution to modern parasitology.

2. Materials and Methods

The development of the MLens 1.0.0 beta was carried out in Belém, PA, Brazil or processing light microscopy images, aimed at detecting two genera of myxozoans via the You Only Look Once (YOLOv5) algorithm, was performed in three distinct phases: dataset collection, project dataset training, and implementation. Notably, all stages were conducted at the Laboratory of Morphomolecular Integration and Technologies (LIMT) of the Federal Rural University of the Amazon.

2.1. Cnidaria: Myxozoa

Myxozoa (Cnidaria: Endocnidoxoa) represents one-fifth of the known cnidarian species [25,31] worldwide, spanning freshwater and marine environments [31,32]. Many infections are asymptomatic and are considered harmless. However, some species have high pathogenic potential, causing disease and significant mortality in parasitized organisms. This can lead to substantial economic losses in the fishing and aquaculture sector. Additionally, immunocompromised individuals who consume raw fish protein from infected fish may experience harm caused by these parasites, although there is no evidence that humans are hosts in any stage of the myxozoan life cycle. However, myxozoan spores were detected in human feces [33].
Myxozoans exhibit a complex life cycle that involves alternating between invertebrate and vertebrate hosts. The latter serve as intermediate hosts, whereas invertebrates act as definitive hosts, particularly those from the classes Oligochaeta and Polychaeta, which belong to the phylum Annelida [7]. Transmission between hosts occurs through two morphologically distinct forms: the myxospore, which develops in vertebrates, and the actinospore, which is found in invertebrates. Typically, this transmission occurs via spores that disperse in water, as shown in the work of Fiala et al. [2] and Figure 1.
Okamura et al. [25] reported that myxozoans exhibit pronounced structural simplification, along with miniaturization, as a result of their parasitic niche. Okamura et al. [7] reported that these parasites obligatorily undergo complex life cycles, requiring both invertebrate and vertebrate hosts. This extensive group of endoparasites has complex, obligatory microscopic, and heterogeneous life cycles that are dependent on both invertebrate and vertebrate hosts [25,34]. These parasites have the potential to cause significant environmental and economic damage. Although they are commonly found in fish, they can also infect reptiles, amphibians, waterfowl, and even small mammals, as shown in the work of Eiras et al. [26].

2.2. Data Acquisition

The image collection system used in this study belongs to the LIMMT. The team imaged all parasites of the genera Henneguya and Myxobolus found during necropsies between 2022 and 2024 via smartphones, as exemplified in the pipeline (Figure 2). In total, 125 images with dimensions of 2160 × 3840 pixels were collected and subdivided into 12 smaller images with dimensions of 640 pixels. Therefore, the sample comprised 12,000 images divided into two genera, Myxobolus and Henneguya, covering various species of Amazonian fish and different tissues, such as gills, liver, and skeletal muscle. The images were uniformly sized at 640 pixels each and were in JPEG format.

2.3. Data Preprocessing

To enhance the generalizability of the model for identifying microparasites in microscopy images, several image processing techniques were employed, as shown in Figure 3. As discussed by Zhao et al. [20], rotations were applied at angles of 90°, 180°, and 270°. Additionally, horizontal and vertical flips were employed on the original images and at 90° clockwise, as demonstrated in the works of Bydder et al. [35], Zendehdel et al. [22], and Ghose et al. [36], thus avoiding redundancies. These transformations provided increased diversity in the dataset because, after subdividing the original images, each subunit resulted in eight new unique images. This helped the model recognize patterns in multiple orientations and reflections, thereby improving its robustness and accuracy in detecting these genera in various scenarios.

2.4. Labeling of Myxozoans

The Roboflow platform for CV was used to demarcate the regions of interest, that is, to label the parasites in the form of bounding boxes. As emphasized by Noever and Noever [16], the essence of this task was the accuracy of labeling the objects of interest, which, in turn, depended on human performance. As illustrated in Figure 4, the quality and consistency of these labels were crucial because they served as the foundation for the learning process of the model for identifying and locating objects in new images.
This requirement was met by carefully creating bounding boxes, precisely outlining the location of the myxozoans, and assigning each genus a specific label, as shown in Table 1. This enabled the generation of label files in the “.txt” format linked to the source image. These files included the category (in this case, the genus) and bounding box coordinate information (X: x-coordinate center of the bounding box; Y: y-coordinate center of the bounding box; W: width of the bounding box; and H: height of the bounding box), as illustrated in Table 1.
The model was trained after appropriately labeling the data and organizing the training and validation sets. MLens was subsequently implemented in a real-world context by performing myxozoan detection on the basis of the meticulously provided information in this process. Table 2 shows the categories listed in Table 1.
Furthermore, this study aimed to obtain reliable evidence that the labeled genera Henneguya and Myxobolus presented a numerical balance. The analysis included an exploration of the characteristics of the labels used and the associated bounding boxes (Figure 5—statistical analysis of labels: gender histogram and density map of bounding box sizes). These genera exhibited a quantitative balance. However, the number of parasites in question tended to be very small compared with the image scale, which could complicate detection.

2.5. Quantitative Characteristics of the Dataset

The images were randomly distributed for training and validation at a ratio of 80:20. However, when making such divisions, obtaining a balance between the number of samples of each genus in the training and test sets was important to ensure maximum consistency in data distribution.

2.6. Experimental Configuration

To ensure the rigor of this study, Google Colaboratory was used for the training and validation phases of the model. This free cloud server was designed to foster research and education in machine learning and includes integrated libraries, such as TensorFlow, Matplotlib, Keras, and OpenCV, eliminating the need for complex local installations, as demonstrated by Ray et al. [37]. Additionally, Colab facilitates synchronization with Google Drive, allowing the storage of large data volumes without consuming disk space. It also promotes mobility, collaboration, and project sharing, requiring only an email account, making it convenient for community collaboration.
This platform provides hardware adapted for machine learning, including an Intel Xeon CPU; an NVIDIA Tesla® L4 GPU with 24 Gigabytes (GB) of GDDR6 memory and 7680 CUDA cores (NVIDIA, Santa Clara, CA, USA); and 78 GB of storage and 51 GB of RAM, which is supported by CUDA v12.2. The deep learning frames PyTorch v2.2.1 and Python v3.10.12 were used for development. Finally, an interactive dashboard was created via Streamlit v1.19.0.

2.7. YOLOv5

YOLOv5 has become an essential algorithm in contemporary object detection, balancing speed and execution accuracy, as demonstrated in the works of Liu et al. [38] and Li et al. [21]. Object detection is a fundamental digital field. It involves identifying the objects present in an image or frame and determining their exact locations. This type of activity is more complex than object classification, which only recognizes the presence of objects without indicating their positions in an image, as shown by Xu et al. [39]. Moreover, objects cannot be classified multiple times, increasing the complexity of the detection process.
The range of YOLOv5 is noteworthy. Compared with previous versions, YOLOv3 and YOLOv4, YOLOv5 boasts higher accuracy and excels in terms of lightness, speed of object detection, and inference speed. These characteristics are crucial for embedded applications that require real-time detection, as demonstrated by Xu et al. [39], Zhang et al. [40], and Badgujar et al. [41]. These findings demonstrated that this algorithm can substantially improve the detection and identification of myxozoans via light microscopy.
The YOLOv5 architecture includes various variants; however, in this study, only the following four variants were used: YOLOv5n, YOLOv5s, YOLOv5m, and YOLOv5l. As discussed by Elshahawy et al. [27], the recommended steps for machine training were followed and documented in a notebook hosted on Google Colab. However, its ability and adaptability to other applications using different datasets are relatively simple, flexible, and open.

2.8. Evaluation Metrics

In machine learning, the interpretation of appropriate metrics ensures the accurate assessment of a model. However, these metrics can vary from situation to situation, causing some difficulty in comparison and potentially leading to erroneous conclusions. In this study, the precision (P), recall, F1_Score, and mAP were used as evaluation metrics [42,43].
According to Shen et al. [28], the mAP metric is the average precision calculated across all genera for all categories (N). Notably, an mAP of 0.5 is an intersection over the union (IoU) threshold of 0.5. Furthermore, mAP0.5:0.95 represents the average across IoU thresholds from 0.5 to 0.95, with an interval of 0.05. The formulas for precision, recall, mAP, and F1_Score are presented in the following section.

3. Results and Discussion

Standard hyperparameters were used: an initial learning rate (lr0) of 0.01 and a final learning rate (lrf) of 0.01. Furthermore, an SGD with a momentum of 0.937 and weight_decay of 0.0005 was used to control the regularization and avoid overfitting. A batch size of 16 was selected to balance the training speed and model accuracy [44], with training lasting for 300 epochs to allow the models to clearly absorb patterns in the data, and the image dimensions were set to 640 pixels. In the first three epochs, a warmup_momentum of 0.8 and warmup_bias_lr of 0.1 were applied.
The emergence of various object detection models, such as the YOLO series, revolutionized the field of CV. Each variant of these networks has unique characteristics in terms of GPU memory requirements [29,45,46,47]. The results were derived from the detection training of the Henneguya and Myxobolus genera via the YOLOv5n, YOLOv5s, YOLOv5m, and YOLOv5l variants configured with the parameters listed in Table 1, as exemplified by Huang et al. [48] and Du [30].
In particular, the YOLOv5 model has drawn significant attention because of its high performance and efficiency. In Figure 6, the YOLOv5l model requires 11 GB of GPU memory to stabilize its training process, whereas the YOLOv5m model requires 7.1 GB. Interestingly, the YOLOv5s variant requires only 4 GB of GPU memory, and the most resource-efficient model, YOLOv5n, uses only 2.4 GB of GPU memory until the completion of training [47,49].
The most plausible explanation for the difference in GPU memory requirements can be attributed to the inherent architectural characteristics of each variant. As discussed by Hasan et al. [47], the YOLOv5n model, which is the smallest in the series, is more suitable for deployment on mobile hardware platforms, such as those embedded in vehicles, owing to its compact size and low memory demand.
Nevertheless, the analysis of the training time and model performance is a common consideration in machine learning. Larger models, such as YOLOv5l, are often more accurate in detection but are associated with robust computational resources and longer training times [50]. Fu et al. [51] highlighted that smaller variants such as YOLOv5n are more efficient and suitable for implementation on resource-constrained platforms while accepting some precision trade-offs.
The variants differed in their training completion times, as shown in Figure 6. Notably, YOLOv5n, the shortest, took 400.46 min, followed by YOLOv5s (508.85 min) and YOLOv5m (870.15 min), and YOLOv5l, which was the most complex, required 1336.78 min. Additionally, the model sizes, presented in Figure 6, corresponded to their complexities: YOLOv5n resulted in a 3.9 MB model, YOLOv5s in a 13.8 MB model, YOLOv5m in a 40.3 MB model, and YOLOv5l in a 92.9 MB model. These differences clearly indicate that model complexity influences training time and hardware demand, as mentioned by Tong and Yiquan [50] and Cai et al. [52].
The training performance curves of the models are shown in Figure 7a–d. Ma et al. [53] reported that such metrics are evaluation criteria for models that allow for the understanding of their advantages and disadvantages and comparison of different models to determine the most suitable model for real-world situations. Without such metrics, verifying the progress or superiority of one model over others would be impossible [54,55].
According to Figure 7a,c, the precision and recall rates of all the models started to increase significantly from epoch 2. Initially, at epoch 0, the precision and recall varied according to the complexity of the model: YOLOv5l started with the highest precision at 59.6% and recall at 68.7%; YOLOv5m had the highest initial precision, reaching 69.6% and a recall of 68.1%, whereas YOLOv5n had the lowest values (41.5% and 58.6%, respectively). All the models achieved accelerated gains in precision and recall for up to approximately epoch 10.
The mAP scores, shown in Figure 7b, demonstrated a similar trend. The YOLOv5m and YOLOv5l models begin with higher mAP_50 scores (66.8% and 71.8%, respectively); however, the YOLOv5l model outperforms YOLOv5m (93.4% and 92.6%, respectively) only at approximately epoch 18. However, until epoch 10, all the variants rapidly evolve in terms of mAP_50.
When the mAP50:95 metric was evaluated via Figure 7d, an interesting dynamic was observed between the YOLOv5l and YOLOv5m variants. The training started with YOLOv5m at an mAP50:95 of 38.7%, followed by YOLOv5l at 36.5%. Both methods achieved similar performances until epoch 9, when they reached mAP50:95 values of 64.5% and 66.3%, respectively. However, as the training progressed, YOLOv5l showed consistent improvement, achieving a remarkable score of 89.2% at the end of the training. However, YOLOv5m maintained considerable performance, scoring 86%. In contrast, YOLOv5s reached 80.6%, and YOLOv5n, despite improving over the epochs, lagged behind with only 73.5% mAP50:95.
Figure 8a–c illustrates the training losses, highlighting the effectiveness of YOLOv5 variants over 300 epochs in detecting myxozoans. Initially, all the models experienced abrupt losses, which decreased rapidly during the first 50 epochs and then gradually stabilized. Among them, YOLOv5n, being the least complex, showed greater losses and resulted in greater losses than did other more complex models, such as YOLOv5s, YOLOv5m, and YOLOv5l, which showed a better fit to the training data. This difference underscores the superior ability of more robust models to capture data patterns.
However, Figure 9a–c indicates losses during validation, which were not discrepant and followed a trend similar to that during training, with the losses rapidly decreasing in the first 50 epochs. The YOLOv5n model again presented the highest final losses, whereas the more complex models achieved better validation results.
These results reflected the superior generalization capacity of the more complex models, which is essential for effective performance on new data, as shown in Table 3, with precision, recall, mAP:50, and mAP:50–95 for the validation set.
Figure 10a shows precision and recall, while Figure 10b highlights precision and confidence, the performance of the models in distinguishing the genera Henneguya, Myxobolus, and background. Overall, the YOLOv5l model achieved a robust accuracy of 99.3% for Henneguya, indicating that almost all of its predictions were correct, even with high recall. For Myxobolus, a precision of 96.6% demonstrated high performance, though slightly lower than that of Henneguya. The average precision for all classes (mAP: 0.5) of 97.9% highlights the consistency of the model. In the recall–confidence curve, even with high confidence, the model maintains recall, indicating that its predictions are reliable.
The confusion matrices shown in Figure 11a–d for the YOLOv5n, YOLOv5s, YOLOv5l, and YOLOv5m variants provide an analysis of the performance for the detection of Henneguya, Myxobolus, and background. Furthermore, YOLOv5n correctly identified 4.89k (TP) instances of Henneguya, with 396 false positives (FPs), and 6.70k TPs for Myxobolus, with 954 FPs. Although it performs well in detection, the number of FPs can still be reduced. Additionally, YOLOv5s maintained balance, identifying 5.13k TPs for Henneguya with 274 FPs and 6.78k TPs for Myxobolus with 648 FPs. A notable reduction in false positives was observed compared with YOLOv5n.
Finally, YOLOv5m stood out with 5.23k TP for Henneguya and only 153 FPs and 6.82k TP for Myxobolus with 534 FPs, showing a good balance between precision and discrimination capability. However, YOLOv5l demonstrated significant superiority, with 5.29k TP for Henneguya and only 130 FPs and 6.83k TP for Myxobolus, with 433 FPs, indicating its higher precision.
Figure 12 provides a qualitative analysis of the detection results for the genera Myxobolus and Henneguya for some images not used in the proposed dataset. As depicted in the images, myxozoans of various shapes and positions were observed, though belonging to the genera Myxobolus and Henneguya, which were still detected with high precision. This demonstrates that the MLens can perform robust detections.

4. Conclusions

In the present study, a diverse collection of myxozoan images was obtained from the genera Henneguya and Myxobolus. This collection aimed to cover various situations and scenarios in light microscopy, providing a practical dataset for use in CV. Simultaneously, different versions of the YOLOv5 network were implemented to detect tar-infected parasites, which significantly improved the myxozoan detection accuracy.
Four networks were tested to automate the detection and differentiation of myxozoans and obtained satisfactory results, achieving an average precision of 97.9% for all genera, with a recall of 96.7% and an F1_Score of 97%. These results demonstrated the potential applicability of the MLens in real-world scenarios, indicating that machine learning in parasitology offers considerable benefits, making microparasite detection processes more efficient and less labor intensive.
From a forward-looking perspective, further advancements in web and mobile applications for the detection of myxozoans and parasites aimed at increasing the number of identifiable genera with continuous improvements in precision, accuracy, and real-time performance are expected. This is an important field for future research, including topics such as enhancing the quality and diversity of datasets and addressing the challenges in detecting multiple small targets in complex scenarios.
The MLens showed excellent results as a beta version, and further medium-term training, fine tuning of hyperparameters, and expansion of the collection are needed. This is the first software developed for the detection of myxozoans, which will hopefully serve as a gateway to the field of parasitology associated with AI. Building upon the success of this study, future work aims to develop more specific versions of the application, explore novel applications of the MLens, and refine its use for broader and more specialized parasitological diagnoses, fostering innovation in the intersection of artificial intelligence and parasitology.

Author Contributions

Conceptualization, G.S.C., J.L.S.-N., M.V.O.d.S. and A.d.S.L.d.S.; methodology, G.S.C., J.L.S.-N. and M.V.O.d.S.; software, G.S.C., J.L.S.-N. and M.V.O.d.S.; validation, G.S.C., J.L.S.-N. and M.V.O.d.S.; formal analysis, G.S.C., J.L.S.-N. and M.V.O.d.S.; investigation, G.S.C., J.L.S.-N. and M.V.O.d.S.; resources, G.S.C., J.L.S.-N., M.V.O.d.S. and A.d.S.L.d.S.; data curation, G.S.C., J.L.S.-N. and M.V.O.d.S.; writing—original draft preparation, G.S.C., J.L.S.-N. and M.V.O.d.S.; writing—review and editing, G.S.C., J.L.S.-N., M.V.O.d.S. and K.C.X.; visualization, G.S.C., J.L.S.-N., M.V.O.d.S. and K.C.X.; supervision, G.S.C., J.L.S.-N. and M.V.O.d.S.; project administration, G.S.C., J.L.S.-N. and M.V.O.d.S.; funding acquisition, A.d.S.L.d.S. All authors have read and agreed to the published version of the manuscript.

Funding

The APC for this study was funded by the Postgraduate Program in Society, Environment, and Quality of Life (PPGSAQ), Federal University of Western Pará (UFOPA), Santarém, PA, Brazil.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The final product, consolidated with the data supporting the results reported in this study, is openly available at https://github.com/gscproseg/master.git (accessed on 1 November 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lom, J.; Dyková, I. Myxozoan genera: Definition and notes on taxonomy, life-cycle terminology and pathogenic species. Folia Parasitol. 2006, 53, 1–36. [Google Scholar] [CrossRef]
  2. Fiala, I.; Bartošová-Sojková, P.; Whipps, C.M. Classification and phylogenetics of Myxozoa. In Myxozoan Evolution, Ecology and Development; Okamura, B., Gruhl, A., Bartholomew, J.L., Eds.; Springer: Cham, Switzerland, 2015; pp. 85–110. [Google Scholar] [CrossRef]
  3. Jones, S.R.M.; Bartholomew, J.L.; Zhang, J.Y. Mitigating Myxozoan disease impacts on wild fish populations. In Myxozoan Evolution, Ecology and Development; Okamura, B., Gruhl, A., Bartholomew, J., Eds.; Springer: Cham, Switzerland, 2015. [Google Scholar] [CrossRef]
  4. Lom, J.; Arthur, J.R. A guideline for the preparation of species descriptions in Myxosporea. J. Fish Dis. 1989, 12, 151–156. [Google Scholar] [CrossRef]
  5. Carvalho, A.A.; Videira, M.N.; Bittencourt, L.S.; Araújo, P.G.; Ferreira, R.L.S.; Tavares, J.C.; Matos, E.R. Infection of Henneguya sp. on the gills of Metynnis lippincottianus from Curiaú River, in eastern Amazon region (Brazil). Braz. J. Vet. Parasitol. 2020, 29, 003320. [Google Scholar] [CrossRef] [PubMed]
  6. Zago, A.C.; Vieira, D.H.M.D.; Franceschini, L.; Silva, R.J. Morphological, ultrastructural, and molecular analysis of a new species of Myxobolus (Cnidaria, Myxosporea) parasitizing Apareiodon piracicabae (Characiformes, Parodontidae) from Brazil. Parasitol. Int. 2022, 88, 102556. [Google Scholar] [CrossRef] [PubMed]
  7. Okamura, B.; Gruhl, A.; Bartholomew, J.L. An Introduction to Myxozoan Evolution, Ecology and Development. In Myxozoan Evolution, Ecology and Development; Okamura, B., Gruhl, A., Bartholomew, J.L., Eds.; Springer: Cham, Switzerland, 2015. [Google Scholar] [CrossRef]
  8. Jerônimo, G.T.; da Cruz, M.G.; Bertaglia, E.d.A.; Furtado, W.E.; Martins, M.L. Fish parasites can reflect environmental quality in fish farms. Rev. Aquac. 2022, 14, 1558–1571. [Google Scholar] [CrossRef]
  9. Lauringson, M.; Kahar, S.; Veevo, T.; Silm, M.; Philpott, D.; Svirgsden, R.; Rohtla, M.; Päkk, P.; Gross, R.; Kaart, T.; et al. Spatial and intrahost distribution of myxozoan parasite Tetracapsuloides bryosalmonae among Baltic sea trout (Salmo trutta). J. Fish Dis. 2023, 46, 1073–1083. [Google Scholar] [CrossRef] [PubMed]
  10. De Moraes, A.L.Z.; Barbosa, L.V.F.; Del Grossi, V.C.D. Artificial Intelligence and Human Rights: Contributions to a Regulatory Framework in Brazil; Editora Dialética: São Paulo, Brazil, 2022; ISBN 9786525253725. [Google Scholar]
  11. Siqueira-Batista, R.; Vitorino, R.R.; Gomes, A.P.; Oliveira, A.D.P.; Ferreira, R.D.S.; Esperidião-Antonio, V.; Santana, L.A.; Cerqueira, F.R. Artificial neural networks and medical education. Rev. Bras. De Educ. Médica 2014, 38, 548–556. [Google Scholar] [CrossRef]
  12. Chen, L.; Li, S.; Bai, Q.; Yang, J.; Jiang, S.; Miao, Y. Review of image classification algorithms based on convolutional neural networks. Remote Sens. 2021, 13, 4712. [Google Scholar] [CrossRef]
  13. Yonck, R. Heart of the Machine: Our Future in a World of Artificial Emotional Intelligence; Arcade: New York, NY, USA, 2020; ISBN 9781628727374. [Google Scholar]
  14. Rahimzadeh, M.; Attar, A. A modified deep convolutional neural network for detecting COVID-19 and pneumonia from chest X-ray images based on the concatenation of Xception and ResNet50v2. Inform. Med. Unlocked 2020, 19, 100360. [Google Scholar] [CrossRef] [PubMed]
  15. Yaacoub, J.P.; Noura, H.; Salman, O.; Chealb, A. Security analysis of drones systems: Attacks, limitations, and recommendations. Internet Things 2020, 11, 100218. [Google Scholar] [CrossRef] [PubMed]
  16. Noever, D.; Noever, S.E.; Miller, M. Hunting with machine vision. arXiv 2021. [Google Scholar] [CrossRef]
  17. Carvalho, A.C.P. Inteligência artificial: Riscos, benefícios e uso responsável. Estud. Avançados 2021, 35, 21–36. [Google Scholar] [CrossRef]
  18. Candiotto, K.B.B.; Karasinski, M. Inteligência artificial e os riscos existenciais reais: Uma análise das limitações humanas de controle. Filos. Unisinos 2022, 23, e23307. [Google Scholar] [CrossRef]
  19. Keszthelyi, S.; Pónya, Z.; Csóka, A.; Bázar, G.; Morschhauser, T.; Donkó, T. Non-destructive imaging and spectroscopic techniques to investigate the hidden-lifestyle arthropod pests: A review. J. Plant Dis. Prot. 2020, 127, 283–295. [Google Scholar] [CrossRef]
  20. Zhao, Z.; Tang, J.; Zhang, Z.; Li, L.; Ding, Y. When self-supervised learning meets scene classification: Remote sensing scene classification based on a multitask learning framework. Remote Sens. 2020, 12, 3276. [Google Scholar] [CrossRef]
  21. Li, R. Artificial Intelligence Revolution: How AI Will Change Our Society, Economy, and Culture; Skyhorse Publishing: New York, NY, USA, 2020; ISBN 9781510753006. [Google Scholar]
  22. Zendehdel, N.; Chen, H.; Leu, M.C. Real-time tool detection in smart manufacturing using you-only-look-once (YOLO) v5. Manuf. Lett. 2023, 35, 1052–1059. [Google Scholar] [CrossRef]
  23. Sichman, J.S. Artificial intelligence and society: Advances and risks. Estud. Avançados 2021, 35, 37–50. [Google Scholar] [CrossRef]
  24. Lee, K.W. Augmenting or Automating? Breathing Life into the Uncertain Promise of Artificial Intelligence. Ph.D. Thesis, New York University, New York, NY, USA, 2022. [Google Scholar]
  25. Okamura, B.; Hartigan, A.; Naldoni, J. Extensive uncharted biodiversity: The parasite dimension. Integr. Comp. Biol. 2018, 58, 1132–1145. [Google Scholar] [CrossRef] [PubMed]
  26. Eiras, J.C.; Barman, G.D.; Chanda, S.; Panigrahi, A.K. An update of the species of Myxosporea (Cnidaria, Myxozoa) described from Indian fish. J. Parasit. Dis. 2023, 47, 12–36. [Google Scholar] [CrossRef] [PubMed]
  27. Elshahawy, M.; Elnemr, A.; Oproescu, M.; Schiopu, A.G.; Elgarayhi, A.; Elmogy, M.M. Early melanoma detection based on a hybrid YOLOv5 and ResNet technique. Diagnostics 2023, 13, 2804. [Google Scholar] [CrossRef]
  28. Shen, L.; Lang, B.; Song, Z. Ca-YOLO: Model optimization for remote sensing image object detection. IEEE Access 2023, 11, 26438–26450. [Google Scholar] [CrossRef]
  29. Wang, J.; Chen, Y.; Dong, Z.; Gao, M. Improved YOLOv5 network for real-time multiscale traffic sign detection. Neural Comput. Appl. 2023, 35, 7853–7865. [Google Scholar] [CrossRef]
  30. Du, J. Understanding of object detection based on CNN family and YOLO. J. Phys. Conf. Ser. 2018, 1004, 012029. [Google Scholar] [CrossRef]
  31. Kumar, S.; Arif, T.; Ahamad, G.; Chaudhary, A.A.; Khan, S.; Ali, M.A.M. An efficient and effective framework for intestinal parasite egg detection using YOLOv5. Diagnostics 2023, 13, 2978. [Google Scholar] [CrossRef] [PubMed]
  32. Giribet, G.; Edgecombe, G. The Invertebrate Tree of Life; JSTOR: New York, NY, USA, 2020. [Google Scholar] [CrossRef]
  33. Zatti, S.A.; Marinho, A.M.R.; Adriano, E.A.; Maia, A.A.M. Integrative taxonomy reveals a panmictic population of Henneguya longisporoplasma n. sp. (Cnidaria: Myxozoa) in the Amazon Basin. Acta Parasitol. 2022, 67, 1644–1656. [Google Scholar] [CrossRef]
  34. Okamura, B.; Gruhl, A. Myxozoa + Polypodium: A common route to endoparasitism. Trends Parasitol. 2016, 32, 268–271. [Google Scholar] [CrossRef] [PubMed]
  35. Bydder, M.; Rahal, A.; Fullerton, G.D.; Cooper, T.G. The magic angle effect: A source of artifacts, a determinant of image contrast, and a technique for imaging. J. Magn. Reson. Imaging Off. J. Int. Soc. Magn. Reson. Med. 2007, 25, 290–300. [Google Scholar] [CrossRef]
  36. Ghose, P.; Ghose, A.; Sadhukhan, D.; Pal, S.; Mitra, M. Improved polyp detection from colonoscopy images using finetuned YOLO-v5. Multimed. Tools Appl. 2024, 83, 42929–42954. [Google Scholar] [CrossRef]
  37. Ray, S.; Alshouiliy, K.; Agrawal, D.P. Dimensionality reduction for human activity recognition using Google Colab. Information 2020, 12, 6. [Google Scholar] [CrossRef]
  38. Liu, Q.; Zhang, Y.; Yang, G. Small unopened cotton boll counting by detection with MRF-YOLO in the wild. Comput. Electron. Agric. 2023, 204, 107576. [Google Scholar] [CrossRef]
  39. Xu, X.; Zhang, X.; Zhang, T. Lite-YOLOv5: A lightweight deep learning detector for on-board ship detection in large-scene Sentinel-1 SAR images. Remote Sens. 2022, 14, 1018. [Google Scholar] [CrossRef]
  40. Zhang, J.; Chen, Z.; Yan, G.; Wang, Y.; Hu, B. Faster and lightweight: An improved YOLOv5 object detector for remote sensing images. Remote Sens. 2023, 15, 4974. [Google Scholar] [CrossRef]
  41. Badgujar, C.M.; Poulose, A.; Gan, H. Agricultural object detection with You Only Look Once (YOLO) algorithm: A bibliometric and systematic literature review. Comput. Electron. Agric. 2024, 223, 109090. [Google Scholar] [CrossRef]
  42. Wu, D.; Jiang, S.; Zhao, E.; Liu, Y.; Zhu, H.; Wang, W.; Wang, R. Detection of Camellia oleifera fruit in complex scenes via YOLOv7 and data augmentation. Appl. Sci. 2022, 12, 11318. [Google Scholar] [CrossRef]
  43. Zhang, Y.; Wang, H.; Xu, R.; Yang, X.; Wang, Y.; Liu, Y. High-precision seedling detection model based on a multiactivation layer and depth-separable convolution using images acquired by drones. Drones 2022, 6, 152. [Google Scholar] [CrossRef]
  44. Isa, I.S.; Rosli, M.S.A.; Yusof, U.K.; Maruzuki, M.I.F.; Sulaiman, S.N. Optimizing the hyperparameter tuning of YOLOv5 for underwater detection. IEEE Access 2022, 10, 52818–52831. [Google Scholar] [CrossRef]
  45. Benjumea, A.; Aduen, I.; Cuzzolin, F.; Bradley, A. YOLO-Z: Improving small object detection in YOLOv5 for autonomous vehicles. arXiv 2021. [Google Scholar] [CrossRef]
  46. Wang, A.; Peng, T.; Cao, H.; Xu, Y.; Wei, X.; Cui, B. Tia-YOLOv5: An improved YOLOv5 network for real-time detection of crop and weed in the field. Front. Plant Sci. 2022, 13, 1091655. [Google Scholar] [CrossRef] [PubMed]
  47. Hasan, H.; Saad, F.; Ahmed, S.; Mohammed, N.; Farook, T.H.; Dudley, J. Experimental validation of computer-vision methods for the successful detection of endodontic treatment obturation and progression from noisy radiographs. Oral Radiol. 2023, 39, 683–698. [Google Scholar] [CrossRef]
  48. Huang, R.; Pedoeem, J.; Chen, C. YOLO-LITE: A real-time object detection algorithm optimized for non-GPU computers. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 2503–2510. [Google Scholar] [CrossRef]
  49. Redmon, J. YOLO: Real-Time Object Detection. Available online: http://pjreddie.com/yolo/ (accessed on 1 November 2024).
  50. Tong, K.; Wu, Y. Deep learning-based detection from the perspective of small or tiny objects: A survey. Image Vis. Comput. 2022, 123, 104471. [Google Scholar] [CrossRef]
  51. Fu, X.; Wei, G.; Yuan, X.; Liang, Y.; Bo, Y. Efficient YOLOv7-Drone: An enhanced object detection approach for drone aerial imagery. Drones 2023, 7, 616. [Google Scholar] [CrossRef]
  52. Cai, H.; Shangguan, H.; Wang, L. PC-Yolo: Enhanced YOLOv5-based defect detection system with improved partial convolution for ham sausage inspection. In Proceedings of the Fourth International Conference on Computer Vision and Pattern Analysis (ICCPA 2024), Anshan, China, 17–19 May 2024; pp. 451–459. [Google Scholar] [CrossRef]
  53. Ma, S.; Lu, H.; Liu, J.; Zhu, Y.; Sang, P. LAYN: Lightweight Multi-Scale Attention YOLOv8 Network for Small Object Detection. IEEE Access 2024, 12, 29294–29307. [Google Scholar] [CrossRef]
  54. Geng, X.; Su, Y.; Cao, X.; Li, H.; Liu, L. YOLOFM: An improved fire and smoke object detection algorithm based on YOLOv5n. Sci. Rep. 2024, 14, 4543. [Google Scholar] [CrossRef] [PubMed]
  55. Xiong, C.; Zayed, T.; Abdelkader, E.M. A novel YOLOv8-GAMWise-IoU model for automated detection of bridge surface cracks. Constr. Build. Mater. 2024, 414, 135025. [Google Scholar] [CrossRef]
Figure 1. Life cycle of myxosporeans. A—Annelid or Ectoprocta (definitive host); B—Actinosporidian; C—Fish (intermediate host); D—Myxosporean.
Figure 1. Life cycle of myxosporeans. A—Annelid or Ectoprocta (definitive host); B—Actinosporidian; C—Fish (intermediate host); D—Myxosporean.
Parasitologia 05 00050 g001
Figure 2. Pipeline for data collection and dataset creation: Schematic representation of the steps involved in collecting parasitological data from fish, obtaining microscopic images, and labeling objects of interest for model training.
Figure 2. Pipeline for data collection and dataset creation: Schematic representation of the steps involved in collecting parasitological data from fish, obtaining microscopic images, and labeling objects of interest for model training.
Parasitologia 05 00050 g002
Figure 3. Object parameters in images—coordinates and dimensions (x, y, width, height) of objects detected in an image.
Figure 3. Object parameters in images—coordinates and dimensions (x, y, width, height) of objects detected in an image.
Parasitologia 05 00050 g003
Figure 4. Statistical Analysis of Labels—Histogram of Genus and Density Map of Bounding Box Sizes.
Figure 4. Statistical Analysis of Labels—Histogram of Genus and Density Map of Bounding Box Sizes.
Parasitologia 05 00050 g004
Figure 5. Statistical analysis of labels: histogram of genus (a) and density map of bounding box sizes (b).
Figure 5. Statistical analysis of labels: histogram of genus (a) and density map of bounding box sizes (b).
Parasitologia 05 00050 g005
Figure 6. YOLOv5 (s, n, m and l), training time (TT), model size (TM), and GPU memory usage (UM), features normalized by the number of epochs.
Figure 6. YOLOv5 (s, n, m and l), training time (TT), model size (TM), and GPU memory usage (UM), features normalized by the number of epochs.
Parasitologia 05 00050 g006
Figure 7. YOLOv5 models—(a) precision across epochs, (b) recall across epochs, (c) mAP@0.5 across epochs, and (d) mAP@0.5:0.95 across epochs.
Figure 7. YOLOv5 models—(a) precision across epochs, (b) recall across epochs, (c) mAP@0.5 across epochs, and (d) mAP@0.5:0.95 across epochs.
Parasitologia 05 00050 g007
Figure 8. YOLOv5 models—(a) training box loss across epochs, (b) training classification loss across epochs, (c) training object loss across epochs.
Figure 8. YOLOv5 models—(a) training box loss across epochs, (b) training classification loss across epochs, (c) training object loss across epochs.
Parasitologia 05 00050 g008
Figure 9. YOLOv5 models—(a) validation box loss across epochs, (b) validation classification loss across epochs, (c) validation object loss across epochs.
Figure 9. YOLOv5 models—(a) validation box loss across epochs, (b) validation classification loss across epochs, (c) validation object loss across epochs.
Parasitologia 05 00050 g009
Figure 10. Performance of YOLOv5 models in distinguishing between Henneguya, Myxobolus, and Background. (a) Precision and recall, showing the balance between true positive detections and model robustness, analyzed alongside the YOLOv5l model confusion matrix Figure 11c. (b) Precision and confidence, illustrating the model’s certainty in detecting each genus relative to the background, for the YOLOv5l model.
Figure 10. Performance of YOLOv5 models in distinguishing between Henneguya, Myxobolus, and Background. (a) Precision and recall, showing the balance between true positive detections and model robustness, analyzed alongside the YOLOv5l model confusion matrix Figure 11c. (b) Precision and confidence, illustrating the model’s certainty in detecting each genus relative to the background, for the YOLOv5l model.
Parasitologia 05 00050 g010
Figure 11. Confusion matrix for parameter detection with YOLOv5—performance of different versions of the YOLOv5 model (n, s, l, m) in detecting the Henneguya, Myxobolus, and background categories.
Figure 11. Confusion matrix for parameter detection with YOLOv5—performance of different versions of the YOLOv5 model (n, s, l, m) in detecting the Henneguya, Myxobolus, and background categories.
Parasitologia 05 00050 g011
Figure 12. MLens-processed images: accurate detection and counting of parasites from the Henneguya and Myxobolus genus.
Figure 12. MLens-processed images: accurate detection and counting of parasites from the Henneguya and Myxobolus genus.
Parasitologia 05 00050 g012
Table 1. Labeling scheme in text file—X and Y are the center coordinates of the target, and W and H are the width and height of the target, respectively.
Table 1. Labeling scheme in text file—X and Y are the center coordinates of the target, and W and H are the width and height of the target, respectively.
GeneraXYWH
10.725000000.696093750.117187500.79609375
10.831250000.784375000.073437500.48437500
00.441406250.647656250.106250000.74765625
Table 2. Information on each of the genera labeled for YOLO: Henneguya and Myxobolus.
Table 2. Information on each of the genera labeled for YOLO: Henneguya and Myxobolus.
Category Information Genera Labeled for YOLOv5
Henneguya0 Myxobolus1
Table 3. YOLOv5 model performance by each genus.
Table 3. YOLOv5 model performance by each genus.
ModelGenusImagesParasitesPrecisionRecallmAP:50mAP:50–95
YOLOv5lAll200012,3860.9740.9670.9800.892
Henneguya200053960.9940.9700.9930.905
Myxobolus200069900.9540.9650.9660.879
YOLOv5mAll200012,3860.9720.9570.9760.860
Henneguya200053960.9950.9510.9900.876
Myxobolus200069900.9490.9620.9620.844
YOLOv5nAll200012,3860.9280.8950.9430.727
Henneguya200053960.9760.8520.9470.74
Myxobolus200069900.8800.9380.9390.713
YOLOv5sAll200012,3860.9480.9350.9640.954
Henneguya200053960.9850.9150.9740.815
Myxobolus200069900.9110.9540.9540.780
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Carneiro, G.S.; Xavier, K.C.; Sindeaux-Neto, J.L.; Lima da Silva, A.d.S.; Oliveira da Silva, M.V. MLens: Advancing the Real-Time Detection, Identification, and Counting of Pathogenic Microparasites Through a Web Interface. Parasitologia 2025, 5, 50. https://doi.org/10.3390/parasitologia5040050

AMA Style

Carneiro GS, Xavier KC, Sindeaux-Neto JL, Lima da Silva AdS, Oliveira da Silva MV. MLens: Advancing the Real-Time Detection, Identification, and Counting of Pathogenic Microparasites Through a Web Interface. Parasitologia. 2025; 5(4):50. https://doi.org/10.3390/parasitologia5040050

Chicago/Turabian Style

Carneiro, Gustavo Souza, Karoliny Caldas Xavier, José Ledamir Sindeaux-Neto, Alanna do Socorro Lima da Silva, and Michele Velasco Oliveira da Silva. 2025. "MLens: Advancing the Real-Time Detection, Identification, and Counting of Pathogenic Microparasites Through a Web Interface" Parasitologia 5, no. 4: 50. https://doi.org/10.3390/parasitologia5040050

APA Style

Carneiro, G. S., Xavier, K. C., Sindeaux-Neto, J. L., Lima da Silva, A. d. S., & Oliveira da Silva, M. V. (2025). MLens: Advancing the Real-Time Detection, Identification, and Counting of Pathogenic Microparasites Through a Web Interface. Parasitologia, 5(4), 50. https://doi.org/10.3390/parasitologia5040050

Article Metrics

Back to TopTop