Abstract
Algal blooms have seriously affected the production and life of people and real-time detection of algae in water samples is a powerful measure to prevent algal blooms. The traditional manual detection of algae with a microscope is extremely time-consuming. In recent years, although there have been many studies using deep learning to classify and detect algae, most of them have focused on the relatively simple task of algal classification. In addition, some existing algal detection studies not only use small datasets containing limited algal species, but also only prove that object detection algorithms can be applied to algal detection tasks. These studies cannot implement the real-time detection of algae and timely warning of algal blooms. Therefore, this paper proposes an efficient self-organized detection system for algae. Benefiting from this system, we propose an interactive method to generate the algal detection dataset containing 28,329 images, 562,512 bounding boxes and 54 genera. Then, based on this dataset, we not only explore and compare the performance of 10 different versions of state-of-the-art object detection algorithms for algal detection, but also tune the detection system we built to its optimum state. In practical application, the system not only has good algal detection results, but also can complete the scanning, photographing and detection of a 2 cm × 2 cm, 0.1 mL algal slide specimen within five minutes (the resolution is 0.25886 m/pixel); such a task requires a well-trained algal expert to work continuously for more than three hours. The efficient algal self-organized detection system we built makes it possible to detect algae in real time. In the future, with the help of IoT, we can use various smart sensors, actuators and intelligent controllers to achieve real-time collection and wireless transmission of algal data, use the efficient algal self-organized detection system we built to implement real-time algal detection and upload the detection results to the cloud to realize timely warning of algal blooms.
1. Introduction
Algal blooms such as red tide [1,2] and cyanobacterial blooms [3] are occurring more and more frequently around the world and they not only devastate ecosystems but also seriously damage human health. For example, the blooms of dinoflagellate karenia brevis are thought to be capable of killing large numbers of fish and causing significant economic losses [4]. In 2005, the bloom of karenia in the Gulf of Mexico resulted in more than a month of benthic mortality [5]. In 2007, a cyanobacteria bloom in Wuxi, Jiangsu Province, caused a drinking water crisis for 2 million people [6], and a large cyanobacteria bloom took place in the western part of Lake Erie, Ohio, which cut off water supplies to over 500,000 people in 2014 [7]. Therefore, real-time monitoring of algae is necessary to prepare for and even prevent algal blooms before they occur, and how to quickly detect algae from acquired sample images is the core issue of real-time algal monitoring. The traditional manual detection of algae with a microscope is extremely time-consuming and not time-efficient.
In recent years, convolutional neural networks (CNNs) have developed rapidly in the fields of image classification, object detection and semantic segmentation, and have achieved great success in many fields, such as face recognition [8] and autonomous driving [9,10]. CNNs are mainly used for feature extraction through convolution operation to achieve the purpose of images classification or objection detection. Due to the fast speed of convolution operation, CNNs are very effective in dealing with large-scale datasets and under the drive of a certain scale of data, CNNs can achieve unexpected good results. Therefore, more and more algal researchers have begun to apply CNNs to the classification and detection of algae. Moreover, the acquisition and collection of algal images are difficult due to their obvious regional nature. Most algal researchers use data augmentation methods to expand the numbers of algal images [11,12,13,14]. These extended algal datasets can easily enable classification algorithms to achieve accuracy of more than 99% [11,13,15] and average precision of more than 80% [15,16], which leads to the classification and detection performance of CNN on algal dataset not being able to be well mined. Furthermore, most algal researchers focus on algal classification and there are few studies on algal detection. The reason is that the algal classification task requires the CNN to achieve the correct output of the category for the input image containing a single algal species, while the algal detection task requires the CNN to achieve the correct output of the category and location for the input image containing multiple algal species. Compared with algal classification, algal detection requires more complex dataset and algorithms [17]. The existing literature on algal detection not only uses a small dataset containing limited algal species, but also only proves that object detection algorithms can be applied to the algal detection task. In other words, the existing studies cannot realize the real-time algal detection and timely warning before algal blooms occur.
Therefore, based on the object detection algorithm, we construct an efficient self-organized detection system for algae. In order to better build and debug the algal self-organized detection system, we use an interactive method to generate the algal detection dataset containing 28,329 images, 562,512 bounding boxes and 54 genera. Based on this algal detection dataset, we explore and compare 10 different versions of state-of-the-art object detection algorithms. The experimental results show that under the same dataset and training conditions, YOLOv7 has the best detection results. We will consider replacing the YOLOv5 embedded in the algal self-organized detection system with YOLOv7. In practical application, the system not only has good algal detection results, but also can complete the scanning, photographing and detection of a 2 cm × 2 cm, 0.1 mL algal slide specimen within five minutes (resolution is 0.25886 m/pixel); such a task requires a well-trained algal expert to work continuously for more than three hours.
At present, the collection and transmission of algal data depend on manual completion, which hinders the real-time detection of algae. The Internet of Things (IoT), which aims to enable ubiquitous wireless connections among various smart sensors, actuators and intelligent controllers and then integrate their functions to realize the mutual sharing and interaction of information, can be used to collect and transmit algal data in real-time. With the help of the IoT and the algal self-organized detection system built by us, real-time detection of algae can be well realized in the future. At the same time, we can upload the results to the cloud through the IoT, which is used as the basis for algal bloom warning. The contribution of this paper can be stated as follows:
- An algal self-organized detection system is established, which not only guarantees the good detection results, but also realizes the rapid detection of algae;
- An interactive method for generating an algal detection dataset is proposed;
- The detection performance of 10 different versions of state-of-the-art object detection algorithms is compared on the algal detection dataset.
2. Related Works
Object detection algorithm: As we know, object detection is one of the core research contents of computer vision. Its purpose is to predict the position of objects in a given image and label each object with a corresponding category. In the early stage, object detection is divided into three steps: generating region proposal, extracting features and region classification [18]. The most representative algorithm of this method is the deformable part-based model (DPM) algorithm [19] extended on histogram of oriented gradients (HOG) [20], which won the 2007, 2008, 2009 detection challenge on Pascal VOC dataset [21].
After CNN achieved a higher accuracy of image classification [22], a wave of research on object detection using CNN was set off. Currently, object detection based on deep learning has formed two factions: two-stage detection and one-stage detection. In the two-stage detection framework, the first stage generates the candidate region proposals and uses CNN for feature extraction. Then, the second stage uses a specific classifier to predict the categories of the candidate region proposals. The most representative algorithm is R-CNN [23] and its series of optimization deformation algorithms, such as Fast R-CNN [24], Faster-RCNN [25], Feature Pyramid Network (FPN) [26]. In the one-stage detection framework, all locations of the whole image are regarded as potential objects by default, the bounding boxes and categories of the objects are predicted simultaneously on the feature map. Its representative algorithm is You Only Look Once (YOLO). Since Joseph et al. [27] first proposed the YOLO algorithm in 2015, there have been seven versions of YOLO, namely YOLOv1-YOLOv7 [27,28,29,30,31,32,33].
CNN application in algal identification: Recently, with the remarkable success of CNN in various industries, a large number of algal researchers using CNN to identify algae have emerged. Pedraza et al. [11] first applied CNN to the classification of 80 diatoms, using the data augmentation approach to expand the dataset to over 160,000 samples and finally achieved an overall accuracy of 0.99 in AlexNet. Park et al. [12] used the neural architecture search (NAS) technology to design a CNN suitable for algal image classification and obtained an F1-score of 0.95 on eight algal genera. In addition, they also discussed the impact of data enhancement on classification. Several repeated experimental results show that the classification results after data enhancement are lower than those using the original dataset. Yadav et al. [13] used data augmentation techniques to expand the algal dataset 100 images to 80,000 images. Based on the expanded dataset of 16 algal families, ResNeXt was modified and a classification accuracy of 0.9997 was finally achieved. Xu et al. [14] expanded 13 algal species through data enhancement, forming a relatively balanced dataset among different algal species. Based on this dataset, they designed a new CNN algorithm, which obtained the lowest classification probability of 0.939.
While the above studies mainly focus on the classification of algae, the detection of algae has also attracted the attention of algal researchers. Samantaray et al. [34] proposed a viable algal monitoring system that uses transfer learning techniques to test three object detection algorithms, Faster R-CNN, Single Shot Detector (SSD) and Region-based Fully Convolutional Networks (R-FCN), on two datasets of hundreds of ground algal images and aerial algal images. The final monitoring system chooses the more robust, accurate and faster reasoning R-FCN algorithm. Baek et al. [16] used R-CNN and CNN to detect and count five cyanobacteria species and the average precision values of the final detection were between 0.89 and 0.929. Qian et al. [35] proposed a new object detection network based on faster R-CNN and tested the new network on the algal dataset containing 27 genera, achieving 0.7464 mean average precision (mAP). Park et al. [17] compared YOLOv3 with YOLOv4 on a dataset of 437 images containing 30 algal genera and showed that YOLOv4 performed better. Salido et al. [15] proposed a low-cost automated digital microscopy platform for the automatic classification and detection of diatoms. On a dataset containing 80 species of algae, they achieved a detection accuracy of 0.86 using YOLO and a classification accuracy of 0.9951 using AlexNet. Ali et al. [36] applied deep convolutional generative adversarial neural (DC-SGAN) to expand the dataset containing four types of algae and carried out comparative experiments on YOLOv3, YOLOv4 and YOLOv5 on the expanded dataset. The comparison results show that YOLOv5 has the best performance.
Different from the above studies, based on the object detection algorithm, we build the efficient algal self-organized detection system, which can automatically scan the algal slide specimens and realize the algal self-organized detection. The system is designed to achieve real-time detection of algae, so as to give a timely warning of algal blooms.
3. Material and Methods
In this section, we first introduce the built algal self-organized detection system. Then, we present the data acquisition and pre-processing, give the specific steps of interactive algal detection dataset generation method and show the detailed information of the algal detection dataset we have established. Finally, the evaluation standard of object detection is proposed.
3.1. Algal Self-Organized Detection System
The algal self-organized detection system is mainly composed of the Algae-Hub Algae Artificial Intelligence Analyser (AH-20-S, Jiangsu Metabio Science & Technology Co., Ltd., Wuxi, China), data analysis workstation, display, mouse and keyboard. The system we built and the functions of its major components are shown in Figure 1.
Figure 1.
The demonstration picture of the algal self-organized detection system and the functions of its major components.
The Algae-Hub consists of an objective lens and an imaging camera, accepting 20 mm × 20 mm, 0.1 mL algal specimens in a slide. The scan magnification of the Algae-Hub is 20× or 40×, the resolution is 0.25886 m/pixel and the scanning speed is less than 120 s. It can realize automatic focus or manually fine-tune the focus. In addition, the embedded camera is a 5-megapixel CMOS camera.
The data analysis workstation is a computer installed with a win10 operating system, and also installed with Artificial Intelligence Analyser professional analysis software. It analyzes the images generated by the Algae-Hub independently. The methods of analyzing the images include the visual method, the diagonal method, the lattice method and the whole section method. By default, the entire image is evenly divided into 100 square grids. The visual method analyzes a selected square grid, the diagonal method analyzes the diagonal square grids of the 100 square grids, the lattice method selects the square grids of some rows for analysis and the whole section method analyzes the entire image. After the analysis is completed, the workstation displays the name and number of algal species identified in the selected square grids and saves the image detection results in the workstation. For the detection results, we can view, verify and modify. Once we have optimized the system, we just put in algal slide specimens or algal images, choose an analysis method and the system can realize self-organized algal detection. In addition, the workstation integrates functions such as statistical algal density and distribution. We can also continue to add features as needed in the future.
In the context of the IoT, we focus on using various smart sensors, actuators and intelligent controllers to achieve real-time acquisition and sharing of algal data in the future. Combined with the efficient algal self-organized detection system we built, the algal self-organized detection system under the IoT is formed to implement the real-time detection of algae and the interaction of detection results, so as to realize a timely warning of algal blooms. In order to better display the specific workflow of the algal self-organized detection system under the IoT in the future, we present the schematic diagram in Figure 2.
Figure 2.
The schematic diagram of the algal self-organized detection system under the IoT.
It is worth noting that the core of the detection system is the object detection algorithm integrated in the system. As we know, although the two-stage detection method has good detection performance, the detection speed is far slower than that of the one-stage detection method. Therefore, we chose the YOLO series in the one-stage detection method to accomplish the task of real-time object detection. While there are many object detection algorithms in the YOLO series, YOLOv5 is one of the most stable and widely used. Initially, the algal self-organized detection system we built was based on YOLOv5, whose performance meets our requirements.
3.2. Dataset Acquisition and Pre-Processing
We collect water samples through several pilot sites in Taihu Lake in Wuxi, China. For the water samples with algal cells gathered together, the algal cells are scattered as much as possible by shaking or ultrasonic crushing, while for the water samples with large density of algal cells, appropriate dilution is carried out.
The processed water samples are made into slide specimens and the algal images are automatically scanned and saved using our proposed algal self-organized detection system. It is worth noting that the algal detection dataset built by us is generated interactively and the specific steps are as follows:
- (i)
- For the slide specimens, the algal self-organized detection system automatically scans to obtain algal images.
- (ii)
- The algal images acquired in (i) are cross-labeled by 15 algal experts using labelImg annotation software with reference to the VOC dataset format and the annotation files are saved. Then, the initial algal detection dataset is generated.
- (iii)
- The algal detection dataset generated in (ii) is trained by the object detection algorithm to obtain the optimal weight and the optimal weight is imported into the algal self-organized detection system.
- (iv)
- For the new slide specimens or images, the algal self-organized detection system implements automatic scanning, analysis and export of the analysis result images.
- (v)
- The algal images analyzed in (iv) are manually checked and combine with the algal detection data produced in (ii) to generate a new algal detection dataset.
- (vi)
- Repeat (iii)–(v) to finally obtain the algal detection dataset used in this paper.
Among the above steps to interactively generate the algal dataset, the dataset in (ii) is used to start and tune the algal self-organized detection system we have built. New samples are detected in (iv) to expand and enrich the algal detection dataset. The manual check in (v) is to eliminate the detection errors produced by the algal self-organized detection system. Through interactive data generation, we can quickly obtain a considerable number of algal detection datasets with relatively accurate labels and bounding boxes. Based on this dataset, we can adjust the algal self-organized detection system to the optimal state, so as to detect algae more quickly and accurately.
For our algal dataset, each image has a width and height of pixels between 1536 and 1984. Due to the inconsistent number of bounding boxes for various algae, we select algae with more than 200 bounding boxes as the final algal detection dataset. Therefore, we obtain an algal detection dataset consisting of 28,329 images, 562,512 bounding boxes and 54 genera. We randomly split the dataset into three parts, training set, validation set and test set and their ratio is 7:2:1. The specific information of the datasets is presented in Table 1.
Table 1.
Statistics of the algal detection dataset.
As shown in Table 1, we give the names of 54 algae that reach the genus level, the number of images for each algal species and the number of bounding boxes for each algal species after data division. Note that there are multiple algal genera in an image, the sum of the images of each algal genus is not the total number of algal images. To better display the algal detection dataset constructed in this paper, we randomly select 12 images, draw their bounding boxes and categories and show them in Figure 3.
Figure 3.
Twelve algal detection images with annotation information. Images are randomly selected from the algal detection dataset generated via the interactive method. Each image corresponds to an annotation file, which stores the location information and the category of algae. We use red rectangles and blue fonts to display the bounding boxes and categories on the chosen images.
3.3. Detection Evaluation
As we know, the object detection goal is to find the location of the object in the image and give the corresponding label. For the predicted object locations, we use intersection over union (IOU) to evaluate the accuracy, which is calculated by the following formula:
where represents the area of the predicted bounding box and presents the area of the ground truth bounding box [21]. When the value of the IOU between the predicted bounding box and the ground truth bounding box is greater than the preset IOU threshold, the location detection is correct. Otherwise, it is regarded as missed detection.
For the predicted labels, average precision (AP) is used for evaluation. AP is the area under the precision–recall curve calculated by interpolation. The precision (P) and recall (R) are calculated as follows:
where TP is true positive, representing the number of true positive samples that are correctly predicted. FP is false positive, representing the number of true positive samples that are incorrectly predicted. FN is false negative, representing the number of true negative samples that are incorrectly predicted.
In image classification, positive samples refer to the samples of the current category, while negative samples refer to the remaining samples that do not belong to the current category. In object detection, the definition of positive and negative samples is more complex and even varies in different periods. For example, YOLOv3 [29] uses the dual IOU thresholds strategy; that is, the thresholds are 0.3 and 0.7 respectively. For a predicted bounding box, if its IOU with the ground truth bounding box is greater than 0.7, it is considered as a positive sample; if its IOU with the ground truth bounding box is between 0.3 and 0.7, it is ignored. If its IOU with the ground truth bounding box is less than 0.3, it is considered as a negative sample. YOLOv4 [30] states that for a predicted bounding box, if its IOU with the ground truth bounding box is greater than the preset threshold, it is a positive sample. Otherwise, it is a negative sample. In general, we use the mean AP (mAP) of the average over all objects as the indicator to evaluate the prediction label.
In order to take into account both the evaluation of the predicted location and label, mAP with an IOU of 0.5 (mAP@.5) and an average mAP of 10 different IOU thresholds with a step size of 0.05 between 0.5 and 0.95 (mAP@.5:.95) is commonly used to evaluate the performance of the object detection algorithm. Under the same dataset and training environment, the larger the mAP@.5 and mAP@.5:.95 values, the better the detection performance of the object detection algorithm.
4. Experiments
Recently, YOLOv6 and YOLOv7 appeared, which performs better than YOLOv5 on the COCO dataset. Therefore, this section gives the parameters and equipment for experiment implementation, as well as the detection results of 10 different versions of YOLOv5, YOLOv6 and YOLOv7 on the algal detection dataset.
4.1. Implementation
The experiments in this paper are based on pytorch 1.12, running on an ubuntu operating system with Intel(R) Core(TM) i9-12900k, a 3090 24GB GPU. Before providing the data to the network, we only convert the annotations to the format that YOLO needs. We use the default hyperparameter settings of YOLOv5, YOLOv6 and YOLOv7. We set the image size to pixels and the batch size to 24. Each model is trained from scratch for 300 epochs and does not use pre-trained weights. When testing the training model, the confidence threshold is set to 0.001 for YOLOv5 and YOLOv7 and 0.03 for YOLOv6. All the IOU thresholds are set to 0.6.
4.2. Detection Results
Based on the algal detection dataset we built, 10 different versions of YOLOv5, YOLOv6 and YOLOv7 are trained on the training set and verified on the validation set, respectively. After the training, the parameter model with the best performance on the validation set is selected for the final test on the test set. Table 2 shows the detection results.
Table 2.
Comparison of detection results on algal dataset.
It can be seen from Table 2 that under the same input condition of pixels, under the same detection model framework, the larger the model, the higher the mAP@.5 and mAP@.5:.95 and the smaller the frames per second (FPS). Hence, YOLOv5-L, YOLOv6-M and YOLOv7 achieve the highest mAP@.5 and mAP@.5:.95 with their respective frameworks. Among 10 different versions of YOLOv5, YOLOv6 and YOLOv7, YOLOv7 has the highest mAP@.5, but YOLOv5 has the highest mAP@.5:.95. Compared with YOLOv5-L and YOLOv6-M, YOLOv7 has the fastest frames per second (FPS). Therefore, we are considering replacing the YOLOv5-L algorithm embedded in the algal self-organized detection system with YOLOv7. Moreover, from Table 2, we can see that the detection speed (FPS) of 10 object detection algorithms is far higher than that of manual algal detection, which can well realize real-time algal detection. In order to better present the detection effect of the trained model in the test set, we set the confidence threshold to 0.25 and the IOU threshold to 0.45. Then, we randomly select four images and present them together with the corresponding manually annotated image, detection results of YOLOv5-L, YOLOv6-M and YOLOv7 in Figure 4.
Figure 4.
Schematic of four randomly selected images in the test set and the corresponding detection results. The four images in each row from left to right are the original image, the corresponding manually annotated image and the final detection results of YOLOv5-L, YOLOv6-M and YOLOv7.
From Figure 4, YOLOv5-L, YOLOv6-M and YOLOv7 can detect the algae that are not manually labeled although there are few missed algae. This fully proves that the object detection algorithm can surpass the manual algal detection method in algal detection. Meanwhile, Table 2 and Figure 4 also fully demonstrate the high efficiency of the algal self-organized detection system built by us based on YOLOv5.
5. Conclusions
In this paper, we first present the algal self-organized detection system we built. Then, we used an interactive method to generate an algal detection dataset containing 28,329 images, 562,512 bounding boxes and 54 genera. Based on the algal detection dataset, 10 different versions of YOLOv5, YOLOv6 and YOLOv7 were compared. The experimental results show that under the same dataset and training conditions, the detection performance of YOLOv7 is better than that of YOLOv5-L and YOLOv6-M. We will consider using YOLOv7 to replace YOLOv5-L embedded in the algal self-organized detection system. Meanwhile, in practical applications, the efficient algal self-organized detection system embedded with YOLOv5-L can realize rapid and accurate detection of algae. However, the current system relies on manual data acquisition and transmission, which hinders real-time detection of algae. In the future, we will focus on using intelligent sensors to realize real-time collection and sharing of algal data and combine with the algal self-organized detection system we have built to form an algal self-organized detection system under the IoT, so as to realize timely warning of algal blooms. This is one of our future major works.
Author Contributions
Conceptualization, X.G. and C.M.; methodology, B.S. and J.Z.; software, X.G.; validation, C.M., B.S. and J.Z.; formal analysis, C.M.; investigation, B.S.; resources, C.M.; data curation, J.Z.; writing—original draft preparation, X.G.; writing—review and editing, B.S.; visualization, J.Z.; supervision, C.M.; project administration, J.Z.; funding acquisition, C.M. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Science and Technology Development Fund, Macau SAR (No. 0019/2021/A1).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Due to non-disclosure agreements, the dataset in this paper are not publicly available. The dataset can be made available to real algal researchers, but only if they sign non-disclosure agreements.
Conflicts of Interest
The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.
References
- Harvell, C.D.; Kim, K.; Burkholder, J.; Colwell, R.R.; Epstein, P.R.; Grimes, D.J.; Hofmann, E.E.; Lipp, E.; Osterhaus, A.; Overstreet, R.M.; et al. Emerging marine diseases–climate links and anthropogenic factors. Science 1999, 285, 1505–1510. [Google Scholar] [CrossRef] [PubMed]
- Van Dolah, F.M. Marine algal toxins: Origins, health effects and their increased occurrence. Environ. Health Perspect. 2000, 108, 133–141. [Google Scholar] [CrossRef] [PubMed]
- Landsberg, J.H. The effects of harmful algal blooms on aquatic organisms. Rev. Fish. Sci. 2002, 10, 113–390. [Google Scholar] [CrossRef]
- Davis, C.C. Gymnodinium brevis sp. nov., a cause of discolored water and animal mortality in the Gulf of Mexico. Bot. Gaz. 1948, 109, 358–360. [Google Scholar] [CrossRef]
- Heil, C.A.; Steidinger, K.A. Monitoring, management and mitigation of Karenia blooms in the eastern Gulf of Mexico. Harmful Algae 2009, 8, 611–617. [Google Scholar] [CrossRef]
- Qin, B.; Zhu, G.; Gao, G.; Zhang, Y.; Li, W.; Paerl, H.W.; Carmichael, W.W. A drinking water crisis in Lake Taihu, China: Linkage to climatic variability and lake management. Environ. Manag. 2010, 45, 105–112. [Google Scholar] [CrossRef]
- Paerl, H.W.; Gardner, W.S.; McCarthy, M.J.; Peierls, B.L.; Wilhelm, S.W. Algal blooms: Noteworthy nitrogen. Science 2014, 346, 175. [Google Scholar] [CrossRef]
- Sun, Y.; Liang, D.; Wang, X.; Tang, X. Deepid3: Face recognition with very deep neural networks. arXiv 2015, arXiv:1502.00873. [Google Scholar]
- Li, B.; Ouyang, W.; Sheng, L.; Zeng, X.; Wang, X. Gs3d: An efficient 3d object detection framework for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1019–1028. [Google Scholar]
- Feng, D.; Haase-Schütz, C.; Rosenbaum, L.; Hertlein, H.; Glaeser, C.; Timm, F.; Wiesbeck, W.; Dietmayer, K. Deep multi-modal object detection and semantic segmentation for autonomous driving: Datasets, methods and challenges. IEEE Trans. Intell. Transp. Syst. 2020, 22, 1341–1360. [Google Scholar] [CrossRef]
- Pedraza, A.; Bueno, G.; Deniz, O.; Cristóbal, G.; Blanco, S.; Borrego-Ramos, M. Automated diatom classification (Part B): A deep learning approach. Appl. Sci. 2017, 7, 460. [Google Scholar] [CrossRef]
- Park, J.; Lee, H.; Park, C.Y.; Hasan, S.; Heo, T.Y.; Lee, W.H. Algal morphological identification in watersheds for drinking water supply using neural architecture search for convolutional neural network. Water 2019, 11, 1338. [Google Scholar] [CrossRef]
- Yadav, D.; Jalal, A.; Garlapati, D.; Hossain, K.; Goyal, A.; Pant, G. Deep learning-based ResNeXt model in phycological studies for future. Algal Res. 2020, 50, 102018. [Google Scholar] [CrossRef]
- Xu, L.; Xu, L.; Chen, Y.; Zhang, Y.; Yang, J. Accurate Classification of Algae Using Deep Convolutional Neural Network with a Small Database. ACS ES&T Water 2022, 2, 1921–1928. [Google Scholar]
- Salido, J.; Sánchez, C.; Ruiz-Santaquiteria, J.; Cristóbal, G.; Blanco, S.; Bueno, G. A low-cost automated digital microscopy platform for automatic identification of diatoms. Appl. Sci. 2020, 10, 6033. [Google Scholar] [CrossRef]
- Baek, S.S.; Pyo, J.; Pachepsky, Y.; Park, Y.; Ligaray, M.; Ahn, C.Y.; Kim, Y.H.; Chun, J.A.; Cho, K.H. Identification and enumeration of cyanobacteria species using a deep neural network. Ecol. Indic. 2020, 115, 106395. [Google Scholar] [CrossRef]
- Park, J.; Baek, J.; Kim, J.; You, K.; Kim, K. Deep Learning-Based Algal Detection Model Development Considering Field Application. Water 2022, 14, 1275. [Google Scholar] [CrossRef]
- Wu, X.; Sahoo, D.; Hoi, S.C. Recent advances in deep learning for object detection. Neurocomputing 2020, 396, 39–64. [Google Scholar] [CrossRef]
- Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D.; Ramanan, D. Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1627–1645. [Google Scholar] [CrossRef]
- Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
- Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
- Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
- Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
- Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
- Jocher, G.; Chaurasia, A.; Stoken, A.; Borovec, J.; Kwon, Y.; Michael, K.; Fang, J. ultralytics/yolov5: v6.2-YOLOv5 Classification Models, Apple M1, Reproducibility, ClearML and Deci.ai integrations. GitHub 2022. [Google Scholar] [CrossRef]
- Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
- Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
- Samantaray, A.; Yang, B.; Dietz, J.E.; Min, B.C. Algae detection using computer vision and deep learning. arXiv 2018, arXiv:1811.10847. [Google Scholar]
- Qian, P.; Zhao, Z.; Liu, H.; Wang, Y.; Peng, Y.; Hu, S.; Zhang, J.; Deng, Y.; Zeng, Z. Multi-target deep learning for algal detection and classification. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 1954–1957. [Google Scholar]
- Ali, S.; Khan, Z.; Hussain, A.; Athar, A.; Kim, H.C. Computer Vision Based Deep Learning Approach for the Detection and Classification of Algae Species Using Microscopic Images. Water 2022, 14, 2219. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).