Next Article in Journal
Molecular Characterization and Expression Response of Ghrelin, GLP-1 and PYY to Fasting, Dietary Lipid, and Fatty Acids in Silver Pomfret (Pampus argenteus)
Next Article in Special Issue
Individual Fish Echo Detection Method Based on Peak Delay Estimation and Instantaneous Frequency Characterization
Previous Article in Journal
Uptake Routes and Biodistribution of Polystyrene Nanoplastics on Zebrafish Larvae and Toxic Effects on Development
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Diseased Fish Detection in the Underwater Environment Using an Improved YOLOV5 Network for Intensive Aquaculture

1
Xianning Academy of Agriculture Sciences, Xianning 437100, China
2
Xianning Branch, Hubei Academy of Agricultural Sciences, Xianning 437100, China
3
Nanjing Institute of Agricultural Mechanization, Ministry of Agriculture and Rural Affairs, Nanjing 210049, China
*
Author to whom correspondence should be addressed.
Fishes 2023, 8(3), 169; https://doi.org/10.3390/fishes8030169
Submission received: 14 February 2023 / Revised: 13 March 2023 / Accepted: 20 March 2023 / Published: 21 March 2023
(This article belongs to the Special Issue Current Research on Fish Tracking Technology)

Abstract

:
In intensive aquaculture, the real-time detection and monitoring of common infectious disease is an important basis for scientific fish epidemic prevention strategies that can effectively reduce fish mortality and economic loss. However, low-quality underwater images and low-identification targets present great challenges to diseased fish detection. To overcome these challenges, this paper proposes a diseased fish detection model, using an improved YOLOV5 network for aquaculture (DFYOLO). The specific implementation methods are as follows: (1) the C3 structure is used instead of the CSPNet structure of the YOLOV5 model to facilitate the industrial deployment of the algorithm; (2) all the 3 × 3 convolutional kernels in the backbone network are replaced by a convolutional kernel group consisting of parallel 3 × 3, 1 × 3 and 3 × 1 convolutional kernels; and (3) the convolutional block attention module is added to the YOLOV5 algorithm. Experimental results in a fishing ground showed that the DFYOLO is better than that of the original YOLOV5 network, and the average precision was improved from 94.52% to 99.38% (when the intersection over union is 0.5), for an increase of 4.86%. Therefore, the DFYOLO network can effectively detect diseased fish and is applicable in intensive aquaculture.

1. Introduction

China is the largest consumer of freshwater fish in the world [1,2]. Wild-caught freshwater fish make up a large portion of freshwater fish consumption [3]. Overfishing due to consumption of wild-caught freshwater fish destroys biodiversity [4]. In order to reduce the ecological damage caused by fishing, in 2020, the Chinese Ministry of Agriculture and Rural Affairs announced the start of a 10-year fishing ban on the Yangtze River [4]. With the implementation of the fishing ban, China’s freshwater fish consumer market is more dependent on artificial culture. During freshwater aquaculture, fish feed or fertilizer is put into the water to increase freshwater fish production, but residual fertilizer, fish feces and other excreta can cause water eutrophication and lead to eco-catastrophe, such as red tides, so this type of freshwater aquaculture was banned in natural waters. In this context, the proportion of intensive aquaculture will further increase [5,6]. One study shows that intensive aquaculture systems will have to predominate [7]. The high-density stocking and feeding method will inevitably face the problem of excessive multiplication of viruses, bacteria and fungi, and the accumulation of nitrogen and phosphorus, especially ammonia and nitrite nitrogen concentrations, making farmed fish more susceptible to various diseases. One study showed that more than half of production losses in aquaculture are caused by diseases [8]. At present, fish disease detection mainly relies on manual methods. However, as light is refracted from the air into water, the human eye observes the health of fish with difficulty, leading to failure of timely drug administration or adjustment of breeding programs, missing the best treatment period, and resulting in huge economic losses. This is, thus, the technical bottleneck of intensive aquaculture. Therefore, it is important to study automatic fish disease identification and analysis methods.
With the development of deep learning technology, many researchers have proposed to identify animal behavior, and animal and plant body surface characteristics based on video image analysis, so as to identify animals in heat, hunger and disease, and plant maturity and disease, and great progress has been made in recent years. Chen et al. [9] proposed a fish species identification system for fish markets that combines state-of-the-art instance-segmentation methods, with ResNet-based classification. Rauf et al. [10] proposed a deep learning framework based on a CNN approach for fish species recognition, which has been shown to achieve state-of-the-art performance through experimental comparisons with other deep learning frameworks. Qi et al. [11] proposed a novel lightweight convolutional neural network for medicinal chrysanthemum detection. Måløy et al. [12] proposed a two-stream recurrent network (DSRN) to automatically capture the spatio-temporal behavior of salmonids as they swim, and the model achieved a prediction accuracy of 80%. Zhang et al. [13] proposed an automated fish population counting method based on machine vision and a new hybrid deep neural network model to count farmed Atlantic salmon. Labao and Naval [14] proposed a fish detection system consisting of an ensemble of region-based convolutional neural networks, that can detect and count fish objects under various benthic backgrounds and illumination conditions. The rapid development of intelligence has been applied in aquaculture.
In order to reduce the losses caused by fish infectious diseases, the detection method should have a high level of real-time. The You Only Look Once (YOLO) series algorithm, now updated to the fifth generation, is the mainstream single-stage detection algorithm with high detection accuracy and fast detection speed, which is widely used in various target detection tasks. Roy et al. [15] proposed a mango detection framework based on an improved YOLOv4 algorithm, by including DenseNet in the backbone. Fan et al. [16] proposed a real-time apple defects inspection method based on a simplified YOLOV4 algorithm, using channel pruning and layer pruning methods. Qi et al. [17] proposed a highly fused and lightweight deep learning architecture, based on YOLO, for tea chrysanthemum detection. Ge et al. [18] proposed the UW_YOLOv3 lightweight model to solve the problems of calculating energy consumption and storage resource limitations in underwater application scenarios. Cai et al. [19] proposed a fish detection method combining YOLOv3 and MobileNetv1, which was used to detect the number of fish within real breeding farms. Li et al. [20] proposed a real-time fish detection method based on YOLO-V3-Tiny-MobileNet, using YOLO-V3-Tiny as the baseline, and combined with MobileNet, the method can provided timely warning to fishing vessel operators. Hu et al. [21] proposed an uneaten feed pellet detection model, using an improved YOLO-V4 network. Jalal et al. [22] proposed a hybrid solution combining optical flow and Gaussian mixture models with YOLO deep neural network, which achieved 95.47% and 91.2% F-scores for fish detection, while the accuracy of fish species classification was 91.64% and 79.8%, respectively. Wen et al. [23] proposed a lightweight YOLOv4 detection algorithm improved by multi-scale feature fusion for dense silkworm detection. Zhou et al. [24] proposed a marine biometric identification method based on image enhancement and improved YOLOv4 algorithm. Prasetyo et al. [25] fish body part detection method, based on YOLOV4-tiny with a wing convolution layer. Abinaya et al. [26] proposed a YOLOv4 based fish segmentation detection method to estimate the fish biomass in an occulted environment. Wang et al. [27] proposed a neural network based on improved YOLOV5 and SiamRPN++, to detect and track the abnormal behavior of porphyry seabream. YOLO has been used successfully in many applications for target detection in aquaculture. However, the original YOLO algorithm still has limitations for the detection of fish infectious diseases due to the small difference between diseased and normal fish.
To solve the above problems, this paper proposes a YOLOV5m network-based diseased fish detection method for intensive aquaculture. The recognition algorithm replaces all the 3 × 3 convolutional kernels in the backbone network with a convolution kernel group, and adds a convolutional block attention module (CBAM), which effectively improves the detection accuracy of diseased fish. The next sections of this paper are organized as follows: Section 2 introduces the dataset, the improved algorithm and network; Section 3 and Section 4 present the results and discussion; and Section 5 is the conclusion.

2. Materials and Methods

2.1. Dataset

Fish image data were collected in the aquaculture base of Xianning Academy of Agricultural Sciences, Hubei Province. The shooting scene is as shown in Figure 1.
The acquisition equipment consisted of the GO POR8 motion camera (GO POR, San Mateo, CA, USA), of which the two-dimensional RGB image resolution was 5650 × 4238 pixels, focal length was 2 mm, and exposure time was 1/200 s. In order to reduce the probability of network model overfitting caused by insufficient diversity of training samples, strong light and weak light at acquisition time were distinguished by 500–1000 lux and 50–499 lux of underwater light intensity of the day, respectively. Under strong light and weak light, fish feature images of 600 mm, 1200 mm and 1800 mm depth of water body were collected. The dimensions of the barrel and image acquisition are shown in Figure 1. In order to increase the diversity of the samples, the image samples included two species of fish, red tilapia and micropterus salmoides, including the different conditions of fish density, growth phase, temperature and water quality, as well as light conditions such as light and backlight. Figure 2 shows a group of typical images of fish in complex underwater environments.
The complex underwater environment of the captive barrel (Fisheries College of Huazhong Agricultural University, Wuhan, China) and the various underwater electronic sensors can interfere with the acquisition and transmission of the images, making them noisy. In order to improve the recognition efficiency of diseased fish, the images need to be processed for noise reduction. Analysis of the image noise shows that the probability density function obeys a Gaussian distribution, and the common processing methods are Gaussian filtering or bilateral filtering. Gaussian filtering has good performance in low-pass filtering algorithms, but it only takes into account the spatial position of the pixels, and the filtering results in the loss of edge information. The edges in this case are the main areas of color in the image (turbidity of the water, individual body colors of red tilapia, etc.), and bilateral filtering solves this problem by adding an additional weight to the Gaussian filter.
Photographs taken between 1 July and 31 October were manually checked and invalid photographs of poor quality were removed. Including 12,045 photographs of rainbow snapper and 12,317 photographs of micropterus salmoides, 24,362 images were obtained. As the photographs were taken underwater over a certain period of time, there was turbidity in the water as the weather and the environment in the captive barrel changed, a phenomenon that can be likened to taking photographs in the air due to haze, resulting in poor image characteristics and affecting image quality. Therefore, it is necessary to use a dark channel defogging method on the collected samples, to saturate the color and increase the brightness of the photographs to facilitate the identification of lesions on the surface of the fish. The expression for haze removal using dark channel [28] is shown in Equation (1).
J ( x ) = I ( x ) A t ( x ) + A
when the transmittance, t ( x ) , is too small, the value of J ( x ) will be large and a threshold t ( 0 ) is set to limit J ( x ) , typically taking the value 0.1. Then, there is the Equation (2), shown below.
J ( x ) = I ( x ) A m a x ( t ( x ) , t 0 ) + A   ,
where x is the input pixels, J(x) is the recovered haze-free image, A is the atmospheric illumination (maximum pixel value), I(x) is the existing images (to be foggy) and t(x) is the medium transmission rate.
The pathological analysis showed that the main disease of red tilapia was ocular streptococcal infection, which caused congestion, protrusion and ulceration of pus in the eyes of red tilapia. The main disease of micropterus salmoides was Nocardia infection, which caused obvious redness, inflammation, ulceration and necrosis on the surface of the body of both fish. A comparison of the symptoms of the two species is shown in Figure 3.
To further clarify the trend of fish infection with culture time, 25 images of red tilapia and micropterus salmoides were selected for tagging each day, and a total of 6000 images were tagged, with 50% of each of the two species. The labelled images were divided into a train set and a validation set according to 80% and 20%, and another 1000 images were selected from the images taken as a test set, with 50% of each species. The number of disease and illness levels for each fish species in the captive barrel was counted, as shown in Table 1. Three thousand images of red tilapia were marked, of which 785 had eye infections and 213 died, while the remaining healthy fish were marked as normal. Of the 3000 images of micropterus salmoides, a total of 2080 fish had surface decay and the remaining healthy fish were marked as normal.

2.2. The Diseased Fish Identification Algorithms

The YOLOV5 network is a current excellent one-stage target detection algorithm, of which the structure is divided into three parts: backbone, neck and head. The main component of YOLOV5 includes Focus, SPP and PANet, where Focus improves the receptive field through slice and convolution operations, SPP separates the important features of the context through max pool and convolution operations, and PANet obtains the strong feature map by integrating different level features.
There are four types of YOLOV5: V5s, V5m, V5l and V5x. Taking speed and accuracy into account, this paper proposes a diseased fish detection model based on YOLOV5m (DFYOLO), the network structure is shown in Figure 4.
The main model improvements of this study are as follows:
  • The implementation of lightweight computing is key to the industrial deployment of algorithms. In order to achieve lightweight computing, the C3 component is used instead of the CSPNet [29] component of the YOLOV5 model in this study. As shown in Figure 5, compared to the CSPNet component, the C3 component removes one convolution operation after the skip connect, and splices directly with another branch of the input image after one convolution operation. The C3 component has a similar effect to the CSPNet component, but with a simpler structure.
2.
As fish activity photos are taken underwater, the scenes in the photos often change with water quality, lighting and fish conditions, resulting in varying degrees of variation and obscuration of fish features, and the original YOLOV5 backbone network is unable to extract clear features. In order to reduce the impact of underwater complexities, this study improves the generalization ability of the network by expanding the training samples and replacing all 3 × 3 convolutional kernels in the backbone network with convolutional kernel group (Conv KG), to enhance the network’s ability to extract features from the photographed fish. Conv KG consists of three parallel 3 × 3, 1 × 3 and 3 × 1 convolutional kernels, which convolve the input image in the same steps to produce feature maps of the same size and number of channels, respectively, and the corresponding feature maps are summed to obtain the output features, as shown in Figure 6. The three parallel convolutional kernels enhance the network’s ability to extract fish surface features.
P K ( 1 ) + P K ( 2 ) + P K ( 3 ) = P ( K ( 1 ) + K ( 2 ) + K ( 3 ) ) = P K ,  
where P is the input image, K(1), K(2) and K(3) are the 3 × 3, 1 × 3 and 3 × 1 convolution kernels, K is the equivalent convolution kernel, ∗ is the convolution operation.
Therefore, the three parallel convolutional kernels are equivalent to a new 3 × 3 convolutional kernel with different weights. Compared to the 3 × 3 kernels before the replacement, the trained equivalent convolutional kernel can enhance the extraction of fish surface features, without increasing the computational effort in the train set.
3.
Due to the variable underwater environment and the large variation in the number of fish in fish photos, which interferes with the original YOLOV5 detection algorithm, this study adds the attention mechanism module, convolutional block attention module (CBAM), to the YOLOV5 network (Figure 7). CBAM is a simple, but effective feed-forward convolutional neural network attention module, which combines the channel attention module (CAM) and spatial attention module (SAM) [30]. Given an intermediate feature map, our module sequentially inferred the attention of an image along two independent dimensions, channel and space, and then multiplied the attention map by the input feature map for adaptive feature refinement. Because CBAM is a lightweight, general-purpose module, it can be seamlessly integrated into any convolutional neural network architecture with negligible overhead, and can be trained end-to-end with the underlying convolutional neural network.
The overall CBAM process is shown in Equations (4) and (5):
F = M C ( F ) F
F = M S ( F ) F   ,
where F R C × H × W is the input feature map, M C R C × 1 × 1 is the CAM weight data, which is 1 × 1 × C, F is the CAM output, M S R 1 × H × W is the spatial attention module weight data, which is 2 × H × W and F is the CBAM output.
The channel attention module mainly detects the contour characteristics of the fish and obtains the main contents of the detection target. Its calculation method is as follows:
M C ( F ) = σ ( W 1 ( W 0 ( F a v g c ) ) + W 1 ( W 0 ( F m a x c ) ) )
In the formula, σ represents the Sigmoid function, W 0 R c / r × c , W 1 R c × c / r , two inputs share the weights W 0 and W 1 , the ReLU activation function is followed by the W 0 , F a v g c , F m a x c , representing features generated using average-pooled and max-pooled, H is the height, W is the width, C is the number of channels, and r is the rate of reduction.
The introduction of the spatial attention module is to further improve the detection accuracy of the target fish, and accurately locate the position of target fish. The calculation method is as follows:
M S ( F ) = σ ( f 7 × 7 ( [ F a v g s ; F m a x s ] ) ) ,  
where F a v g s and F m a x s represent the features generated using average-pooled and max-pooled and f 7 × 7 represents a convolution operation with the filter size of 7 × 7.

2.3. Experimental Environment

In this study, the Darknet framework was used to improve the YOLOV5 network. The experimental environment is shown in Table 2.

2.4. Performance Evaluation Metrics

In this study, precision (P), recall (R) and mean average precision (mAP) are used as evaluation indicators for DFYOLO. Precision is the probability of an actual diseased fish being predicted in all samples predicted to be diseased, which represents the precision of the predicted outcome for positive samples. Recall is the probability of being predicted as a diseased fish among the actual diseased fish samples, and it represents the overall prediction accuracy. It is calculated as shown by Equations (8) and (9).
p r e c i s i o n = T P T P + F P
r e c a l l = T P T P + F N   ,
where TP (true positives) is a sample that is correctly identified as abnormal behavior; FN (false negatives) is a sample that is mistaken as a background; TN (true negatives) is a sample that is correctly identified as a background; FP (false positives) is a sample that is misidentified as abnormal behavior.
Average precision (AP) refers to the area under the P–R curve, for which the calculation formula is shown by Equation (10). AP50 is the mean of precision under different recall values when IOU (intersection over union) = 0.5. AP50:95 is the mean of the ten values of AP50, AP55 …, AP90, AP95, for which the calculation formula is shown by Equation (11). The mAP refers to the average of the AP values of the two species, micropterus salmoides (MS) and red tilapia (RT). The calculation formula is shown by Equation (12).
A P = 0 1 P ( R ) d R  
A P 50 : 95 = 1 10 ( A P 50 + A P 55 + + A P 90 + A P 95 )
m A P = 1 2 ( A P M S + A P R T )

3. Results

3.1. Training Result

The training results of DFYOLO are shown in Figure 8. After 475 times epoch, the loss value dropped to 0.027. The accuracy of DFYOLO was 99.75%, the recall was 99.31% and the mAP50 was 99.38%. From the training results, the model has convergence with high accuracy and check-all rates, as well as low loss value.

3.2. Identification Results under Various Varieties

The images of red tilapia and micropterus salmoides were randomly collected at the same light intensity and water depth, and entered into DFYOLO. The monitoring results are shown in Figure 9. When red tilapia were close to the camera and the sides and their heads were fully displayed, no missed detections occurred; the micropterus salmoides that had their side torsos fully displayed in the image were able to be detected in their entirety, with no missed detections.

3.3. Identification Results under Various Environments

The images of red tilapia and micropterus salmoides were randomly selected at 500–1000 lux (strong light) and 50–499 lux (weak light), respectively, and input into DFYOLO. The monitoring results are shown in Figure 10, and there were no missed detections, indicating that DFYOLO can adapt to the different lighting conditions common in the breeding process.
The monitoring results for different water depths are shown in Figure 11. The images taken at 600 mm depth had the best detection results, because there were fewer fish in the upper layer and fewer fish adhered to the images; the images taken at 1200 mm depth had no missed detection, which was slightly less effective than those taken at 600 mm depth; the images taken at 1800 mm depth had missed detection of individual fish, as shown by the pink dashed circle in Figure 11, indicating that DFYOLO had better detection at different shooting depths. This shows that DFYOLO has a good monitoring effect at different shooting heights.

3.4. Contrast to Mainstream Target Detection Networks

In order to verify the detection performance of the improved algorithm in this paper, the DFYOLO proposed in this study was compared with the mainstream target detection models SSD, Faster-R CNN, YOLOv3 and YOLOv4, and the five metrics of Occupy Memory, FPS, Recall, Precision and mAP50, were used to evaluate and compare the mainstream detection algorithms. The comparison experimental parameters are shown in Table 3.
The analysis in Table 4 shows that when the IOU = 0.5, the recall of the DFYOLO algorithm was 99.31%, which is 5.55% higher than the original YOLOV5 model; the precision was 99.75%, which is 5.39% higher than the original YOLOV5 model; the mAP50 was 99.38%, which is 5.86% higher than the original YOLOV5 algorithm; the mAP50:95 was 88.09%, which is 9.56% higher than the original YOLOV5 algorithm; the model occupied 0.9 MB less memory; while ensuring high accuracy detection, the FPS of the model did not show a significant decrease.

3.5. Ablation Experiment

An ablation experiment is a commonly used experimental method in deep learning, which proves the necessity of this module by removing the effect of single- or multiple-improved methods. If the results of an ablation experiment are worse or the performance is significantly reduced, the improved method is reasonable.
The results of the ablation experiment are shown in Table 5, where it can be seen that bilateral filtering and haze removal improved the quality of underwater photos, and slightly improved the model detection. Image transformation can reduce the computation time of the algorithm, Conv KG enhances the feature extraction of the fish surface, CBAM focuses on “what the fish is?” and “where the fish are? is?”. The above methods improved the detection speed, precision recall and mAP of the model.

3.6. Comparison of before and after Improvement

Randomly selected photographs of red tilapia and micropterus salmoides were entered into the original YOLOV5 and DFYOLO models, as shown in Figure 12. The thin line part is the recognition result of the original model, the thick line is the recognition result of the DFYOLO, and the fish missed by the original model are only marked by the dashed line. By comparing the images, the DFYOLO had a higher detection efficiency, more accurate detection parts and a lower missed detection rate.

4. Discussion

4.1. Robustness of the Method

Although our work is an exploratory experiment, it aims to explore whether recognition algorithms can be used for disease recognition in fish. Nevertheless, we simulated the actual production situation to improve the robustness of our method. We chose to use a large intensive culture barrel for our experiment. First, we took time series of photographs of fish at different stages of growth. Secondly, we also selected two typical varieties (Figure 9), different light conditions (Figure 10) and different water depths (Figure 11), which were as close to the real environment as possible. In addition, the experimental design added some algae, so that we could see that all the photos were yellowish-green. Our models still work well in this underwater environment.

4.2. Comparison with Other Methods

At present, there are also many underwater fish tracking and identification methods used [31,32]. They generally use RCNN or CNN, which has a high recognition accuracy and a low technical threshold. However, due to its structure, the recognition speed is not fast enough. Intensive farming requires relatively fast recognition speed, because more fish targets appear per unit image area. Therefore, this study preferentially selects YOLO as the basic algorithm. According to the data in Table 4, YOLOv5 has an obvious speed advantage compared with other algorithms in actual use. After some improvements, the DFYOLO sacrifices some of its speed advantage to reduce memory usage, but still achieves 93.21FPS.

4.3. Contribution to the Detection of Fish Diseases

In recent years, with the rapid development of artificial intelligence, the application of computer vision model in fishery industry is more and more popular. This study is concerned with the use of computer vision technology for real-time detection of fish body surface to estimate the disease and degree of fish, which can improve the management efficiency of freshwater fish culture. The difference between underwater fish disease recognition and common target recognition is that the features of fish body surface are more homogenous and smaller than those of common targets. On the other hand, the underwater shooting environment is much more complex than that in the air. In addition, the most important point is that fish have a large degree of overlap in space, so it is easy to introduce redundant features and affect the accuracy of body surface disease recognition. In this study, the image noise reduction is carried out by the method of bilateral filtering, which reduces the interference of other electrical equipment to the underwater camera. The application of haze removal using dark channel reduces the influence of turbid water on the image quality, and C3 replacing the original CPSNet reduces the memory occupied by the model, while CABM (convolutional block attention module) is introduced to avoid the influence of a large amount of redundant information on fish recognition. According to the data in Table 5, compared with the original algorithm before improvement (YOLO v5), our model has great advantages in speed and accuracy, achieving a real-time detection speed and accuracy of 99.75%.

5. Conclusions

In order to detect fish health in real time, this paper proposes a DFYOLO network based on YOLOV5m, to detect the proportion of diseased fish in the shoal of fish.
The captive barrel has an automatic aerator, sewage discharge and other electrical equipment, generating the electromagnetic interference in the process of camera shooting. In this study, the form of bilateral filtering is adopted to reduce the noise of the images taken. In addition, intensive culture will lead to turbidity of water and unclear surface features of fish body. If the unprocessed image is directly input into the algorithm, the recognition effect will be greatly reduced. The addition of haze removal using dark channel can effectively reduce the influence of water turbidity on photo quality. C3 replaces CSPNet, which makes the whole algorithm structure simple and efficient. The introduction of Conv KG reduces the memory of the whole algorithm. The application of CABM makes the whole algorithm combined-target detection and feature recognition, and achieves good results in fish detection and body surface feature recognition. The experimental results showed that DFYOLO not only achieves the highest mAP50 (99.38%), which is 4.86% higher than YOLOv5m, but also keeps a high inference speed (93.21FPS) and occupies less memory (13.6 MB). This method will provide a theoretical basis for the development of intelligent fishery.
Although this study has realized the identification and statistics of fish disease, the proposed method still has some limitations.
  • The algorithm used in this paper is a target recognition algorithm. It does not actually support multi-target dynamic tracking. As a result, it is unable to track fish movements, and the information collected is too simple to provide a comprehensive picture of their health.
  • The dataset used in this study for two common fish diseases is not sufficient for practical application. More species of fish disease behavior should be collected to supplement the dataset.
  • In future work, we will use stereo and multispectral cameras to capture images, to reduce the impact of illumination, water quality and refraction on image quality. In addition, we have found a number of methods regarding geometric features that can be used to describe the size, weight and swimming speed of the fish, to construct a more accurate estimation model of the health of the fish.

Author Contributions

Conceptualization, Z.W., H.L. and G.Z.; methodology, X.Y. and L.W.; software, Z.W.; validation, W.Z.; writing—original draft preparation, Z.W. writing—review and editing, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by the Hubei science and technology service fishery industry chain “515” action, Agricultural Science and Technology Innovation Program of Chinese Academy of Agricultural Sciences (ASTIP, CAAS) and Xianning Academy of Agricultural Sciences subject group leader responsibility system project (XNNK20210602).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We greatly appreciate the careful and precise reviews by the anonymous reviewers and editors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ai, L.; Ma, B.; Shao, S.; Zhang, L.; Zhang, L. Heavy metals in Chinese freshwater fish: Levels, regional distribution, sources and health risk assessment. Sci. Total Environ. 2022, 853, 158455. [Google Scholar] [CrossRef]
  2. Fang, J.; Fabinyi, M. Characteristics and Dynamics of the Freshwater Fish Market in Chengdu, China. Front. Sustain. Food Syst. 2021, 5, 638997. [Google Scholar] [CrossRef]
  3. Fluet-Chouinard, E.; Funge-Smith, S.; McIntyre, P.B. Global hidden harvest of freshwater fish revealed by household surveys. Proc. Natl. Acad. Sci. USA 2018, 115, 7623–7628. [Google Scholar] [CrossRef] [Green Version]
  4. Zhang, Y. Freshwater biodiversity conservation in China: Progress in the Yangtze River basin. Aquat. Conserv.-Mar. Freshw. Ecosyst. 2022, 32, 1565–1570. [Google Scholar] [CrossRef]
  5. Santos, L.; Ramos, F. Antimicrobial resistance in aquaculture: Current knowledge and alternatives to tackle the problem. Int. J. Antimicrob. Agents 2018, 52, 135–143. [Google Scholar] [CrossRef]
  6. Bulfon, C.; Volpatti, D.; Galeotti, M. Current research on the use of plant-derived products in farmed fish. Aquac. Res. 2015, 46, 513–551. [Google Scholar] [CrossRef]
  7. Boyd, C.E.; D’Abramo, L.R.; Glencross, B.D.; Huyben, D.C.; Juarez, L.M.; Lockwood, G.S.; McNevin, A.A.; Tacon, A.G.J.; Teletchea, F.; Tomasso, J.R.; et al. Achieving sustainable aquaculture: Historical and current perspectives and future needs and challenges. J. World Aquac. Soc. 2020, 51, 578–633. [Google Scholar] [CrossRef]
  8. Assefa, A.; Abunna, F. Maintenance of Fish Health in Aquaculture: Review of Epidemiological Approaches for Prevention and Control of Infectious Disease of Fish. Vet. Med. Int. 2018, 2018, 5432497. [Google Scholar] [CrossRef]
  9. Chen, C.-H.; Chen, L.-H.; Chen, C.-Y. Automatic Fish Segmentation and Recognition in Taiwan Fish Market using Deep Learning Techniques. J. Imaging Sci. Technol. 2021, 65, 40403-1–40403-10. [Google Scholar] [CrossRef]
  10. Rauf, H.T.; Lali, M.I.U.; Zahoor, S.; Shah, S.Z.H.; Rehman, A.U.; Bukhari, S.A.C. Visual features based automated identification of fish species using deep convolutional neural networks. Comput. Electron. Agric. 2019, 167, 105075. [Google Scholar] [CrossRef]
  11. Qi, C.; Chang, J.; Zhang, J.; Zuo, Y.; Ben, Z.; Chen, K. Medicinal Chrysanthemum Detection under Complex Environments Using the MC-LCNN Model. Plants 2022, 11, 838. [Google Scholar] [CrossRef]
  12. Maloy, H.; Aamodt, A.; Misimi, E. A spatio-temporal recurrent network for salmon feeding action recognition from underwater videos in aquaculture. Comput. Electron. Agric. 2019, 167, 105087. [Google Scholar] [CrossRef]
  13. Zhang, S.; Yang, X.; Wang, Y.; Zhao, Z.; Liu, J.; Liu, Y.; Sun, C.; Zhou, C. Automatic Fish Population Counting by Machine Vision and a Hybrid Deep Neural Network Model. Animals 2020, 10, 364. [Google Scholar] [CrossRef] [Green Version]
  14. Labao, A.B.; Naval, P.C., Jr. Cascaded deep network systems with linked ensemble components for underwater fish detection in the wild. Ecol. Inform. 2019, 52, 103–121. [Google Scholar] [CrossRef]
  15. Roy, A.M.; Bhaduri, J. Real-time growth stage detection model for high degree of occultation using DenseNet-fused YOLOv4. Comput. Electron. Agric. 2022, 193, 106694. [Google Scholar] [CrossRef]
  16. Fan, S.; Liang, X.; Huang, W.; Zhang, V.J.; Pang, Q.; He, X.; Li, L.; Zhang, C. Real-time defects detection for apple sorting using NIR cameras with pruning-based YOLOV4 network. Comput. Electron. Agric. 2022, 193, 106715. [Google Scholar] [CrossRef]
  17. Qi, C.; Gao, J.; Pearson, S.; Harman, H.; Chen, K.; Shu, L. Tea chrysanthemum detection under unstructured environments using the TC-YOLO model. Expert Syst. Appl. 2022, 193, 116473. [Google Scholar] [CrossRef]
  18. Ge, H.; Dai, Y.; Zhu, Z.; Liu, R. A Deep Learning Model Applied to Optical Image Target Detection and Recognition for the Identification of Underwater Biostructures. Machines 2022, 10, 809. [Google Scholar] [CrossRef]
  19. Cai, K.W.; Miao, X.Y.; Wang, W.; Pang, H.S.; Liu, Y.; Song, J.Y. A modified YOLOv3 model for fish detection based on MobileNetv1 as backbone. Aquac. Eng. 2020, 91, 102117. [Google Scholar] [CrossRef]
  20. Li, J.; Zhu, K.; Wang, F.; Jiang, F. Deep neural network-based real time fish detection method in the scene of marine fishing supervision. J. Intell. Fuzzy Syst. 2021, 41, 4527–4532. [Google Scholar] [CrossRef]
  21. Hu, X.; Liu, Y.; Zhao, Z.; Liu, J.; Yang, X.; Sun, C.; Chen, S.; Li, B.; Zhou, C. Real-time detection of uneaten feed pellets in underwater images for aquaculture using an improved YOLO-V4 network. Comput. Electron. Agric. 2021, 185, 106135. [Google Scholar] [CrossRef]
  22. Jalal, A.; Salman, A.; Mian, A.; Shortis, M.; Shafait, F. Fish detection and species classification in underwater environments using deep learning with temporal information. Ecol. Inform. 2020, 57, 101088. [Google Scholar] [CrossRef]
  23. Wen, C.; Wen, J.; Li, J.; Luo, Y.; Chen, M.; Xiao, Z.; Xu, Q.; Liang, X.; An, H. Lightweight silkworm recognition based on Multi-scale feature fusion. Comput. Electron. Agric. 2022, 200, 107234. [Google Scholar] [CrossRef]
  24. Zhou, J.; Yang, Q.; Meng, H.; Gao, D. An underwater target recognition method based on improved YOLOv4 in complex marine environment. Syst. Sci. Control. Eng. 2022, 10, 590–602. [Google Scholar] [CrossRef]
  25. Prasetyo, E.; Suciati, N.; Fatichah, C. Yolov4-tiny with wing convolution layer for detecting fish body part. Comput. Electron. Agric. 2022, 198, 107023. [Google Scholar] [CrossRef]
  26. Abinaya, N.S.; Susan, D.; Sidharthan, R.K. Deep learning-based segmental analysis of fish for biomass estimation in an occulted environment. Comput. Electron. Agric. 2022, 197, 106985. [Google Scholar] [CrossRef]
  27. Wang, H.; Zhang, S.; Zhao, S.; Wang, Q.; Li, D.; Zhao, R. Real-time detection and tracking of fish abnormal behavior based on improved YOLOV5 and SiamRPN plus. Comput. Electron. Agric. 2022, 192, 106512. [Google Scholar] [CrossRef]
  28. He, K.; Sun, J.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar] [CrossRef]
  29. Wang, C.Y.; Liao, H.Y.M.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H.; IEEE Comp, S.O.C. CSPNet: A New Backbone that can Enhance Learning Capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 1571–1580. [Google Scholar]
  30. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  31. Li, D.; Su, H.; Jiang, K.; Liu, D.; Duan, X. Fish Face Identification Based on Rotated Object Detection: Dataset and Exploration. Fishes 2022, 7, 219. [Google Scholar] [CrossRef]
  32. Deng, Y.; Tan, H.; Tong, M.; Zhou, D.; Li, Y.; Zhu, M. An Automatic Recognition Method for Fish Species and Length Using an Underwater Stereo Vision System. Fishes 2022, 7, 326. [Google Scholar] [CrossRef]
Figure 1. The experimental setup of image acquisition.
Figure 1. The experimental setup of image acquisition.
Fishes 08 00169 g001
Figure 2. Example images in the underwater environment.
Figure 2. Example images in the underwater environment.
Fishes 08 00169 g002
Figure 3. Typical images of fish infection. Note: The infected fish in the figure are inside the Ground truth box, and normal fish are found outside the Ground truth box.
Figure 3. Typical images of fish infection. Note: The infected fish in the figure are inside the Ground truth box, and normal fish are found outside the Ground truth box.
Fishes 08 00169 g003
Figure 4. Algorithm structure and workflow of DFYOLO. Note: Focus is the slice operation, and it can double down sample feature graph without information loss; SPP is a spatial pyramid pool structure, and it effectively avoids the problems of image distortion caused by image region cropping and scaling operation.
Figure 4. Algorithm structure and workflow of DFYOLO. Note: Focus is the slice operation, and it can double down sample feature graph without information loss; SPP is a spatial pyramid pool structure, and it effectively avoids the problems of image distortion caused by image region cropping and scaling operation.
Fishes 08 00169 g004
Figure 5. Schematic diagram of the component replacement. Note: The C3 component replaces the CSPNet component.
Figure 5. Schematic diagram of the component replacement. Note: The C3 component replaces the CSPNet component.
Fishes 08 00169 g005
Figure 6. Improved convolution module.
Figure 6. Improved convolution module.
Fishes 08 00169 g006
Figure 7. Convolutional block attention module.
Figure 7. Convolutional block attention module.
Fishes 08 00169 g007
Figure 8. Training results.
Figure 8. Training results.
Fishes 08 00169 g008
Figure 9. Monitoring effect of different varieties.
Figure 9. Monitoring effect of different varieties.
Fishes 08 00169 g009
Figure 10. Monitoring effect at different light conditions.
Figure 10. Monitoring effect at different light conditions.
Fishes 08 00169 g010
Figure 11. Monitoring effect at different water depth.
Figure 11. Monitoring effect at different water depth.
Fishes 08 00169 g011
Figure 12. Comparison of the test image results.
Figure 12. Comparison of the test image results.
Fishes 08 00169 g012
Table 1. Various state statistics of the fish in the dataset.
Table 1. Various state statistics of the fish in the dataset.
Collection NameRed TilapiaMicropterus Salmoides
Eye infections7850
Surface decay02080
Died2130
Normal2002920
Summation30003000
Table 2. The experimental environment.
Table 2. The experimental environment.
ConfigurationParameter
CPUIntel Core i7-9700 K
GPUNvidia GeForce RTX 3080 Ti × 2
Operating systemWindows 10
Accelerated environmentCUDA10.2 CUDNN7.6.5
Development environmentVisual Studio 2020
LibraryOpencv3.4.0
Note: CPU manufactured by Intel Corporation, Santa Clara, CA, USA; GPU manufactured by Nvidia Corporation, Santa Clara, CA, USA; the operating system is manufactured by Microsoft Corporation, Albuquerque, NM, USA.
Table 3. Parameter settings for different detection networks.
Table 3. Parameter settings for different detection networks.
MethodBatch SizeLearning RateEpochMomentumWeight Decay
SSD160.0014750.90.0005
Faster-RCNN160.0014750.90.0005
YOLOv3160.0014750.90.0005
YOLOv4160.0014750.90.0005
YOLOV5160.014750.90.0005
DFYOLO160.014750.90.0005
Table 4. The detection results of different detection networks.
Table 4. The detection results of different detection networks.
MethodOccupy
Memory/MB
FPSRecall/%Precision/%mAP50/%mAP50:95/%
SSD63.269.7373.0176.8168.2442.03
Faster-RCNN77.857.9783.2886.1879.8751.14
YOLOv362.169.4486.1282.3890.9941.10
YOLOv422678.2493.8493.5893.3959.53
YOLOV514.596.4393.7694.3694.5278.53
DFYOLO13.693.2199.3199.7599.3888.09
Table 5. The results of an ablation test.
Table 5. The results of an ablation test.
Baseline NetworkBilateral FilteringHaze RemovalImage
Transformation
Conv KGCBAMRecall/%Precision/%mAP50/%mAP50:95/%
YOLOV5 93.7694.3694.5278.53
YOLOV5 95.8194.4094.5879.28
YOLOV5 95.8594.8696.0282.12
YOLOV5 95.7593.4295.4182.07
YOLOV5 97.2198.2498.3587.98
YOLOV5 97.1898.8598.1686.06
YOLOV599.3199.7599.3888.09
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Liu, H.; Zhang, G.; Yang, X.; Wen, L.; Zhao, W. Diseased Fish Detection in the Underwater Environment Using an Improved YOLOV5 Network for Intensive Aquaculture. Fishes 2023, 8, 169. https://doi.org/10.3390/fishes8030169

AMA Style

Wang Z, Liu H, Zhang G, Yang X, Wen L, Zhao W. Diseased Fish Detection in the Underwater Environment Using an Improved YOLOV5 Network for Intensive Aquaculture. Fishes. 2023; 8(3):169. https://doi.org/10.3390/fishes8030169

Chicago/Turabian Style

Wang, Zhen, Haolu Liu, Guangyue Zhang, Xiao Yang, Lingmei Wen, and Wei Zhao. 2023. "Diseased Fish Detection in the Underwater Environment Using an Improved YOLOV5 Network for Intensive Aquaculture" Fishes 8, no. 3: 169. https://doi.org/10.3390/fishes8030169

Article Metrics

Back to TopTop