Next Article in Journal
Optical Fiber Ball Resonator Sensor Spectral Interrogation through Undersampled KLT: Application to Refractive Index Sensing and Cancer Biomarker Biosensing
Next Article in Special Issue
A Novel Complete-Surface-Finding Algorithm for Online Surface Scanning with Limited View Sensors
Previous Article in Journal
Sensorless Estimation Based on Neural Networks Trained with the Dynamic Response Points
Previous Article in Special Issue
Sensor-Enabled Multi-Robot System for Automated Welding and In-Process Ultrasonic NDE
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of the Deep-Sea Plankton Community in Marine Ecosystem with Underwater Robotic Platform

1
School of Information Science and Engineering, Shandong University, Jinan 266237, China
2
Shenzhen Research Institute, Shandong University, Shenzhen 518000, China
3
China National Deep Sea Center, Qingdao 266237, China
4
IETR-INSA UMR-CNRS 6164, 35708 Rennes, France
5
Centre for Excellence in Signal and Image Processing, Department of Electronic and Electrical Engineering, University of Strathclyde, Glasgow G1 1XQ, UK
*
Authors to whom correspondence should be addressed.
Sensors 2021, 21(20), 6720; https://doi.org/10.3390/s21206720
Submission received: 16 August 2021 / Revised: 24 September 2021 / Accepted: 5 October 2021 / Published: 10 October 2021
(This article belongs to the Special Issue Robotic Non-destructive Testing)

Abstract

:
Variations in the quantity of plankton impact the entire marine ecosystem. It is of great significance to accurately assess the dynamic evolution of the plankton for monitoring the marine environment and global climate change. In this paper, a novel method is introduced for deep-sea plankton community detection in marine ecosystem using an underwater robotic platform. The videos were sampled at a distance of 1.5 m from the ocean floor, with a focal length of 1.5–2.5 m. The optical flow field is used to detect plankton community. We showed that for each of the moving plankton that do not overlap in space in two consecutive video frames, the time gradient of the spatial position of the plankton are opposite to each other in two consecutive optical flow fields. Further, the lateral and vertical gradients have the same value and orientation in two consecutive optical flow fields. Accordingly, moving plankton can be accurately detected under the complex dynamic background in the deep-sea environment. Experimental comparison with manual ground-truth fully validated the efficacy of the proposed methodology, which outperforms six state-of-the-art approaches.

1. Introduction

Plankton are organisms that live in oceans and fresh water [1] that play an important role in the material and energy recycling within the marine food chain [2]. The study of plankton community and plankton itself is indispensable for understanding of marine resources and the impacts of climate change on ecosystems [3]. In addition, the number of plankton is a key indicator of carbon and energy cycling [4], and of great significance to species diversity and ecosystem diversity [5]. From the early 19th century to date, many examples of large-scale sensor equipment were used to solve the challenge of getting reliable high-resolution estimates of plankton abundance at depth [6]. Acoustic and optical techniques for the in-situ observation of zooplankton are currently popularly used for plankton distribution assessment. Although acoustic-based observation has outstanding advantages of high observation frequency, it has inaccurate quantification and usually requires the combination of optical image analysis or other traditional sampling of zooplankton. In recent years, a series of advances were made in computer vision [7], including hyperspectral imaging [8], principal component analysis of images [9,10], and deep learning [11,12,13] for image classification [14]. As marine plankton is small and uneven in size, it is difficult to describe it quantitatively, such as with inventory and abundance statistics.
At present, a lot of plankton detection methods are proposed that often rely heavily on the use of sophisticated underwater instruments. J. Craig et al. [15,16] constructed an ICDeep system, based on the Image Intensified Charge Coupled Device (ICCD) camera, to assess the quantity of low-light bioluminescent sources in the marine environment. Philips et al. [17] created a marine biological detector, where a Scientific CMOS (SCMOS) camera was used to image the organisms before conducting statistical analysis of the plankton abundance. With the development of the computer vision, multitarget tracking-enabled automatic analysis was gradually applied to this field [18]. Kocak et al. [19] proposed to use the active contour (snake) models to segment, label, and track images of the snake model for the classification of the plankton. Luca et al. [20] also presented an automatic plankton counting method, which mainly used the interframe difference and the intersection of the bounding boxes to perform multitarget matching. The aforementioned methods achieved some results in automatic analysis and counting. However, there are still some challenges due to the particularity and complexity of plankton’s own form and passive movement mode. Applying machine vision techniques to underwater images or videos is a feasible way to study plankton at present. Underwater plankton imaging has the capacity to detect patterns of the plankton distributions that we would be unable to be tackled by sampling with nets. [21]. Therefore, we consider applying machine vision technology to underwater images or videos is currently a feasible method for studying plankton.
Underwater robots play an important role in various video surveillance tasks including data collection. A mobile robot that can be fixed on a rotatable axis would be advantageous because it provides 360 visual coverage instead of using a fixed image camera installed in a predetermined direction. These mobile robots capture unprecedented shots of marine life in dangerous environments inaccessible to humans. A submarine can push and control the underwater robot to complete the collection of deep-sea data and store the data in the computer for analysis. Some underwater robots are shown in Figure 1.
In this paper, we propose a deep-sea plankton detection method based on the Horn–Schunck (HS) optical flow [22]. The optical flow is the instantaneous velocity of the pixel movement of the moving object on the image plane. The advantage of the optical flow method is that the motion vectors can be estimated by the optical flow vector accurately. In this way, one can detect the plankton and easily analyze statistically its volume using image processing and machine vision. The research on plankton can be specifically divided into density, position, number, individual and total volume, etc. In the case where the spatial position of plankton does not coincide in two consecutive frames, the presence or absence of plankton should be determined according to the following conditions: the time gradient maps at the plankton’s location in two consecutive optical flow fields will be opposite to each other, and the horizontal and vertical gradients of the plankton at that location are equal and their direction is the same. Since the connected components are marked as the location of plankton, the number of connected components can be regarded as the number of plankton. By using this method, we firstly count the number of plankton in the video, followed by a statistical analysis. Various comparative experiments are carried out to benchmark with other methods to fully demonstrate the effectiveness of the proposed methodology.

2. The Proposed Method

2.1. Principle

The deep ocean floor is clear and suitable for video acquisition with active lighting. During the video acquisition process, the camera position and shooting angle change with the movement of the submersible, making the plankton detection task a moving target detection problem under complex and dynamic backgrounds. Two consecutive optical flow field matrices derived from three consecutive video frames in a video are employed. For fast-moving plankton (plankton does not overlap in space in two consecutive frames), the two consecutive optical flow values at the position where the plankton is located are opposite. In practice, the amount of grayscale change is often close to 0. Therefore, the two consecutive optical flows are approximately opposite to each other, and we discuss this situation by setting two thresholds in the experiment section. We use this property to map out the location of the plankton. Figure 2, hereafter provides an overview of the proposed method, which consists of three modules.
As shown in the Module 1 of Figure 2, grayscale images are obtained by weighting three channels of the input frames. In module 2, three convolution operations are performed on two consecutive frames to produce three different gradients(see Figure 3), which correspond to three different convolution kernels. The details of the convolution process are shown in Figure 3 to illustrate this process. We find that the time gradients of the two optical flow fields derived from three consecutive frames of images are opposite in numerical value and direction in the corresponding positions of plankton in the middle frame. In the following description, the time gradients of the two consecutive optical flow fields are represented by t and t . The horizontal gradients of the two consecutive optical flow fields derived from three consecutive frames are equal in magnitude and direction in the corresponding positions of plankton in the middle frame. Similarly, the vertical gradients are also equal. In the following description, the horizontal gradients of the two optical flow fields are represented by x and x , the vertical gradients are y and y . Finally, Module 3 is for dual thresholding, which is explained separately when discussing the parameters later.

2.2. Proof

In the HS optical flow method, the constraint equation of optical flow can be established as Equation (2) according to the premise of the optical flow method: invariance of gray level [22]. Three first-order differences are used to replace the horizontal, vertical, and time gradients. Let the gray value at plankton’s position in the middle frame be I x , y , t , where the subscripts x and y are the pixel index, and t is the time index. The position of plankton changes with the movement of ocean current and the camera lens. As shown in Figure 4, the plankton is small-sized, so its position in frame t doesn’t overlap in frame t + 1. When it changes from position 1 to position 2, the gray value corresponding to position 2 of plankton at frame t − 1 is the background gray value I x , y , t 1 . In a similar way, when the position of plankton changes from position 2 to position 3, the gray value corresponding to position 2 at frame t + 1 becomes the background gray value I x , y , t + 1 . Based on the characteristics of deep-sea underwater video, the background around the plankton is invariant in time, i.e.,:
I x , y , t 1 = I x , y , t + 1
x u + y v + t = 0
The time gradients at the plankton’s positions in the two adjacent optical flow fields are:
t = 1 2 ( I x , y , t I x , y , t 1 + I x + 1 , y , t I x + 1 , y , t 1 )
t = 1 2 ( I x , y , t + 1 I x , y , t + I x + 1 , y , t + 1 I x + 1 , y , t )
Based on Equation (1), the background gray value I x , y , t 1 = I x , y , t + 1 , t = t , the time gradients of the two optical flow fields derived from three consecutive frames of images are opposite in the corresponding positions of plankton in the middle frame.
The horizontal gradients of the plankton’s location in the two optical flow fields are:
x = 1 2 ( I x + 1 , y , t I x , y , t + I x + 1 , y , t 1 I x , y , t 1 )
x = 1 2 ( I x + 1 , y , t + 1 I x , y , t + 1 + I x + 1 , y , t I x , y , t )
The same way, based on Equation (1), we can get that x = x , i.e., the horizontal gradients of the two optical flow fields derived from three consecutive frames are equal in the corresponding positions of plankton in the middle frame. In the same way, we can get y = y .
In fact, in the process of proof, the time and space gradients are estimated in a 2 × 2 × 2 cubic neighborhood by taking the mean.
Then, we iterate n times for gray gradient relaxation by setting the initial conditions as v 0 = v 0 = 0 and u 0 = u 0 = 0 .
Δ = ( x u n + y v n + t α 2 + x 2 + y 2 )
u n + 1 = u n x Δ
v n + 1 = v n y Δ
The parameter α 2 reflects the smoothness constraints of the HS optical flow algorithm; Δ is an iteration factor in the process of the iterative algorithm; x and y are the horizontal and vertical gradients, and u and v are the horizontal and vertical optical flow field matrices, respectively.
The relationships of Equations (7)–(9) are represented by a series, where the number of iterations is n. Let’s substitute Equations (7)–(9), the new formulas are as follows:
u n + 1 = u n x ( x u n + y v n + t α 2 + x 2 + y 2 )
v n + 1 = v n y ( x u n + y v n + t α 2 + x 2 + y 2 )
where u n + 1 and u n are two horizontal optical flow fields before and after the n-th iteration, v n + 1 and v n are two vertical optical flow fields before and after the n-th iteration. We can derive u n + 1 = u n + 1 , v n + 1 = v n + 1 . When n = 0, we have:
v 1 = v 0 y ( x u 0 + y v 0 + t α 2 + x 2 + y 2 )
v 1 = v 0 y ( x u 0 + y v 0 + t α 2 + x 2 + y 2 )
v 1 and v 1 are the two consecutive vertical optical flow field at the first iteration. If the time gradients of the last two optical flow fields are opposite, that is t = t , we can get: v 1 = v 1 .
When n = k , v k + 1 = v k + 1 . That is, Equations (14) and (15) are opposite:
v k + 1 = v k y ( x u k + y v k + t α 2 + x 2 + y 2 )
v k + 1 = v k y ( x u k + y v k + t α 2 + x 2 + y 2 )
where v k + 1 and v k + 1 represent the previous and the next vertical optical flow field matrix at the (k + 1)th iteration, respectively.
When n = k + 1 , we can show that v k + 2 = v k + 2
v k + 2 = v k + 1 y ( x u k + 1 + y v k + 1 + t α 2 + x 2 + y 2 )
v k + 2 = v k + 1 y ( x u k + 1 + y v k + 1 + t α 2 + x 2 + y 2 )
By adding Equations (16) and (17), and substituting v k + 1 = v k + 1 , x = x and y = y into Equation (16) and Equation (17), respectively, we have:
v k + 2 = v k + 2
Therefore, for fast-moving plankton, the values of the vertical optical flow field matrices of the space position where the plankton is located are opposite from each other: v = v , and the same applies horizontally: u = u .

2.3. The Volume of Plankton

Based on the above proof, one can calculate the number of pixels where plankton is located, and then multiply the actual size of a pixel to obtain the area of plankton. The resolution of the known image is h e i g h t × w i d t h . According to camera internal reference, the actual range of our field of view is about W m by H m. The calculation of the actual area is given by:
S = N × ( W / w i d t h ) × ( H / h e i g h t )
where N is the number of pixels, and S is the corresponding actual surface. A method of approximate calculation is adopted here. Firstly, we can get the radius of a circle that has the same area as the plankton, and then calculate the volume of the sphere based on that radius. The advantage of this method is that we can get the 3D volume of an irregular object only by its area [23]. In addition, we can predict the type of plankton based on the estimated size, laying the foundation for the later identification of plankton types. The volume can be calculated by:
V = 4 3 × π 1 2 × S 3 2
The proposed method adds its own theoretical innovation on the basis of the original optical flow method and was proved mathematically. In this way, the complexity and passive motion patterns of plankton are well-solved, and the accuracy improves as the above problems are solved.

3. Experimental Results and Analysis

The data capture was provided by the China National Deep Sea Center. The data set was obtained by an underwater robotic nondestructive testing system carried by a deep-sea manned submersible. The camera’s technical specifications are: resolution: 1080i HDTV; minimum illumination: 2l ux; optical zoom: 10 times; digital zoom: 12 times; aperture range: 3.2 mm–32 mm; video aspect ratio: 16:9 or 4:3. In this study, three six-minute videos of the plankton community from appearing to disappearing from the screen were selected, which were obtained from a submarine on the western Pacific sea mountain slope, and the diving depths are 2741.88 m and 5555.68 m, corresponding to 76 and 77 dives, respectively. The reason why the three videos are selected is that plankton appeared more frequently in them. Due to the complexity of the deep-sea environment and the irregular camera movement, the background is complex and dynamic. In this case, using high-precision image processing technology to study the plankton community from appearing to disappearing from the screen can effectively distinguish sedimentary clouds and plankton community in images. Examples of deep-sea plankton images are shown in Figure 5 and the details of data set including diving number, date, diving time, longitude, latitude and depth are shown in Table 1.

3.1. Number and Volume of Plankton

Processing the recorded video of a complete plankton community from appearing to disappearing from the screen, the results obtained are shown in Figure 6. Figure 6a shows the variation of the number of plankton in three six-minute videos, and Figure 6b shows the variation of the volume of the corresponding three videos. The process of plankton appearing in front of the camera to disappearing is shown in Figure 6c,d. In the first 30 s of Figure 6c, the amount of plankton is small and the detection results are more accurate. We can see that the amount of plankton rises in the last 30 s of Figure 6c. For dense particle clouds, overlap, and hence, occlusion occurs frequently, which leads to relatively low average accuracy and recall rates.
The actual volume curve of plankton in the video is shown in Figure 6b,d. We can see that the volume curve and the quantity curve of plankton generally follow the same trend. At the 40th second in Figure 6c, the plankton community moves away from the camera and then comes back, resulting in a smaller scene and a smaller overall volume due to perspective. So, we can see that the volume curve goes down and then goes up from Figure 6d.

3.2. Comparison with Six Target Detection Methods

The proposed method is compared with six state-of-the-art methods for performance evaluation. The results are shown in Figure 7, where Figure 7a represents some original images of the video, including sediment clouds, plankton, and uneven backgrounds. Top-Hat transform [24] is used to detect the location of the plankton in the image as shown in Figure 7b, the weakness of this algorithm is that there are some missed cases. Figure 7c and Figure 7d show the detection results of the frame difference method [25] and the motion estimation and image matching method [26], respectively. We show the result from the scan line marking method [27] in Figure 7e results from the simple block-based sum of absolute differences flow (SD) method [28], and the Lucas–Kanade (LK) optical flow method [29] are given in Figure 7f,g. The weakness of the above three methods is that there are a few false positives, and both Figure 7c,e detected the sediment cloud in the background by mistake. The result of Figure 7h is obtained using the proposed method. After comparing with the manual ground truth, we find that the plankton detected by the proposed method is more consistent with the original image in Figure 7a.
We take 20 images of the video, and the data are cleaned by manual counting to get the ground-truth. Then, we compare the number of plankton, recall rate, precision rate, and F 1 - s c o r e of the seven methods. When using 10 frames in the first 30 s of the video, the amount of plankton is small and the detection results are more accurate, the average accuracy rate is 0.901, the average recall rate is 0.955, and F 1 - s c o r e is 0.927. In addition, the equations and related symbols are shown in Table 2 and Equations (21)–(23). The results are shown in Table 3 and Table 4. Taking 10 frames in the last 30 s of the video, the amount of plankton is high. For dense particle clouds, overlap can easily occur, and hence, occlusion occurs frequently, so the average accuracy and recall rates are relatively low, i.e., 0.895 and 0.943, respectively, and the F 1 - s c o r e is 0.918, The results are shown in Table 5 and Table 6. In addition, we randomly selected 10 frames from the video for testing. The experimental results are shown in Table 7 and Table 8. The performance of the proposed method is still very good. We use bold font to highlight the best results in each category in Table 4, Table 6, Table 8 and Table 9.
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l

3.3. Discussion of Parameters

For each imaging system, there is a depth of field within which the closest field objects and farthest field objects are all in focus. If we deploy the system in air, the light intensity for the near field object and far field object should not be different in theory. However, when deployed in seawater, the light intensity changes as the light propagates in the water from near-field to far-field because of scattering caused by seawater and particles in the seawater. Therefore, during the experiment, there are two situations that need to be discussed. Firstly, ’grayscale invariance’ is one of the prerequisites of the HS optical flow method, but in actual operation, the amount of grayscale change is often close to 0 but not equal to 0. Therefore, the threshold β 1 is set to handle this situation, as shown in Equation (24).
u + u < β 1 o r v + v < β 1
Secondly, when there is no plankton and the optical flow happens to be small, if the values of the optical flow are not the opposite but the sum still conforms to Equation (24), the threshold β 2 needs to be set to solve this situation, as shown in Equation (25).
u u > β 2 o r v v > β 2
The best threshold value is obtained by traversing the range value, the scope of β 1 is 0.05 to 0.35 , step size is 0.05 , the scope of β 2 is 3–9, and the step length is 1. Then, the original images and all those resulting from different thresholds are represented by vectors. At last, we calculate the cosine similarity between two images, that is the calculation of cosine distance between two vectors; the larger the cosine distance between the two vectors, the more similar the two images are. The results are shown in Table 9.
As shown in Figure 8, Figure 8a is the original image, Figure 8c represents the result of using the threshold β 2 , and the one without the threshold β 2 is shown in Figure 8b.

3.4. Time Complexity Comparison

The time complexity comparison of the proposed method and six state-of-the-art methods is provided in Table 10. We select a one-minute video of 1440 frames and calculate the computation time to measure the time complexity of difference methods. Although the proposed method doesn’t have great advantage in term of the time complexity, it outperforms other methods in accurate detection of plankton. In terms of the detection efficiency, some experimental comparisons were carried out. Based on the same one-minute video, the computation time and recall rate of the following four different strategies are compared, respectively. We sample pixels at intervals of 1, and take interval frames from full sequence at intervals of 1 frame. According to the results shown in Table 11, the interval between pixels has a weak influence on the error of the result, where the recall rate, precision rate, and F 1 - s c o r e are the closest to the original image’s result, and the detection efficiency is improved by greatly reducing the calculation time.

4. Conclusions

Detection of plankton plays an important role in the exploration and research of deep-sea areas. Variations in the quantity and spatial distribution of plankton determine the function of the entire marine ecosystem. In this paper, we introduce a method for deep-sea plankton community detection in marine ecosystem with an underwater robotic platform. Compared with that of traditional methods, our method simultaneously improves the precision and recall of plankton detection. The obtained results and the proved theory provide a scientific basis for studying the material cycle and energy flow of deep-sea ecosystems. For our future work, with a view to strengthening the proposed solution, we aim to improve our plankton detection approach, and then conduct studies for plankton recognition and identification of their species.

Author Contributions

Conceptualization, J.W. and Z.D.; methodology, J.W. and M.Y.; writing—original draft preparation, J.W., K.K. and D.W.; visualization, J.R. and K.K.; writing—review and editing J.W., K.K., Q.Z. and J.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key R&D Program of China under Grant No. 2018YFC0831503.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank National Deep Sea Center for the provision of the video material used in the experimental validation of the method.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Brierley, A. Plankton. Curr. Biol. 2017, 27, 478–483. [Google Scholar] [CrossRef]
  2. Wang, K.; Razzano, M.; Mou, X. Cyanobacterial blooms alter the relative importance of neutral and selective processes in assembling freshwater bacterioplankton community. Sci. Total Environ. 2020, 706, 135724. [Google Scholar] [CrossRef]
  3. Forster, D.; Qu, Z.; Pitsch, G.; Bruni, E.P.; Kammerlander, B.; Proschold, T.; Sonntag, B.; Posch, T.; Stoeck, T. Lake Ecosystem Robustness and Resilience Inferred from a Climate-Stressed Protistan Plankton Network. Microorganisms 2021, 9, 549. [Google Scholar] [CrossRef] [PubMed]
  4. Henrik, L.; Torkel, G.; Benni, W. Planton community structure and carbon cycling on the western coast of Greenland during the stratified summer situation. II.Heterotrophic dinoflagellates and ciliates. Aquat. Microb. Ecol. 1999, 16, 217–232. [Google Scholar]
  5. Lu, H.; Wang, D.; Li, J.; Li, X.; Kim, H.; Serikawa, S.; Humar, I. Conet: A cognitive ocean network. IEEE Wirel. Commun. 2019, 26, 90–96. [Google Scholar] [CrossRef] [Green Version]
  6. Wiebe, P.; Benfield, M. From the Hensen net toward four-dimensional biological oceanography. Prog. Oceanogr. 2003, 56, 7–136. [Google Scholar] [CrossRef]
  7. Yan, Y.; Ren, J.; Sun, G.; Zhao, H.; Han, J.; Li, X.; Marshall, S.; Zhan, J. Unsupervised image saliency detection with Gestalt-laws guided optimization and visual attention based refinement. Pattern Recogn. 2018, 79, 65–78. [Google Scholar] [CrossRef] [Green Version]
  8. Tschannerl, J.; Ren, J.; Yuen, P.; Sun, G.; Zhao, H.; Yang, Z.; Wang, Z.; Marshall, S. MIMR-DGSA: Unsupervised hyperspectral band selection based on information theory and a modified discrete gravitational search algorithm. Inform. Fusion 2019, 51, 189–200. [Google Scholar] [CrossRef] [Green Version]
  9. Zabalza, J.; Ren, J.; Yang, M.; Zhang, Y.; Wang, J.; Marshall, S.; Han, J. Novel Folded-PCA for improved feature extraction and data reduction with hyperspectral imaging and SAR in remote sensing. ISPRS J. Photogramm. 2014, 93, 112–122. [Google Scholar] [CrossRef] [Green Version]
  10. Zheng, Q.; Tian, X.; Yang, M.; Wu, Y.; Su, H. PAC-Bayesian framework based drop-path method for 2D discriminative convolutional network pruning. Multidim. Syst. Sign. Process. 2020, 31, 793–827. [Google Scholar] [CrossRef]
  11. Zheng, Q.; Yang, M.; Tian, X.; Jiang, N.; Wang, D. A Full Stage Data Augmentation Method in Deep Convolutional Neural Network for Natural Image Classification. Discrete Dyn. Nat. Soc. 2020, 2020, 1–11. [Google Scholar] [CrossRef]
  12. Zheng, Q.; Yang, M.; Yang, J.; Zhang, Q.; Zhang, X. Improvement of Generalization Ability of Deep CNN via Implicit Regularization in Two-Stage Training Process. IEEE Access 2018, 6, 15844–15869. [Google Scholar] [CrossRef]
  13. Zheng, Q.; Tian, X.; Jiang, N.; Yang, M. Layer-wise learning based stochastic gradient descent method for the optimization of deep convolutional neural network. J. Intell. Fuzzy Syst. 2019, 37, 5641–5654. [Google Scholar] [CrossRef]
  14. Zheng, Q.; Yang, M.; Zhang, Q.; Zhang, X. Fine-grained image classification based on the combination of artificial features and deep convolutional activation features. In Proceedings of the 2017 IEEE/CIC International Conference on Communications in China, Qingdao, China, 22–24 October 2019; pp. 1–6. [Google Scholar]
  15. Craig, J. Distribution of Deep-Sea Bioluminescence across the Mid-Atlantic Ridge and Mediterranean Sea: Relationships with Surface Productivity, Topography and Hydrography. Ph.D. Thesis, University of Aberdeen, Aberdeen, UK, 2012. [Google Scholar]
  16. Craig, J.; Priede, I.G.; Aguzzi, J.; Company, J.B.; Jamieson, A.J. Abundant bioluminescent sources of low-light intensity in the deep Mediterranean Sea and North Atlantic Ocean. Mar. Biol. 2015, 162, 1637–1649. [Google Scholar] [CrossRef]
  17. Phillips, B.; Gruber, D.; Vasan, G.; Pieribone, V.; Sparks, J.; Roman, C. First Evidence of Bioluminescence on a "Black Smoker" Hydrothermal Chimney. Oceanography 2016, 29, 10–12. [Google Scholar] [CrossRef] [Green Version]
  18. Kydd, J.; Rajakaruna, H.; Briski, E.; Bailey, S. Examination of a high resolution laser optical plankton counter and FlowCAM for measuring plankton concentration and size. J. Sea Res. 2018, 133, 2–10. [Google Scholar] [CrossRef] [Green Version]
  19. Kocak, D.M.; Da Vitoria Lobo, N.; Widder, E.A. Computer vision techniques for quantifying, tracking, and identifying bioluminescent plankton. IEEE J. Oceanic Eng. 2002, 24, 81–95. [Google Scholar] [CrossRef]
  20. Mazzei, L.; Marini, S.; Craig, J.; Aguzzi, J.; Fanelli, E.; Priede, I.G. Automated Video Imaging System for Counting Deep-Sea Bioluminescence Organisms Events. In Proceedings of the 2014 ICPR Workshop on Computer Vision for Analysis of Underwater Imagery, Stockholm, Sweden, 24 August 2014; pp. 57–64. [Google Scholar]
  21. Schmid, M.; Cowen, R.; Robinson, K.; Luo, J.; Briseno-Avena, C.; Sponaugle, S. Prey and predator overlap at the edge of a mesoscale eddy: Fine-scale, in-situ distributions to inform our understanding of oceanographic processes. Sci. Rep. 2020, 10, 1–16. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Horn, B.K.P.; Schunck, B.G. Determining optical flow. Artif. Intell. 1981, 17, 185–203. [Google Scholar] [CrossRef] [Green Version]
  23. Iversen, M.H.; Nowald, N.; Ploug, H.; Jackson, G.A.; Fischer, G. High resolution profiles of vertical particulate organic matter export off Cape Blanc, Mauritania: Degradation processes and ballasting effects. Deep Sea Res. Part I Oceanogr. Res. Pap. 2010, 57, 771–784. [Google Scholar] [CrossRef]
  24. He, F.; Hu, Y.; Wang, J. Texture Detection of Aluminum Foil Based on Top-Hat Transformation and Connected Region Segmentation. Adv. Mater. Sci. Eng. 2020, 3, 1–7. [Google Scholar] [CrossRef] [Green Version]
  25. Wei, H.; Peng, Q. A block-wise frame difference method for real-time video motion detection. Int. J. Adv. Robot. Syst. 2018, 15, 1–13. [Google Scholar] [CrossRef] [Green Version]
  26. Yang, S.; Shi, X.; Zhang, G.; Lv, C.; Yang, X. Measurement of 3-DOF Planar Motion of the Object Based on Image Segmentation and Matching. Nanomanuf. Metrol. 2019, 2, 124–129. [Google Scholar] [CrossRef]
  27. Wang, W.; Wang, W.; Yan, Y. A Scan-Line-Based Hole Filling Algorithm for Vehicle Recognition. Adv. Mater. Res. 2011, 179, 92–96. [Google Scholar] [CrossRef]
  28. Lu, N.; Wang, J.; Wu, Q. An optical flow and inter-frame block-based histogram correlation method for moving object detection. Int. J. Model. Identif. Control 2010, 10, 87–93. [Google Scholar] [CrossRef]
  29. Mohiuddin, K.; Alam, M.; Das, A.; Munna, T.; Allayear, S.; Ali, H. Haar Cascade Classifier and Lucas-Kanade Optical Flow Based Realtime Object Tracker with Custom Masking Technique. Adv. Inf. Commun. Netw. 2019, 2, 398–410. [Google Scholar]
Figure 1. Underwater robot pattern: submarine can push and control underwater robot to complete collection of deep-sea data, then store data in computer for analysis.
Figure 1. Underwater robot pattern: submarine can push and control underwater robot to complete collection of deep-sea data, then store data in computer for analysis.
Sensors 21 06720 g001
Figure 2. Main processing blocks of proposed algorithm. Module 1 is for preprocessing, whilst Module 2 performs 3D convolution on video frame to extract dense optical flow. Module 3 is a dual threshold setting to determine whether a plankton is contained at a specific location or not(see Section 3.3).
Figure 2. Main processing blocks of proposed algorithm. Module 1 is for preprocessing, whilst Module 2 performs 3D convolution on video frame to extract dense optical flow. Module 3 is a dual threshold setting to determine whether a plankton is contained at a specific location or not(see Section 3.3).
Sensors 21 06720 g002
Figure 3. Three convolution kernels corresponding in time and space. Two consecutive frames are used to form a 3D matrix whose size is ( h e i g h t + 1 ) × ( w i d t h + 1 ) × 2 . Size of filter is 2 × 2 × 2 . Result of each operation is gradient of the pixel at upper-left corner of convolution kernel.
Figure 3. Three convolution kernels corresponding in time and space. Two consecutive frames are used to form a 3D matrix whose size is ( h e i g h t + 1 ) × ( w i d t h + 1 ) × 2 . Size of filter is 2 × 2 × 2 . Result of each operation is gradient of the pixel at upper-left corner of convolution kernel.
Sensors 21 06720 g003
Figure 4. Position of plankton in three consecutive frames.
Figure 4. Position of plankton in three consecutive frames.
Sensors 21 06720 g004
Figure 5. Example images of deep sea plankton.
Figure 5. Example images of deep sea plankton.
Sensors 21 06720 g005
Figure 6. (a) Number of plankton in three six-minute videos. (b) Total volume of plankton in three six-minute videos. (c) Number of plankton in a period. (d) Volume of plankton in a period.
Figure 6. (a) Number of plankton in three six-minute videos. (b) Total volume of plankton in three six-minute videos. (c) Number of plankton in a period. (d) Volume of plankton in a period.
Sensors 21 06720 g006
Figure 7. Location of plankton detected with seven different methods: (a) original image; (b) Top-Hat transform; (c) frame difference method; (d) motion estimation and image match; (e) scan line marking method; (f) simple block-based sum of absolute differences flow (SD); (g) Lucas–Kanade (LK) optical flow method, and (h) proposed method.
Figure 7. Location of plankton detected with seven different methods: (a) original image; (b) Top-Hat transform; (c) frame difference method; (d) motion estimation and image match; (e) scan line marking method; (f) simple block-based sum of absolute differences flow (SD); (g) Lucas–Kanade (LK) optical flow method, and (h) proposed method.
Sensors 21 06720 g007
Figure 8. Comparison with or without threshold: (a) original image; (b) one without threshold β 2 , and (c) result of using threshold β 2 .
Figure 8. Comparison with or without threshold: (a) original image; (b) one without threshold β 2 , and (c) result of using threshold β 2 .
Sensors 21 06720 g008
Table 1. Details of datasets including diving number, date, diving time, longitude, latitude, and depth.
Table 1. Details of datasets including diving number, date, diving time, longitude, latitude, and depth.
Diving NumberDateDiving TimeLongitudeLatitudeDepth
7617 July 20148.95 h155.32 E–155.34 E15.50 N–15.52 N2741.88 m
7721 July 201410.33 h154.58 E–154.59 E15.70 N–15.72 N5555.68 m
Table 2. Confusion Matrix.
Table 2. Confusion Matrix.
RelevantNonrelevant
RetrievedTrue Positives (TP)False Positives (FP)
Not RetrievedFalse Negatives (FN)True Negatives (TN)
Table 3. In first 30 s, comparison of number of detected plankton using seven methods and Ground-Truth.
Table 3. In first 30 s, comparison of number of detected plankton using seven methods and Ground-Truth.
The Ten Frames:12345678910Meanstd
Top-Hat101291517181822242216.74.9
Frame difference2022182221171915171418.52.7
Image match2624232124171515161619.74.1
Scan line marking11789109109989.01.1
SD1713111113131210121012.21.9
LK1714131313131211111012.71.8
Proposed method1614121212111210111012.01.7
Ground-Truth14131212111211910911.31.6
Table 4. In first 30 s, comparison of recall rate, precision rate, and F 1 - s c o r e of seven methods.
Table 4. In first 30 s, comparison of recall rate, precision rate, and F 1 - s c o r e of seven methods.
The Ten
Frames:
12345678910Average
Top-HatPrecision0.90.850.890.730.590.610.560.360.420.410.632
Recall0.640.920.670.920.910.920.910.89110.878
F 1 0.750.880.760.810.720.730.690.510.590.580.73
Frame
difference
Precision0.650.550.610.50.520.650.560.530.590.570.573
Recall0.930.920.920.9210.9210.8910.890.939
F 1 0.770.690.730.650.680.760.720.660.740.690.712
Image
match
Precision0.540.540.490.520.460.650.670.530.560.560.552
Recall110.920.9210.920.910.890.910.946
F 1 0.70.70.640.660.630.760.770.660.690.720.7
Scan line
marking
Precision0.820.860.880.780.90.890.90.890.890.880.869
Recall0.640.46580.580.820.670.820.890.80.780.704
F 1 0.720.60.70.670.860.760.860.890.840.830.778
SDPrecision0.760.920.910.910.770.850.830.80.750.80.83
Recall0.930.920.830.830.910.920.910.890.90.890.893
F 1 0.840.920.870.870.830.880.870.840.820.840.86
LKPrecision0.760.860.850.850.850.850.830.730.820.80.82
Recall0.930.930.920.9210.920.910.890.90.890.921
F 1 0.840.890.880.880.920.880.870.80.860.840.868
Proposed
method
Precision0.810.93110.9210.830.80.820.90.901
Recall0.9311110.920.910.890.910.955
F 1 0.870.96110.960.960.870.840.860.950.927
Table 5. In last 30 s, comparison of number of detected plankton using seven methods and Ground-Truth.
Table 5. In last 30 s, comparison of number of detected plankton using seven methods and Ground-Truth.
The Ten Frames:12345678910Meanstd
Top-Hat221717161312131291814.93.6
Frame difference2828302430152828272426.24.2
Image match2222252631273228312426.83.5
Scan line marking1315141116151915151314.62.0
SD1621222323232223201721.02.4
LK1622232223232123201821.12.4
Proposed method1521222122212122191620.02.4
Ground-Truth1919191821182121181518.91.8
Table 6. In last 30 s, comparison of recall rate, precision rate, and F 1 - s c o r e of seven methods.
Table 6. In last 30 s, comparison of recall rate, precision rate, and F 1 - s c o r e of seven methods.
The Ten
Frames:
12345678910Average
Top-HatPrecision0.770.940.940.930.920.920.920.9210.770.903
Recall0.890.840.840.830.570.610.570.520.50.930.762
F 1 0.830.890.890.880.70.730.70.660.670.840.827
Frame
difference
Precision0.640.640.60.710.650.930.710.710.590.580.676
Recall0.950.950.950.940.950.780.950.950.880.930.923
F 1 0.760.760.740.810.770.850.810.810.710.710.78
Image
match
Precision0.820.820.720.650.670.630.630.710.550.580.678
Recall0.950.950.950.940.950.940.950.950.940.930.945
F 1 0.880.880.820.770.790.750.760.810.690.710.789
Scan line
marking
Precision0.920.930.9310.940.930.950.930.930.920.938
Recall0.630.740.680.610.710.780.860.670.780.80.726
F 1 0.750.820.790.760.810.850.90.780.850.860.82
SDPrecision0.940.850.820.740.870.740.910.870.850.820.841
Recall0.790.940.940.940.950.940.950.950.940.930.893
F 1 0.860.890.880.830.910.830.930.910.890.870.88
LKPrecision0.940.820.780.770.870.740.950.870.850.780.837
Recall0.790.940.950.940.950.940.950.950.940.930.928
F 1 0.860.880.860.850.910.830.950.910.890.850.88
Proposed
method
Precision10.850.820.860.910.810.90.910.950.940.895
Recall0.790.950.9510.950.940.90.95110.943
F 1 0.880.90.880.920.930.870.90.930.970.970.918
Table 7. Comparison of number of detected plankton from 10 randomly selected frames.
Table 7. Comparison of number of detected plankton from 10 randomly selected frames.
The Ten Frames:12345678910
Top-Hat19341474533002
Frame difference28126010159143151381012
Image match7812912106120161511914
Scan line marking81191519974223
SD17219508312415119
LK163190086121151110
Proposed method941050718917115
Ground-Truth87941667626115
Table 8. Comparison of recall, precision, and F1-score of detected plankton from 10 randomly selected frames.
Table 8. Comparison of recall, precision, and F1-score of detected plankton from 10 randomly selected frames.
The Ten
Frames:
12345678910Average
Top-HatPrecision0.950.881.000.940.890.671.000.000.001.000.733
Recall0.210.321.000.660.531.000.500.000.000.400.462
F 1 0.340.471.000.780.660.800.670.000.000.570.529
Frame
difference
Precision0.280.350.10.380.490.130.380.130.100.330.267
Recall0.920.961.000.910.921.000.831.001.000.800.934
F 1 0.430.510.180.540.640.230.520.230.180.470.393
Image
match
Precision0.900.700.080.570.580.130.330.090.110.290.378
Recall0.800.961.000.910.921.000.831.001.000.800.922
F 1 0.850.810.150.700.710.230.470.170.200.430.472
Scan line
marking
Precision0.940.731.000.880.710.291.000.500.500.670.722
Recall0.860.931.000.680.921.000.671.001.000.400.846
F 1 0.900.821.000.770.800.450.800.670.670.500.738
SDPrecision0.470.460.000.720.571.001.001.001.000.560.678
Recall0.930.950.000.910.930.500.831.001.001.000.805
F 1 0.620.620.000.800.710.670.911.001.000.720.705
LKPrecision0.500.470.000.720.571.001.001.001.000.500.676
Recall0.940.950.000.940.910.500.831.001.001.000.807
F 1 0.650.630.000.820.700.670.911.001.000.670.705
Proposed
method
Precision0.880.860.000.890.811.000.861.001.001.000.830
Recall0.950.960.000.950.950.501.001.001.001.000.831
F 1 0.910.910.000.920.870.670.921.001.001.000.820
Table 9. Select best threshold by comparing cosine distance between two vectors.
Table 9. Select best threshold by comparing cosine distance between two vectors.
Threshold:0.050.100.150.200.250.300.35
30.0500.0590.0650.0690.0720.0740.076
40.0480.0570.0620.0660.0690.0720.074
50.0470.0550.0610.0650.0680.0700.072
60.0460.0540.0590.0760.0690.0660.070
70.0450.0530.0580.0620.0650.0670.069
80.0440.0520.0570.0610.0640.0660.068
90.0440.0510.0560.0600.0630.0650.067
Table 10. Time complexity comparison of proposed method and 6 state-of-the-art methods in a 1-min video of 1440 frames.
Table 10. Time complexity comparison of proposed method and 6 state-of-the-art methods in a 1-min video of 1440 frames.
Top-HatFrame DifferenceImage MatchScan Line MarkingSDLKProposed Method
3478 s176 s13,149 s4476 s989 s1070 s1112 s
Table 11. Time complexity comparison of different sampling in a 1-min video of 1440 frames.
Table 11. Time complexity comparison of different sampling in a 1-min video of 1440 frames.
A Total of 116 PlanktonTake Interval Frames from full Sequence
QuantityPrecisionRecall F 1 Calculation Time
Pixel interval sampling810.860.60.71137 s
All the pixels300.830.230.36618 s
A Total of 116 PlanktonFull Sequence
QuantityPrecisionRecall F 1 Calculation Time
Pixel interval sampling1100.950.910.93436 s
Full sequence1130.970.950.961112 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, J.; Yang, M.; Ding, Z.; Zheng, Q.; Wang, D.; Kpalma, K.; Ren, J. Detection of the Deep-Sea Plankton Community in Marine Ecosystem with Underwater Robotic Platform. Sensors 2021, 21, 6720. https://doi.org/10.3390/s21206720

AMA Style

Wang J, Yang M, Ding Z, Zheng Q, Wang D, Kpalma K, Ren J. Detection of the Deep-Sea Plankton Community in Marine Ecosystem with Underwater Robotic Platform. Sensors. 2021; 21(20):6720. https://doi.org/10.3390/s21206720

Chicago/Turabian Style

Wang, Jiaxing, Mingqiang Yang, Zhongjun Ding, Qinghe Zheng, Deqiang Wang, Kidiyo Kpalma, and Jinchang Ren. 2021. "Detection of the Deep-Sea Plankton Community in Marine Ecosystem with Underwater Robotic Platform" Sensors 21, no. 20: 6720. https://doi.org/10.3390/s21206720

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop