Next Article in Journal
Energy Harvesting Techniques for Wireless Sensor Networks/Radio-Frequency Identification: A Review
Previous Article in Journal
Optimal Incentive Contract for Sales Team with Loss Aversion Preference
Article Menu
Issue 7 (July) cover image

Export Article

Symmetry 2019, 11(7), 866;

Segmentation of Laterally Symmetric Overlapping Objects: Application to Images of Collective Animal Behavior
Institute of Complex Systems, South Bohemian Research Center of Aquaculture and Biodiversity of Hydrocenoses, Kompetenzzentrum MechanoBiologie in Regenerativer Medizin, Faculty of Fisheries and Protection of Waters, University of South Bohemia in České Budějovice, Zámek 136, 373 33 Nové Hrady, Czech Republic
Author to whom correspondence should be addressed.
Received: 5 June 2019 / Accepted: 27 June 2019 / Published: 3 July 2019


Video analysis is currently the main non-intrusive method for the study of collective behavior. However, 3D-to-2D projection leads to overlapping of observed objects. The situation is further complicated by the absence of stall shapes for the majority of living objects. Fortunately, living objects often possess a certain symmetry which was used as a basis for morphological fingerprinting. This technique allowed us to record forms of symmetrical objects in a pose-invariant way. When combined with image skeletonization, this gives a robust, nonlinear, optimization-free, and fast method for detection of overlapping objects, even without any rigid pattern. This novel method was verified on fish (European bass, Dicentrarchus labrax, and tiger barbs, Puntius tetrazona) swimming in a reasonably small tank, which forced them to exhibit a large variety of shapes. Compared with manual detection, the correct number of objects was determined for up to almost 90 % of overlaps, and the mean Dice-Sørensen coefficient was around 0.83 . This implies that this method is feasible in real-life applications such as toxicity testing.
image segmentation; lateral objects; 2D vision

1. Introduction

The study of collective behavior is a challenging task for any method of individual object tracking. Since the artificial marking of individuals, e.g., by an electronic device, can affect their behavior [1], it is necessary to use nonintrusive methods. Due to its excellent spatial and time resolution, video analysis is the most prominent among these methods. However, the disadvantage of this method is that it only creates a 2D projection of observed objects that are, in reality, always 3D. Indeed, the 2D projection of the whole space unavoidably leads to overlapping of objects in the image. Moreover, information about the shapes and textures of objects is irrecoverably lost. This can be crucial in tracking of individual objects, especially if the density of the objects is high [2]. This holds mainly for video analysis of typical collective behavior like fish schooling, bird flocking, or crowds of people, when the individuals are overlapping most of the time [3,4,5].
The most commonly used solutions of the problem of multiple object tracking using machine vision systems still ignore the overlaps. Nevertheless, the overlaps can be typically detected from morphological parameters (e.g., area, perimeter, eccentricity, and their combinations) of the image binary mask, and the trajectories of the individuals are then reconstructed using track extrapolation [6], particle filter [7], or, more frequently, Kalman movement prediction [8,9]. These methods work well for less crowded scenes where, in addition, the mobility of the individuals is low. Solutions that are directly aimed at observing the interactions in groups usually utilize texture matching before and after collision [10] instead of movement prediction. This introduces some robustness into the approach but important information about the movement is not utilized.
None of these methods mentioned above can track individuals in very dense scenes, where the objects are overlapping most of the time. For these complicated cases, the method of rigid pattern matching can be used [11]. This method requires the shape of the object to be invariant, which is rarely the case in practical applications, and thus the detection rate of this method is low. Most of the more advanced solutions include complex nonlinear optimizations to fit a set of ellipsoids around a fish’s central line in 3D [12]. This approach works for dense scenes but the model used contains many degrees of freedom, which leads to significant inaccuracies, and requires high-resolution imaging and a high frame rate. In addition, this method is not single-image and requires knowledge of the number of objects and the previous state of the system. But the previous state of the system can unavoidably lead to the error propagation and the number of objects is not always known, for example, in cases where the observed volume contains hideouts or is not fully open. A method which is reliable under such conditions is crucial for the study of collective behavior, since many interesting behavioral patterns are observed mainly in complex environments with obstacles, hideouts, and inanimate models [13,14].
In this paper, for the very first time, we propose a robust, nonlinear, optimization-free, and pose-invariant way of solving laterally symmetric, overlapping objects. The method is truly single-image and—as verified on image sets of fish schools—tolerant of severe data corruption and low image resolution.

2. Binary Image Data Acquisition

To describe and validate the method, two species of fish were video-observed and analyzed: Relatively large European bass (Dicentrarchus labrax); and tiger barb (Puntius tetrazona), a popular aquarium fish.

2.1. Design of Experiments on Fish

Two experiments were conducted with European bass. In the first experiment, only one individual was recorded in a relatively small (d = 3 m) circular tank by an IR camera with 1280 × 1024 resolution. The scene was illuminated by an 830 nm, 60 mW IR diode placed above the tank. The required set of individual object fingerprints was collected during this experiment. In the second experiment, 20 individuals of European bass swam together in the same tank. We note that the usage of an IR camera is not essential to the method, as the same results may be obtained with an ordinary video camera. The contours of the detected individuals, as well as of the overlaps, are very noisy due to the dependency of the fish texture coloration in the IR at the depth in which the fish are swimming. The average length and width of the fish projected on the camera was 190 px and 110 px (3 μ m 2 /px), respectively, in 12-bit depth.
For tiger barb, only one experiment was conducted, with 6 fish individuals swimming in a small (375 × 210 mm 2 ) tank. All sides of the aquarium were observed simultaneously using a mirror system. Data collection was done separately for each view of the aquarium. The typical projected size of the fish in the bottom and top views of the tank was 50 × 30 px (5 μ m 2 /px), in 12-bpc color depth.
All experimental manipulations were conducted according to the principles of the Institutional Animal Care and Use Committee (IACUC) of the University of South Bohemia in České Budějovice, Faculty of Fisheries and Protection of Waters, Vodňany, Czech Republic, based on the EU harmonized animal welfare act of the Czech Republic. The described research was reviewed and approved by the Ministry of Education, Youth and Sports Committee (MSMT–8792/2017–2). The original color and grayscale datasets are available upon request from the authors.

2.2. Acquisition of Binary Masks of Fish in Overlap

For both datasets, foreground detection was performed using the Gaussian Mixture Model [15,16] with 8 Gaussians. Two consecutive image dilations [17] with a 1 px diamond structural element fixed the detection artifacts.
After that, automatic classification was used for distinguishing binary masks of individual fish from binary masks of overlapping fish. In the case of European bass, it was assumed that the analyzed object was a single fish if the area of the produced mask obeyed:
A E ( A s ) ± 3 2 s t d ( A b ) ,
where A is the area of mask, A b are the areas of all detected blobs, s t d is the standard deviation, and E is the expected value. If the mask area fell outside this range, the object was classed as overlapping individuals.
For tiger barb we used a more advanced classification method. We assume that a bivariate distribution of area and perimeters of the binary masks is a mixture of two Gaussian mixtures. The first component (peak) of the mixture corresponds to the images of fish individuals, and has a lower mean value and standard deviation, whereas the second component can be related to the overlapping. After fitting, a posterior probability for each of the images was calculated. We treated the binary image as a fish individual or a fish overlap if the corresponding probability was greater than 0.5.
The binary masks with the overlapped fish were used as the input data for the proposed algorithm (Figure 1), described in detail in Section 3.

3. Segmentation of Binary Images of Symmetric Objects in Overlap

3.1. Morphological Fingerprinting

Most fish s p e c i e s (and other vertebrates) are bilaterally symmetric [18], i.e., only the transverse plane divides them into two asymmetrical parts. This allows one to describe the morphology of a fish according to the distances of the boundaries of its body from its central line. To proceed with this approach, all binary masks must first adopt the same orientation. There are two degrees of freedom in the orientation of the fish image—rotation around the center and reflection. The angle of the fish binary mask is defined as the angle between the ellipsoidal fit of its major axis and the o x axis. To ensure the same orientation of all images, the algorithm rotates the image to the zero angle. This results in horizontal orientation of the fish and implies two cases where the fish head is located in either the right or left half of the image. Flipping the images in order to locate the center of mass of the binary mask in the left part of the image completes the standardization of the images, so that the masks all have the same orientation.
After this ‘normalization’, the positions of central lines are determined from the distance transform. This technique converts the binary silhouette of the normalized fish image to a matrix of the minimal distances from corresponding pixels to the contour of the mask. To do this, an algorithm with time linear complexity [19] was used. The central line was defined for the whole fish silhouette in such a way that the X coordinates correspond to the orders of the columns in the distance matrix, and the Y coordinates are defined as the indices of the rows with the minima in the relevant columns. The obtained line was fitted by a low-rank (2–4) polynomial and corresponds to the intuitive estimation of the fish line of symmetry. Its length reflects the true fish length, independent of its pose. To determine the shape, the polynomial fitting of the central line was divided into N e equidistant points (50 points in our datasets). In order to reduce the computational complexity, piecewise approximation in these equidistant points substituted solving of the integral equations. The perpendicular distances from these points to the contour define the shape of the fish (Figure 2a,f). As with the length of the central line, these distances are pose-invariant and reflect the morphological properties of the fish. The average fingerprint can be easily calculated from the relevant mean values. Moreover, this approach to fingerprinting is reversible, and the silhouette of the fish can be constructed for any, even previously unseen, poses (Figure 2e).
The set of pose-invariant fingerprints is an alternative to the rigid pattern used in other models, and reflects the strong dependence of fish proportions on the length of the fish central line. From the biological point of view, this length corresponds to the fish size [20].

3.2. Dynamic Pattern Extraction

The reconstruction of the fish shape using a line of symmetry requires preliminary image treatments. The foreground detection used to obtain the binary mask for fingerprinting can give plausible but not reliable results. These irrelevant fingerprints (1% of the lowest and highest lines) were eliminated from the dataset. The resultant dependency of the shape on the length was smoothed using a robust k-dimensional spline [21]. The obtained data was stored in the form of an F × P matrix, where F is the number of remaining lengths and P is the count of the points per central line. Together with this matrix, the values of lengths are stored in a column vector.
For a given length of the central line, this form of data storage allows fast interpolation of the fish shape. In order to eliminate undesirable oscillations, which may produce negative distances, a linear interpolation was used.

3.3. Position Guess

Searching for fish individuals in overlap is a challenging task, and implies a complex nonlinear optimization which can be avoided if a set of possible solutions is known (Figure 1a,b). To obtain these solutions, we assume that the fish central lines coincide with the fish image skeletons. The positions and strength of edges are sensitive to the smoothness of the skeleton. Due to this fact, we applied a balanced skeletonization [22]. To ensure the same level of detail (LoD) for different overlaps, two kinds of skeleton special points—fork and end—were defined. The LoD denotes a number of these points. The subroutine of the equalization process reduced the skeletonization threshold by a fixed step until the LoD exceeded the desired LoD. The optimal count of skeleton points is dependent on the image size. We incorporate it into the heuristic formula T = 8 + 0.5 m i n ( W , H ) , where W is the image width and H is the image height.
Not all variants of the central lines are relevant for the computation, and so overly short or overly long central lines were removed. The range of acceptable lengths was defined as
L e = M d n ( l i n e s ) ± U · s t d ( l i n e s ) ,
where l i n e s are the lengths of the central lines in the training set and U = 1.5 is related to the dispersion of fish sizes in the experiment. The central line is assumed to be polynomial with coordinates of a high uniqueness (>0.8) along at least one dimension, i.e., 80% of points form the polynomial (nonconstant) function.
The variants of the central line that fulfilled the above-mentioned criterion underwent polynomial fitting of order 4. Whereas the nearly vertical lines could not be fitted well by a polynomial, their transposed representations could. To determine the correct orientation, the original and the transposed data were fitted by polynomial and evaluated for the minimum of the standard deviation in the fitting discrepancy.

3.4. Solution Search

The most accurate fits were found when we applied a simple greedy search to the known set of solutions (Figure 1c). However, first, two variables had to be calculated: The unknown count of objects and the cost function of the greedy search. The unknown count of objects was resolved by introducing a stop criterion. As can be understood intuitively, this criterion is related to the optimal coverage of overlaps by the reconstructed fingerprints. The introduction of this criterion requires the introduction of weights for overlapping pixels. The weights can vary from 1 (absence of a reconstructed object in the vicinity of the corresponding contour pixel) to 0 (ideal coincidence). The stop criterion was defined as
i s E n o u g h = 1 N n = 1 N [ W n = 1 ] < R ,
where N is the count of pixels in an overlapping contour, W n is the weight of the n-th pixel in the contour, and R is the robustness of the method. We used R = 0.22 , which means the algorithm will stop when at least 78 % of the contour points are covered. Higher or lower values of this parameter provide more false negative or positive results, respectively.
To evaluate the accuracy of the selected central lines, the shape of the object had to be reconstructed from each central line. The procedure is inverse to the fingerprint acquisition but starts similarly, by dividing the polynomial representation of the central line into N e equidistant (along the curve) points. The reconstructed shape is composed of points located within a certain perpendicular distance from the node points along the central line (Figure 2e). The distances are defined by the (interpolated) reference fingerprint, with the same length of the central line as the selected solution. The data for the interpolation was taken from the initial fingerprint set.
The cost function reflects two main measures of the optimization process: How accurately the reconstructed object fits to the contour and how accurately it covers the overlap. Let us denote the first class of cost functions as local costs; and the latter class as global costs. The global cost is defined as a median of distances between intact points of weights equal to 1 in an overlapping contour and the reconstructed fingerprint contour:
g l o b a l = M d n n = 1 N m = 1 M [ W n = 1 ] · ( | D m C n | ) ,
where N is the number of pixels in the overlapping contour, M is the doubled (below and above) number of equidistant points, W n is the weight for the n-th pixel of the contour, D m denotes the distance from the central line to the m-th point of the reconstructed object contour, and C n the distance to the n-th point of the overlapping contour. The main goal in introducing the global cost function is to eliminate solutions that fit fish-like objects that are, in fact, only a small part of the overlaps, rather than the full overlap itself.
The local cost is based on the distances from the central line to the overlapping contour. The measurement procedure bears similarities to the method of fingerprint reconstruction, and includes dividing the central line into N e equidistant points and measurement of the perpendicular distances to the overlapping contour. These distances were compared with the reference. In making this comparison, two issues must be addressed—ambiguity in fish orientation and excessively large distances (from central lines) at the places where the objects are in overlap. The uncertainty in the orientation was eliminated by the comparison of the local cost for the reference distances and ‘flipped’ reference distances (about the ‘y-axis’ of the fitted polynomial) when searching for the polynomial fit that minimized the local cost function. To address the problem of excessively large distances, distances are marked as too large if they are more than twice longer than the reference. The excessively large distances are then eliminated from the comparison by replacing them with the reference. The normalized cross-correlation with zero lag was used as the measure of similarity. To handle the broken distances, the correlation was decreased about the relative (to the number of equidistant points) number of broken distances. The local cost was then defined as
l o c a l = max s c o r r ( D , R s ) 1 M m = 1 M [ D m > 2 R m ] ,
where D is a set of distances measured from the central line to the overlapping contour, R m is the reference distance, M is the number of doubled distances, and s is the forward or backward orientation of the reference distance.
The global and local cost functions are required simultaneously and the resultant cost contains their product. The remaining issue in searching for solutions—the uniqueness of the solutions—can be resolved by introducing a degree of uniqueness. The degree of uniqueness is defined as the median of the weights of points of overlapping contour which are the closest to the solution:
u n i q = M d n ( W n ) , n = a r g min m | D m C n | ,
where a r g is the index of the value. The measure of the uniqueness is maximized. This intuitively corresponds to the idea that the solution is correct if it is localized near the points where there are no other solutions.
Combining the above, the total cost was defined as
c o s t = l o c a l · u n i q 1 + g l o b a l ,
where g l o b a l and l o c a l are global and local costs, respectively, as defined in Equations (4) and (5); and u n i q is a degree of solution uniqueness, as introduced in Equation (6).
Searching for solutions includes three main steps: Finding the solution of the minimal cost, renewing the weights, and checking the stopping criterion. To determine the weights, discrepancies between the solution and contour are calculated. The discrepancies are defined as the mean of the distances between the points of solution and the nearest 3 points of the overlapping contour. We denote this parameter as f u z z i n e s s . All weights where the minimal distance from the corresponding points to the solution contour was less than three f u z z i n e s s e s were divided by the term 3 × f u z z i n e s s . This feedback was aimed to eliminate coincident solutions.

4. Results and Discussion

The morphological fingerprinting technique demonstrates great robustness and stability for both testing datasets: High-resolution (in terms of pixels per object) but noisy images of European bass; and low-resolution but smooth tiger barb images. The mean shapes of the fish in the corresponding datasets are shown in Figure 3. The mean fingerprint of a single European bass is presented for 174 images (Figure 3c). As can be easily seen, for European bass, the standard deviation of the distances shows peaks in two high-variability regions, which are the fins and the tail. This splits the European bass image into three regions of stability along the central line. These regions may be used in one of the possible applications of the method—fingerprinting of individuals. The mean fingerprint (1182 images) of tiger barb (Figure 3a) is presented for six fish individuals. Here, the standard deviation is nearly uniform along the central line which may be interpreted as being due to large differences in the shapes of the fish individuals. In addition, the distances are significantly more noisy than for European bass. This can be ascribed to the usage of the lowest possible image resolution that was still suitable for fingerprinting of the fish.
The standard deviation of the curvature of the mean central line is significantly more interesting than the curvature itself. It shows points where the fish does or does not bend. The standard deviation of the curvature for both species (Figure 3b,d) has two extrema. The minimum corresponds to the position of the fish skull and implies the inability of the fish to bend at this point. The maximum corresponds to the fish inflection point, which is virtually impossible to locate, even by visual inspection of the images. As seen in the plots of standard deviations of the curvature (Figure 3b,d), the studied species’ bodies are of significantly different absolute sizes and proportions.
A further benefit of the technique is that it enables one to solve the overlapping. In order to evaluate the method accuracy, the results were compared with manually segmented overlaps (147 images of European bass and 187 images of tiger barbs). As the main measures of accuracy, we used the correct determination of the amount of overlapping objects and the deviation in the centroids of objects. The secondary measures of quality were the deviation in the object orientation and the similarity of the computer-aided contours to the manually segmented contours. As the measures of the contour similarity, the Dice-Sørensen coefficient [23], the Jaccard coefficient [24], and the boundary F1 score [25] were chosen. For both series, the best matching between the manual segmentation and automatic method was obtained for the classical Dice-Sørensen coefficient. The results of similarity in detection of fish individuals between the proposed method and the manual segmentation is summarized in Table 1.
The discussed method shows great potential and, despite the low quality of the foreground detection, found the correct amount of fish in almost 90 % of cases for tiger barb. The mean error in the determination of centroids was 5.9 % (in units of fish length), and the angular discrepancy was 4.2 . The Dice-Sørensen coefficient was 0.82 , which we interpret as a good result because the proposed method is not a segmentation method but rather a reconstruction method and deals with lost and irrecoverable information. For the European bass dataset, the correct amount of fish was determined in 83 % of cases, centroids error was 5.8 % , the Dice-Sørensen coefficient was 0.83 , and the angular discrepancy was 5.7 . The boundary F1 score is lower for European bass, which can be related to the different level of detail. In other words, the images of tiger barb are significantly smaller and their contours are more convex. Angle discrepancies demonstrate the same tendency. Moreover, in the case of extremely low-resolution images, the method works well for dorsal and ventral views of the fish (Figure 4). Therefore, let us note that downscaling the images before processing might be a beneficial strategy for potential process speed-up.
The mean time of calculation per overlap was approx. 5 s. The code is written as a prototype in MATLAB and is not fully optimized. A substantial amount of time is consumed by the self-overhead of functions, and thus the method can be significantly optimized.
The binary datasets (for fingerprinting the tiger barb, we present another, shorter, dataset than shown in the text), all codes, verification tools, and the GUI for the fingerprinting and collision solving are available in the supplementary materials [26].

5. Conclusions

The observation of animal movement using a digital camera is a general approach both in the contemporary ethology research as well as in many practical applications such as observation of fish breeding tanks, cages, or cattle pounds or barns. The presented paper deals with the solution of one of the most complex problems of automated interpretation of such an observation, the division of overlapping animal bodies in a digital image in video-tracking systems. As follows from the literature review, this topic has not been so far sufficiently solved.
We present an algorithm providing a robust solution applicable in a wide range of cases found in practice, and show its application in a study of fish school behavior: (1) In a simple observation of relatively large European bass in a recirculating tank by an IR camera (obtained 147 binary images of overlaps); as well as (2) in a very sophisticated ethological setup of an aquarium surrounded by four mirrors and recorded by a fast-speed color camera which provides maximal technically available information on the experiment on tiny tiger barbs (obtained 187 binary images of overlaps). Even in the latter, technically best defined, example, cases of overlapping which are needed to be solved remain. This demonstrate that such an algorithm is unavoidable.
The solving of overlaps itself is a nontrivial task, especially if objects are nonconvex and have no strict shape. Our method combines only known morphological properties and symmetries of objects with empirical features of image skeletons. Such an approach attains a significant efficiency (hardly attainable by any model-free approach) both in quality and speed. The segmentation outputs highly correlate with manual/visual object detection (correspondence in up to 90% cases, similarity indices up to 0.83), mainly in case of relatively small objects of interest as tiger barbs are. Of course, the quality of the proposed segmentation method in general depends on the quality of the binary masks. The method has been applied to the dorsal and ventral sides of the fish but is supposed to work particularly with semisymmetrical objects such as lateral sides of fish. The same approach with minor changes may be applied (with even greater efficacy than in the case of fish) to other elongated organisms that frequently overlap, e.g., worms and snakes. We believe that the developed method can greatly improve existing systems of tracking and will initiate the development of new ones which will facilitate the processing of previously intractable data.

Author Contributions

Conceptualization, K.L.; data curation, M.S. and K.L.; formal analysis, K.L. and R.R.; funding acquisition, D.Š.; investigation, K.L.; methodology, K.L.; resources, D.Š.; software, K.L.; supervision, D.Š. and R.R.; validation, R.R.; writing—original draft, K.L.; writing—review and editing, R.R., M.S., and D.Š.


This work was supported by the Ministry of Education, Youth and Sports of the Czech Republic—projects CENAKVA (LM2018099) and the CENAKVA Centre Development (No. CZ.1.05/2.1.00/19.0380)—and from the European Regional Development Fund in frame of the project Kompetenzzentrum MechanoBiologie (ATCZ133) in the Interreg V-A Austria–Czech Republic programme. The work was further financed by the TA CR Gama PoC 02-22-Štys sub-project and by the GAJU 013/2019/Z project.


The authors would like to thank Luke Shaw for editing and proofreading the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

List of Symbols

a r g index of the value in computation of the degree of uniqueness
Aarea of binary mask
A b area of all detected binary masks, including fish overlaps
c o s t total cost, i.e., total accuracy of the reconstruction method
C n distance from the central line to the n-th point of the overlapping contour
ddiameter of the fish circular tank
D set of distances from the central line to the reconstructed object contour
D m distance from the central line to the m-th point of the reconstructed object contour
E expected value
f u z i n e s s discrepancy between the solution and the contour
Fnumber of relevant lengths of the fingerprint central line
g l o b a l global cost, i.e., median of distances between intact points of W n = 1 in an overlapping contour and the reconstructed fingerprint contour
H image height
i s E n o u g h stop criterion in calculation of the unknown number of objects in the solution search
Kminimum from the set D of distances from the central line to the reconstructed object contour
L e range of acceptable lengths of the central lines in the training set
l i n e s set of lengths of the central lines in the training set
l o c a l local cost, i.e., comparison of the perpendicular distances D m with the overlapping contour with the reference length R m
morder of the doubled equidistant points
Mdoubled (below and above) number of equidistant points on the fingerprint central line
norder of the pixel in the contour
Ncount of pixels in an overlapping contour
N e number of equidistant points on the polynomial representation of the central line
Pcount of the points per fingerprint central line in dynamic pattern extraction
R m reference distance in calculation of the local cost
R s reference distance of the s-th, forward or backward, orientation
R robustness of the solution search method
sorientation, forward or backward, of the reference distance
Toptimal count of skeleton points, i.e., the level of detail (LoD)
u n i q degree of solution uniqueness of the reconstruction method
Udispersion of fish sizes in the experiment
W image width
W n weight of the n-th pixel in the contour


  1. Dennis, R.L.; Newberry, R.C.; Cheng, H.W.; Estevez, I. Appearance matters: Artificial marking alters aggression and stress. Poultry Sci. 2008, 87, 1939–1946. [Google Scholar] [CrossRef] [PubMed]
  2. Delcourt, J.; Denoel, M.; Ylieff, M.; Poncin, P. Video multitracking of fish behaviour: A synthesis and future perspectives. Fish. Fish. 2012, 14, 186–204. [Google Scholar] [CrossRef]
  3. Li, T.; Chang, H.; Wang, M.; Ni, B.; Hong, R.; Yan, S. Crowded scene analysis: A survey. IEEE Trans. Circuits Syst. Video Technol. 2015, 25, 367–386. [Google Scholar] [CrossRef]
  4. Kok, V.J.; Lim, M.K.; Chan, C.S. Crowd behavior analysis: A review where physics meets biology. Neurocomputing 2016, 177, 342–362. [Google Scholar] [CrossRef]
  5. Jolles, J.W.; Boogert, N.J.; Sridhar, V.H.; Couzin, I.D.; Manica, A. Consistent individual differences drive collective behavior and group functioning of schooling fish. Curr. Biol. 2017, 27, 2862–2868. [Google Scholar] [CrossRef] [PubMed]
  6. Delcourt, J.; Becco, C.; Vandewalle, N.; Poncin, P. A video multitracking system for quantification of individual behavior in a large fish shoal: Advantages and limits. Beh. Res. Methods 2009, 41, 228–235. [Google Scholar] [CrossRef] [PubMed]
  7. Morais, E.; Campos, M.; Padua, F.; Carceroni, R. Particle filter-based predictive tracking for robust fish counting. In Proceedings of the IEEE XVIII Brazilian Symposium on Computer Graphics and Image Processing, Natal, Brazil, 9–12 October 2005. [Google Scholar] [CrossRef]
  8. Raj, A.; Sivaraman, A.; Bhowmick, C.; Verma, N.K. Object tracking with movement prediction algorithms. In Proceedings of the 2016 IEEE 11th International Conference on Industrial and Information Systems (ICIIS), Roorkee, India, 3–4 December 2016. [Google Scholar] [CrossRef]
  9. Tang, W.Q.; Jiang, Y.L. Target tracking of the robot fish based on adaptive fading Kalman filtering. In Proceedings of the 2013 IEEE International Conference on Mechatronic Sciences, Electric Engineering and Computer (MEC), Shengyang, China, 20–22 December 2013. [Google Scholar] [CrossRef]
  10. Perez-Escudero, A.; Vicente-Page, J.; Hinz, R.C.; Arganda, S.; de Polavieja, G.G. idTracker: Tracking individuals in a group by automatic identification of unmarked animals. Nat. Methods 2014, 11, 743–748. [Google Scholar] [CrossRef] [PubMed]
  11. Terayama, K.; Habe, H.; aki Sakagami, M. Multiple fish tracking with an NACA airfoil model for collective behavior analysis. IPSJ Trans. Comp. Vis. Appl. 2016, 8. [Google Scholar] [CrossRef]
  12. Butail, S.; Paley, D.A. Three-dimensional reconstruction of the fast-start swimming kinematics of densely schooling fish. J. R. Soc. Interface 2011, 9, 77–88. [Google Scholar] [CrossRef] [PubMed]
  13. Saverino, C.; Gerlai, R. The social zebrafish: Behavioral responses to conspecific, heterospecific, and computer animated fish. Behav. Brain Res. 2008, 191, 77–87. [Google Scholar] [CrossRef] [PubMed]
  14. Calfee, R.D.; Puglis, H.J.; Little, E.E.; Brumbaugh, W.G.; Mebane, C.A. Quantifying fish swimming behavior in response to acute exposure of aqueous copper using computer assisted video and digital image analysis. J. Vis. Exp. 2016, 53477. [Google Scholar] [CrossRef] [PubMed]
  15. Stauffer, C.; Grimson, W. Adaptive background mixture models for real-time tracking. In Proceedings of the 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149), Fort Collins, CO, USA, 23–25 June 1999. [Google Scholar] [CrossRef]
  16. KaewTraKulPong, P.; Bowden, R. An improved adaptive background mixture model for real-time tracking with shadow detection. In Video-Based Surveillance Systems; Springer: New York, NY, USA, 2002; pp. 135–144. [Google Scholar] [CrossRef]
  17. van den Boomgaard, R.; van Balen, R. Methods for fast morphological image transforms using bitmapped binary images. CVGIP Graph. Models Image Process. 1992, 54, 252–258. [Google Scholar] [CrossRef]
  18. Holló, G. A new paradigm for animal symmetry. Interface Focus 2015, 5, 20150032. [Google Scholar] [CrossRef] [PubMed]
  19. Maurer, C.; Qi, R.; Raghavan, V. A linear time algorithm for computing exact Euclidean distance transforms of binary images in arbitrary dimensions. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 265–270. [Google Scholar] [CrossRef]
  20. Mojekwu, T.O.; Anumudu, C.I. Advanced techniques for morphometric analysis in fish. J. Aquac. Res. Dev. 2015, 06, 354. [Google Scholar] [CrossRef]
  21. Garcia, D. Robust smoothing of gridded data in one and higher dimensions with missing values. Comput. Stat. Data Anal. 2010, 54, 1167–1178. [Google Scholar] [CrossRef] [PubMed]
  22. van Eede, M.; Macrini, D.; Telea, A.; Sminchisescu, C.; Dickinson, S. Canonical skeletons for shape matching. In Proceedings of the 18th IEEE International Conference on Pattern Recognition, Hong Kong, China, 20–24 August 2006. [Google Scholar] [CrossRef]
  23. Dice, L.R. Measures of the amount of ecologic association between species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  24. Jaccard, P. The distribution of the flora in the alpine zone. New Phytol. 1912, 11, 37–50. [Google Scholar] [CrossRef]
  25. Csurka, G.; Larlus, D.; Perronnin, F. What is a good evaluation measure for semantic segmentation? In Proceedings of the British Machine Vision Conference, Bristol, UK, 9–13 September 2013; BMVA Press: Durham, UK, 2013; pp. 32.1–32.11. [Google Scholar] [CrossRef]
  26. Lonhus, K.; Štys, D.; Saberioon, M.; Rychtáriková, R. Segmentation of laterally symmetric overlapping objects: Application to images of collective animal behavior. arXiv 2018, arXiv:1811.01373. [Google Scholar]
Figure 1. Algorithm for segmentation of fish overlaps. (a) Phase of the position guess which includes choosing of the desired level of details, construction of all possible central lines, and filtering of unsuitable central lines out. (b) Fingerprint reconstruction followed by measurement of the fitting accuracy. (c) Greedy search for finding the best solution among the precalculated ones.
Figure 1. Algorithm for segmentation of fish overlaps. (a) Phase of the position guess which includes choosing of the desired level of details, construction of all possible central lines, and filtering of unsuitable central lines out. (b) Fingerprint reconstruction followed by measurement of the fitting accuracy. (c) Greedy search for finding the best solution among the precalculated ones.
Symmetry 11 00866 g001
Figure 2. The procedure for the segmentation of fish (European bass) overlaps. (a) The process of obtaining a fish fingerprint from contours. (b) The original image of fish overlapping. (c) The skeletonization with desired level of detail of a fish overlap, segmented from image (b). (d) The fish fingerprint. (e) Reconstruction of the fish shape from an arbitrary pose. (f) The final solution.
Figure 2. The procedure for the segmentation of fish (European bass) overlaps. (a) The process of obtaining a fish fingerprint from contours. (b) The original image of fish overlapping. (c) The skeletonization with desired level of detail of a fish overlap, segmented from image (b). (d) The fish fingerprint. (e) Reconstruction of the fish shape from an arbitrary pose. (f) The final solution.
Symmetry 11 00866 g002
Figure 3. Mean fingerprints of (a) 6 tiger barbs (for 1182 images) and of (c) a European bass individual (for 174 images). The red lines highlight regions of low variability. (b,d) The standard deviations of the central line curvatures for the relevant species.
Figure 3. Mean fingerprints of (a) 6 tiger barbs (for 1182 images) and of (c) a European bass individual (for 174 images). The red lines highlight regions of low variability. (b,d) The standard deviations of the central line curvatures for the relevant species.
Symmetry 11 00866 g003
Figure 4. Solving of overlapping of extremely low-resolution objects. (a) The bottom view of the tank where the fish are strictly symmetrical. (b,c) The side views of the tank, where the objects are not symmetrical, but the algorithm can still provide reasonable results.
Figure 4. Solving of overlapping of extremely low-resolution objects. (a) The bottom view of the tank where the fish are strictly symmetrical. (b,c) The side views of the tank, where the objects are not symmetrical, but the algorithm can still provide reasonable results.
Symmetry 11 00866 g004
Table 1. Similarity in fish detection between the proposed reconstruction method and the manual segmentation of the fish overlaps.
Table 1. Similarity in fish detection between the proposed reconstruction method and the manual segmentation of the fish overlaps.
SeriesImgs.Mean BFJACDiceCount [%]Centroid e. [%]Orient. [ ]
T. barb187 0.91 ± 0.10 0.71 ± 0.09 0.82 ± 0.07 89.30 5.97 ± 3.78 4.16 ± 7.42
E. bass147 0.45 ± 0.18 0.72 ± 0.13 0.83 ± 0.10 84.35 5.52 ± 4.30 5.69 ± 11.94
Imgs.—number of images with fish overlaps in the series; Mean BF—boundary F1 score (contour matching score); JAC—Jaccard similarity coefficient; Dice—Dice-Sørensen similarity coefficient; Count—amount of images correctly determined using the automatic method (in % of the whole image series); Centroid e.—mean error in the determination of centroids (in % of fish length); Orient.—angular discrepancy (in angular ).

© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (
Symmetry EISSN 2073-8994 Published by MDPI AG, Basel, Switzerland RSS E-Mail Table of Contents Alert
Back to Top