Next Article in Journal
A Simulator and First Reinforcement Learning Results for Underwater Mapping
Next Article in Special Issue
Image Semantic Segmentation of Underwater Garbage with Modified U-Net Architecture Model
Previous Article in Journal
Optimization of Printed Polyaniline Composites for Gas Sensing Applications
Previous Article in Special Issue
A Novel Detection Refinement Technique for Accurate Identification of Nephrops norvegicus Burrows in Underwater Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Impact of Data Augmentations on Deep Learning-Based Marine Object Classification in Benthic Image Transects

Biodata Mining Group, Bielefeld University, P.O. Box 100131, 33501 Bielefeld, Germany
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(14), 5383; https://doi.org/10.3390/s22145383
Submission received: 16 May 2022 / Revised: 15 July 2022 / Accepted: 18 July 2022 / Published: 19 July 2022
(This article belongs to the Special Issue Advanced Sensor Applications in Marine Objects Recognition)

Abstract

:
Data augmentation is an established technique in computer vision to foster the generalization of training and to deal with low data volume. Most data augmentation and computer vision research are focused on everyday images such as traffic data. The application of computer vision techniques in domains like marine sciences has shown to be not that straightforward in the past due to special characteristics, such as very low data volume and class imbalance, because of costly manual annotation by human domain experts, and general low species abundances. However, the data volume acquired today with moving platforms to collect large image collections from remote marine habitats, like the deep benthos, for marine biodiversity assessment and monitoring makes the use of computer vision automatic detection and classification inevitable. In this work, we investigate the effect of data augmentation in the context of taxonomic classification in underwater, i.e., benthic images. First, we show that established data augmentation methods (i.e., geometric and photometric transformations) perform differently in marine image collections compared to established image collections like the Cityscapes dataset, showing everyday traffic images. Some of the methods even decrease the learning performance when applied to marine image collections. Second, we propose new data augmentation combination policies motivated by our observations and compare their effect to those proposed by the AutoAugment algorithm and can show that the proposed augmentation policy outperforms the AutoAugment results for marine image collections. We conclude that in the case of small marine image datasets, background knowledge, and heuristics should sometimes be applied to design an effective data augmentation method.

1. Introduction

Underwater imaging with autonomous or remotely operated vehicles such as AUV (autonomous underwater vehicles [1]) or ROV (remotely operated vehicles [2]) allows visual assessments of large remote marine habitats through large image collections with 10 2 10 4 images collected in one dive. One of the main purposes of these image collections is to provide valuable information about the biodiversity in marine life. To detect and classify species or morphotypes in these images, machine learning, has been proposed with some early promising results in the last decade obtained with pre-deep learning methods such as support vector machines (SVM) [3,4,5] or, more recently, using convolutional neural networks (CNN) [6,7,8,9,10,11,12,13,14,15,16,17,18,19]. The application of CNN for taxonomic classification problems in this marine image domain features some characteristics that separate this field of research from the majority of CNN applications in the context of human civilization (like traffic image classification/identification, quality control, observation, manufacturing). First, the process of image collection is very expensive as it involves costs for ship cruises, ROV/AUV hardware, operator personnel, advanced planning, maneuvering, and special camera equipment. Second, the number of domain experts that are required for collecting taxonomic labels to build training and test data is limited. In addition, the task of taxonomic classification in the images is very difficult as the domain experts usually have only one single image of an organism, in contrast to the traditional approach of collecting samples from the seafloor and intense visual inspection of the specimen in the laboratory. Another problem is that due to the natural structure of food webs and communities, some species are rather rare by nature and appear in just a few images and some species can be observed in hundreds of images, which leads to a strong class imbalance. Similar problems can be observed in aerial images collected for environmental monitoring. As a consequence, the field of marine image classification may sometimes require individual approaches to pattern recognition problems shaped by the characteristics listed above.
Data augmentation (DA) is a standard approach to overcome training problems caused by limitations in training data and over-fitting. In the case of image classification, the image augmentation algorithms can be broadly classified as deep learning approaches, for example, adversarial training [20], neural style transfer [21], feature space augmentation [22,23], generative adversarial networks (GAN) [24,25], meta-learning (AutoAugment [26], Smart Augmentation [27]), and basic image manipulation augmentations, for instance, geometric transformations (horizontal flipping, vertical flipping, random rotation, shearing), color space transformations (contrast modulation, brightness adjustment, hue variety), random erasing, mixing images [28,29], and kernel filters [30]. Deep learning approaches and basic image manipulation augmentations do not form a mutually exclusive dichotomy. In this work, we are mainly curious about the effectiveness of the most broadly used and readily available basic image manipulation operations in marine images. However, there is little literature available on methodological approaches to (i) select one or more data augmentation method(s) for a given image domain, or (ii) employ (or improve) the effect of DA in the marine image domain in particular. While single individual successful applications of DA have been reported already [14,17,19,31] also some examples have been reported on unsuccessful DA applications leading to decreasing performance [31,32,33].
Recently, a small number of concepts have been proposed for combinations of DA methods. Shorten and Khoshgoftaar [33] describes that it is important to consider the ‘safety’ of augmentation, and this is somewhat domain-dependent, providing a challenge for developing generalizable augmentation policies. Cubuk et al. have met the challenge of developing generalizable augmentation policies in their work, proposing the AutoAugment algorithm [26] to search for augmentation policies from a dataset automatically. Based on this work, Fast AutoAugment [34] optimized the search strategy, which speeds up the search time. One study was conducted by Shijie et al. [35], which compared the performance evaluation of some DA methods and their combinations on the CIFAR-10 and ImageNet datasets. They found four individual methods generally perform better than others, and some appropriate combinations of methods are slightly more effective than the individuals. However, although all these approaches provide highly valuable new methods, a domain-specific investigation of DA effectiveness is missing, especially for special domains like medical imaging, aerial images, digital microscopy, or -like in this case- benthic marine images.
As already explained, the marine imaging domain challenges deep learning applications with a permanent lack of annotated data. On the other hand, marine benthic images often are recorded with a downward-looking camera (sometimes referred to as “ortho-images”) that feature a higher degree of regularity (for instance, regarding illumination or camera—object distance). This can also increase the potential of machine learning-based classifications in benthic images when compared to everyday benchmark image data in computer vision documenting, for instance, different aspects of human activities (e.g., traffic scenes, social activities). Such images somehow constitute the mainstream in CNN application domains and show lesser irregularities in this regard due to changing weather, light condition, camera distance to the object, etc. Thus, we hypothesize that marine images may show special characteristics regarding the effectiveness of particular DA methods. These characteristics need to be thoroughly investigated as DA can be considered one of the most promising tools to overcome problems in missing or unbalanced training data in marine imaging. To show the effect of different DA methods in the context of deep learning classification in marine images, we first report results from exhaustive comparative experiments using single DA methods. Based on our findings, we propose different combinations of augmentation methods, referred to as augmentation policies, and can show a significant improvement for our marine-domain datasets.

2. Materials and Methods

2.1. Data Sets

To show the effect of different DA approaches, we conduct experiments with two marine-domain datasets. The Porcupine Abyssal Plane (PAP) which is in the northwest Atlantic Ocean, to the southwest of the United Kingdom [36,37,38] and the Clarion Clipperton Zone (CCZ) which is located in the Pacific Ocean and is known for its rich deposits of manganese nodules [39]. They are collected with AUVs in several 1000 m depths and show deep-sea benthos with different species. In addition to these two marine-domain datasets, we conduct a series of data augmentation experiments on the Cityscapes dataset [40], referred to as CSD, collected from annotated traffic videos in urban street scenes.

2.1.1. PAP

In this work, we choose the following four categories (see Figure 1) to form Γ PAP for our experiments to ensure that a more trustworthy number of test samples are left.
Figure 2 reveals the structure of the Γ PAP dataset. For each class we randomly sample 300 images from Γ PAP as Γ PAP train , 200 images as Γ PAP validation , and the rest as Γ PAP test . Γ PAP = Γ PAP train Γ PAP validation Γ PAP test . Additionally, as shown in Figure 3 we sample 50, 100, and 200 images randomly from Γ PAP train as Γ PAP train 50 , Γ PAP train 100 , Γ PAP train 200 , respectively, to investigate the effect of train set size on DA performance. Sample sizes of each class are shown in Table 1.

2.1.2. CCZ

Similarly, we choose the two most abundant categories (shown in Figure 4) to form the Γ CCZ dataset for the experiments in this work. Sample sizes of each class in the Γ CCZ dataset are shown in Table 2.

2.1.3. CSD

We choose two classes (see Figure 5) from the vehicle group of the CSD to generate the dataset Γ CSD (shown in Table 3) for our experiments in this work.

2.2. Model and Evaluation Criteria

In this work, we use a MobileNet-v2 which is pre-trained on the ImageNet [41] to investigate several augmentation policies. Image data are resized to 224 px × 224 px and normalized based on the ImageNet dataset. We use Adam as the optimizer in our experiments and set a learning rate of 1 × 10 4 accompanied by a step decay with a step size of 1 and a gamma of 0.1. The loss function used in this work is cross-entropy loss. For each epoch, we perform a train and a validation phase, and compute the prediction accuracy A c c e , j = n e , j N e , j , where n e , j and N e , j stand for the number of correct predictions and the total number of samples in phase j { train , validation } of epoch e, respectively. In the test phase, since the sample sizes of each class in the test set are not consistent, we compute the prediction accuracy A c c k = n k N k for each class, where n k and N k stand for the number of correct predictions for class k and the sample sizes of class k in the test set, respectively. In each experiment we record two trained-models corresponding to the highest A c c e , train and the highest A c c e , validation in all epochs and apply them to the test set separately. The two inference results are averaged to obtain the average accuracy A A k = 1 2 ( A c c k A c c e , train highest + A c c k A c c e , validation highest ) for each class. As the last step, we compute the mean average accuracy m A A = 1 K k A A k with K the number of classes as the prediction accuracy in the test phase. The flowchart of our work is shown in Figure 6.

2.3. Methods

In this work, we investigate the data augmentations implemented in torchvision of PyTorch during the training progress. We describe the data augmentation in the way that a training image X i is given defined by X i = f ( x i ) with x i as the original image with index i and f ( ) as the transformation function executing the augmentation method. In this work we apply RandomRotation f RR ( x i , d ) to rotate the image randomly within the angle range represented by d, RandomVerticalFlip f RVF ( x i , p ) to vertically flip the given image randomly with a given probability p, RandomHorizontalFlip f RHF ( x i , p ) to horizontally flip the given image randomly with a given probability p, RandomAffine f RA ( x i , t , s ) to randomly affine transformation translate and shear of the image keeping center invariant according to the parameters t and s. Color transformations f CT ( x i , b , c , s , h ) is used to randomly change the brightness, contrast, saturation, and hue of an image according to the values of parameters b, c, s, h, respectively.
We investigate the performance of DA methods on Γ PAP and Γ CCZ , determining the four best-performing ones. We propose six DA combination policies and apply them to Γ PAP , Γ CCZ and Γ CSD . To avoid randomness affecting the results during the experiments, we set a seed for fixing the following random number generators: CUDA, NumPy, Python, PyTorch, and cudnn.

3. Results

The experimental results are represented by the change of m A A of applying DA and without applying DA. We generate a heatmap for each experiment based on the change of m A A , with blue indicating positive increments and orange indicating decrease. The darker the color, the greater the change.

3.1. Experiment A: Performance of Data Augmentations on Γ PAP

We apply a series of DA methods and parameters to Γ PAP train 50 , Γ PAP train 100 , and Γ PAP train 200 to observe the performance of different DA and different parameters. The seed is fixed to 350 in all experiments to exclude the interference of the seed with the experiment. The results of Γ PAP train 50 and Γ PAP train 200 are shown in Table 4 and Table 5, and the results of Γ PAP train 100 are shown in Appendix A.
Table 4 reveals the performance of different DA methods and parameters when applied on Γ PAP train 50 with setting seed to 350. From this heatmap, we can see that almost all of the DA methods and parameters used in our experiments perform well. The best-performing DA methods and parameters on Γ PAP train 50 are RandomRotation with a parameter of 100°, Contrast with a parameter of 5, Brightness with a parameter of 2, and Shear with a parameter of (−30°, 30°, −30°, 30°). By applying these four, a significant improvement can be achieved on Γ PAP train 50 .
When increasing the number of training images to 200 per class, we can see from Table 5 that the effect of DA on the improvement of average accuracy further diminishes. The best-performing four DA methods are still the same and the best-performing parameters of RandomRotation, Brightness, Contrast, and Shear are 170°, 1.9, 4.5, (−40°, 40°, −40°, 40°), respectively.
In Experiment A, we use Γ PAP validation and Γ PAP test as validation set and test set. The four best-performing DA methods are RandomRotation, Brightness, Contrast, and Shear, regardless of whether the training set is Γ PAP train 50 , Γ PAP train 100 , or Γ PAP train 200 . As the size of training samples increases, the best-performing parameters of the DA vary. The increment of m A A shows an almost proportional trend to the magnitude of the parameters of RandomRotation, Brightness, Contrast, and Saturation. The performance of Shear reveals that the parameters that introduce more deformation can yield a better augmentation effect. However, it can be shearing in one direction at a bigger angle, or shearing in two directions at one angle.
We also conduct experiments on Γ PAP train 50 when setting the seed to 3500, finding that the heatmap shows a similar trend to that with the seed set to 350. The results are shown in Appendix B.

3.2. Experiment B: Performance of Data Augmentations on Γ CCZ

In Experiment B, we apply the same DA methods and parameters to Γ CCZ to verify whether the observations we obtained in Experiment A can be seen on another marine-domain dataset as well. The experimental results with setting seed to 350 are shown in Table 6 and Table 7, and the results of Γ CCZ train 50 with setting seed to 3500 are shown in Appendix B.
The heatmap shown in Table 6 presents the performance of different DA methods and parameters when applied on Γ CCZ train 50 with setting seed to 350. We can find that the best-performing DA methods on Γ CCZ train 50 are the same as the observations from Experiment A, which are RandomRotation, Brightness, Contrast, and Shear. On Γ CCZ train 50 , RandomRotation shows the best results with parameter 150°. Brightness, Contrast, and Shear work best with parameters 1.9, 4.5, and (−40°, 40°, −40°, 40°), respectively.
Table 7 shows the performance of different DA methods and parameters when applied on Γ CCZ train 100 with setting seed to 350. We increase the number of training data from 50 even to 100 per class. The effect of DA methods, except for Saturation, becomes weaker or has a negative effect as the number of training samples is increased, which is the same as shown in Table 5. The four most effective DA methods are still RandomRotation, Brightness, Contrast, and Shear, whose best-performing parameters are 170°, 2, 5, and 40°, respectively.
Experiment B investigates the effect of different DA methods and different parameters on Γ CCZ with different training set sizes and shows similar results to the observations on Γ PAP . We find that RandomRotation, Contrast, Brightness, and Shear are always the four best-performing DA methods on both Γ PAP and Γ CCZ marine-domain dataset. The effect of RandomRotation, Brightness, and Contrast becomes more significant as the parameters increase. Similarly, as the amount of training data is increased, almost all DA methods’ effects are diminished.

3.3. Experiment C: Performance of Data Augmentations on Γ CSD

We conduct research on Γ CSD that is different from the marine domain to demonstrate that the effect of DA is domain-dependent. The seed is set to 350 in Experiment C, and the experimental results of Γ CSD train 50 and Γ CSD train 150 are shown in Table 8 and Table 9. The results of Γ CSD train 100 are shown in Appendix A.
Table 8 illustrates the effect of different DA methods and different parameters on the average accuracy improvement of Γ CSD train 50 . From this heatmap, we can see that RandomRotation no longer works as well as it did on the marine-domain datasets, and only when the parameters are small does it improve the average accuracy a little. Similarly, Shear with parameters of big degrees decreases the m A A . In addition, RandomVerticalFlip is also no longer suitable for this dataset.
When training samples are supplemented to 150 per class, we can see from Table 9 that RandomRotation with a parameter of big degrees and RandomVerticalFlip still have a negative impact. The effect of Brightness, Contrast, and Saturation on Γ CSD train 150 performs well with the value of parameters increasing, which is similar to that on Γ PAP train 50 and Γ CCZ train 50 .
Experiment C investigated the performance of different DA methods and parameters on Γ CSD with different training set sizes. Overall, Saturation shows an effect almost proportional to the value of parameters on all three datasets. The performance of Contrast and Brightness improve with increasing training data size. RandomRotation can slightly increase m A A with parameters smaller than 10°, but have increasingly negative effects on m A A as the parameters become larger. RandomHorizontalFlip and Hue can slightly improve m A A , while RandomVerticalFlip and Translate almost always reduce m A A . The effect of Shear is no longer as shown in Experiment A and Experiment B, showing a negative effect at large deformation angles.
To compare the different DA methods’ impact on the three different datasets more intuitively, the percentage change of classification m A A is plotted for different parameters, which are shown in Figure 7 and Figure 8 and Appendix C. In each plot, the x-axis represents parameters, the y-axis represents the change of m A A , and circles, stars, triangle stand for Γ PAP , Γ CCZ , and Γ CSD , respectively.

3.4. Experiment D: Data Augmentation Policies

According to the above experimental results, the four best-performing DA methods on both Γ PAP and Γ CCZ datasets are RandomRotation, Contrast, Brightness, and Shear. In most cases, the best performances are shown when the parameter of RandomRotation is around 150°, the parameters of Contrast and Brightness are around 2 and 5, respectively, and the parameters of Shear are those that produce larger deformations (we experiment with the parameters 40). To verify the effect of the combination of these DA methods, we proposed DA combination policies in Table 10, where the function of RBC_1 indicates that RandomRotation 150°, Brightness 1.9, and Contrast 5 are applied to the training images successively.
The performance of our policies on Γ PAP , Γ CCZ , and Γ CSD are shown in Table 11, where AA_IP and AA_CP represent ImageNet policy and CIFAR-10 policy proposed by AutoAugment [26], respectively. It shows that all our policies trained on Γ PAP train 50 , Γ PAP train 100 , Γ PAP train 200 , Γ PAP train , and Γ CCZ train 50 can outperform AA_IP and AA_CP policies which are proposed by AutoAugment [26]. RBC_3, RBC_5, and RBCS trained on Γ CCZ train 100 and Γ CCZ train can also outperform AA_IP and AA_CP. In contrast to the effect of our policies on Γ PAP and Γ CCZ , these policies have a negative effect on the Γ CSD dataset. Here the AutoAugment policies outperform our policies.

4. Discussion

In this paper, we could show a clear domain dependence for the application of augmentation. While experiments A and B applying augmentation to marine data show similar results, experiment C applied to the more established everyday traffic data shows different trends. The same observation applies to experiment D when comparing auto augmentation policies fit on everyday data (ImageNet and CIFAR-10) on the one hand with our manual combination policies on the other hand. Here the AutoAugment policies work best on the traffic data while leading to results inferior to the policies proposed by our experiments (RBC _ * ). From Experiment A (marine), we can find that the effect of DA diminishes as the number of samples in the training set increases. This is because the additional training images allow the model to learn more features, which also indicates that DA is an effective way to address the lack of training data. The results of Experiment A and Experiment B (marine) indicate that for the marine domain, with increasing training sample size and different parameter choices, some DA methods show the possibility of decreasing m A A (e.g., flip, translate), but RandomRotation, Brightness, Contrast, and Shear always show good results. This may be due to the natural variation regarding the orientation and position of the underwater target objects relative to the camera. Besides, the light during underwater data acquisition has a significant effect on the image data. Overall, experiments A and B show similar trends. However, due to the insufficient sample size of Γ CCZ test , the results of Experiment B are not as reliable as those of Experiment A.
In Experiment C (traffic), the effect of RandomRotation, RandomVerticalFlip, and Shear have a significantly different effect than in Experiment A and Experiment B, by often even decreasing the performance. This is likely caused by the fact that a VerticalFlip or a Rotation at a large angle is unrealistic under the urban traffic domain. We can find that on the Γ CSD dataset, the color transformations perform better than the geometric transformations.
From Table 11, we can see that the performance of our policies can reach or outperform that of transferring the policies proposed by AutoAugment [26] to marine data. However, our policies have a negative effect on the traffic dataset. As can be seen in Table 4 and Table 5, and the experimental result of Γ PAP train 100 in Appendix A, the maximum increments of m A A are 8.02%, 5.61%, and 3.29% on Γ PAP train 50 , Γ PAP train 100 , and Γ PAP train 200 datasets, respectively. After applying our policies, the increments of m A A in the Γ PAP train dataset all exceed the maximum increments of applying a single DA method, while AA-IP and AA-CP are not as effective as the best single DA method on Γ PAP train 100 and Γ PAP train 200 . A similar performance can be observed on Γ CCZ . It indicates that for our marine-domain data, RandomRotation works very well. However, they have the opposite effect on traffic-domain dataset Γ CSD , revealing that DA methods have a strong domain dependency.

5. Conclusions

We have shown that in our work, we could observe a clear difference in the effects of DA applied to our domain-specific marine dataset or the more established everyday urban traffic dataset. Therefore we propose to use DA with lower expectations, especially when applied to image domains that differ from the mainstream image domains CNNs are applied to. As we can show good results with a customized DA policy, we conclude that DA can definitely be the tool of choice, especially for small training sets, but increased efforts are required to make data augmentation more adaptive and domain aware.

Author Contributions

Conceptualization, M.T., D.L. and T.W.N.; methodology, M.T., D.L. and T.W.N.; formal analysis, M.T., D.L. and T.W.N.; investigation, M.T.; resources, D.L. and T.W.N.; data curation, M.T.; writing—original draft preparation, M.T. and T.W.N.; writing—review and editing, M.T., D.L. and T.W.N.; supervision, D.L. and T.W.N.; project administration, T.W.N.; funding acquisition, T.W.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by BMBF (Cosemio FKZ 03F0812C) and the BMWi (Isymoo FKZ 0324254D). We acknowledge the financial support of the German Research Foundation (DFG) and the Open Access Publication Fund of Bielefeld University for the article processing charge.

Data Availability Statement

The data are available on request from the corresponding author.

Acknowledgments

We acknowledge the UK Natural Environment Research Council that funded the Autonomous Ecological Surveying of the Abyss (AESA) project (Grant Reference NE/H021787/1), Climate Linked Atlantic Section Science (CLASS) program (Grant Reference NE/R015953/1) and Seabed Mining And Resilience To EXperimental impact (SMARTEX) project (Grant Reference NE/T003537/1). We also acknowledge the European Union Seventh Framework Programme (FP7/2007-2013) under the MIDAS (Managing Impacts of Deep-seA resource exploitation) project, grant agreement 603418.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AUVAutonomous Underwater Vehicles
ROVRemotely Operated Vehicles
SVMSupport Vector Machines
CNNConvolutional Neural Networks
DAData Augmentation
GANGenerative Adversarial Networks
PAPPorcupine Abyssal Plane
CCZClarion Clipperton Zone
CSDCityscapes Dataset

Appendix A. Performance of DA Methods on Γ PAP train 100 and Γ CSD train 100

Table A1. Performance of DA methods on Γ PAP train 100 . This table indicates the augment effect of different DA methods and parameters applied on Γ PAP train 100 when setting seed to 350. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy). The four DA methods with the best effect are still RandomRotation, Brightness, Contrast, and Shear. These four methods can obtain their best enhancement effect on Γ PAP train 100 with parameters 160°, 2, 5, and (−20°, 20°, −20°, 20°), respectively. The trend of the effect of different parameters is similar to that of Γ PAP train 50 , but with fewer increments for the average accuracy.
Table A1. Performance of DA methods on Γ PAP train 100 . This table indicates the augment effect of different DA methods and parameters applied on Γ PAP train 100 when setting seed to 350. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy). The four DA methods with the best effect are still RandomRotation, Brightness, Contrast, and Shear. These four methods can obtain their best enhancement effect on Γ PAP train 100 with parameters 160°, 2, 5, and (−20°, 20°, −20°, 20°), respectively. The trend of the effect of different parameters is similar to that of Γ PAP train 50 , but with fewer increments for the average accuracy.
Random RotationBrightnessContrastSaturationRandom
VerticalFlip
Random
HorizontalFlip
2.64%0.12.86%0.12.27%0.11.72%0.12.76%0.11.90%
3.37%0.22.82%0.22.10%0.22.73%0.22.31%0.22.10%
3.40%0.33.57%0.33.17%0.32.70%0.32.52%0.31.89%
3.37%0.43.34%0.42.70%0.42.32%0.42.83%0.42.45%
3.34%0.54.02%0.53.23%0.52.41%0.51.98%0.52.11%
3.79%0.64.22%0.64.02%0.62.92%0.62.52%0.62.17%
3.13%0.74.43%0.73.46%0.72.00%0.72.37%0.72.24%
3.73%0.84.03%0.83.51%0.82.75%0.82.07%0.82.44%
2.96%0.94.11%0.94.84%0.93.15%0.92.21%0.91.54%
10°2.98%14.65%14.45%13.16%11.14%10.09%
20°3.66%1.14.69%1.34.33%1.53.68%Shear
30°4.06%1.24.17%1.54.54%22.97%
40°4.71%1.34.41%1.84.48%33.48%2.01%
50°4.94%1.44.91%24.65%43.62%10°3.30%
60°5.21%1.54.91%2.54.37%53.58%20°3.45%
70°4.83%1.64.74%34.50%64.03%30°4.14%
80°4.49%1.75.08%3.54.11%74.05%40°3.26%
90°5.07%1.84.79%44.50%84.17%(0°, 0°, −5°, 5°)2.91%
100°5.00%1.94.38%4.54.88%93.39%(0°, 0°, −10°, 10°)3.31%
110°5.37%25.08%55.40%103.94%(0°, 0°, −20°, 20°)3.37%
120°4.72%HueTranslate(0°, 0°, −30°, 30°)3.20%
130°4.92%(0°, 0°, −40°, 40°)4.27%
140°4.80%0.13.00%(0.1, 0.1)3.23%(−5°, 5°, −5°, 5°)2.77%
150°4.76%0.23.38%(0.2, 0.2)3.06%(−10°, 10°, −10°, 10°)3.81%
160°5.61%0.32.98%(0.3, 0.3)2.16%(−20°, 20°, −20°, 20°)4.54%
170°5.43%0.42.63%(0.4, 0.4)2.19%(−30°, 30°, −30°, 30°)4.03%
180°5.17%0.53.12%(0.5, 0.5)−0.17%(−40°, 40°, −40°, 40°)4.45%
Table A2. Performance of DA methods on Γ CSD train 100 . This heatmap shows the results when we increase the training set samples to 100 per class. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy). We can see that RandomRotation and RandomVerticalFlip have a negative impact. Saturation with a parameter of 9 and Brightness with a parameter of 1.7 are the two most effective DA methods and parameters.
Table A2. Performance of DA methods on Γ CSD train 100 . This heatmap shows the results when we increase the training set samples to 100 per class. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy). We can see that RandomRotation and RandomVerticalFlip have a negative impact. Saturation with a parameter of 9 and Brightness with a parameter of 1.7 are the two most effective DA methods and parameters.
Random RotationBrightnessContrastSaturationRandom
VerticalFlip
Random
HorizontalFlip
0.14%0.10.13%0.10.10%0.10.09%0.1−1.16%0.10.24%
0.15%0.20.23%0.20.05%0.20.08%0.2−1.42%0.20.41%
0.23%0.30.23%0.30.08%0.30.18%0.3−1.56%0.30.50%
0.05%0.40.31%0.40.16%0.40.15%0.4−2.02%0.40.42%
0.38%0.50.26%0.50.22%0.50.26%0.5−1.95%0.50.50%
0.51%0.60.27%0.60.25%0.60.07%0.6−2.13%0.60.47%
0.50%0.7−0.08%0.7−0.23%0.70.08%0.7−2.53%0.70.34%
0.65%0.80.03%0.8−1.36%0.80.30%0.8−2.43%0.80.20%
0.55%0.90.60%0.90.13%0.90.21%0.9−2.10%0.90.24%
10°−0.01%10.66%10.19%10.24%1−7.55%10.27%
20°−0.01%1.10.88%1.30.41%1.50.37%Shear
30°−0.17%1.20.71%1.50.50%20.53%
40°−0.20%1.30.45%1.80.45%30.63%0.36%
50°0.10%1.40.29%20.62%41.02%10°0.44%
60°−0.45%1.50.75%2.50.70%51.27%20°0.11%
70°−0.56%1.60.70%30.52%61.14%30°−0.03%
80°−3.18%1.70.95%3.5−0.31%71.24%40°0.01%
90°−2.44%1.80.57%40.31%81.30%(0°, 0°, −5°, 5°)0.29%
100°−2.68%1.90.58%4.50.83%91.41%(0°, 0°, −10°, 10°)0.33%
110°−2.11%20.32%5−0.36%101.35%(0°, 0°, −20°, 20°)0.28%
120°−2.34%HueTranslate(0°, 0°, −30°, 30°)0.48%
130°−1.83%(0°, 0°, −40°, 40°)0.56%
140°−3.11%0.10.50%(0.1, 0.1)−0.16%(−5°, 5°, −5°, 5°)0.53%
150°−3.36%0.20.57%(0.2, 0.2)−0.44%(−10°, 10°, −10°, 10°)0.37%
160°−3.38%0.30.69%(0.3, 0.3)−0.34%(−20°, 20°, −20°, 20°)0.50%
170°−3.22%0.40.52%(0.4, 0.4)−2.34%(−30°, 30°, −30°, 30°)0.35%
180°−3.12%0.50.52%(0.5, 0.5)−2.76%(−40°, 40°, −40°, 40°)−0.21%

Appendix B. Performance of DA Methods on Γ PAP train 50 and Γ CCZ train 50 with Setting Seed to 3500

Table A3. Performance of DA methods on Γ PAP train 50 with setting seed to 3500. This heatmap shows the effect of different DA methods and parameters on the average accuracy improvement of Γ PAP train 50 when setting seed to 3500. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy). DA methods show a similar trend to that at seed is 350.
Table A3. Performance of DA methods on Γ PAP train 50 with setting seed to 3500. This heatmap shows the effect of different DA methods and parameters on the average accuracy improvement of Γ PAP train 50 when setting seed to 3500. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy). DA methods show a similar trend to that at seed is 350.
Random RotationBrightnessContrastSaturationRandom
VerticalFlip
Random
HorizontalFlip
3.26%0.11.76%0.11.06%0.10.75%0.12.26%0.11.60%
4.34%0.22.57%0.21.52%0.20.88%0.22.64%0.21.75%
4.96%0.33.19%0.32.12%0.31.32%0.32.87%0.32.00%
5.21%0.43.16%0.41.24%0.41.63%0.42.91%0.42.08%
5.45%0.53.59%0.52.98%0.51.63%0.52.99%0.51.97%
5.58%0.63.94%0.63.29%0.61.94%0.62.79%0.61.93%
5.65%0.74.16%0.73.61%0.72.16%0.72.72%0.71.88%
5.68%0.84.70%0.84.01%0.82.11%0.82.74%0.81.67%
5.75%0.95.30%0.94.43%0.92.51%0.92.55%0.91.21%
10°5.58%15.76%14.84%12.48%1−0.51%1−0.36%
20°5.61%1.15.34%1.34.83%1.52.69%Shear
30°5.81%1.25.34%1.54.83%23.00%
40°6.10%1.36.17%1.84.77%33.17%3.36%
50°6.25%1.45.97%25.12%43.50%10°3.68%
60°6.25%1.56.10%2.55.78%54.02%20°4.30%
70°6.71%1.65.92%35.49%64.20%30°5.79%
80°7.21%1.76.06%3.55.80%74.25%40°5.65%
90°7.31%1.85.66%45.75%84.22%(0°, 0°, −5°, 5°)4.17%
100°7.59%1.96.13%4.55.12%94.48%(0°, 0°, −10°, 10°)4.60%
110°7.45%26.34%56.70%104.53%(0°, 0°, −20°, 20°)4.98%
120°7.21%HueTranslate(0°, 0°, −30°, 30°)5.59%
130°7.24%(0°, 0°, −40°, 40°)5.98%
140°7.10%0.13.01%(0.1, 0.1)5.50%(−5°, 5°, −5°, 5°)5.32%
150°7.53%0.23.91%(0.2, 0.2)4.20%(−10°, 10°, −10°, 10°)5.07%
160°7.28%0.33.83%(0.3, 0.3)3.31%(−20°, 20°, −20°, 20°)5.52%
170°7.42%0.43.75%(0.4, 0.4)1.54%(−30°, 30°, −30°, 30°)6.32%
180°6.94%0.53.47%(0.5, 0.5)−3.24%(−40°, 40°, −40°, 40°)5.41%
Table A4. Performance of DA methods on Γ CCZ train 50 with setting seed to 3500. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy). This heatmap shows the effect of different DA methods and different parameters on the average accuracy improvement of Γ CCZ train 50 when setting seed to 3500. As with the experimental results on Γ PAP train 50 , the performance of DA is similar under different seeds.
Table A4. Performance of DA methods on Γ CCZ train 50 with setting seed to 3500. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy). This heatmap shows the effect of different DA methods and different parameters on the average accuracy improvement of Γ CCZ train 50 when setting seed to 3500. As with the experimental results on Γ PAP train 50 , the performance of DA is similar under different seeds.
Random RotationBrightnessContrastSaturationRandom
VerticalFlip
Random
HorizontalFlip
4.14%0.12.76%0.12.45%0.11.51%0.13.61%0.12.72%
4.11%0.23.18%0.23.23%0.21.62%0.23.20%0.23.02%
4.60%0.34.03%0.33.23%0.31.21%0.33.23%0.33.10%
4.56%0.44.08%0.43.93%0.42.08%0.43.10%0.43.28%
4.53%0.54.49%0.54.25%0.52.10%0.53.14%0.53.04%
4.51%0.64.75%0.64.42%0.62.12%0.63.11%0.63.12%
4.43%0.74.35%0.74.59%0.72.34%0.72.90%0.73.23%
4.69%0.84.47%0.84.66%0.82.00%0.82.57%0.83.10%
4.97%0.94.86%0.94.79%0.92.11%0.92.32%0.92.81%
10°4.61%14.58%13.44%12.22%1−1.51%10.16%
20°5.28%1.14.45%1.35.09%1.52.73%Shear
30°5.16%1.24.83%1.54.50%22.71%
40°5.22%1.35.08%1.84.85%32.83%4.18%
50°5.44%1.44.86%25.29%42.93%10°4.38%
60°5.58%1.55.33%2.55.64%52.93%20°4.56%
70°5.76%1.65.20%35.75%63.09%30°4.95%
80°5.50%1.74.93%3.55.21%72.78%40°5.21%
90°6.19%1.84.86%45.82%82.52%(0°, 0°, −5°, 5°)4.13%
100°6.31%1.94.89%4.55.38%92.60%(0°, 0°, −10°, 10°)4.49%
110°6.27%24.81%55.48%102.41%(0°, 0°, −20°, 20°)4.22%
120°5.73%HueTranslate(0°, 0°, −30°, 30°)5.00%
130°5.82%(0°, 0°, −40°, 40°)4.77%
140°6.08%0.12.59%(0.1, 0.1)4.30%(−5°, 5°, −5°, 5°)4.98%
150°5.97%0.22.77%(0.2, 0.2)4.07%(−10°, 10°, −10°, 10°)4.97%
160°6.07%0.32.79%(0.3, 0.3)3.54%(−20°, 20°, −20°, 20°)5.29%
170°5.98%0.43.35%(0.4, 0.4)1.94%(−30°, 30°, −30°, 30°)4.84%
180°6.22%0.53.12%(0.5, 0.5)−1.62%(−40°, 40°, −40°, 40°)4.72%

Appendix C. Impacts Comparison of Flip, Translate, and Hue

Figure A1. Impacts comparison of Random Horizontal Flip and Random Vertical Flip. The figure reveals the impacts of the different parameters of Random Horizontal Flip and Random Vertical Flip on Γ PAP , Γ CCZ and Γ CSD classification m A A improvement.
Figure A1. Impacts comparison of Random Horizontal Flip and Random Vertical Flip. The figure reveals the impacts of the different parameters of Random Horizontal Flip and Random Vertical Flip on Γ PAP , Γ CCZ and Γ CSD classification m A A improvement.
Sensors 22 05383 g0a1
Figure A2. Impacts comparison of Translate and Hue. The figure shows the performance of the different parameters of Translate and Hue on the three datasets.
Figure A2. Impacts comparison of Translate and Hue. The figure shows the performance of the different parameters of Translate and Hue on the three datasets.
Sensors 22 05383 g0a2

References

  1. Wynn, R.B.; Huvenne, V.A.; Le Bas, T.P.; Murton, B.J.; Connelly, D.P.; Bett, B.J.; Ruhl, H.A.; Morris, K.J.; Peakall, J.; Parsons, D.R.; et al. Autonomous Underwater Vehicles (AUVs): Their past, present and future contributions to the advancement of marine geoscience. Mar. Geol. 2014, 352, 451–468. [Google Scholar] [CrossRef] [Green Version]
  2. Christ, R.D.; Wernli, R.L., Sr. The ROV Manual: A User Guide for Remotely Operated Vehicles; Butterworth-Heinemann: Oxford, UK, 2013. [Google Scholar]
  3. Langenkämper, D.; Kevelaer, R.V.; Nattkemper, T.W. Strategies for tackling the class imbalance problem in marine image classification. In International Conference on Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2018; pp. 26–36. [Google Scholar]
  4. Wei, Y.; Yu, X.; Hu, Y.; Li, D. Development a zooplankton recognition method for dark field image. In Proceedings of the 2012 5th International Congress on Image and Signal Processing, Chongqing, China, 16–18 October 2012; pp. 861–865. [Google Scholar]
  5. Bewley, M.; Douillard, B.; Nourani-Vatani, N.; Friedman, A.; Pizarro, O.; Williams, S. Automated species detection: An experimental approach to kelp detection from sea-floor AUV images. In Proceedings of the Australasian Conference on Robotics and Automation, Wellingto, New Zealand, 3–5 December 2012. [Google Scholar]
  6. Lu, H.; Li, Y.; Uemura, T.; Ge, Z.; Xu, X.; He, L.; Serikawa, S.; Kim, H. FDCNet: Filtering deep convolutional network for marine organism classification. Multimed. Tools Appl. 2018, 77, 21847–21860. [Google Scholar] [CrossRef]
  7. Li, Y.; Lu, H.; Li, J.; Li, X.; Li, Y.; Serikawa, S. Underwater image de-scattering and classification by deep neural network. Comput. Electr. Eng. 2016, 54, 68–77. [Google Scholar] [CrossRef]
  8. Liu, X.; Jia, Z.; Hou, X.; Fu, M.; Ma, L.; Sun, Q. Real-time marine animal images classification by embedded system based on mobilenet and transfer learning. In Proceedings of the OCEANS 2019-Marseille, Marseille, France, 17–20 June 2019; pp. 1–5. [Google Scholar]
  9. Langenkämper, D.; Simon-Lledó, E.; Hosking, B.; Jones, D.O.; Nattkemper, T.W. On the impact of Citizen Science-derived data quality on deep learning based classification in marine images. PLoS ONE 2019, 14, e0218086. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Langenkämper, D.; Van Kevelaer, R.; Purser, A.; Nattkemper, T.W. Gear-induced concept drift in marine images and its effect on deep learning classification. Front. Mar. Sci. 2020, 7, 506. [Google Scholar] [CrossRef]
  11. Irfan, M.; Zheng, J.; Iqbal, M.; Arif, M.H. A novel feature extraction model to enhance underwater image classification. In International Symposium on Intelligent Computing Systems; Springer: Cham, Switzerland, 2020; pp. 78–91. [Google Scholar]
  12. Mahmood, A.; Bennamoun, M.; An, S.; Sohel, F.A.; Boussaid, F.; Hovey, R.; Kendrick, G.A.; Fisher, R.B. Deep image representations for coral image classification. IEEE J. Ocean. Eng. 2018, 44, 121–131. [Google Scholar] [CrossRef] [Green Version]
  13. Han, F.; Yao, J.; Zhu, H.; Wang, C. Marine organism detection and classification from underwater vision based on the deep CNN method. Math. Probl. Eng. 2020, 2020, 3937580. [Google Scholar] [CrossRef]
  14. Huang, H.; Zhou, H.; Yang, X.; Zhang, L.; Qi, L.; Zang, A.Y. Faster R-CNN for marine organisms detection and recognition using data augmentation. Neurocomputing 2019, 337, 372–384. [Google Scholar] [CrossRef]
  15. Xu, L.; Bennamoun, M.; An, S.; Sohel, F.; Boussaid, F. Deep learning for marine species recognition. In Handbook of Deep Learning Applications; Springer: Berlin/Heidelberg, Germany, 2019; pp. 129–145. [Google Scholar]
  16. Zhuang, P.; Xing, L.; Liu, Y.; Guo, S.; Qiao, Y. Marine Animal Detection and Recognition with Advanced Deep Learning Models. In CLEF (Working Notes); Springer: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  17. Xu, Y.; Zhang, Y.; Wang, H.; Liu, X. Underwater image classification using deep convolutional neural networks and data augmentation. In Proceedings of the 2017 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Xiamen, China, 22–25 October 2017; pp. 1–5. [Google Scholar]
  18. Wang, N.; Wang, Y.; Er, M.J. Review on deep learning techniques for marine object recognition: Architectures and algorithms. Control. Eng. Pract. 2020, 118, 104458. [Google Scholar] [CrossRef]
  19. NgoGia, T.; Li, Y.; Jin, D.; Guo, J.; Li, J.; Tang, Q. Real-Time Sea Cucumber Detection Based on YOLOv4-Tiny and Transfer Learning Using Data Augmentation. In International Conference on Swarm Intelligence; Springer: Cham, Switzerland, 2021; pp. 119–128. [Google Scholar]
  20. Xie, C.; Tan, M.; Gong, B.; Wang, J.; Yuille, A.L.; Le, Q.V. Adversarial examples improve image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–19 June 2020; pp. 819–828. [Google Scholar]
  21. Xiao, Q.; Liu, B.; Li, Z.; Ni, W.; Yang, Z.; Li, L. Progressive data augmentation method for remote sensing ship image classification based on imaging simulation system and neural style transfer. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 9176–9186. [Google Scholar] [CrossRef]
  22. Chu, P.; Bian, X.; Liu, S.; Ling, H. Feature space augmentation for long-tailed data. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 694–710. [Google Scholar]
  23. DeVries, T.; Taylor, G.W. Dataset augmentation in feature space. arXiv 2017, arXiv:1702.05538. [Google Scholar]
  24. Li, C.; Huang, Z.; Xu, J.; Yan, Y. Data Augmentation using Conditional Generative Adversarial Network for Underwater Target Recognition. In Proceedings of the 2019 IEEE International Conference on Signal, Information and Data Processing (ICSIDP), Chongqing, China, 11–13 December 2019; pp. 1–5. [Google Scholar]
  25. Antoniou, A.; Storkey, A.; Edwards, H. Augmenting image classifiers using data augmentation generative adversarial networks. In International Conference on Artificial Neural Networks; Springer: Berlin/Heidelberg, Germany, 2018; pp. 594–603. [Google Scholar]
  26. Cubuk, E.D.; Zoph, B.; Mane, D.; Vasudevan, V.; Le, Q.V. Autoaugment: Learning augmentation strategies from data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 113–123. [Google Scholar]
  27. Lemley, J.; Bazrafkan, S.; Corcoran, P. Smart augmentation learning an optimal data augmentation strategy. IEEE Access 2017, 5, 5858–5869. [Google Scholar] [CrossRef]
  28. Dabouei, A.; Soleymani, S.; Taherkhani, F.; Nasrabadi, N.M. Supermix: Supervising the mixing data augmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 13794–13803. [Google Scholar]
  29. Naveed, H. Survey: Image mixing and deleting for data augmentation. arXiv 2021, arXiv:2106.07085. [Google Scholar]
  30. Kang, G.; Dong, X.; Zheng, L.; Yang, Y. Patchshuffle regularization. arXiv 2017, arXiv:1707.07103. [Google Scholar]
  31. O’Gara, S.; McGuinness, K. Comparing Data Augmentation Strategies for Deep Image Classification; Technological University Dublin: Dublin, Ireland, 2019. [Google Scholar]
  32. Engstrom, L.; Tran, B.; Tsipras, D.; Schmidt, L.; Madry, A. A Rotation and a Translation Suffice: Fooling Cnns with Simple Transformations. arXiv 2018, arXiv:171202779. [Google Scholar]
  33. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  34. Lim, S.; Kim, I.; Kim, T.; Kim, C.; Kim, S. Fast autoaugment. Adv. Neural Inf. Process. Syst. 2019, 32, 6665–6675. [Google Scholar]
  35. Shijie, J.; Ping, W.; Peiyi, J.; Siping, H. Research on data augmentation for image classification based on convolution neural networks. In Proceedings of the 2017 Chinese Automation Congress (CAC), Jinan, China, 20–22 October 2017; pp. 4165–4170. [Google Scholar]
  36. Durden, J.M.; Bett, B.J.; Ruhl, H.A. Subtle variation in abyssal terrain induces significant change in benthic megafaunal abundance, diversity, and community structure. Prog. Oceanogr. 2020, 186, 102395. [Google Scholar] [CrossRef]
  37. Hartman, S.E.; Bett, B.J.; Durden, J.M.; Henson, S.A.; Iversen, M.; Jeffreys, R.M.; Horton, T.; Lampitt, R.; Gates, A.R. Enduring science: Three decades of observing the Northeast Atlantic from the Porcupine Abyssal Plain Sustained Observatory (PAP-SO). Prog. Oceanogr. 2021, 191, 102508. [Google Scholar] [CrossRef]
  38. Morris, K.J.; Bett, B.J.; Durden, J.M.; Huvenne, V.A.; Milligan, R.; Jones, D.O.; McPhail, S.; Robert, K.; Bailey, D.M.; Ruhl, H.A. A new method for ecological surveying of the abyss using autonomous underwater vehicle photography. Limnol. Oceanogr. Methods 2014, 12, 795–809. [Google Scholar] [CrossRef] [Green Version]
  39. Simon-Lledó, E.; Bett, B.J.; Huvenne, V.A.; Schoening, T.; Benoist, N.M.; Jeffreys, R.M.; Durden, J.M.; Jones, D.O. Megafaunal variation in the abyssal landscape of the Clarion Clipperton Zone. Prog. Oceanogr. 2019, 170, 119–133. [Google Scholar] [CrossRef] [PubMed]
  40. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes Dataset for Semantic Urban Scene Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. pp. 3213–3223.
  41. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
Figure 1. Γ PAP dataset. (a) Ophiuroidea; (b) Cnidaria; (c) Amperima; (d) Foraminifera. Reproduced with permission from Henry Ruhl.
Figure 1. Γ PAP dataset. (a) Ophiuroidea; (b) Cnidaria; (c) Amperima; (d) Foraminifera. Reproduced with permission from Henry Ruhl.
Sensors 22 05383 g001
Figure 2. Dataset structure. The datasets were randomly subdivided into train, validation, and test set.
Figure 2. Dataset structure. The datasets were randomly subdivided into train, validation, and test set.
Sensors 22 05383 g002
Figure 3. Construction of the training sets. Experiments were conducted using training sets of different sizes to investigate the influence of training set size on the impact of augmentation.
Figure 3. Construction of the training sets. Experiments were conducted using training sets of different sizes to investigate the influence of training set size on the impact of augmentation.
Sensors 22 05383 g003
Figure 4. Γ CCZ dataset. (a) Sponge; (b) Coral.
Figure 4. Γ CCZ dataset. (a) Sponge; (b) Coral.
Sensors 22 05383 g004
Figure 5. Γ CSD dataset. (a) Car; (b) Bicycle.
Figure 5. Γ CSD dataset. (a) Car; (b) Bicycle.
Sensors 22 05383 g005
Figure 6. Flowchart of the work. Each DA method was applied to each train data and an individual model was trained and optimized using the validation data. The accuracy of the individually trained model was evaluated using the test data.
Figure 6. Flowchart of the work. Each DA method was applied to each train data and an individual model was trained and optimized using the validation data. The accuracy of the individually trained model was evaluated using the test data.
Sensors 22 05383 g006
Figure 7. Impacts comparison of Rotation and Shear. The figure reveals the impacts of the different parameters of Rotation and Shear on Γ PAP , Γ CCZ , and Γ CSD classification performance improvement.
Figure 7. Impacts comparison of Rotation and Shear. The figure reveals the impacts of the different parameters of Rotation and Shear on Γ PAP , Γ CCZ , and Γ CSD classification performance improvement.
Sensors 22 05383 g007aSensors 22 05383 g007b
Figure 8. Impacts comparison of Brightness, Contrast and Saturation. The impacts of Brightness, Contrast and Saturation on improving the classification performance on Γ PAP , Γ CCZ , and Γ CSD are plotted in the figure.
Figure 8. Impacts comparison of Brightness, Contrast and Saturation. The impacts of Brightness, Contrast and Saturation on improving the classification performance on Γ PAP , Γ CCZ , and Γ CSD are plotted in the figure.
Sensors 22 05383 g008aSensors 22 05383 g008b
Table 1. Sample sizes of each category per subset in the Γ PAP dataset.
Table 1. Sample sizes of each category per subset in the Γ PAP dataset.
Γ PAP OphiuroideaCnidariaAmperimaForaminifera
Γ PAP train 50 50505050
Γ PAP train 100 100100100100
Γ PAP train 200 200200200200
Γ PAP train 300300300300
Γ PAP validation 200200200200
Γ PAP test 8883886152022132
Table 2. Sample sizes of category per subset in the Γ CCZ dataset.
Table 2. Sample sizes of category per subset in the Γ CCZ dataset.
Γ CCZ SpongeCoral
Γ CCZ train 50 5050
Γ CCZ train 100 100100
Γ CCZ train 150150
Γ CCZ validation 150150
Γ CCZ test 1236700
Table 3. Sample sizes of each category per subset in the Γ CSD dataset.
Table 3. Sample sizes of each category per subset in the Γ CSD dataset.
Γ CSD CarBicycle
Γ CSD train 50 5050
Γ CSD train 100 100100
Γ CSD train 150150
Γ CSD validation 200200
Γ CSD test 24,3713208
Table 4. Performance of DA methods on Γ PAP train 50 . The impact of DA methods and parameters on classification performance for Γ PAP train 50 is revealed. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Table 4. Performance of DA methods on Γ PAP train 50 . The impact of DA methods and parameters on classification performance for Γ PAP train 50 is revealed. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Random RotationBrightnessContrastSaturationRandom
VerticalFlip
Random
HorizontalFlip
3.82%0.12.79%0.11.97%0.10.92%0.12.68%0.11.65%
5.11%0.23.95%0.22.85%0.21.35%0.23.15%0.21.93%
5.69%0.34.38%0.33.34%0.31.96%0.33.24%0.31.95%
5.89%0.44.40%0.43.71%0.42.18%0.43.51%0.41.99%
6.11%0.54.90%0.54.15%0.52.46%0.53.57%0.52.11%
6.01%0.65.35%0.64.53%0.62.81%0.63.66%0.62.14%
6.03%0.75.61%0.74.66%0.73.11%0.73.57%0.72.11%
6.30%0.85.64%0.85.19%0.83.06%0.83.31%0.81.80%
6.10%0.96.24%0.95.59%0.92.97%0.93.12%0.91.56%
10°6.01%16.82%16.22%12.95%10.34%1–1.16%
20°6.85%1.16.25%1.36.33%1.53.27%Shear
30°6.97%1.26.87%1.55.87%23.29%
40°6.71%1.36.75%1.86.18%34.08%4.05%
50°6.96%1.47.01%26.29%44.67%10°3.92%
60°7.41%1.57.22%2.55.86%54.97%20°4.34%
70°7.17%1.66.56%37.13%65.43%30°4.93%
80°7.39%1.76.95%3.57.00%75.49%40°5.48%
90°7.61%1.87.01%46.88%85.72%(0°, 0°, −5°, 5°)4.20%
100°8.02%1.97.11%4.56.59%95.83%(0°, 0°, −10°, 10°)5.38%
110°7.84%27.74%57.70%105.99%(0°, 0°, −20°, 20°)6.20%
120°7.70%HueTranslate(0°,0°, −30°, 30°)6.06%
130°7.69%(0°, 0°, −40°, 40°)6.47%
140°7.76%0.14.17%(0.1, 0.1)5.59%(−5°, 5°, −5°, 5°)5.89%
150°7.31%0.25.13%(0.2, 0.2)4.72%(−10°, 10°, −10°, 10°)5.75%
160°7.47%0.34.86%(0.3, 0.3)4.21%(−20°, 20°, −20°, 20°)6.22%
170°7.11%0.44.52%(0.4, 0.4)2.53%(−30°, 30°, −30°, 30°)6.68%
180°7.55%0.54.70%(0.5, 0.5)1.40%(−40°, 40°, −40°, 40°)6.58%
Table 5. Performance of DA methods on Γ PAP train 200 . The DA methods and parameters impact on Γ PAP train 200 classification performance is shown. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Table 5. Performance of DA methods on Γ PAP train 200 . The DA methods and parameters impact on Γ PAP train 200 classification performance is shown. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Random RotationBrightnessContrastSaturationRandom
VerticalFlip
Random
HorizontalFlip
1.30%0.11.19%0.10.69%0.10.25%0.10.94%0.10.77%
1.55%0.21.22%0.21.41%0.20.34%0.20.81%0.20.92%
2.18%0.31.38%0.31.41%0.30.42%0.31.01%0.30.80%
2.11%0.41.70%0.41.48%0.40.57%0.40.86%0.40.88%
2.01%0.51.75%0.51.56%0.50.74%0.51.31%0.50.79%
1.94%0.61.85%0.61.76%0.61.11%0.60.89%0.60.82%
2.08%0.72.13%0.71.95%0.71.24%0.70.77%0.70.74%
1.55%0.82.15%0.82.12%0.80.38%0.80.56%0.80.70%
2.18%0.92.59%0.91.82%0.91.09%0.90.78%0.90.62%
10°2.15%12.57%12.05%11.27%1−0.43%1−0.43%
20°2.14%1.12.34%1.31.84%1.50.81%Shear
30°2.53%1.22.43%1.52.15%21.49%
40°2.17%1.32.34%1.82.32%31.75%1.70%
50°2.62%1.42.35%21.81%41.72%10°1.85%
60°2.74%1.52.06%2.51.94%51.80%20°1.78%
70°2.86%1.62.43%32.15%61.85%30°1.53%
80°2.45%1.72.35%3.52.26%71.86%40°2.13%
90°3.06%1.82.47%42.19%81.61%(0°, 0°, −5°, 5°)1.63%
100°3.17%1.92.52%4.52.40%91.76%(0°, 0°, −10°, 10°)2.34%
110°3.09%22.39%52.28%101.72%(0°, 0°, −20°, 20°)1.72%
120°3.25%HueTranslate(0°, 0°, −30°, 30°)2.06%
130°2.96%(0°, 0°, −40°, 40°)2.33%
140°3.14%0.11.45%(0.1, 0.1)2.00%(−5°, 5°, −5°, 5°)2.20%
150°3.27%0.21.61%(0.2, 0.2)1.67%(−10°, 10°, −10°, 10°)2.22%
160°3.08%0.31.94%(0.3, 0.3)1.57%(−20°, 20°, −20°, 20°)2.40%
170°3.29%0.41.73%(0.4, 0.4)0.69%(−30°, 30°, −30°, 30°)2.29%
180°3.23%0.51.72%(0.5, 0.5)0.21%(−40°, 40°, −40°, 40°)2.40%
Table 6. Performance of DA methods on Γ CCZ train 50 . The impact of different DA methods and parameters on Γ CCZ train 50 classification performance is displayed. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Table 6. Performance of DA methods on Γ CCZ train 50 . The impact of different DA methods and parameters on Γ CCZ train 50 classification performance is displayed. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Random RotationBrightnessContrastSaturationRandom
VerticalFlip
Random
HorizontalFlip
3.07%0.11.86%0.12.12%0.10.77%0.12.93%0.12.27%
4.13%0.22.77%0.22.40%0.21.20%0.23.05%0.22.67%
3.99%0.33.18%0.33.08%0.31.26%0.33.24%0.32.60%
4.39%0.44.03%0.44.01%0.41.66%0.43.01%0.42.53%
4.91%0.54.26%0.54.32%0.51.82%0.52.80%0.52.60%
4.98%0.64.50%0.64.88%0.61.37%0.63.04%0.62.85%
5.00%0.74.94%0.74.91%0.71.76%0.73.01%0.72.78%
5.12%0.84.51%0.85.36%0.82.07%0.83.15%0.82.65%
4.97%0.94.44%0.94.77%0.92.08%0.92.63%0.92.31%
10°5.31%15.32%14.97%12.29%1−0.93%1−0.99%
20°5.70%1.15.35%1.34.86%1.52.21%Shear
30°5.54%1.25.19%1.55.06%21.27%
40°6.35%1.35.10%1.85.03%31.67%2.70%
50°6.11%1.45.43%25.15%41.42%10°3.39%
60°5.81%1.55.16%2.55.39%52.09%20°4.95%
70°6.17%1.64.71%34.95%62.20%30°5.55%
80°6.48%1.75.53%3.54.56%72.03%40°5.44%
90°6.69%1.85.09%44.88%82.43%(0°, 0°, −5°, 5°)4.71%
100°6.31%1.95.54%4.55.30%92.19%(0°, 0°, −10°, 10°)5.16%
110°6.57%25.28%55.16%101.85%(0°, 0°, −20°, 20°)4.95%
120°6.85%HueTranslate(0°, 0°, −30°, 30°)5.23%
130°6.66%(0°, 0°, −40°, 40°)5.51%
140°6.31%0.13.18%(0.1, 0.1)4.98%(−5°, 5°, −5°, 5°)5.55%
150°7.14%0.23.96%(0.2, 0.2)4.19%(−10°, 10°, −10°, 10°)5.60%
160°6.95%0.33.98%(0.3, 0.3)3.76%(−20°, 20°, −20°, 20°)5.64%
170°6.80%0.43.48%(0.4, 0.4)3.09%(−30°, 30°, −30°, 30°)5.91%
180°6.73%0.52.95%(0.5, 0.5)−0.15%(−40°, 40°, −40°, 40°)6.11%
Table 7. Performance of DA methods on Γ CCZ train 100 . The effect of different DA methods and different parameters on Γ CCZ train 100 classification m A A improvement is shown. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Table 7. Performance of DA methods on Γ CCZ train 100 . The effect of different DA methods and different parameters on Γ CCZ train 100 classification m A A improvement is shown. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Random RotationBrightnessContrastSaturationRandom
VerticalFlip
Random
HorizontalFlip
1.84%0.10.73%0.10.53%0.10.32%0.11.50%0.10.92%
2.01%0.21.40%0.20.91%0.20.40%0.21.29%0.20.92%
1.81%0.31.59%0.31.18%0.30.60%0.31.26%0.30.99%
2.02%0.41.74%0.41.76%0.41.06%0.41.25%0.41.26%
2.18%0.51.73%0.51.72%0.51.32%0.51.10%0.51.14%
1.82%0.61.65%0.61.87%0.61.29%0.61.17%0.61.24%
1.80%0.72.47%0.72.34%0.71.48%0.70.95%0.71.20%
2.12%0.81.93%0.82.52%0.81.56%0.80.84%0.80.90%
2.26%0.92.13%0.92.55%0.91.91%0.90.73%0.90.87%
10°2.39%12.57%13.08%12.05%1−1.07%1−0.19%
20°2.71%1.12.22%1.33.33%1.52.46%Shear
30°3.19%1.22.83%1.53.22%22.46%
40°2.45%1.32.41%1.83.32%32.40%1.36%
50°2.60%1.42.49%23.00%42.35%10°1.59%
60°2.18%1.52.66%2.53.10%52.13%20°1.96%
70°3.09%1.62.87%33.03%61.88%30°2.37%
80°3.16%1.72.90%3.53.13%71.53%40°2.99%
90°3.76%1.82.43%43.10%81.73%(0°, 0°, −5°, 5°)1.88%
100°3.57%1.92.54%4.53.19%92.00%(0°, 0°, −10°, 10°)1.66%
110°3.68%23.14%53.33%102.18%(0°, 0°, −20°, 20°)1.85%
120°3.66%HueTranslate(0°, 0°, −30°, 30°)2.32%
130°3.56%(0°, 0°, −40°, 40°)2.75%
140°3.32%0.11.40%(0.1, 0.1)1.63%(−5°, 5°, −5°, 5°)2.02%
150°3.17%0.21.68%(0.2, 0.2)1.44%(−10°, 10°, −10°, 10°)2.14%
160°3.74%0.31.69%(0.3, 0.3)1.02%(−20°, 20°, −20°, 20°)2.30%
170°3.93%0.41.57%(0.4, 0.4)−0.66%(−30°, 30°, −30°, 30°)2.16%
180°3.64%0.51.62%(0.5, 0.5)−0.90%(−40°, 40°, −40°, 40°)2.11%
Table 8. Performance of DA methods on Γ CSD train 50 . The impact of different DA methods and parameters on Γ CSD train 50 classification performance, which is very different from the impact on Γ PAP train and Γ CCZ train , is displayed. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Table 8. Performance of DA methods on Γ CSD train 50 . The impact of different DA methods and parameters on Γ CSD train 50 classification performance, which is very different from the impact on Γ PAP train and Γ CCZ train , is displayed. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Random RotationBrightnessContrastSaturationRandom
VerticalFlip
Random
HorizontalFlip
−0.08%0.1−0.12%0.1−0.09%0.10.07%0.1−1.95%0.10.59%
0.32%0.20.06%0.2−0.14%0.20.17%0.2−1.43%0.20.46%
0.63%0.30.12%0.3−0.06%0.30.42%0.3−1.52%0.30.63%
0.39%0.4−0.16%0.4−0.08%0.40.63%0.4−1.39%0.40.68%
0.30%0.5−0.04%0.5−0.16%0.50.62%0.5−1.69%0.50.59%
0.52%0.60.27%0.6−0.06%0.60.69%0.6−1.82%0.60.57%
0.42%0.70.20%0.70.18%0.70.61%0.7−2.07%0.70.56%
0.34%0.80.17%0.8−0.35%0.80.77%0.8−2.49%0.80.51%
0.31%0.9−0.18%0.9−0.08%0.90.75%0.9−2.80%0.90.37%
10°0.19%1−0.19%1−0.56%10.90%1−8.53%1−0.30%
20°−0.67%1.1−0.60%1.3−0.67%1.51.18%Shear
30°−0.98%1.2−0.67%1.5−0.25%21.33%
40°−1.61%1.3−0.33%1.80.02%31.36%−0.14%
50°−0.47%1.40.19%2−0.25%41.30%10°0.25%
60°−0.88%1.50.07%2.5−0.33%51.22%20°−0.06%
70°−1.87%1.6−0.40%3−0.08%61.16%30°−0.42%
80°−2.67%1.70.03%3.50.22%71.26%40°−0.25%
90°−2.35%1.80.11%40.04%81.30%(0°, 0°, −5°, 5°)0.92%
100°−2.79%1.90.57%4.50.59%91.15%(0°, 0°, −10°, 10°)0.89%
110°−2.57%20.40%51.13%101.25%(0°, 0°, −20°, 20°)0.31%
120°−2.94%HueTranslate(0°, 0°, −30°, 30°)0.19%
130°−3.85%(0°, 0°, −40°, 40°)−0.10%
140°−3.81%0.10.53%(0.1, 0.1)−0.38%(−5°, 5°, −5°, 5°)0.37%
150°−3.89%0.20.27%(0.2, 0.2)−0.14%(−10°, 10°, −10°, 10°)0.55%
160°−4.80%0.30.16%(0.3, 0.3)−0.57%(−20°, 20°, −20°, 20°)0.14%
170°−4.31%0.4−0.01%(0.4, 0.4)−2.02%(−30°, 30°, −30°, 30°)−0.37%
180°−4.28%0.50.10%(0.5, 0.5)−2.46%(−40°, 40°, −40°, 40°)−1.18%
Table 9. Performance of DA methods on Γ CSD train 150 . The heatmap shows the DA methods and parameters impact on Γ CSD train 150 classification performance. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Table 9. Performance of DA methods on Γ CSD train 150 . The heatmap shows the DA methods and parameters impact on Γ CSD train 150 classification performance. The displayed percentage values describe the increase/decrease of m A A in percent. The value is color-coded from blue (increase in accuracy) over white (no effect) to orange (decrease in accuracy).
Random RotationBrightnessContrastSaturationRandom
VerticalFlip
Random
HorizontalFlip
0.35%0.10.23%0.10.25%0.10.13%0.1−1.22%0.10.26%
0.55%0.20.40%0.20.46%0.20.24%0.2−1.37%0.20.53%
0.74%0.30.34%0.30.55%0.30.23%0.3−1.78%0.30.50%
0.09%0.40.62%0.40.53%0.40.29%0.4−1.82%0.40.44%
0.55%0.50.73%0.50.62%0.50.21%0.5−2.21%0.50.37%
0.75%0.60.91%0.60.37%0.60.29%0.6−2.31%0.60.65%
0.83%0.70.51%0.70.90%0.70.42%0.7−2.33%0.70.47%
0.55%0.80.86%0.80.59%0.80.49%0.8−2.33%0.80.32%
0.84%0.90.86%0.90.18%0.90.48%0.9−1.64%0.90.57%
10°−0.21%10.81%11.07%10.49%1−8.28%10.48%
20°0.32%1.10.91%1.30.35%1.50.70%Shear
30°0.31%1.21.18%1.51.37%20.81%
40°−0.16%1.31.33%1.80.87%30.82%0.28%
50°−0.18%1.41.11%21.00%40.94%10°0.78%
60°−0.14%1.51.24%2.50.83%51.02%20°0.50%
70°−0.08%1.61.33%30.93%61.09%30°0.17%
80°−0.64%1.70.56%3.50.75%71.27%40°−0.11%
90°−2.72%1.80.96%41.40%81.31%(0°, 0°, −5°, 5°)0.56%
100°−2.48%1.90.96%4.50.73%91.31%(0°, 0°, −10°, 10°)−0.18%
110°−3.08%21.26%51.08%101.11%(0°, 0°, −20°, 20°)0.32%
120°−2.34%HueTranslate(0°, 0°, −30°, 30°)0.31%
130°−2.49%(0°, 0°, −40°, 40°)−0.27%
140°−3.21%0.10.60%(0.1, 0.1)0.13%(−5°, 5°, −5°, 5°)0.87%
150°−3.02%0.20.54%(0.2, 0.2)−1.52%(−10°, 10°, −10°, 10°)0.91%
160°−2.56%0.30.56%(0.3, 0.3)−0.62%(−20°, 20°, −20°, 20°)−0.72%
170°−3.26%0.40.61%(0.4, 0.4)−1.47%(−30°, 30°, −30°, 30°)−0.14%
180°−4.57%0.50.48%(0.5, 0.5)−0.79%(−40°, 40°, −40°, 40°)−0.06%
Table 10. DA combination policies.
Table 10. DA combination policies.
DA PolicyFunction of Each Policy
RBC_1 X i = f C T ( f R R ( x i , d = 150 ° ) , b = 1.9 , c = 5 )
RBC_2 X i = f C T ( f R R ( x i , d = 150 ° ) , b = 2 , c = 5 )
RBC_3 X i = f C T ( f R R ( x i , d = 140 ° ) , b = 2 , c = 5 )
RBC_4 X i = f C T ( f R R ( x i , d = 140 ° ) , b = 1.9 , c = 5 )
RBC_5 X i = f C T ( f R R ( x i , d = 140 ° ) , b = 2 , c = 4.5 )
RBCS X i = f R A ( f C T ( f R R ( x i , d = 150 ° ) , b = 2 , c = 5 ) , s = 40 ° )
Table 11. Performance of policies applied on Γ PAP and Γ CCZ . The best performing policy is highlighted in yellow.
Table 11. Performance of policies applied on Γ PAP and Γ CCZ . The best performing policy is highlighted in yellow.
AA_IPAA_CPRBC_1RBC_2RBC_3RBC_4RBC_5RBCS
Γ PAP train 50 +9.59%+9.43%+10.63%+10.45%+10.49%+10.38%+10.86%+10.57%
Γ PAP train 100 +5.48%+5.07%+6.32%+6.35%+6.26%+6.07%+5.86%+6.14%
Γ PAP train 200 +3.19%+2.89%+3.73%+3.66%+3.72%+3.80%+3.61%+3.66%
Γ PAP train +2.48%+2.66%+2.76%+2.84%+2.80%+2.91%+2.83%+2.88%
Γ CCZ train 50 +6.23%+5.06%+6.78%+6.77%+6.73%+7.31%+7.05%+7.10%
Γ CCZ train 100 +4.26%+3.81%+4.52%+4.08%+4.68%+4.21%+4.74%+4.58%
Γ CCZ train +2.41%+2.04%+2.39%+2.52%+2.50%+2.37%+2.58%+2.46%
Γ CSD train 50 +0.94%+0.92%−1.47%−1.73%−1.54%−0.81%−0.82%−2.67%
Γ CSD train 100 +2.57%+1.06%−2.61%−1.95%−1.99%−2.12%−1.80%−1.53%
Γ CSD train +1.17%+1.44%−1.02%−1.76%−1.31%−2.12%−1.08%−1.65%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tan, M.; Langenkämper, D.; Nattkemper, T.W. The Impact of Data Augmentations on Deep Learning-Based Marine Object Classification in Benthic Image Transects. Sensors 2022, 22, 5383. https://doi.org/10.3390/s22145383

AMA Style

Tan M, Langenkämper D, Nattkemper TW. The Impact of Data Augmentations on Deep Learning-Based Marine Object Classification in Benthic Image Transects. Sensors. 2022; 22(14):5383. https://doi.org/10.3390/s22145383

Chicago/Turabian Style

Tan, Mingkun, Daniel Langenkämper, and Tim W. Nattkemper. 2022. "The Impact of Data Augmentations on Deep Learning-Based Marine Object Classification in Benthic Image Transects" Sensors 22, no. 14: 5383. https://doi.org/10.3390/s22145383

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop