Next Article in Journal
An Automated Image-Based Multivariant Concrete Defect Recognition Using a Convolutional Neural Network with an Integrated Pooling Module
Next Article in Special Issue
Fuzzy Inference and Sequence Model-Based Collision Risk Prediction System for Stand-On Vessel
Previous Article in Journal
A Uniform Magnetic Field Generator Combined with a Thin-Film Magneto-Impedance Sensor Capable of Human Body Scans
Previous Article in Special Issue
Analysis of Marine-Pilot Biometric Data Recordings during Port-Approach Using a Full-Mission Simulator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Seabed Modelling by Means of Airborne Laser Bathymetry Data and Imbalanced Learning for Offshore Mapping

1
Department of Geodesy and Offshore Survey, Maritime University of Szczecin, Żołnierska 46, 71-250 Szczecin, Poland
2
Department of Computer Engineering, Koszalin University of Technology, Sniadeckich 2, 75-453 Koszalin, Poland
3
Department of Geodesy and Geoinformatics, Koszalin University of Technology, Sniadeckich 2, 75-453 Koszalin, Poland
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(9), 3121; https://doi.org/10.3390/s22093121
Submission received: 23 March 2022 / Revised: 15 April 2022 / Accepted: 18 April 2022 / Published: 19 April 2022

Abstract

:
An important problem associated with the aerial mapping of the seabed is the precise classification of point clouds characterizing the water surface, bottom, and bottom objects. This study aimed to improve the accuracy of classification by addressing the asymmetric amount of data representing these three groups. A total of 53 Synthetic Minority Oversampling Technique (SMOTE) algorithms were adjusted and evaluated to balance the amount of data. The prepared data set was used to train the Multi-Layer Perceptron (MLP) neural network used for classifying the point cloud. Data balancing contributed to significantly increasing the accuracy of classification. The best overall classification accuracy achieved varied from 95.8% to 97.0%, depending on the oversampling algorithm used, and was significantly better than the classification accuracy obtained for unbalanced data and data with downsampling (89.6% and 93.5%, respectively). Some of the algorithms allow for 10% increased detection of points on the objects compared to unbalanced data or data with simple downsampling. The results suggest that the use of selected oversampling algorithms can aid in improving the point cloud classification and making the airborne laser bathymetry technique more appropriate for seabed mapping.

1. Introduction

Information on water depth and seabed topography can contribute to improving the safety of maritime transport and to the development of other maritime industries, including the offshore sector. Hydrographic surveying is done systematically all over the world to prepare data for nautical charts, electronic navigation systems, and other databases used in the management of hydrospace and maritime infrastructure. The airborne laser bathymetry (ALB) technique can be a valuable addition to Multibeam Echosounders (MBES) or perhaps an alternative in shallow waters. It has proven to be a large-scale, accurate, rapid, safe, and versatile approach for surveying shallow waters and coastlines where sonar systems are ineffective or impossible to use [1,2,3,4]. Research has shown that ALB can identify similar seafloor features such as MBES systems [5]. However, additional improvements must be done to separate the LiDAR seafloor intensity data from the depth component of the signal waveform. Receiving bathymetric lidar data with unassigned point classes or inaccurate point classification that may not meet industrial or research requirements is not unusual [6]. Studies that have used ALB for depth determination and object detection primarily point to challenges in classifying the resulting point cloud into three basic groups: bottom, water surface, and bottom objects. These issues can be overcome using well-recognized machine learning classification methods.
The main goal of this study is to increase the accuracy of the classification of point clouds measured by an ALB scanner to improve seabed modeling and object detection.
This paper can be considered a novel input to the ALB classification of point clouds with the use of imbalanced learning. To achieve the goal, the study evaluated Multi-Layer Perceptron (MLP) Artificial Neural Network (ANN) with the softmax activation function employing over 50 variants of the oversampling techniques for imbalanced learning. The results confirmed that data balancing had a quantitative impact on classification accuracy, allowing enhanced detection of seabed and bottom based on the ALB data. The classification results indicated that the best overall classification accuracy achieved varied from 95.8% to 97.0% depending on the oversampling algorithm used and was significantly better than the classification accuracy obtained for unbalanced data and data with downsampling (89.6% and 93.5%, respectively). Some of the algorithms allow for 10% increased detection of points on the objects compared to unbalanced data or data with simple downsampling. This study did not develop a new data balancing method or enhance the existing ones.
The classification accuracy of point clouds of all the classes is influenced by class distribution. According to the scanned area in the majority, the laser scanning data are unbalanced, and therefore require remodeling. The ALB data set of shallow waters comprises data on the seabed and a small percentage of data on underrepresented seabed objects. This application necessitates a high rate of accurate detection in the minority class (seabed objects) and a low rate of mistakes in the majority class (seabed or water surface). Different oversampling methods have been analyzed to address this concern [7]. Archaeologists focusing on detecting former field systems from LiDAR data in their research recommend the use of the Synthetic Minority Oversampling Technique (SMOTE) for achieving better results [8]. Balancing the training data for automatic mapping of high-voltage power lines based on the LiDAR data led to an almost 10% increase in accuracy in comparison to imbalanced data [9]. Landslide prediction research based on a set of geomorphological factors revealed that the Support Vector Machine (SVM) model yielded the highest accuracy with the SMOTE data balancing method [10]. The supporting synthetic samples were used in the classification of bottom materials (sand, stones, rocks) performed using ALB. The obtained results were promising but were specific for particular classes [11]. A study on the application of SMOTE for balancing data distribution with land cover mapping using LiDAR data showed increased detection accuracy. The challenges associated with imbalanced classes and low density of LiDAR point clouds in urban areas were also satisfactorily resolved by applying several oversampling methods for the classification and extraction of roof superstructures [12]. Due to its proven advantages in classification, the present study used SMOTE, a method for producing synthetic new data from existing ones, which provided new information and variations to synthetically generated data.
The paper is organized as follows: Section 2 describes the test area and ALB data with features and architecture of ANN. Section 3 presents the results obtained with the proposed approach and a discussion. Finally, Section 4 presents our conclusions.

2. Materials and Methods

2.1. Test Area

The test area is the artificial reef Rosenort on the Baltic Sea. It is located between Markgrafenheide and Graal-Müritz (Germany), approximately 2000 m from the coast, at a water depth of 6 m. The reef is a protected fishery reserve, and thus activities such as angling, fishing, and anchoring are prohibited. The Rosenort reef is divided into four artificially constructed zones. The zones were built from (1) 52-ton concrete tetrapods, (2) 180-ton natural stones, (3) 30 cut reef cones, and (4) six 6-ton concrete tetrapods (Figure 1).

2.2. Point Cloud and Features

The point cloud was collected in September 2013 using an AHAB Chiroptera I scanner, at a flight altitude of 400 m. The Chiroptera I scanner is equipped with two beams and scans in an elliptical shape at an angle of 20° between the scan direction and the nadir. This laser scanner uses a near-infrared (NIR) laser with a wavelength of 1064 nm at a peak measurement frequency of 400 kHz for detecting water surfaces and a green laser with a wavelength of 532 nm at a frequency of 36 kHz for underwater measurements. The horizontal nominal accuracy of the infrared beam and the green beam is 0.2 and 0.75 m, respectively, while the depth of nominal accuracy is 0.15 m. The Secchi depth achieved with the scanner exceeded 1.5 m, and during the measurement, the depth was measured at around 6.3 m. In this study, the point cloud obtained from the green beam (Figure 2) was used for analysis. The density of the point cloud obtained for the test area was 2.6 points on the water surface and 3.3 points on the underwater point (seabed and seabed object).
Scanning with the use of the AHAB Chiroptera I scanner allowed collecting the cloud of spatially coordinated points with their intensities. The analysis of such data, especially the full waveform, can provide additional information on the measured points that can aid in the classification of the acquired data. Five features (U1–U5, Table 1) derived from the full waveform, were used. A well-defined region delineated by a cylinder of a given radius r, which was 5 m (Figure 3), was used to analyze the location of each point along with its neighborhood. Features U6–U15, which describe the geometry of the point cloud, were used in the investigation.

2.3. Architecture of ANN

The raw (unbalanced) ALB data set used for training the ANN consisted of 6198 vectors (Figure 3, data in the black box). Each vector had 18 items describing the values of 15 input attributes (U1–U15, Table 1) and that of three output classes (U38–U40, Table 2). For the error back-propagation method, 80% of these vectors were utilized for training and 20% for validating the ANN.
The three classes were labeled as follows: class 1 (U38)-water surface with 2729 vectors, class 2 (U39)-seabed represented by 3396 vectors, and class 3 (U40)-seabed object containing 73 vectors. Since the classes had a different number of vectors, for training the ANN, the number of vectors in each class was balanced by applying different oversampling algorithms (Table 3, first column). The data set thus prepared, consisted of a different number of vectors (Table 3, last three columns), depending on the algorithm used. Imbalancing of data typically refers to classification tasks where the classes are not equally represented. Several approaches have been proposed for this issue. Among them, SMOTE has been widely used to produce synthetic samples between minority samples in the feature space. This technique improves class imbalance by linear interpolation between the underrepresented class samples [7]. It creates new instances of minority group data, by copying existing data and making minor changes. Moreover, SMOTE is a great tool for amplifying the already existing signal in minority groups without creating new signals for these groups. In general, synthetic samples are generated as a difference between the feature vector (sample) and its randomly chosen nearest neighbor. This difference is multiplied by a random number between 0 and 1 and added to the feature vector considered for creating a new sample —the synthetic one. Several improvements have been proposed for synthetic sample creation algorithms since the introduction of SMOTE. The present work included 53 oversampling methods, and a comparison of their results is provided in this paper. The data were standardized in a later step of data processing.
The ANN used in the experiments is presented in Figure 4. It is an MLP neural network [15], which has 15 inputs (U1–U15), three layers of neurons, and three outputs (U38–U40). The first layer comprises 15 neurons (U16–U30), the second layer has seven neurons (U31–U37), and the third layer has three neurons (U38–U40). Neurons in the previous layer are fully connected with those in the next layer (Figure 4). In the first layer, as well as the second layer from the bottom, all neurons possess a unipolar sigmoidal activation function, while in the last layer, all neurons (U38–U40) possess the soft-max activation function.
The values of the neural network outputs (U38–U40) inform the probability value, which indicates the degree of belonging of a given input vector to each of the three classes (water, seabed, seabed object). The ANN presented in Figure 4 was trained using an error-back propagation algorithm with the learning coefficient ro = 0.01. The maximal number of iterations was 1750.

3. Results and Discussion

The proposed approach was tested for each oversampling method by training the MLP neural network. A random starting point was used in error back-propagation. Consequently, the training procedure was repeated 11 times to obtain reliable results. After completion of each iteration, the data were tested with the dataset, which initially contained 10,612 water surface points, 13,318 seabed points, and 212 seabed object points. The results of the tests are presented in Table 3. The first two columns in the table present the names of oversampling algorithms and the year they were introduced. The next four columns show the best, worst, mean and median values of overall classification accuracy. The last four columns present the number of vectors used for training the MLP neural network.
The overall classification accuracy (Ac [%]) was calculated using the following formula:
A c = ( c o r w a l l w + c o r s a l l s + c o r o a l l o ) × 100 % 3
where corw is the number of input vectors successfully identified as “water surface” in class 1, cors is the number of input vectors successfully identified as “seabed” in class 2, coro is the number of input vectors successfully identified as “seabed object” in class 3, and all{w,s,o} is the total number of vectors in classes 1–3.
The best overall classification accuracy of 97.0% was achieved for the LVQ SMOTE (Learning Vector Quantization based SMOTE) algorithm. The oversampling method generated synthetic samples using codebooks obtained by learning vector quantization [16]. The second algorithm with about 96% overall classification accuracy was ROSE (Random OverSampling Examples). This algorithm works based on smoothed bootstrap resampling from data [17]. The next algorithm with the best results was PDFOS (Probability Density Function Over-Sampling), and its overall classification accuracy was about 95.8%. This algorithm generated synthetic instances as additional training data based on the estimated probability density function [18].
Table 3. Results of classification with balanced learning.
Table 3. Results of classification with balanced learning.
NameYearBestWorstMeanMedianAll VectorsClass 1Class 2Class 3
1SMOTE [7]200293.491.592.792.910,188339633963396
2SMOTE + Tomek [19]200493.091.992.692.610,135339633963343
3SMOTE + ENN [19]200493.590.692.292.19990339633963198
4Borderline-SMOTE1 [20]200593.391.592.292.19191272933963066
5Borderline-SMOTE2 [20]200595.493.494.794.89191272933963066
6SMOTE + LLE [21]200691.188.289.789.710,188339633963396
7Distance-SMOTE [22]200793.591.992.592.510,188339633963396
8Polynomial-SMOTE [23]200891.088.790.390.413,234545833964380
9ADOMS [24]200894.291.493.393.510,188339633963396
10Safe Level SMOTE [25]200966.766.766.766.7657327293396448
11MSMOTE [26]200994.192.092.992.910,188339633963396
12SMOBD [27]201195.092.793.393.010,188339633963396
13SVM balance [28]201294.291.992.792.510,172339633963380
14TRIM SMOTE [29]201292.491.592.092.010,188339633963396
15SMOTE RSB [30]201281.766.771.467.6771633963396924
16ProWSyn [31]201393.690.692.492.510,188339633963396
17SL graph SMOTE [32]201392.191.191.691.69191272933963066
18NRSBoundary SMOTE [33]201392.691.491.891.89191272933963066
19LVQ SMOTE [16]201397.094.796.396.710,188339633963396
20ROSE [17]201496.092.594.695.010,188339633963396
21SMOTE OUT [34]201493.591.292.292.110,188339633963396
22SMOTE Cosine [34]201493.289.691.290.910,188339633963396
23Selected SMOTE [34]201494.992.793.693.610,188339633963396
24LN SMOTE [35]201194.466.786.393.59282339633962490
25MWMOTE [36]201491.590.491.091.010,188339633963396
26PDFOS [18]201495.892.994.694.710,188339633963396
27RWO sampling [37]201493.088.691.091.510,188339633963396
28NEATER [38]201488.075.884.886.58728339633961936
29DEAGO [39]201585.885.885.885.810,188339633963396
30MCT [40]201595.493.594.594.510,188339633963396
31SMOTE IPF [41]201594.192.593.293.410,188339633963396
32OUPS [42]201693.191.492.092.09493339633962701
33SMOTE D [43]201681.478.780.180.110,189339833963395
34CE SMOTE [44]201094.866.786.290.18647272933962522
35Edge Det SMOTE [45]201093.892.693.293.510,188339633963396
36ASMOBD [46] 201288.286.887.487.310,188339633963396
37Assembled SMOTE [47]201393.090.991.691.59191272933963066
38SDSMOTE [48]201494.492.093.493.510,188339633963396
39G SMOTE [49]201494.492.593.293.210,188339633963396
40NT SMOTE [50]201493.792.893.193.110,188339633963396
41Lee [51]201593.892.993.393.310,188339633963396
42MDO [52]201692.190.391.391.410,188339633963396
43Random SMOTE [53]201194.492.593.393.210,188339633963396
44VIS RST [54]201666.766.666.766.7711933963396327
45AND SMOTE [55]201692.090.491.191.010,188339633963396
46NRAS [56]201790.288.589.189.010,188339633963396
47NDO sampling [57]201195.193.694.594.610,189339733963396
48Gaussian SMOTE [58]201792.290.391.191.010,188339633963396
49Kmeans SMOTE [59]201892.190.891.591.610,188339633963396
50Supervised SMOTE [60]201492.891.592.192.110,188339633963396
51SN SMOTE [61]201295.292.393.793.710,188339633963396
52CCR [62]201788.987.188.088.29191272933963066
53ANS [63]201791.388.790.090.19191272933963066
The correctly classified points constituting the seabed object were presented in the 10 confusion matrices formed for:
  • unbalanced data and data with downsampling (Table 4) for comparison [64],
  • four matrices for algorithms with the highest overall classification accuracy (Table 5), and
  • four matrices for algorithms with the highest median overall classification accuracy in 11 repetitions (Table 6).
The overall classification accuracy achieved for unbalanced data was 89.6% and for downsampling data was 93.5% [64]. The downsampling method was used, in which each class was given the same number of vectors, similar to in class 3. The data set, divided into three equal classes, contained a total of 219 input vectors (3 × 79). Downsampling contributed to increasing the overall classification accuracy by 3.9%. The correct classification of points in class 3 also increased by 11.3%.
Table 5 and Table 6 present the confusion matrix for the four algorithms with the best object detection results and four algorithms with the best median values. The correct classification of points in class 3 (seabed object) ranged between 89.7% and 93.4% for the best results of imbalanced learning and between 84.9% and 92.5% for median results. In all cases, an increase in the overall classification accuracy and point detection on the seabed objects was achieved. The water surface was classified with an accuracy of 100% in all algorithms. Two algorithms—Safe Level SMOTE and VIS RST—were found to be ineffective and as a result, none of the points on the objects were detected.
The accuracy of oversampling algorithms was assessed using three accuracy evaluation indices: precision, recall, and F1-score.
Precision refers to the proportion of correctly predicted points on the object to all points on the object, i.e.,
Precision = T P T P + F P  
Recall: refers to the proportion of the correctly predicted points on the object to all points on the object, i.e.,
Recall = T P T P + F N  
F1-score refers to the harmonic mean of precision and recall, i.e.,
F 1 score = 2 T P 2 T P + F P + F N  
where TP, TN, FP, and FN denote true positive, true negative, false positive, and false negative, respectively.
The indices were computed for the median of results. Recall was found to be high for all four algorithms: 0.92 for LVQ SMOTE, 0.86 for ROSE, 0.85 for borderline-SMOTE2, and 0.84 for PDFOS. The F1-score for class 3 was calculated to be 0.54, 0.66, 0.68, and 0.72, respectively. Among the oversampling algorithms, MDO had the best F1-score of 0.75, which was comparable with that of PDFOS. The overall accuracy of the median results of MDO was 91.4, and the confusion matrix of the median results is presented in Table 7.

4. Conclusions

ALB technique follows existing water reservoir measurement patterns. Monitoring the seabed and detection of seabed objects in the coastal zone around ports with heavy vessel traffic help in decreasing the risk of maritime grounding and collision with underwater obstacles, thereby reducing the probability of environmental incidents that can occur due to cargo and fuel leakage or even unexploded ordnance explosion.
This study used a total of 53 oversampling algorithms with imbalanced MLP neural learning for the classification of the ALB data and detection of seabed objects. The results revealed that selected oversampling algorithms classified point clouds better than unbalanced data or data with simple downsampling. The algorithms that produced the best results can be divided into two groups: (1) the algorithms with good recall, which improves the detection of points on objects—LVQ SMOTE and ROSE; and (2) those that improve the general classification with the highest F1-score—MDO and PDFOS. Identifying the oversampling method that gives the best results for object classification and detection is challenging. This is because a good recall is often associated with false classification of points.
As the present study did not cover all the issues related to the subject, future work should focus on using SMOTE methods for improving the detection of underwater objects. Additionally, the possibility of applying SMOTE in deep-sea bottom imaging using MBES would be a topic of interest.

Author Contributions

Conceptualization, T.K., A.S. and A.T.; methodology, T.K. and A.S.; software, T.K. and A.S.; validation T.K. and A.T.; formal analysis, T.K. and A.T.; investigation, T.K. and A.S.; data curation, T.K.; writing—original draft preparation, T.K. and A.T.; writing—review and editing, T.K., A.S., A.T. and T.O.; visualization, T.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

For the data used in this paper, the authors would like to thank the Institute of Photogrammetry and GeoInformation in Hannover.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Muirhead, K.; Cracknell, A.P. Airborne Lidar Bathymetry. Int. J. Remote Sens. 1986, 7, 597–614. [Google Scholar] [CrossRef]
  2. Wang, C.-K.; Philpot, W.D. Using Airborne Bathymetric Lidar to Detect Bottom Type Variation in Shallow Waters. Remote Sens. Environ. 2007, 106, 123–135. [Google Scholar] [CrossRef]
  3. Yeu, Y.; Yee, J.-J.; Yun, H.S.; Kim, K.B. Evaluation of the Accuracy of Bathymetry on the Nearshore Coastlines of Western Korea from Satellite Altimetry, Multi-Beam, and Airborne Bathymetric LiDAR. Sensors 2018, 18, 2926. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Stępień, G.; Tomczak, A.; Loosaar, M.; Ziębka, T. Dimensioning Method of Floating Offshore Objects by Means of Quasi-Similarity Transformation with Reduced Tolerance Errors. Sensors 2020, 20, 6497. [Google Scholar] [CrossRef]
  5. Costa, B.M.; Battista, T.A.; Pittman, S.J. Comparative Evaluation of Airborne LiDAR and Ship-Based Multibeam SoNAR Bathymetry and Intensity for Mapping Coral Reef Ecosystems. Remote Sens. Environ. 2009, 113, 1082–1100. [Google Scholar] [CrossRef]
  6. Jung, J.; Lee, J.; Parrish, C.E. Inverse Histogram-Based Clustering Approach to Seafloor Segmentation from Bathymetric Lidar Data. Remote Sens. 2021, 13, 3665. [Google Scholar] [CrossRef]
  7. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-Sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  8. Herrault, P.-A.; Poterek, Q.; Keller, B.; Schwartz, D.; Ertlen, D. Automated Detection of Former Field Systems from Airborne Laser Scanning Data: A New Approach for Historical Ecology. Int. J. Appl. Earth Obs. Geoinf. 2021, 104, 102563. [Google Scholar] [CrossRef]
  9. Chasco-Hernández, D.; Sanz-Delgado, J.A.; García-Morales, V.; Álvarez-Mozos, J. Automatic Detection of High-Voltage Power Lines in LiDAR Surveys Using Data Mining Techniques. In Advances in Design Engineering; Lecture Notes in Mechanical Engineering; Cavas-Martínez, F., Sanz-Adan, F., Morer Camo, P., Lostado Lorza, R., Santamaría Peña, J., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 568–575. ISBN 978-3-030-41199-2. [Google Scholar]
  10. Al-Najjar, H.A.H.; Pradhan, B.; Sarkar, R.; Beydoun, G.; Alamri, A. A New Integrated Approach for Landslide Data Balancing and Spatial Prediction Based on Generative Adversarial Networks (GAN). Remote Sens. 2021, 13, 4011. [Google Scholar] [CrossRef]
  11. Eren, F.; Pe’eri, S.; Rzhanov, Y.; Ward, L. Bottom Characterization by Using Airborne Lidar Bathymetry (ALB) Waveform Features Obtained from Bottom Return Residual Analysis. Remote Sens. Environ. 2018, 206, 260–274. [Google Scholar] [CrossRef]
  12. Aissou, B.E.; Aissa, A.B.; Dairi, A.; Harrou, F.; Wichmann, A.; Kada, M. Building Roof Superstructures Classification from Imbalanced and Low Density Airborne LiDAR Point Cloud. IEEE Sens. J. 2021, 21, 14960–14976. [Google Scholar] [CrossRef]
  13. Wagner, W.; Ullrich, A.; Ducic, V.; Melzer, T.; Studnicka, N. Gaussian Decomposition and Calibration of a Novel Small-Footprint Full-Waveform Digitising Airborne Laser Scanner. ISPRS J. Photogramm. Remote Sens. 2006, 60, 100–112. [Google Scholar] [CrossRef]
  14. Niemeyer, J.; Rottensteiner, F.; Soergel, U. Contextual Classification of Lidar Data and Building Object Detection in Urban Areas. ISPRS J. Photogramm. Remote Sens. 2014, 87, 152–165. [Google Scholar] [CrossRef]
  15. Shibata, K.; Ikeda, Y. Effect of Number of Hidden Neurons on Learning in Large-Scale Layered Neural Networks. In Proceedings of the 2009 ICCAS-SICE, Fukuoka, Japan, 18–21 August 2009; pp. 5008–5013. [Google Scholar]
  16. Nakamura, M.; Kajiwara, Y.; Otsuka, A.; Kimura, H. LVQ-SMOTE—Learning Vector Quantization Based Synthetic Minority Over–Sampling Technique for Biomedical Data. BioData Min. 2013, 6, 16. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Menardi, G.; Torelli, N. Training and Assessing Classification Rules with Imbalanced Data. Data Min. Knowl. Disc. 2014, 28, 92–122. [Google Scholar] [CrossRef]
  18. Gao, M.; Hong, X.; Chen, S.; Harris, C.J.; Khalaf, E. PDFOS: PDF Estimation Based over-Sampling for Imbalanced Two-Class Problems. Neurocomputing 2014, 138, 248–259. [Google Scholar] [CrossRef]
  19. Batista, G.E.A.P.A.; Prati, R.C.; Monard, M.C. A Study of the Behavior of Several Methods for Balancing Machine Learning Training Data. SIGKDD Explor. Newsl. 2004, 6, 20–29. [Google Scholar] [CrossRef]
  20. Han, H.; Wang, W.-Y.; Mao, B.-H. Borderline-SMOTE: A New Over-Sampling Method in Imbalanced Data Sets Learning. In Advances in Intelligent Computing; Huang, D.-S., Zhang, X.-P., Huang, G.-B., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; pp. 878–887. [Google Scholar]
  21. Wang, J.; Xu, M.; Wang, H.; Zhang, J. Classification of Imbalanced Data by Using the SMOTE Algorithm and Locally Linear Embedding. In Proceedings of the 2006 8th International Conference on Signal Processing, Beijing, China, 16–20 November 2006; Volume 3. [Google Scholar]
  22. Calleja, J.L.; Fuentes, O. A Distance-Based Over-Sampling Method for Learning from Imbalanced Data Sets. In Proceedings of the FLAIRS Conference, Florida, FL, USA, 7–9 May 2007. [Google Scholar]
  23. Gazzah, S.; Amara, N.E.B. New Oversampling Approaches Based on Polynomial Fitting for Imbalanced Data Sets. In Proceedings of the 2008 The Eighth IAPR International Workshop on Document Analysis Systems, Nara, Japan, 16–19 September 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 677–684. [Google Scholar]
  24. Tang, S.; Chen, S. The Generation Mechanism of Synthetic Minority Class Examples. In Proceedings of the 2008 International Conference on Information Technology and Applications in Biomedicine, Shenzhen, China, 30–31 May 2008; pp. 444–447. [Google Scholar]
  25. Bunkhumpornpat, C.; Sinapiromsaran, K.; Lursinsap, C. Safe-Level-SMOTE: Safe-Level-Synthetic Minority Over-Sampling TEchnique for Handling the Class Imbalanced Problem. In Advances in Knowledge Discovery and Data Mining; Theeramunkong, T., Kijsirikul, B., Cercone, N., Ho, T.-B., Eds.; Springer: Berlin/Heidelberg, Germany, 2009; pp. 475–482. [Google Scholar]
  26. Hu, S.; Liang, Y.; Ma, L.; He, Y. MSMOTE: Improving Classification Performance When Training Data Is Imbalanced. In Proceedings of the 2009 Second International Workshop on Computer Science and Engineering, Qingdao, China, 28 October 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 13–17. [Google Scholar]
  27. Cao, Q.; Wang, S. Applying Over-Sampling Technique Based on Data Density and Cost-Sensitive SVM to Imbalanced Learning. In Proceedings of the 2011 International Conference on Information Management, Innovation Management and Industrial Engineering, Shenzhen, China, 26–27 November 2011; Volume 2, pp. 543–548. [Google Scholar]
  28. Farquad, M.A.H.; Bose, I. Preprocessing Unbalanced Data Using Support Vector Machine. Decis. Support Syst. 2012, 53, 226–233. [Google Scholar] [CrossRef]
  29. Puntumapon, K.; Waiyamai, K. A Pruning-Based Approach for Searching Precise and Generalized Region for Synthetic Minority Over-Sampling. In Advances in Knowledge Discovery and Data Mining; Tan, P.-N., Chawla, S., Ho, C.K., Bailey, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 371–382. [Google Scholar]
  30. Ramentol, E.; Caballero, Y.; Bello, R.; Herrera, F. SMOTE-RSB*: A Hybrid Preprocessing Approach Based on Oversampling and Undersampling for High Imbalanced Data-Sets Using SMOTE and Rough Sets Theory. Knowl. Inf. Syst. 2012, 33, 245–265. [Google Scholar] [CrossRef]
  31. Barua, S.; Islam, M.M.; Murase, K. ProWSyn: Proximity Weighted Synthetic Oversampling Technique for Imbalanced Data Set Learning. In Advances in Knowledge Discovery and Data Mining; Pei, J., Tseng, V.S., Cao, L., Motoda, H., Xu, G., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 317–328. [Google Scholar]
  32. Bunkhumpornpat, C.; Subpaiboonkit, S. Safe Level Graph for Synthetic Minority Over-Sampling Techniques. In Proceedings of the 2013 13th International Symposium on Communications and Information Technologies (ISCIT), Surat Thani, Thailand, 4–6 September 2013; pp. 570–575. [Google Scholar]
  33. Hu, F.; Li, H. A Novel Boundary Oversampling Algorithm Based on Neighborhood Rough Set Model: NRSBoundary-SMOTE. Math. Probl. Eng. 2013, 2013, 694809. [Google Scholar] [CrossRef]
  34. Koto, F. SMOTE-Out, SMOTE-Cosine, and Selected-SMOTE: An Enhancement Strategy to Handle Imbalance in Data Level. In Proceedings of the 2014 International Conference on Advanced Computer Science and Information System, Jakarta, Indonesia, 18–19 October 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 280–284. [Google Scholar]
  35. Maciejewski, T.; Stefanowski, J. Local Neighbourhood Extension of SMOTE for Mining Imbalanced Data. In Proceedings of the 2011 IEEE Symposium on Computational Intelligence and Data Mining (CIDM), Paris, France, 11–15 April 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 104–111. [Google Scholar]
  36. Barua, S.; Islam, M.M.; Yao, X.; Murase, K. MWMOTE–Majority Weighted Minority Oversampling Technique for Imbalanced Data Set Learning. IEEE Trans. Knowl. Data Eng. 2014, 26, 405–425. [Google Scholar] [CrossRef]
  37. Zhang, H.; Li, M. RWO-Sampling: A Random Walk over-Sampling Approach to Imbalanced Data Classification. Inf. Fusion 2014, 20, 99–116. [Google Scholar] [CrossRef]
  38. Almogahed, B.A.; Kakadiaris, I.A. NEATER: Filtering of over-Sampled Data Using Non-Cooperative Game Theory. Soft Comput. 2015, 19, 3301–3322. [Google Scholar] [CrossRef]
  39. Bellinger, C.; Japkowicz, N.; Drummond, C. Synthetic Oversampling for Advanced Radioactive Threat Detection. In Proceedings of the 2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 9–11 December 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 948–953. [Google Scholar]
  40. Jiang, L.; Qiu, C.; Li, C. A Novel Minority Cloning Technique for Cost-Sensitive Learning. Int. J. Patt. Recogn. Artif. Intell. 2015, 29, 1551004. [Google Scholar] [CrossRef]
  41. Sáez, J.A.; Luengo, J.; Stefanowski, J.; Herrera, F. SMOTE–IPF: Addressing the Noisy and Borderline Examples Problem in Imbalanced Classification by a Re-Sampling Method with Filtering. Inf. Sci. 2015, 291, 184–203. [Google Scholar] [CrossRef]
  42. Rivera, W.A.; Xanthopoulos, P. A Priori Synthetic Over-Sampling Methods for Increasing Classification Sensitivity in Imbalanced Data Sets. Expert Syst. Appl. 2016, 66, 124–135. [Google Scholar] [CrossRef]
  43. Torres, F.R.; Carrasco-Ochoa, J.A.; Martínez-Trinidad, J.F. SMOTE-D a Deterministic Version of SMOTE. In Pattern Recognition; Martínez-Trinidad, J.F., Carrasco-Ochoa, J.A., Ayala Ramirez, V., Olvera-López, J.A., Jiang, X., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 177–188. [Google Scholar]
  44. Chen, S.; Guo, G.; Chen, L. A New Over-Sampling Method Based on Cluster Ensembles. In Proceedings of the 2010 IEEE 24th International Conference on Advanced Information Networking and Applications Workshops, Perth, Australia, 20–23 April 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 599–604. [Google Scholar]
  45. Kang, Y.-I.; Won, S. Weight Decision Algorithm for Oversampling Technique on Class-Imbalanced Learning. In Proceedings of the ICCAS 2010, Gyeonggi-do, Korea, 27–30 October 2010; pp. 182–186. [Google Scholar]
  46. Wang, S.; Li, Z.; Chao, W.; Cao, Q. Applying Adaptive Over-Sampling Technique Based on Data Density and Cost-Sensitive SVM to Imbalanced Learning. In Proceedings of the The 2012 International Joint Conference on Neural Networks (IJCNN), Brisbane, Australia, 10–15 June 2012; pp. 1–8. [Google Scholar]
  47. Zhou, B.; Yang, C.; Guo, H.; Hu, J. A Quasi-Linear SVM Combined with Assembled SMOTE for Imbalanced Data Classification. In Proceedings of the The 2013 International Joint Conference on Neural Networks (IJCNN), Dallas, TX, USA, 4–9 August 2013; pp. 1–7. [Google Scholar]
  48. Li, K.; Zhang, W.; Lu, Q.; Fang, X. An Improved SMOTE Imbalanced Data Classification Method Based on Support Degree. In Proceedings of the 2014 International Conference on Identification, Information and Knowledge in the Internet of Things, Beijing, China, 17–18 October 2014; pp. 34–38. [Google Scholar]
  49. Sandhan, T.; Choi, J.Y. Handling Imbalanced Datasets by Partially Guided Hybrid Sampling for Pattern Recognition. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; pp. 1449–1453. [Google Scholar]
  50. Xu, Y.H.; Li, H.; Le, L.P.; Tian, X.Y. Neighborhood Triangular Synthetic Minority Over-Sampling Technique for Imbalanced Prediction on Small Samples of Chinese Tourism and Hospitality Firms. In Proceedings of the 2014 Seventh International Joint Conference on Computational Sciences and Optimization, Washington, DC, USA, 4–6 July 2014; pp. 534–538. [Google Scholar]
  51. Lee, J.; Kim, N.; Lee, J.-H. An Over-Sampling Technique with Rejection for Imbalanced Class Learning. In Proceedings of the 9th International Conference on Ubiquitous Information Management and Communication, Bali, Indonesia, 8–10 January 2015; ACM: New York, NY, USA, 2015; pp. 1–6. [Google Scholar]
  52. Abdi, L.; Hashemi, S. To Combat Multi-Class Imbalanced Problems by Means of Over-Sampling Techniques. IEEE Trans. Knowl. Data Eng. 2016, 28, 238–251. [Google Scholar] [CrossRef]
  53. Dong, Y.; Wang, X. A New Over-Sampling Approach: Random-SMOTE for Learning from Imbalanced Data Sets. In Knowledge Science, Engineering and Management; Xiong, H., Lee, W.B., Eds.; Springer: Berlin/Heidelberg, Germany, 2011; pp. 343–352. [Google Scholar]
  54. Borowska, K.; Stepaniuk, J. Imbalanced Data Classification: A Novel Re-Sampling Approach Combining Versatile Improved SMOTE and Rough Sets. In Computer Information Systems and Industrial Management; Saeed, K., Homenda, W., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 31–42. [Google Scholar]
  55. Yun, J.; Ha, J.; Lee, J.-S. Automatic Determination of Neighborhood Size in SMOTE. In Proceedings of the 10th International Conference on Ubiquitous Information Management and Communication, Danang, Viet Nam, 4–6 January 2016; ACM: New York, NY, USA, 2015; pp. 1–8. [Google Scholar]
  56. Rivera, W.A. Noise Reduction A Priori Synthetic Over-Sampling for Class Imbalanced Data Sets. Inf. Sci. 2017, 408, 146–161. [Google Scholar] [CrossRef]
  57. Zhang, L.; Wang, W. A Re-Sampling Method for Class Imbalance Learning with Credit Data. In Proceedings of the 2011 International Conference of Information Technology, Computer Engineering and Management Sciences, Nanjing, China, 24–25 September 2011; Volume 1, pp. 393–397. [Google Scholar]
  58. Lee, H.; Kim, J.; Kim, S. Gaussian-Based SMOTE Algorithm for Solving Skewed Class Distributions. Int. J. Fuzzy Log. Intell. Syst. 2017, 17, 229–234. [Google Scholar] [CrossRef]
  59. Douzas, G.; Bacao, F.; Last, F. Improving Imbalanced Learning through a Heuristic Oversampling Method Based on K-Means and SMOTE. Inf. Sci. 2018, 465, 1–20. [Google Scholar] [CrossRef] [Green Version]
  60. Hu, J.; He, X.; Yu, D.-J.; Yang, X.-B.; Yang, J.-Y.; Shen, H.-B. A New Supervised Over-Sampling Algorithm with Application to Protein-Nucleotide Binding Residue Prediction. PLoS ONE 2014, 9, e107676. [Google Scholar] [CrossRef] [PubMed]
  61. García, V.; Sánchez, J.S.; Martín-Félez, R.; Mollineda, R.A. Surrounding Neighborhood-Based SMOTE for Learning from Imbalanced Data Sets. Prog Artif. Intell. 2012, 1, 347–362. [Google Scholar] [CrossRef] [Green Version]
  62. Koziarski, M.; Wozniak, M. CCR: A Combined Cleaning and Resampling Algorithm for Imbalanced Data Classification. Int. J. Appl. Math. Comput. Sci. 2017, 27, 727–736. [Google Scholar] [CrossRef] [Green Version]
  63. Siriseriwan, W.; Sinapiromsaran, K. Adaptive Neighbor Synthetic Minority Oversampling Technique under 1NN Outcast Handling. Songklanakarin J. Sci. Technol. 2017, 39, 565–576. [Google Scholar] [CrossRef]
  64. Kogut, T.; Slowik, A. Classification of Airborne Laser Bathymetry Data Using Artificial Neural Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1959–1966. [Google Scholar] [CrossRef]
Figure 1. Location of the test area (approximately 25 km north of the city of Rostock in Germany).
Figure 1. Location of the test area (approximately 25 km north of the city of Rostock in Germany).
Sensors 22 03121 g001
Figure 2. Three classes in the ALB point cloud (blue—water surface, class 1; green—seabed, class 2; red—points on the object on the seabed, class 3).
Figure 2. Three classes in the ALB point cloud (blue—water surface, class 1; green—seabed, class 2; red—points on the object on the seabed, class 3).
Sensors 22 03121 g002
Figure 3. Visualization of the cylinder and analyzed points (red—analyzed point, green—points inside the cylinder used to compute the features, grey—other points in the point cloud, r—radius).
Figure 3. Visualization of the cylinder and analyzed points (red—analyzed point, green—points inside the cylinder used to compute the features, grey—other points in the point cloud, r—radius).
Sensors 22 03121 g003
Figure 4. The architecture of ANN.
Figure 4. The architecture of ANN.
Sensors 22 03121 g004
Table 1. Description of features used to train the ANN.
Table 1. Description of features used to train the ANN.
UiDescriptionFormula
U1Amplitude—the maximal peak of the Gaussian curve and is closely associated with the reflectance intensity [13]
U2Echo width—(ω, full width at half maximum)—the width of a Gaussian curve measured between those points on the y-axis which are half the maximal peak, and in the Gaussian function, it is related to standard deviation σ w = 2 2   l n ( 2 )   s (1)
U3Return number (N)
U4Number of returns (Nt)
U5Normalized echo N z = N N t (2)
U6Height difference (dz)— the vertical distance between the examined point zi and the lowest zmin in the cylinder d z = z i z m i n (3)
U7Height variance ( σ 2 ) —a measure of dispersion and is defined as the arithmetic mean of the squares of deviations of individual values z i in the cylinder from the mean value z ¯ σ 2 = 1 n i = 1 n ( z i z ¯ ) 2 (4)
U8Eigenvalue λ1
U9Eigenvalue λ2
U10Eigenvalue λ3
U11Sphericity—a property that describes the convexity or concavity of the analyzed point relative to points inside the cylinder S λ = λ 3 λ 1 (5)
U12Planarity—a characteristic that represents the planar aspect of a point arrangement P λ = λ 2 λ 3 λ 1 (6)
U13Linearity—a characteristic indicating that the distribution of points is linear (continuous). L λ = λ 1 λ 2 λ 1 (7)
U14Eigentropy—defined as entropy computed from eigenvalues E λ = i = 1 3 λ i l n λ i (8)
U15Omnivariance—a property whose low values are associated with flat terrain or linear structures, while high values are associated with point spatial dispersion [14] O λ = i = 1 3 λ i 3 (9)
Table 2. Description of outputs from the ANN.
Table 2. Description of outputs from the ANN.
UiDescription
U38Class 1: “water surface”
U39Class 2: “seabed”
U40Class 3: “seabed object”
Table 4. Confusion matrix of unbalanced data and data with downsampling.
Table 4. Confusion matrix of unbalanced data and data with downsampling.
ClassWater SurfaceSeabedSeabed Object
(Points)(%)(Points)(%)(Points)(%)
Unbalance
Water surface10,6121000000
Seabed0013,05798.02612.0
Seabed object006229.215070.8
Downsampling [64]
Water surface10,6121000000
Seabed0013,11998.51991.5
Seabed object003817.917482.1
Table 5. Confusion matrix of the four algorithms with best object detection.
Table 5. Confusion matrix of the four algorithms with best object detection.
ClassWater SurfaceSeabedSeabed Object
(Points)(%)(Points)(%)(Points)(%)
LVQ SMOTE
Water surface10,6121000000
Seabed0012,98697.53322.5
Seabed object00146.619893.4
ROSE
Water surface10,6121000000
Seabed0013,14998.71691.3
Seabed object002310.818989.2
PDFOS
Water surface10,6121000000
Seabed10.013,14398.71741.3
Seabed object002411.318888.7
Borderline-SMOTE2
Water surface10,6121000000
Seabed60.0513,10498.42081.6
Seabed object002612.318687.7
Table 6. Confusion matrix for the algorithms with the highest median.
Table 6. Confusion matrix for the algorithms with the highest median.
ClassWater SurfaceSeabedSeabed Object
(Points)(%)(Points)(%)(Points)(%)
LVQ SMOTE
Water surface10,6121000000
Seabed0013,00397.63152.4
Seabed object00167.519692.5
ROSE
Water surface10,6121000000
Seabed0013,16098.81581.2
Seabed object002913.718386.3
Borderline-SMOTE2
Water surface10,6121000000
Seabed30.0213,17598.91401.1
Seabed object003114.618185.4
PDFOS
Water surface10,6121000000
Seabed10.0113,21299.21050.8
Seabed object003215.118084.9
Table 7. Confusion matrix of median results for algorithm MDO.
Table 7. Confusion matrix of median results for algorithm MDO.
ClassWater SurfaceSeabedSeabed Object
(Points)(%)(Points)(%)(Points)(%)
MDO
Water surface10,6121000000
Seabed0013,26699.6520.4
Seabed object005425.515874.5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kogut, T.; Tomczak, A.; Słowik, A.; Oberski, T. Seabed Modelling by Means of Airborne Laser Bathymetry Data and Imbalanced Learning for Offshore Mapping. Sensors 2022, 22, 3121. https://doi.org/10.3390/s22093121

AMA Style

Kogut T, Tomczak A, Słowik A, Oberski T. Seabed Modelling by Means of Airborne Laser Bathymetry Data and Imbalanced Learning for Offshore Mapping. Sensors. 2022; 22(9):3121. https://doi.org/10.3390/s22093121

Chicago/Turabian Style

Kogut, Tomasz, Arkadiusz Tomczak, Adam Słowik, and Tomasz Oberski. 2022. "Seabed Modelling by Means of Airborne Laser Bathymetry Data and Imbalanced Learning for Offshore Mapping" Sensors 22, no. 9: 3121. https://doi.org/10.3390/s22093121

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop