Next Article in Journal
Analysis of the Swordfish Xiphias gladius Linnaeus, 1758 Catches by the Pelagic Longline Fleets in the Eastern Pacific Ocean
Next Article in Special Issue
Port State Control Inspections under the Paris Memorandum of Understanding and Their Contribution to Maritime Safety: Additional Risk Classifications and Indicators Using Multivariate Techniques
Previous Article in Journal
Behavior Analysis of a Bucket Foundation with an Offshore Wind Turbine during the In-Water Sinking Process
Previous Article in Special Issue
Air Pollution PM10 Forecasting Maps in the Maritime Area of the Bay of Algeciras (Spain)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Oil Spill Classification Using an Autoencoder and Hyperspectral Technology

by
María Gema Carrasco-García
1,*,
María Inmaculada Rodríguez-García
2,
Juan Jesús Ruíz-Aguilar
2,
Lipika Deka
3,
David Elizondo
3 and
Ignacio José Turias Domínguez
2
1
Department of Industrial and Civil Engineering, Algeciras School of Engineering and Technology (ASET), University of Cádiz, 11202 Algeciras, Spain
2
Department of Computer Science Engineering, Algeciras School of Engineering and Technology (ASET), University of Cádiz, 11202 Algeciras, Spain
3
School of Computer Science and Informatics, Faculty of Computing, Engineering and Media, De Montfort University (DMU), Leicester LE1 9BH, UK
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2024, 12(3), 495; https://doi.org/10.3390/jmse12030495
Submission received: 31 January 2024 / Revised: 12 March 2024 / Accepted: 14 March 2024 / Published: 15 March 2024

Abstract

:
Hyperspectral technology has been playing a leading role in monitoring oil spills in marine environments, which is an issue of international concern. In the case of monitoring oil spills in local areas, hyperspectral technology of small dimensions is the ideal solution. This research explores the use of encoded hyperspectral signatures to develop automated classifiers capable of discriminating between polluted and clean water and distinguishing between various types of oil. The overall objective is to leverage these classifiers to be able to improve the performance of conventional systems that rely solely on hyperspectral imagery. The acquisition of the hyperspectral signatures of water and hydrocarbons was carried out with a spectroradiometer. The range of the spectroradiometer used in this study covers the ranges between [350–1000] (visible near-infrared) and [1000–2500] (short-wavelength infrared). This gives detailed information regarding the targets of interest. Different neural autoencoders (AEs) have been developed to reduce inputs into different dimensions, from 1 to 15. Each of these encoded sets was used to train decision tree (DT) classifiers. The results are very promising, as they show that the AE models encoded data with correlation coefficients above 0.95. The classifiers trained with the different sets provide accuracies close to 1.

1. Introduction

Hyperspectral technology has emerged as a leader in the field of monitoring, covering a broad spectrum of uses thanks to the remarkable advances and increasing availability that have been experienced in recent years [1]. This progress has consolidated the position of hyperspectral technology as a fundamental tool in various sectors. In particular, its relevance in oil spill monitoring stands out, where its ability to analyse and detect subtleties in spectral characteristics allows for more accurate and efficient monitoring [2].
Interest in oil spill monitoring dates back to 1954, when the first International Convention for the Prevention of Marine Pollution by Oil Spills (OILPOL) was enacted. This event marked the beginning of global efforts to address the negative impacts of oil spills on marine ecosystems. As the first marine pollutants to receive significant attention, oil spills led to the implementation of specific measures for their monitoring and reduction. This pioneering approach laid the foundation for future international conventions and protocols, such as The International Convention for the Prevention of Pollution from Ships (MARPOL), underlining the importance of global collaboration in safeguarding and conserving the oceans from oil pollution.
Nowadays, oil spill monitoring is even more important since the 2030 Agenda, resulting from the COP27, establishes Goal 14 “life underwater”, as a specific target for protecting 30% of the seas and oceans by 2030 [3]. Their significant consequences and far-reaching impacts [4], both in the marine domain and human activities [5,6], as well as the oceanic economy [7], place oil spills in the spotlight as one of the 32 main issues to be tackled by science before 2030 [8].
All the efforts made over the years have succeeded in reducing large (>700 t) and medium spills (700–7 t) to marginal amounts in European waters. However, small oil spills still represent significant volumes.
Visual identification of oil spills is often made difficult by their intrinsic characteristics, such as their wide dispersion and sometimes low density, the variability in lighting conditions, or the influence of environmental factors such as waves and wind. Some studies, such as Li’s [9,10], develop wind prediction models to determine the dispersion conditions of oil spills. However, the development of advanced monitoring systems to achieve accurate and real-time detection is required, thus ensuring an effective and prompt response to oil pollution incidents in marine environments to minimise the adverse effects associated with oil spills. This is even more important when it comes to controlling oil spills in local areas with a high risk of incidents, such as waters in port areas and/or bunkering and refineries.
Hyperspectral technology of small dimensions has emerged as an innovative solution to overcome these challenges due to the combination of the spectral resolution of the hyperspectral technology with its high spatial resolution and flexibility for true continuous surveillance of local areas, as it can be installed on unmanned aerial vehicles (UAVs). The distinctive optical characteristics of materials, coupled with advancements in technology and the accessibility of sensors [1], place hyperspectral imaging of small dimensions as a solution for the monitoring of water pollution in local areas.
Hundreds of narrow and contiguous wavelength bands within the visible and infrared wavelength bands provide detailed spectral data. This rich and complete optical information defines quasi-continuous spectral signatures (Figure 1) that allow for identifying subtle differences between two similar targets at first sight. This capacity proves invaluable for the prediction of the thickness of oil film [11,12] as well as differentiating amongst seawater and oil films [13,14], among various other applications.
The development of a model based on hyperspectral signatures to control oil spills in local areas could help to improve the performance rates of a conventional system based solely on hyperspectral images. There is a wide range of techniques that can be used to classify targets according to the information embedded within the hyperspectral signatures [15,16,17]. Machine learning techniques can be applied to problems involving hyperspectral signatures since they represent one-point information without any other extra information. DTs are well-known classifiers, which have been extensively utilised in diverse fields [18], as well as in remote sensing [19,20].
A prior analysis of the relevance of hyperspectral signatures is convenient. This study provides information about the importance or contribution of each spectral band in the ability to distinguish or characterise different elements or phenomena in hyperspectral data. Hyperspectral data are characterised by the large volume of associated information, leading to high dimensionality. An analysis of the relevance of signatures allows for dimensionality reduction while retaining crucial information, thus simplifying the analysis. Furthermore, identifying the most relevant bands enhances the accuracy in classifying objects or materials, and by working only with the most relevant bands, the computational load can be reduced, thus improving analysis efficiency [21].
There are several research studies in which Principal Component Analysis (PCA) has been used successfully as a feature selection method for spectral responses [22,23]. However, the non-linear nature of spectral data [24] renders AE, also known as nonlinear-PCA, a much more convenient feature selection method in the remote sensing field [24,25]. AE has also been demonstrated to be a powerful tool for interpreting better and handling different types of data. Rodríguez-García et al. [26] apply AE for the further forecasting of SO2, PM10, and NO2 concentration. In the same way, Loy_Benitez et al. [27] use a variant of AE to develop a soft sensor validation technique. AE has also been employed in the field of chemistry, to create molecular representations [28], as well as in general industries, such as in intelligent fault diagnosis [29].
This research aims to stack a shallow AE and a DT classifier to create a model for the discrimination of hyperspectral signatures and to determine the dimensional reduction that provides the most optimal classification. Different shallow AEs will be tested to train different classifiers with each new projected dataset. A rigorous and comparative evaluation of the developed models will be conducted to determine their effectiveness in automatic classification. Multiple comparison techniques, such as the Bonferroni method and the Friedman test, will be used for the identification of differences in these methods. They will also be used to determine which were statistically superior. In Section 2, the preparation of samples is discussed. This section also discusses the method used for data collection and the shallow AE as well as DT classifier techniques and the multiple comparison methods including the Friedman test and the Bonferroni method. Section 3 shows the main results of this research. A discussion of the key findings of this work is provided in Section 4. A summary of the main ideas of this work is presented in the Conclusions Section.

2. Materials and Methods

The specific materials, tools, and equipment used for this research will be outlined here. Furthermore, the step-by-step procedures followed during the experiments will be explained, as well as the methodology.

2.1. Sample Preparation

This study was carried out under lab conditions in order to obtain an initial approach to the problem in a controlled environment.
Samples of both water only and hydrocarbon and water mixtures were included as part of the dataset samples.
Real hydrocarbons that are susceptible to spills in marine areas were used to measure the performance level of the proposed method including the following: (a) diesel oil of vessels, (b) C10, and (c) high-sulphur fuel oil. Vessels and bunkering, among others, are possible sources of this type of oil entering the marine environment. These three hydrocarbons present different characteristics. Diesel oil is a hydrocarbon with a very low density and a yellowish colour when it is in high concentration. When it is spread over the entire surface, its appearance resembles that of water, forming rainbows. On the other hand, high-sulphur fuel oil is a very high-density hydrocarbon. Its colour is black and hardly spreads across the water’s surface. The oil C10 is an intermediate hydrocarbon, although is more like diesel oil than fuel oil. It is orange in colour when present in high concentrations but turns yellow as it spreads.

2.2. Dataset Acquisition

Hyperspectral signatures of water and the different oils and their boundaries were acquired with a spectroradiometer (Figure 2). This spectrometer was designed to enable the measurements of emitted, transmitted, absorbed, and reflected electromagnetic energy (EM) within the ranges of [350–1000 nm] (visible-near infrared) and [1000–2500 nm] (short-wavelength infrared). This instrument produces quasi-continuous signatures, as its bandwidth is less than 4 nm in some ranges. This electromagnetic information is collected by a set of optical fibres arranged in a gun-like handle. In this way, the hyperspectral signatures are point responses of the target to be measured.
Signature acquisitions were performed passively with the sun as the light source, ensuring that all spectral bands were available (Figure 3). The truthfulness of measurements was ensured by following the recommended techniques [30].
Reflectance was the optical property used to perform this research. To calculate the proportion of light that is reflected by the target, a previous measurement is required to determine the maximum quantity of light that is possible to reflect, i.e., 100% reflectance. For this purpose, a Lab spectralon board was used as a standard reference. This board acts as a white reference, providing a level of reflectance of almost 100% of the instrument’s range of work. Reference measurements were conducted at 5 min to ensure the precision of the signatures, including an initial measurement taken at the start of the experiment and the beginning of each hydrocarbon tested.
Two hundred and eighty sample spectrums were gathered. They included 40 water samples, 40 diesel oil samples of vessels, 40 samples of C-10, and 40 samples of fuel. A total of 40 samples of every boundary were also collected. All spectrums collected by the instrument correspond to the mean of 50 measurements.

2.3. Pre-Processing

The information provided by hyperspectral technology requires pre-processing before any subsequent analysis. In this initial stage, a specific part of the spectrum is selected to constrain the data. This reduction in the spectral range is necessary so that the results of the research can be applied as additional and complementary models to aid (such as a supplementary expert system) in the future when working with images acquired with a hyperspectral camera working in the VNIR range.
In addition, a normalisation process is carried out to counteract discrepancies in reflectance percentages, which can arise due to small variations in light intensity resulting from the position. This step is essential to ensure proper consistency and comparability in the data, establishing a solid basis for further analysis.
Figure 4 shows the final appearance of the hyperspectral trend signatures grouped by type of targets including the following: (a) diesel on water, (b) fuel oil on water, (c) C10 on water, (d) water, (e) diesel–water boundary, (f) fuel oil–water boundary, and (g) C10–water boundary, once the spectral range was reduced and the signatures were normalised. Each curve represents the spectral response of each measured point of the different targets.

2.4. Shallow Autoencoder

Once the pre-processing of the spectral signatures is completed, the first phase of the analysis begins, which consists of carrying out a relevance analysis using a shallow autoencoder. This step is crucial to identify and select the most informative features of the spectrum, allowing for a more efficient and compact representation of the data and effectively preparing it for more advanced analysis and subsequent modelling processes.
AE [31] is a well-suited technique for this task because, as a neural network model, it can learn complex patterns within the non-linear characteristics of spectral data [32]. AE is a particular case of pattern recognition where the outputs to be assigned are the inputs themselves. Both inputs and targets are identical data, and the outputs of the hidden layer serve as a non-linear encoded representation of the data. Consequently, the dimensions of the newly projected data are dictated by the number of hidden units used within the final hidden layer of the network model. In this work, compression was performed in one step so that the neural network consisted of a single hidden layer, as Figure 5 shows, as a shallow AE configuration.
The backpropagation neural network (BPNN) [33] stands out as the predominant neural model for pattern recognition. This is because it is trained based on the error that is propagated backwards to adjust the weight values of every neuron. Information is transmitted through a fully interconnected network starting at the input layer and moving towards the output one. The backpropagation method can be used to solve function approximation problems [34].
Different AE configurations were tested using different numbers of hidden units within the hidden layer. These ranged from a minimum of 1 to a maximum of 15 units to obtain new projected data in these dimensions. This approach allowed us to explore a diverse range of configurations, evaluating how the variation in the number of neurons affects the quality and efficiency of the compressed representations of the spectral data. Each configuration was trained 20 times, and the original database was randomly organised in each replicate. This process was intended to obtain the AE with the highest levels of generalisation capabilities and robustness. The number of hidden units was determined according to the desired dimension of the projected new database. The selection of the autoencoder for each dimension was based on the correlation coefficients obtained between the predicted and original signatures. Figure 6 shows the scheme used in the autoencoder calculation to obtain each of the new databases.
Following the assessment of the trained autoencoders, the original database was encoded in 15 different ways, resulting in new sets projected into spaces ranging from 1 dimension to 15 dimensions.

2.5. Decision Tree Classifier

DTs are supervised learning models that are inspired by a tree structure. Every internal node refers to a decision taken off a feature. Every branch represents a decision output. Finally, every leaf of the tree refers to a label value. In other words, a DT divides a dataset into smaller subsets based on the relevant features, intending to predict the target variable [35] accurately.
A total of 15 DT classifiers were developed, one for each new projected dataset (different units from within the hidden layer). Following the previous step, each classifier was trained 20 times, changing the inputs and hyperparameters each time to maximise the generalisation and robustness. First, input data were randomly organised to generate distinct training and test sets (70%–30%) for network training and testing purposes. The hyperparameters of the classifier were optimised at each attempt to achieve the best configuration for the DT classifier, minimising the cross-validation loss with which the training set was trained. The optimised hyperparameters were all eligible parameters, as listed below: (1) the maximum number of decision divisions (or branch nodes) that are integers; (2) the minimum number of leaf node observations, and (3) the split criterion. The classifier resulting from each iteration was evaluated using the test set, assessing metrics such as accuracy, sensitivity, specificity, and precision. The scheme followed in the calculation of the classifiers is presented in Figure 7.

2.6. Statistical Comparison

The developed models were rigorously evaluated and compared to determine their goodness of automatic classification and to assess which dimension of the projected data provides the best classifiers.
The Friedman [36] test was performed to statistically analyse the differences in the group means of accuracy, precision, sensitivity, and specificity obtained for the 20 replicates of the 15 classifiers. This test assesses whether there are significant differences between batches, i.e., whether the dimension on which to project the original dataset is significant.
Precision, sensitivity, and specificity parameters were evaluated for water classification, as this is the most relevant classification that the model must successfully achieve. The ultimate aim of the classified models is to be able to discriminate between polluted and clean water, representing a tool for environmental water monitoring.
The Bonferroni method [37] is a multiple comparison procedure used to identify differences in models and determine which are the best based on a statistical analysis. The accuracy levels obtained amongst the classifiers were evaluated.

3. Results

After performing the explained methodology, in which the pre-processed dataset of hyperspectral signatures of water, hydrocarbons on water, and their boundaries, was projected into new spaces of different dimensions, and their respective classifiers were trained, the main outcomes were summarised, highlighting the most significant results.

3.1. Feature Selection: Shallow Autoencoder

Table 1 shows the results obtained after applying a shallow AE using a different number of hidden units to the pre-processed dataset of the hyperspectral signatures of water, oils on water, and their boundaries. This table represents the mean correlation coefficient of 20 attempts of each AE. Standard deviations of each correlation coefficient are also provided. This allows us to analyse the uniformity or disparity between the attempts of each AE.
In addition to the mean values of the correlation coefficients, the maximum values obtained in each AE are also of interest. Figure 8 shows the values of the maximum correlation coefficients achieved in each AE. This representation provides a visual understanding of the evolution of the correlation coefficients regarding the number of hidden units used in the AE models.

3.2. Classification: Decision Trees

The mean results for accuracy, sensitivity, specificity, and precision of the 15 classifiers for the classification of water, the different oils on water, and their boundaries with the water are shown in Table 2. The best result is highlighted in bold.
Several statistical tests and methods were carried out in order to evaluate the different classifiers and to determine the best dimensional reduction to perform the model. This statistical analysis was carried out to evaluate the various classification parameters and to determine the presence of significant differences in these parameters as a function of the size of the inputs with which the classifier was trained.
Sensitivity, specificity, and accuracy parameters are related to the classification of water, which is the main target for identification by the models. This is essential for effective discrimination between contaminated and clean water.
Table 3 presents the p-values produced after performing the Friedman test at the levels of sensitivity, generalisation, specificity, and precision of the 20 replicates of the 15 classifiers. Three different Friedman tests were carried out to enable the evaluation of the classification models without accounting for the impact of the encoded dataset that provided the worst coding. The first column shows the p-values resulting from the Friedman test and the classification parameter values of the 20 replicates of the 15 classifiers. The second column shows the p-values obtained after performing the Friedman test on the classifiers trained with the set of dimensions 2 to 15, leaving out the first dimension, which was the one that gave the worst correlation coefficient in the coding. The procedure was repeated to calculate the p-values of the third column, with the exclusion of the two first dimensions.
The accuracy levels after applying the Bonferroni method are presented in Table 4. Column one specifies the dimension of the inputs, i.e., the dimension to which the original database was coded, with which the classifier was trained. The second column presents the mean value of the evaluated parameter of the 20 repetitions of the classifier. The remaining columns indicate the classifiers for which there are no significant differences according to the Bonferroni method.
The statistical comparison indicates that the best classifier is the one using six-dimensional encoded data as input (highlighted in bold), with a mean accuracy of 0.9577 for the 20 replicates. The effectiveness of this classifier in the discrimination of the different targets, including water, diesel on water, C10 on water, fuel oil on water, and their boundaries, is presented with the confusion matrix of the test group (30% of the input) in Figure 9. The best replicate of the classifier trained with six variables as input, with an accuracy of 0.9881, only misclassifies one hyperspectral signature of the C10–water boundary, being identified as diesel on water, achieving a perfect detection of polluted water.
In addition to this, Figure 10 shows a visual representation of the C10 experiment in water. This figure displays a subset of the 120 samples belonging to hyperspectral signatures representing the following three classes: water (40 samples), C10 (40 samples), and the C10–water boundary (40 samples). All the polluted water samples were correctly classified, and only 1% of the water boundary samples were misclassified.

4. Discussion

The results presented in Table 1 about the different codifications of the original database offer two interpretations. On the one hand, regarding the means of the correlation coefficients, the larger the size of the projected space, the higher the correlation coefficient obtained. AE with more neurons keeps more original information. This finding is to be expected, and this is the reason for the subsequent analysis of the classifiers: to determine the optimal dimensionality reduction. On the other hand, regarding the standard deviations of these means, their low values show that there are no major differences in coding ability between the replicates of each classifier. Consequently, the replicate with the highest correlation coefficient of each classifier was selected to encode the original database, resulting in 15 encoded datasets.
There is one more interpretation that can be made of the feature selection task regarding the highest correlation coefficients for each classifier. The visual representation displayed in Figure 7 shows that dimensions larger than three do not provide a significant difference in the encoded process. However, the determination of the optimal dimension for the classification task with which to code the original database was performed through rigorous statistical analysis.
The results of the classifiers in Table 2 show that DT classifiers have a very good predictive capability. They reach values of accuracies close to 1, and they also present high levels of specificity and precision in the classification of water.
The Friedman test yields really interesting results (Table 3). There are significant differences among the classifiers when all 15 classifiers are tested, regardless of the classification parameter evaluated. This difference is most pronounced when it comes to accuracy. As the tests are reduced to the classifiers trained with the best-encoded inputs, the p-values increase. This increase in p-values in the second and third columns means that the classifiers trained with the 1-D and 2-D sets have associated poor performance. Nevertheless, p-values remain low in terms of accuracy.
This conclusion is clearly shown in the results of the Bonferroni method (Table 4). Inputs of 1-D and 2-D provide the classifiers with the worst accuracies and, moreover, no other is significantly similar to them. At the other extreme are the trained classifiers with encoded inputs to 6 and 13 variables. Both have the same mean accuracy, the highest of the 15 classifiers. Furthermore, the Bonferroni method explains that there are no inputs of dimension less than six that generate a classifier statistically similar to this one. Consequently, the optimal encoding corresponds to reducing the original database to six variables, i.e., an AE with six neurons in its hidden layer.
The best replicate of the trained classifier with six variables achieves an accuracy of 0.9881, misclassifying only one hyperspectral signature of the C10–water boundary as diesel out of the eighty-four hyperspectral signatures in the test dataset (Figure 9 and Figure 10). This means that a classifier with only six variables as input can perfectly classify 100% of polluted water.

5. Conclusions

In conclusion, the results of this research underline the effectiveness of AE in extracting key features from hyperspectral data and reducing their dimensionality. Remarkable correlation coefficients above 0.99 were achieved using AEs with only four neurons in their hidden layer.
Furthermore, it was proven that a basic classifier, such as DT, with a reduced dataset, can perform classifiers with accuracies very close to 1.
In addition to this, through the statistical analysis of the computed classifiers, it was demonstrated that more information is not associated with better classification results.
This knowledge could lay the foundation for future research in which the developed machine learning classifier based on hyperspectral signatures of water and hydrocarbons on water can be applied as a complementary and combined model for the real-time detection of oil in water in hyperspectral imagery of local areas.
The main disadvantage of using a spectroradiometer as prior knowledge of the different targets, such as water and hydrocarbons, lies in the fact that this instrument cannot be used for continuous monitoring of water because of its dimensions. However, the information derived from hyperspectral signatures can enhance the performance of a traditional system that relies exclusively on an imaging system with a hyperspectral camera on board a UAV, and it could also make computational improvements.

Author Contributions

Conceptualisation, M.G.C.-G., I.J.T.D. and J.J.R.-A.; data curation, M.G.C.-G., I.J.T.D. and L.D.; formal analysis, M.G.C.-G. and I.J.T.D.; funding acquisition, I.J.T.D.; investigation, M.G.C.-G. and M.I.R.-G.; methodology, M.G.C.-G., M.I.R.-G. and I.J.T.D.; project administration, J.J.R.-A. and I.J.T.D.; software, M.G.C.-G., L.D., D.E. and I.J.T.D.; resources M.G.C.-G., M.I.R.-G. and I.J.T.D.; supervision J.J.R.-A., D.E. and I.J.T.D.; validation, M.G.C.-G.; visualisation, M.G.C.-G.; writing—original draft, M.G.C.-G. and I.J.T.D.; writing—review and editing, M.G.C.-G., J.J.R.-A., L.D., D.E. and I.J.T.D. All authors have read and agreed to the published version of this manuscript.

Funding

This work is part of the research project EQC2028-004520-P “Smart Cities LAB”, FCTA2020-03 “Automatic identification of marine spills by using machine learning and hyperspectral technology on UVAs”, and “Control hiperespectral de vertidos de hidrocarburos en aguas marinas y fluviales con machine learning” supported by CEI·MAR. This research was supported by “Plan Propio de la Universidad de Cádiz”.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors do not have any relevant conflicts of interest to declare regarding the content of this article.

References

  1. Leifer, I.; Lehr, W.J.; Simecek-Beatty, D.; Bradley, E.; Clark, R. State of the art satellite and airborne marine oil spill remote sensing: Application to the BP Deepwater Horizon oil spill. Remote Sens. Environ. 2012, 124, 185–209. [Google Scholar] [CrossRef]
  2. Mohammadiun, S.; Hu, G.; Gharahbagh, A.A.; Li, J.; Hewage, K.; Sadiq, R. Intelligent computational techniques in marine oil spill management: A critical review. J. Hazard. Mater. 2021, 419, 126425. [Google Scholar] [CrossRef] [PubMed]
  3. European Commission, Press Release. Ocean Biodiversity: Global Agreement on Protection and Sustainable Use of Resources and Biodiversity in High Seas. Available online: https://ec.europa.eu/commission/presscorner/detail/es/ip_23_1382 (accessed on 6 March 2023).
  4. EMSA; EEA. EUROPEAN Maritime Transport Environmental Report 2021; Publications Office of the European Union: Luxembourg, 2021. [Google Scholar]
  5. Aguilera, F.; Méndez, J.; Pásaroa, E.; Laffona, B. Review on the effects of exposure to spilled oils on human health. J. Appl. Toxicol. 2010, 30, 291–301. [Google Scholar] [CrossRef] [PubMed]
  6. Ugwu, C.F.; Ogba, K.T.U.; Ugwu, C.S. Ecological and Economic Costs of Oil Spills in Niger Delta, Nigeria. In Economic Effects of Natural Disasters: Theoretical Foundations, Methods, and Tools; Academic Press: Cambridge, UK, 2021; pp. 439–455. [Google Scholar] [CrossRef]
  7. Cirer-Costa, J.C. Tourism and its hypersensitivity to oil spills. Mar. Pollut. Bull. 2015, 91, 65–72. [Google Scholar] [CrossRef]
  8. Yang, J.; Ma, Y.; Hu, Y.; Jiang, Z.; Zhang, J.; Wan, J.; Li, Z. Decision Fusion of Deep Learning and Shallow Learning for Marine Oil Spill Detection. Remote Sens. 2022, 14, 666. [Google Scholar] [CrossRef]
  9. Li, Y.; Huang, W.; Lyu, X.; Liu, S.; Zhao, Z.; Ren, P. An adversarial learning approach to forecasted wind field correction with an application to oil spill drift prediction. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102924. [Google Scholar] [CrossRef]
  10. Li, Y.; Lyu, X.; Ren, P. Oil Spill Timely Backtracking Oriented by Wind Field Correction with Self-Attention Temporal Convolutional Networks. IEEE J. Ocean. Eng. 2024, 49, 114–132. [Google Scholar] [CrossRef]
  11. Jiang, Z.; Ma, Y.; Yang, J. Inversion of the thickness of crude oil film based on an OG_CNN model. J. Mar. Sci. Eng. 2020, 8, 653. [Google Scholar] [CrossRef]
  12. Lu, Y.; Tian, Q.; Wang, X.; Zheng, G.; Li, X. Determining oil slick thickness using hyperspectral remote sensing in the Bohai Sea of China. Int. J. Digit. Earth 2013, 6, 76–93. [Google Scholar] [CrossRef]
  13. Deepthi; Thomas, T. Spectral similarity algorithm-based image classification for oil spill mapping of hyperspectral datasets. J. Spectr. Imaging 2020, 9, a14. [Google Scholar] [CrossRef]
  14. El-Rahman, S.A.; Zolait, A.H.S. Hyperspectral image analysis for oil spill detection: A comparative study. Int. J. Comput. Sci. Math. 2018, 9, 103–121. [Google Scholar] [CrossRef]
  15. Lv, W.; Wang, X. Overview of Hyperspectral Image Classification. J. Sens. 2020, 2020, 4817234. [Google Scholar] [CrossRef]
  16. Wambugu, N.; Chen, Y.; Xiao, Z.; Tan, K.; Wei, M.; Liu, X.; Li, J. Hyperspectral image classification on insufficient-sample and feature learning using deep neural networks: A review. Int. J. Appl. Earth Obs. Geoinf. 2021, 105, 102603. [Google Scholar] [CrossRef]
  17. Qu, S.; Li, X.; Gan, Z. A Review of Hyperspectral Image Classification Based on Joint Spatial-spectral Features. J. Phys. Conf. Ser. 2022, 578, 435–456. [Google Scholar] [CrossRef]
  18. Maxwell, A.E.; Warner, T.A.; Fang, F. Implementation of machine-learning classification in remote sensing: An applied review. Int. J. Remote Sens. 2018, 39, 2784–2817. [Google Scholar] [CrossRef]
  19. Uys, A.; Steyn, M.; Botha, D. Decision tree analysis for age estimation in living individuals: Integrating cervical and dental radiographic evaluations within a South African population. Int. J. Leg. Med. 2024. [Google Scholar] [CrossRef] [PubMed]
  20. Maurya, K.; Mahajan, S.; Chaube, N. Decision tree (DT) and stacked vegetation indices based mangrove and non-mangrove discrimination using AVIRIS-NG hyperspectral data: A study at Marine National Park (MNP) Jamnagar, Gulf of Kutch. Wetl. Ecol. Manag. 2023, 31, 805–823. [Google Scholar] [CrossRef]
  21. Jarocińska, A.; Kopeć, D.; Kycko, M.; Piórkowski, H.; Błońska, A. Hyperspectral vs. Multispectral data: Comparison of the spectral differentiation capabilities of Natura 2000 non-forest habitats. ISPRS J. Photogramm. Remote Sens. 2022, 184, 148–164. [Google Scholar] [CrossRef]
  22. Carrasco-Garcia, M.G.; Rodríguez-García, M.I.; Ruiz-Aguilar, J.J.; González-Enrique, J.; Turias-Dominguez, I.J. Characterisation of oil spills using hyperspectral technology and feature selection. In Proceedings of the XV Congreso de Ingeniería del Transporte (CIT 2023), La Laguna, Spain, 14–16 June 2023. [Google Scholar]
  23. Carrasco-García, M.G.; Rodríguez-García, M.I.; González-Enrique, J.; Ruiz-Aguilar, J.J.; Turias-Domínguez, I.J. Hyperspectral technology for oil spills characterisation by using feature selection. Transp. Res. Procedia 2023, 71, 117–123. [Google Scholar] [CrossRef]
  24. Scholz, M.; Vigário, R. Nonlinear PCA: A new hierarchical approach. In Proceedings of the European Symposium on Artificial Neural Networks, Bruges, Belgium, 24–26 April 2002. [Google Scholar]
  25. Dong, G.; Liao, G.; Liu, H.; Kuang, G. A Review of the Autoencoder and Its Variants: A Comparative Perspective from Target Recognition in Synthetic-Aperture Radar Images. IEEE Geosci. Remote Sens. Mag. 2018, 6, 44–68. [Google Scholar] [CrossRef]
  26. Rodríguez-García, M.-I.; González-Enrique, J.; Ruiz Aguilar, J.J.; Turias Domínguez, I.J. Forecasting of SO2, PM10, and NO2 concentrations in the Bay of Algeciras (Spain) using autoencoders. Cybern. Syst. 2024, accepted. [Google Scholar]
  27. Loy-Benitez, J.; Heo, S.K.; Yoo, C.K. Soft sensor validation for monitoring and resilient control of sequential subway indoor air quality through memory-gated recurrent neural networks-based autoencoders. Control Eng. Pract. 2020, 97, 104330. [Google Scholar] [CrossRef]
  28. Wigh, D.S.; Goodman, J.M.; Lapkin, A.A. A review of molecular representation in the age of machine learning. Wiley Interdiscip. Rev. Comput. Mol. Sci. 2022, 12, e1603. [Google Scholar] [CrossRef]
  29. Yang, Z.; Xu, B.; Luo, W.; Chen, F. Autoencoder-based representation learning and its application in intelligent fault diagnosis: A review. Measurement 2022, 189, 110460. [Google Scholar] [CrossRef]
  30. Goetsz, A.F.H. Making Accurate Field Spectral Reflectance Measurements-LR; ASD Inc.: Boulder, CO, USA, 2012. [Google Scholar]
  31. Kramer, M.A. Autoassociative neural networks. Comput. Chem. Eng. 1992, 16, 313–328. [Google Scholar] [CrossRef]
  32. Abiodun, O.I.; Jantan, A.; Omolara, A.E.; Dada, K.V.; Mohamed, N.A.; Arshad, H. State-of-the-art in artificial neural network applications: A survey. Heliyon 2018, 4, 938. [Google Scholar] [CrossRef] [PubMed]
  33. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning internal representations by error propagation. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  34. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1980, 2, 359–366. [Google Scholar] [CrossRef]
  35. Quinlan, J.R. Simplifying decision trees. Int. J. Man-Mach. Stud. 1987, 27, 221–234. [Google Scholar] [CrossRef]
  36. Milton, F. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. J. Am. Stat. Assoc. 1937, 32, 675–701. [Google Scholar]
  37. Bonferroni, C.E. Teoria Statistica delle Classi e Calcolo delle Probabilità; R. Istituto Superiore di Scienze Economiche e Commerciali di Firenze: Florence, Italy, 1936. [Google Scholar]
Figure 1. Representation of a hyperspectral signature.
Figure 1. Representation of a hyperspectral signature.
Jmse 12 00495 g001
Figure 2. This figure shows the visual appearance of the different hydrocarbons used in this research including (a) diesel oil of vessels; (b) C10, and (c) high-sulphur fuel oil.
Figure 2. This figure shows the visual appearance of the different hydrocarbons used in this research including (a) diesel oil of vessels; (b) C10, and (c) high-sulphur fuel oil.
Jmse 12 00495 g002
Figure 3. This figure shows the visual appearance of the different hydrocarbons used in this research. (a) Water hyperspectral signatures acquisition and (b) high-sulphur fuel oil hyperspectral signatures acquisition.
Figure 3. This figure shows the visual appearance of the different hydrocarbons used in this research. (a) Water hyperspectral signatures acquisition and (b) high-sulphur fuel oil hyperspectral signatures acquisition.
Jmse 12 00495 g003
Figure 4. Normalised hyperspectral signatures after reducing the spectral range at which the image sensors work (i.e., our camera Cubert Ultris X20 plus works in this range). (a) Diesel on water, (b) fuel oil on water, (c) C10 on water, (d) water, (e) diesel–water boundary, (f) fuel oil–water boundary, and (g) C10–water boundary.
Figure 4. Normalised hyperspectral signatures after reducing the spectral range at which the image sensors work (i.e., our camera Cubert Ultris X20 plus works in this range). (a) Diesel on water, (b) fuel oil on water, (c) C10 on water, (d) water, (e) diesel–water boundary, (f) fuel oil–water boundary, and (g) C10–water boundary.
Jmse 12 00495 g004
Figure 5. Structure of the shallow AE with the hidden layer composed of two units. The neural network topology is made of an input layer, a single hidden layer, and an output layer. Both the input and the output layers are identical. The hidden layer output is the new projected data, where the dimension is determined by the size of this hidden layer.
Figure 5. Structure of the shallow AE with the hidden layer composed of two units. The neural network topology is made of an input layer, a single hidden layer, and an output layer. Both the input and the output layers are identical. The hidden layer output is the new projected data, where the dimension is determined by the size of this hidden layer.
Jmse 12 00495 g005
Figure 6. Calculation scheme of AEs to obtain the new projected data from 1 to 15 dimensions.
Figure 6. Calculation scheme of AEs to obtain the new projected data from 1 to 15 dimensions.
Jmse 12 00495 g006
Figure 7. DT calculation scheme used to obtain the classifiers for each new projected dataset from 1 to 15 dimensions.
Figure 7. DT calculation scheme used to obtain the classifiers for each new projected dataset from 1 to 15 dimensions.
Jmse 12 00495 g007
Figure 8. Representation of the maximum correlation coefficients of 20 replicates of each AE. R is the measurement of the capability to reproduce the input in the output.
Figure 8. Representation of the maximum correlation coefficients of 20 replicates of each AE. R is the measurement of the capability to reproduce the input in the output.
Jmse 12 00495 g008
Figure 9. Confusion matrix of the best classifier trained with 6 variables as input.
Figure 9. Confusion matrix of the best classifier trained with 6 variables as input.
Jmse 12 00495 g009
Figure 10. Visual representation of the classification in the C10 experiment with samples belonging to hyperspectral signatures representing the following three classes: water (40 samples), C10 (40 samples), and the C10–water boundary (40 samples). The classifier only fails to identify one signature in the C10–water boundary.
Figure 10. Visual representation of the classification in the C10 experiment with samples belonging to hyperspectral signatures representing the following three classes: water (40 samples), C10 (40 samples), and the C10–water boundary (40 samples). The classifier only fails to identify one signature in the C10–water boundary.
Jmse 12 00495 g010
Table 1. Mean and standard deviation results using twenty replicates from the fifteen AEs.
Table 1. Mean and standard deviation results using twenty replicates from the fifteen AEs.
Dimension of Projected Datar (Mean)Std Deviation
10.90877.54 × 10−7
20.96310.000471
30.98680.001809
40.99000.000368
50.99161.19 × 10−5
60.99270.000162
70.99350.000112
80.99380.000163
90.994240.000108
100.994500.000115
110.994729.65 × 10−5
120.994920.000130
130.995126.19 × 10−5
140.995300.000164
150.995390.000151
Table 2. Averages for accuracy, sensitivity, specificity, and precision of the 15 classifiers.
Table 2. Averages for accuracy, sensitivity, specificity, and precision of the 15 classifiers.
DimensionAccuracySensitivitySpecificityPrecision
10.84520.70500.97230.7842
20.70830.77500.97360.8078
30.90060.77000.98040.8599
40.89940.81000.97160.8121
50.92080.82000.97700.8381
60.95770.82500.99730.9771
70.94820.78500.97640.8393
80.93210.81500.97300.8139
90.94350.85500.97640.8496
100.92920.85000.97230.8138
110.94230.81000.97160.8208
120.93270.79500.97030.8089
130.95770.89000.98310.8893
140.93390.79500.97770.8508
150.93980.78000.98040.8578
Table 3. The Friedman test p-values, evaluating null hypothesis in the following three cases: (1) p-values obtained by applying the Friedman test to all the projected datasets (from dim. 1 to 15), (2) p-values obtained by applying the Friedman test to the projected dataset from dim. 2 to 15, and (3) p-values obtained by applying the Friedman test to the projected dataset from dim. 3 to 15.
Table 3. The Friedman test p-values, evaluating null hypothesis in the following three cases: (1) p-values obtained by applying the Friedman test to all the projected datasets (from dim. 1 to 15), (2) p-values obtained by applying the Friedman test to the projected dataset from dim. 2 to 15, and (3) p-values obtained by applying the Friedman test to the projected dataset from dim. 3 to 15.
ParameterD1–D15
p-Value
D2–D15
p-Value
D3–D15
p-Value
Accuracy4.86 × 10−261.95 × 10−209.42 × 10−13
Sensitivity0.0070910.0893220.108005
Specificity0.0001320.0002160.000274
Precision3.74 × 10−50.0001130.000231
Table 4. The Bonferroni method for accuracy. Evaluation of the significant differences among the different dimensional reductions regarding accuracy. C1: Dimension of the inputs. C2: Mean value of the evaluated parameter of the 20 repetitions of the classifier. Remaining columns: Classifiers for which there are no significant differences according to the Bonferroni method.
Table 4. The Bonferroni method for accuracy. Evaluation of the significant differences among the different dimensional reductions regarding accuracy. C1: Dimension of the inputs. C2: Mean value of the evaluated parameter of the 20 repetitions of the classifier. Remaining columns: Classifiers for which there are no significant differences according to the Bonferroni method.
DMeanNon-Significant Differences
20.70832
10.8452 1
40.8994 435108
30.9006 43510812
50.9208 4351081214151197
100.9292 4351081214151197613
80.9321 4351081214151197613
120.9327 351081214151197613
140.9339 51081214151197613
150.9399 51081214151197613
110.9423 51081214151197613
90.9435 51081214151197613
70.9482 51081214151197613
60.9577 1081214151197613
130.9577 1081214151197613
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Carrasco-García, M.G.; Rodríguez-García, M.I.; Ruíz-Aguilar, J.J.; Deka, L.; Elizondo, D.; Turias Domínguez, I.J. Oil Spill Classification Using an Autoencoder and Hyperspectral Technology. J. Mar. Sci. Eng. 2024, 12, 495. https://doi.org/10.3390/jmse12030495

AMA Style

Carrasco-García MG, Rodríguez-García MI, Ruíz-Aguilar JJ, Deka L, Elizondo D, Turias Domínguez IJ. Oil Spill Classification Using an Autoencoder and Hyperspectral Technology. Journal of Marine Science and Engineering. 2024; 12(3):495. https://doi.org/10.3390/jmse12030495

Chicago/Turabian Style

Carrasco-García, María Gema, María Inmaculada Rodríguez-García, Juan Jesús Ruíz-Aguilar, Lipika Deka, David Elizondo, and Ignacio José Turias Domínguez. 2024. "Oil Spill Classification Using an Autoencoder and Hyperspectral Technology" Journal of Marine Science and Engineering 12, no. 3: 495. https://doi.org/10.3390/jmse12030495

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop