A New Remote Hyperspectral Imaging System Embedded on an Unmanned Aquatic Drone for the Detection and Identification of Floating Plastic Litter Using Machine Learning

: This paper presents a new Remote Hyperspectral Imaging System (RHIS) embedded on an Unmanned Aquatic Drone (UAD) for plastic detection and identiﬁcation in coastal and freshwater environments. This original system, namely the Remotely Operated Vehicle of the University of Littoral C ô te d’Opale (ROV-ULCO), works in a near-ﬁeld of view, where the distance between the hyperspectral camera and the water surface is about 45 cm. In this paper, the new ROV-ULCO system with all its components is ﬁrstly presented. Then, a hyperspectral image database of plastic litter acquired with this system is described. This database contains hyperspectral data cubes of different plastic types and polymers corresponding to the most-common plastic litter items found in aquatic environments. An in situ spectral analysis was conducted from this benchmark database to characterize the hyperspectral reﬂectance of these items in order to identify the absorption feature wavelengths for each type of plastic. Finally, the ability of our original system RHIS to automatically recognize different types of plastic litter was assessed by applying different supervised machine learning methods on a set of representative image patches of marine litter. The obtained results highlighted the plastic litter classiﬁcation capability with an overall accuracy close to 90%. This paper showed that the newly presented RHIS coupled with the UAD is a promising approach to identify plastic waste in aquatic environments.


Introduction
The high and rapidly increasing levels of plastic litter in aquatic environments represent a serious environmental problem at a global scale, negatively affecting aquatic life and biodiversity, ecosystems, livelihoods, fisheries, maritime transport, recreation, tourism, and economies.To address this problem, the research community is always looking for novel devices, tools, and methods to detect, identify, and quantify plastic litter more rapidly and efficiently [1][2][3][4].Monitoring methods such as visual counting or sampling using nets are labor-intensive, whereas current remote observation (from spaceborne or airborne platforms) has some limitations in detecting and identifying plastic litter.Therefore, it is necessary to develop an innovative remote sensing system able to automatically detect and identify plastic litter in order to study pollution sources properly, to improve the survey assessments, and to support the implementation of mitigation measures [1][2][3][4][5].
resolution [8,10,16].Moreover, satellites and aerial images need to apply atmosph rection methods on the data to extract the hyperspectral reflectance.Conceptual framework for remote marine litter detection from [22] with differe devices (satellite, airborne, drone, etc.) and with the new system (ROV-ULCO) propose paper.
One way to overcome such limitations is to develop a hyperspectral sys works in the near-field and acquires hyperspectral images with a high spatial re and a high number of spectral bands.In this context, our contribution, propose paper, was the development of a new Remote Hyperspectral Imaging System (RH bedded on an Unmanned Aquatic Drone (UAD), namely the Remotely Operated of the University of Littoral Côte d'Opale (ROV-ULCO), as shown in Figure 1 knowledge, such a system is a real technological innovation that has never been p in the literature and, therefore, constitutes the novelty of our work.Our main o were (1) to develop new technologies for the perception and detection of plastic w to reduce time consumption during a study, and (3) to generalize an accurate tool tify, qualify, and identify polymers of floating plastic litter.This study, thus, con to the ongoing research efforts to develop new tools and methodologies for plas detection.
Although the use of hyperspectral imaging provides a large amount of info the problem of marine litter detection and identification is still complex due to the of different types of marine waste, especially for plastic materials present in aqua ronments, and the difficulty to recognize their nature by image analysis because high shape, size, opacity, and polymer variabilities [10,11,24].Indeed, floating pla can be perceived differently depending on its position, its orientation, or its speed of the camera and the lighting device, which can generate shadows and specula tion, depending on the opacity of its material or depending on whether its surface wet or dry and mixed with other materials.In order to reproduce these different sc it is important to first carry out experiments under laboratory-controlled conditio thermore, dealing with hyperspectral data is computationally expensive, and it  [22] with different remote devices (satellite, airborne, drone, etc.) and with the new system (ROV-ULCO) proposed in this paper.
One way to overcome such limitations is to develop a hyperspectral system that works in the near-field and acquires hyperspectral images with a high spatial resolution and a high number of spectral bands.In this context, our contribution, proposed in this paper, was the development of a new Remote Hyperspectral Imaging System (RHIS) embedded on an Unmanned Aquatic Drone (UAD), namely the Remotely Operated Vehicle of the University of Littoral Côte d'Opale (ROV-ULCO), as shown in Figure 1.To our knowledge, such a system is a real technological innovation that has never been presented in the literature and, therefore, constitutes the novelty of our work.Our main objectives were (1) to develop new technologies for the perception and detection of plastic waste, (2) to reduce time consumption during a study, and (3) to generalize an accurate tool to quantify, qualify, and identify polymers of floating plastic litter.This study, thus, contributes to the ongoing research efforts to develop new tools and methodologies for plastic litter detection.
Although the use of hyperspectral imaging provides a large amount of information, the problem of marine litter detection and identification is still complex due to the number of different types of marine waste, especially for plastic materials present in aquatic environments, and the difficulty to recognize their nature by image analysis because of their high shape, size, opacity, and polymer variabilities [10,11,24].Indeed, floating plastic litter can be perceived differently depending on its position, its orientation, or its speed in front of the camera and the lighting device, which can generate shadows and specular reflection, depending on the opacity of its material or depending on whether its surface is either wet or dry and mixed with other materials.In order to reproduce these different scenarios, it is important to first carry out experiments under laboratory-controlled conditions.Furthermore, dealing with hyperspectral data is computationally expensive, and it is quite challenging to collect and manually label data for all types of existing plastic marine litter [3,5,10,11].
In this study, our general focus aimed to assess the capability of the proposed ROV-ULCO system to automatically recognize different types of plastic waste in aquatic environments with classical machine learning methods.For this purpose, waste samples representative of the most-common marine litter items found in the coastal environment were collected.This collection contained different plastic types (HDPE, LDPE, PET, PP, PVC, PS, etc.) and other materials such as wood, paper, rubber, and vegetation.Hyperspectral images of these waste items were then acquired by the RHIS system in laboratory-controlled conditions to build a benchmark database.RHIS provides high-spatial-resolution images and hyperspectral data cubes that cover the NIR (900 nm) to SWIR (1700 nm) range of the electromagnetic spectrum.From this database, an in situ spectral analysis was carried out to check the compliance of the spectra with the literature.Standard machine learning methods were then applied to evaluate the plastic waste recognition performance of our system.
The second section of the paper first presents the ROV-ULCO system with all its components, as well as the collected waste samples used in the experiments.This section describes how the waste samples are scanned by the ROV-ULCO to provide hyperspectral images of the proposed benchmark database.Two kinds of datasets were then derived from this database.The first dataset was constituted by the mean spectral reflectance computed over various parts of each marine litter sample observed under different conditions.This dataset was used to conduct an in situ spectral analysis in order to characterize each type of marine litter by a reference hyperspectral reflectance, as described in Section 3.This analysis aimed to compare the absorption features of each reference hyperspectral reflectance with the literature and thus confirmed which wavelengths were the most-efficient to discriminate the different plastic types.The second dataset was a set of same-sized image patches manually selected from the whole hyperspectral images and labeled with the ground-truth of each available marine litter category.Section 4 presents the experiments conducted with this dataset in order to assess the ability of our original RHIS to recognize different categories of marine litter.In this section, several supervised machine learning models, such as K-Nearest Neighbors (KNNs), Support Vector Machines (SVMs), and Artificial Neural Networks (ANNs), were trained on a set of training image patches.Then, the best-trained models were used to evaluate their performance on the testing image patches of the marine litter so that the testing patches were independent of the training ones to reproduce realistic conditions.Finally, the conclusion highlights that the ROV-ULCO is a promising approach to detect and identify plastic litter in aquatic environments.

Materials
This section describes the materials used in this study.The new remote hyperspectral imaging ROV-ULCO system is first presented with all its components in Section 2.1.The experimental setup to conduct the acquisitions with this system is then described in Section 2.2.This setup aimed to create a benchmark hyperspectral image database of plastic litter in a controlled laboratory environment, which reproduces different real situations.This database is presented in Section 2.3.

The ROV-ULCO System
The ROV-ULCO system, illustrated in Figure 2, is constituted of two subsystems: a new Remote Hyperspectral Imaging System (RHIS) and an Unmanned Aquatic Drone (UAD).
The UAD is an aquatic surface drone, named Jellyfishbot, specifically designed for removing floating debris [25].It is equipped with two propulsions, which are located under the two floating parts, and a remote control and communication system.The aquatic drone can reach a top speed of 2 knots and can have autonomy in terms of power that is greater than 2 h.The UAD was tailored at the University of Littoral (ULCO) to enable plastic material sampling in different water bodies, even in confined and hard-to-reach areas such as small waterways, estuaries, or rivers [25].The UAD is an aquatic surface drone, named Jellyfishbot, specifically designed for removing floating debris [25].It is equipped with two propulsions, which are located under the two floating parts, and a remote control and communication system.The aquatic drone can reach a top speed of 2 knots and can have autonomy in terms of power that is greater than 2 h.The UAD was tailored at the University of Littoral (ULCO) to enable plastic material sampling in different water bodies, even in confined and hard-to-reach areas such as small waterways, estuaries, or rivers [25].
The RHIS is connected at the front of the UAD in order to push it at the water surface.This original imaging system was structured around the following main outside elements (Figure 3):  The RHIS is connected at the front of the UAD in order to push it at the water surface.This original imaging system was structured around the following main outside elements (Figure 3): The system enclosures are certified waterproof up to a 1.5 m depth at atmospheric pressure and for 30 min with an Ingress Protection (IP) rating of IP67.
The usage of the ROV-ULCO is limited to the observation of marine litter floating on the surface of the water.The current prototype is not designed to detect occluded objects The system enclosures are certified waterproof up to a 1.5 m depth at atmospheric pressure and for 30 min with an Ingress Protection (IP) rating of IP67.
The usage of the ROV-ULCO is limited to the observation of marine litter floating on the surface of the water.The current prototype is not designed to detect occluded objects that are not visible at the first observation layer or that are completely submerged under the water, which absorbs SWIR light depending on the depth.
The ROV-ULCO can travel more than 2 km with 1 h of autonomy at a maximum speed of 2 knots, before returning to replace the interchangeable batteries and continue the observation.
Hyperspectral images are stored only when marine litter is present under the RHIS.The recorded images are then processed offline to recognize the type of waste observed by the camera.
The Resonon PIKA-NIR-320 is a line-scan (also called push-broom) hyperspectral camera that covers the NIR to SWIR spectral range (900-1700 nm) with 164 spectral bands.The total number of spectral channels delivered by this camera is actually 168, with bands extending beyond both edges of the spectral range.Its resolution is 320 spatial pixels per line with a pixel size of 30 µm, and its line rate reaches up to 520 Hz.The main characteristics of the PIKA-NIR-320 hyperspectral camera are described in the Supplementary Materials (Datasheet S1) (now referred to as Pika IR hyperspectral camera: https://resonon.com/Pika-IR, accessed on 28 March 2023).This line-scan imager collects data one line at a time, and a two-dimensional image is completed by assembling line-by-line the multiple line-images acquired successively as the object is translated.To obtain hyperspectral data, signals from each pixel of a line-image enter at the same time into a spectrometer, which provides the spectrum of incoming light intensity as a function of wavelength for every pixel of the image.The two-dimensional image thus-acquired can be interpreted as a stack of single-band grayscale images, called a data cube, where each image of the stack corresponds to a different wavelength.This hyperspectral camera is provided with the SpectrononPro software, version 3.4.4(Spectronon software, Hyperspectral Software: https://resonon.com/software,accessed on 28 March 2023) to acquire data.
In order to ensure the stability of the system, the RHIS was designed in such a way that the center of gravity of the camera is as close as possible to the surface of the water.This is the reason why the camera is positioned horizontally with its optical axis parallel to the water surface.The ROV-ULCO is remotely controlled so that marine litter floating on the water is scanned by the RHIS.The latter operates in a very near-field to detect floating marine litter, where the distance between the hyperspectral camera and the water surface is about 45 cm (see the 3D views in Figure S1 of the Supplementary Materials for more details).The waste scrolling under the RHIS is illuminated by a waterproof lighting device protected from ambient light by a plate system.This device consists of a ramp of three halogen lamps, whose light spectrum covers the sensitivity range of the camera (900-1700 nm), in front of a diffuser.The light reflected by the illuminated surface hits an optical mirror oriented at 45 • , to move towards the objective lens and then on to the camera sensor parallel to the surface water.In order to cover a field of view corresponding to the distance between the two floats of the ROV-ULCO, the focal length of the objective lens is equal to 12 mm.The f-number of the objective lens was set to f/2 to let in a sufficient quantity of light for image acquisition without causing too much optical distortion.The length of the field of view is about 30 cm with this setting.
Although the scanning area covered by the proposed system is less than the area observed by other platforms equipped with hyperspectral imaging systems such as satellites, airborne vehicles, and drones, it overcomes their limitations in the spatial and spectral resolutions and enables observations of areas not visible by these other platforms.Moreover, no atmospheric correction of the hyperspectral data is needed.Another advantage of our system is that it can work night and day because it is completely independent of solar illumination and isolated from light noise.Finally, it can be easily used as a portable laboratory imaging system.

Experimental Setup
The two main objectives of the experimental setup were, first, to calibrate in situ the RHIS before being used in aquatic environments and, secondly, to characterize the main types of marine litter so that they can be recognized automatically.For this purpose, different real situations encountered on the water surface were reproduced in a laboratory environment.
To simulate an aquatic environment, a black PVC plastic container (dimensions of 45 × 32 × 10 cm 3 ) was filled with clear seawater.Such a black container was used to hold the seawater and the objects because it has negligible reflectance values compared to the reflectance values of the observed objects over the NIR-SWIR spectrum, while water absorbs infrared light.The hyperspectral camera was positioned at 45 cm of the water surface level to reproduce similar conditions to the real situation where the RHIS works on the water surface to detect the plastic type of the floating objects.For the linear motion simulation, a linear translation stage (linear scanner) was used to move the black container (Figure 4).An embedded computer with the Spectronon Pro software was used to calibrate the camera, focus the objective lens, capture the hyperspectral data cubes, control the integration time and the frame rate of the camera, and drive the motor for the linear scanner.Although hyperspectral cameras are spectrally calibrated, they usually provide raw data, which need to be calibrated to obtain the absolute reflectance of the scanned objects.For this purpose, both the instrument sensor response and illumination functions were considered to correct the acquired images.This calibration, also named flat field correction, was, thus, performed by a dark correction followed by a response correction.For the dark correction, the dark reference was captured by completely closing the aperture of the camera, leading to no light striking the sensor, resulting in a true dark reflectance.For the response correction, a Spectralon ® white diffuse reflectance standard was used as a white reference with a reflectivity of 1 in all the wavelengths.The integration time was adjusted to maximize the apparent reflectance of the Spectralon calibration panel.For all acquisitions, the camera parameters were fixed so that the frame rate supported the integration time required for the illumination and sample brightness, as follows:
The speed of the linear translation stage was then adjusted to maintain a unity aspect ratio so that the observed objects were not distorted in the acquired images.We obtained hyperspectral images with the spatial resolution given by the ratio between the line of An embedded computer with the Spectronon Pro software was used to calibrate the camera, focus the objective lens, capture the hyperspectral data cubes, control the integration time and the frame rate of the camera, and drive the motor for the linear scanner.Although hyperspectral cameras are spectrally calibrated, they usually provide raw data, which need to be calibrated to obtain the absolute reflectance of the scanned objects.For this purpose, both the instrument sensor response and illumination functions were considered to correct the acquired images.This calibration, also named flat field correction, was, thus, performed by a dark correction followed by a response correction.For the dark correction, the dark reference was captured by completely closing the aperture of the camera, leading to no light striking the sensor, resulting in a true dark reflectance.For the response correction, a Spectralon ® white diffuse reflectance standard was used as a white reference with a reflectivity of 1 in all the wavelengths.The integration time was adjusted to maximize the apparent reflectance of the Spectralon calibration panel.For all acquisitions, the camera parameters were fixed so that the frame rate supported the integration time required for the illumination and sample brightness, as follows:
The speed of the linear translation stage was then adjusted to maintain a unity aspect ratio so that the observed objects were not distorted in the acquired images.We obtained hyperspectral images with the spatial resolution given by the ratio between the line of view length and the camera resolution.
Although the hyperspectral camera Resonon PIKA-NIR320 covers the spectral range from 900 to 1700 nm with 168 spectral bands, the 13 first spectral bands and the 12 last spectral bands provide too noisy and distorted information to be exploited.For this reason, these 25 spectral bands were neglected by reducing the number of bands from 168 to 143, with an interval of the wavelength from 949.2 nm to 1650.8 nm instead of the real interval of the wavelength from 886.3 nm to 1711.4 nm.
This setup was used to acquire hyperspectral images of different waste samples in order to spectrally characterize them, compare their spectrum with the state-of-the-art to validate the proposed RHIS, on the one hand, and prove the RHIS's ability to recognize the different plastic litter, on the other hand.The database built for this purpose is presented in the next subsection.

Benchmark Image Database
In our study, plastic waste samples were collected from estuarine and coastal beaches along the Eastern English Channel French coast.In addition, some virgin plastics were used to expand the plastic library, which led to a set of plastic objects that contained and represented all plastic types.A categorized overview of these plastics objects and their types is shown in Table 1.They were divided into the following categories: (1) HDPE, (2) LDPE, (3) PET, (4) PP, (5) PVC, (6) PS, ( 7) PolyURethane (PUR), ( 8) PolyOxyMethylene (POM), (9) and ABS.A tenth category of non-plastic materials found in the aquatic environment (wood, vegetation, cardboard, clear seawater, etc.) and named "Other" was also added.We can notice in Table 1 that the number of objects was different depending on the plastic type.This variation was representative of the diversity of products made with each type of plastic.
The polymer that constituted each type of plastic object presented in Table 1 was then identified by a Macro-Raman Spectrometer (MacroRAM, Horiba Scientific, France, Palaiseau) using a laser of a 785 nm wavelength with a power of 7-450 mW and a fixed grating of 685 gr.mm −1 with a spectral range from 100-3400 cm −1 [26,27].This spectrometer was equipped with a CCD detector for a spectral resolution of 8 cm −1 at 914 nm.The signal acquisition and processing were realized with Labspec software and its identification using the KnowItAll software (KnowItAll, BioRad ® ) and the free-access spectra libraries of Horiba (Raman-Forensic-Horiba) and SLoPP/SLoPP-E.These identifications served as a ground-truth to label the data of our benchmark waste hyperspectral image database.
To create a hyperspectral image database of real marine litter, we performed 39 acquisitions with all the plastic objects presented in Table 1 under several positions using the RHIS in a controlled environment.Similar to [28], two cases were studied: dry and wet objects.The same object was scanned three times under several positions (face up or face down, side up or side down, etc.) and different views of its presence as an object floating (dry and wet) on seawater.We, thus, obtained a database of hyperspectral images which contained hyperspectral data cubes of the nine plastic types (HDPE, LDPE, PET, PP, PVC, PS, PUR, POM, ABS) and the "Other" category.
Two different datasets were then derived from this benchmark image database in order to carry out the two experiments presented in Sections 3 and 4, respectively.The first dataset consisted of reference mean hyperspectral reflectance spectra, and it was used to perform a spectral data analysis (Section 3).The second dataset was a labeled waste image patch dataset, which was used to assess the classification performance of the proposed system (Section 4).

Spectral Data Analysis
This section presents an in situ spectral analysis conducted with the proposed RHIS in order to characterize the hyperspectral reflectance of plastic litter samples (Section 3.1).This analysis aimed to demonstrate that the obtained spectral reflectance of each plastic type was similar to those existing in the literature and, thus, confirm which wavelengths were most-efficient in discriminating between plastic types (Section 3.2).

Spectral Reflectance Dataset
Using the new RHIS and Spectronon Pro software, the reflectance spectra were defined as references to characterize and verify the spectral reflectance of each object category.For this purpose, each hyperspectral data cube image was visualized (Figure 5) and different Regions Of Interest (ROIs) were then selected with various sizes and numbers depending on the size of each object.Large-area objects allowed selecting a large size and/or a large number of ROIs, while small-area objects limited the size and/or the number of ROIs.Each ROI was labeled by its type of plastic or Other.For each waste sample presented in Table 1, the following steps were applied: 1. Manual selection of a Region Of Interest (ROI) with a random size; 2. Computation and plotting of the hyperspectral reflectance spectrum of the selected ROI as the mean value over all pixels in the ROI; 3. Choice of another ROI of the same object present in another acquisition; 4. Return to Step 2, and repeat this process until a significative number of spectra are computed depending on the size of the object.
All selected ROIs represented two possible cases (dry and wet) of the plastic object in seawater.
For example, Figure 5 shows each of the hyperspectral reflectance spectra computed from five ROI extracted from the LDPE plastic object "Blue toothpaste tube".
The so-computed spectral data were quantized by an integer whose maximum value depended on the bit depth.Each hyperspectral reflectance spectrum was then normalized by using the reflectance factor given for each acquisition so that all the reflectance values belonged to the interval [0, 1].Finally, a reference mean hyperspectral reflectance spectrum of each plastic object was calculated as the average of the normalized mean hyperspectral reflectance spectra of the selected ROIs from this object.
For example, Figure 6 displays the reference mean hyperspectral reflectance spectrum (called mean spectra) of the LDPE plastic object "Blue Toothpaste Tube" in the case where it was dry.This figure shows that the different spectra of the same object were close to each other and its reference mean hyperspectral reflectance spectrum can be used as a spectral signature of the type of plastic.Figure 7 displays the reference mean hyperspectral reflectance spectra of different dry objects (green net rope, yellow rope, etc.) of the same plastic type HDPE.This figure shows that the shape of the different spectra was For each waste sample presented in Table 1, the following steps were applied: 1.
Manual selection of a Region Of Interest (ROI) with a random size; 2.
Computation and plotting of the hyperspectral reflectance spectrum of the selected ROI as the mean value over all pixels in the ROI; 3.
Choice of another ROI of the same object present in another acquisition; 4.
Return to Step 2, and repeat this process until a significative number of spectra are computed depending on the size of the object.
All selected ROIs represented two possible cases (dry and wet) of the plastic object in seawater.
For example, Figure 5 shows each of the hyperspectral reflectance spectra computed from five ROI extracted from the LDPE plastic object "Blue toothpaste tube".
The so-computed spectral data were quantized by an integer whose maximum value depended on the bit depth.Each hyperspectral reflectance spectrum was then normalized by using the reflectance factor given for each acquisition so that all the reflectance values belonged to the interval [0, 1].Finally, a reference mean hyperspectral reflectance spectrum of each plastic object was calculated as the average of the normalized mean hyperspectral reflectance spectra of the selected ROIs from this object.
For example, Figure 6 displays the reference mean hyperspectral reflectance spectrum (called mean spectra) of the LDPE plastic object "Blue Toothpaste Tube" in the case where it was dry.This figure shows that the different spectra of the same object were close to each other and its reference mean hyperspectral reflectance spectrum can be used as a spectral signature of the type of plastic.Figure 7 displays the reference mean hyperspectral reflectance spectra of different dry objects (green net rope, yellow rope, etc.) of the same plastic type HDPE.This figure shows that the shape of the different spectra was similar with the absorption features located at the same wavelengths.The level of each spectrum along the reflectance axis can vary depending on the color and opacity of the plastic under consideration.
Remote Sens. 2023, 15, x FOR PEER REVIEW 13 of 26 similar with the absorption features located at the same wavelengths.The level of each spectrum along the reflectance axis can vary depending on the color and opacity of the plastic under consideration.Finally, a dataset of 382 reference mean hyperspectral reflectance spectra was computed (104 for HDPE; 26 for LDPE, 34 for PET, 129 for PP, 18 for PVC, 22 for PS, 13 for PUR, 4 for POM, 9 for ABS, 23 for Other).An overview of this dataset can be found in the Supplementary Materials (Spreadsheet S1).Finally, a dataset of 382 reference mean hyperspectral reflectance spectra was computed (104 for HDPE; 26 for LDPE, 34 for PET, 129 for PP, 18 for PVC, 22 for PS, 13 for PUR, 4 for POM, 9 for ABS, 23 for Other).An overview of this dataset can be found in the Supplementary Materials (Spreadsheet S1).
Remote Sens. 2023, 15, x FOR PEER REVIEW 13 of 26 similar with the absorption features located at the same wavelengths.The level of each spectrum along the reflectance axis can vary depending on the color and opacity of the plastic under consideration.Finally, a dataset of 382 reference mean hyperspectral reflectance spectra was computed (104 for HDPE; 26 for LDPE, 34 for PET, 129 for PP, 18 for PVC, 22 for PS, 13 for PUR, 4 for POM, 9 for ABS, 23 for Other).An overview of this dataset can be found in the Supplementary Materials (Spreadsheet S1).

Comparison with the State-of-the-Art
In this section, the reference mean hyperspectral reflectance spectra presented in Section 3.1 are analyzed to identify the absorption feature wavelengths for each type of plastic and are compared to those of the literature.
Figures 8 and 9 present the reflectance spectra of six objects whose plastic type was HDPE, LDPE, PET, PP, PVC, and PS, respectively.Two spectra are presented for each object depending on whether it was dry or wet.In each case, the absorption feature wavelengths identified in the literature are highlighted in blue to be compared to our reflectance spectra.The objects analyzed in this study are listed below: 1.
HDPE plastic type-green net rope: The dry HDPE had five visible absorption features at wavelengths of 1222 nm, 1400 nm, 1425 nm, 1445 nm, and 1550 nm [3,8].The most-important absorption feature is at a wavelength of 1222 nm, which is in close correspondence with the work of Tasseron et al. [3].Similar results also appear in Figure 7 for dry objects.The wet green net rope of the HDPE plastic type had an attenuated spectral reflectance.However, the main absorption feature (1222 nm) remained visible.

2.
LDPE plastic type-blue toothpaste tube: The dry LDPE plastic had two absorption features at wavelengths of 1222 nm and 1400 nm, which is in close correspondence with Tasseron et al. [3].Similar results also appear in Figure 6 for dry objects.The reflectance spectrum of the wet blue toothpaste tube of LDPE plastic was also attenuated.The two main absorption features of polyethene plastics (HDPE and LDPE) found in this study were centered on wavelengths of 1222 nm and 1400 nm, which is very similar to the absorption features described by Tasseron et al. [3].

3.
PET plastic type-transparent tomato packaging: The dry PET spectral reflectance decreased with increasing wavelengths.The wet semi-transparent packaging of PET plastic type had an attenuated reflectance spectrum.The spectral shape of the transparent PET type found by Tasseron et al. [3] was similar to the spectral shape found in this study.No absorption feature can be highlighted for this type of plastic, whose reflectance was further reduced due to the transparency of the object, which reflected a small amount of light.4.
PP plastic type-red rope: The dry PP had four visible absorption features at wavelengths of 1200 nm, 1222 nm, 1405 nm, and 1650 nm.The most-important absorption features were usually at wavelengths of 1222 nm, 1405 nm, and 1650 nm.The absorption features of PPplastics found in this study were centered on wavelengths of 1222 nm and 1405 nm, which is in close correspondence with Tasseron et al. [3] and Moshtaghi et al. [6].The wet red rope plastic of the PP plastic type had an attenuated reflectance spectrum.

5.
PVC plastic type-semi-transparent packaging: The dry PVC had two small absorption features at wavelengths of 1200-1202 nm and 1400-1405 nm.The wet semitransparent packaging of the PVC plastic type had an attenuated reflectance spectra, but it was similar to the dry spectra.The transparency of this object generated lowlevel spectra along the reflectance axis since the light rays were transmitted through the material to be partly absorbed by the water.Although this specific type of plastic packaging tends to float on water, other PVC objects are rarely found floating due to the high density of this polymer relative to water and was, therefore, not considered by Tasseron et al. [3].6.
PS plastic type-pink parfum cap: The dry PS had three important absorption features at wavelengths of 1148 nm, 1212 nm, and 1420 nm.The wet pink parfum cap of the PS plastic type had an attenuated reflectance spectrum.Polystyrene was characterized by two distinct absorption features at 1150 and 1450 nm by Tasseron et al. [3].This study using the new RHIS revealed the presence of absorption features in the reference mean hyperspectral reflectance spectra of different plastic types in the NIR- This study using the new RHIS revealed the presence of absorption features in the reference mean hyperspectral reflectance spectra of different plastic types in the NIR-SWIR range centered on wavelengths of: 1148 nm, 1200 nm, 1212 nm, 1222 nm, 1400 nm, 1405 nm, 1420 nm, 1425 nm, 1445 nm, 1550 nm, and 1650 nm.These results, which were in correspondence with the results obtained by Tasseron et al. [3] and Moshtaghi et al. [6], confirmed that the RHIS was able to characterize each plastic type by a spectral signature.

Plastic Litter Recognition Using Machine Learning
This section aims to show that the new RHIS is able to automatically recognize the plastic type of the observed objects by hyperspectral image analysis.To evaluate the recognition performances, it was necessary to dispose of a ground-truth where the category of each analyzed data is known.From the benchmark database presented in Section 2.3, a dataset of image patches was, thus, built, where each patch was labeled by a class of plastic or other (Section 4.1).This dataset was then used to apply classical supervised machine learning methods in order to classify the images of waste samples (Section 4.2).
All calculations were performed with Matlab ® R2021b and a Windows 10TM computer with an Intel(R) Core(TM) i9-9880H CPU with 2.30 GHz, 32 GB RAM, and an Nvidia ® Quadro RTX 3000 graphics card with 16 GB GDDR5X memory.

Waste Image Patch Dataset
To build the waste image patch dataset, hyperspectral data cubes were extracted from the benchmark image database presented in Section 2.3.Manually labelling all the pixels of these images was a laborious task, which was not easily feasible.Each hyperspectral data cube was, therefore, divided into patches of size 16 × 16 × 143.The small size of certain plastic objects (Figure 10) led us to choose this patch size.Representative patches of each type of waste were then manually selected from different objects present in the images.Each selected patch was finally manually labeled according to its class, namely its type of plastic or other.SWIR range centered on wavelengths of: 1148 nm, 1200 nm, 1212 nm, 1222 nm, 1400 nm, 1405 nm, 1420 nm, 1425 nm, 1445 nm, 1550 nm, and 1650 nm.These results, which were in correspondence with the results obtained by Tasseron et al. [3] and Moshtaghi et al. [6], confirmed that the RHIS was able to characterize each plastic type by a spectral signature.

Plastic Litter Recognition using Machine Learning
This section aims to show that the new RHIS is able to automatically recognize the plastic type of the observed objects by hyperspectral image analysis.To evaluate the recognition performances, it was necessary to dispose of a ground-truth where the category of each analyzed data is known.From the benchmark database presented in Section 2.3, a dataset of image patches was, thus, built, where each patch was labeled by a class of plastic or other (Section 4.1).This dataset was then used to apply classical supervised machine learning methods in order to classify the images of waste samples (Section 4.2).
All calculations were performed with Matlab ® R2021b and a Windows 10TM computer with an Intel(R) Core(TM) i9-9880H CPU with 2.30 GHz, 32 GB RAM, and an Nvidia ® Quadro RTX 3000 graphics card with 16 GB GDDR5X memory.

Waste Image Patch Dataset
To build the waste image patch dataset, hyperspectral data cubes were extracted from the benchmark image database presented in Section 2.3.Manually labelling all the pixels of these images was a laborious task, which was not easily feasible.Each hyperspectral data cube was, therefore, divided into patches of size 16 × 16 × 143.The small size of certain plastic objects (Figure 10) led us to choose this patch size.Representative patches of each type of waste were then manually selected from different objects present in the images.Each selected patch was finally manually labeled according to its class, namely its type of plastic or other.In order to assess the machine learning model performances, two subsets of image patches representative of the ten classes were created: a training subset for the model learning and a testing subset for the model evaluation.The training and testing image patches were extracted from different original images so that they were as independent as possible and represented a realistic situation in seawater with all its challenges.
For the training and the testing subsets, a total of 788 and 312 image patches of size 16 × 16 × 143 were, respectively, selected from the images of the benchmark database.Table 2 shows, in the first column, the different considered classes (nine plastic types and one class "other").The number of patches that were used for training and testing are shown in the second and third column, respectively.This number depended on the number and size of the available objects for each type of plastic.An overview of the representative patches counted per plastic object can be found in the Supplementary Materials (File S1).

Waste Image Patch Classification
In this section, three well-known supervised machine learning methods are applied to classify the image patches of the dataset presented in the previous section.
As can be observed in Table 2, the number of examples for each class varied.Some classes were represented with a small number of samples for classification, and the difference between the number of patches for the plastic types PUR, POM, PS, and ABS, and the Other types was significant.A class imbalance usually makes it harder to identify (and, hence, classify) a minority class.In our case, the plastic types of PUR, POM, PS, and ABS were minority classes.Imbalanced classification is a challenge for predictive modeling because most machine learning algorithms used for classification are designed around the assumption of an equal number of examples for each class.To take these limitations into account, three classical supervised machine learning methods were chosen: 1.
For each method, there are parameters to be optimized to determine the best tuning of the classification model (classifier) by using the training image patches (learning stage) and then to evaluate its performance with the testing image patches (prediction stage).To fine-tune the classifier parameters with the challenge of class imbalance, Hyper-Parameter (HP) optimization methods (Bayesian [34,35] and random research [36]) offer the possibility to automatically select a classification model with an optimized tuning.
For these experiments, the mean hyperspectral reflectance associated with each patch of size (16 × 16 × 143) was calculated as the mean value over all its 256 (16 × 16) pixels, leading to a vector of size 1 × 143 (number of spectral bands).
In this study, high-definition hyperspectral images were used to classify patches according to different plastic types.However, the high dimension of hyperspectral images often causes computational complexity and the curse of dimensionality.In many cases, it is not necessary to process the hyperspectral information of all spectral bands since many spectral bands are highly correlated.Thus, it is required to remove redundant spectral bands in order to decrease the computational complexity and improve classification performance.Among the many dimensionality reduction methods used for this purpose, Principal Component Analysis (PCA) is a well-known preprocessing step in hyperspectral image analysis [37,38].PCA linearly transforms the initial feature space, whose axes correspond to the input spectral bands, and generates a new feature subspace, where the axes are called principal components, in order to remove redundant dimensions.
The main following stages are, therefore, proposed for the plastic classification: 1. Dimensionality reduction by feature extraction [38]: PCA was applied on the training subset, and different dimensions of the resulting feature subspace were considered for the next stage.

2.
Learning stage: The KNN, SVM, and ANN classification models were applied with optimization methods (Bayesian and random research) of hyper-parameter tuning to determine the best validation accuracy.In order to protect against overfitting, a five-fold cross-validation was considered.This scheme partitions the training subset into five disjoint folds.Each fold was used once as a validation fold, and the others formed a set of training folds.For each validation fold, the classification model was trained using the training folds, and the classification accuracy was assessed using the validation fold.The average accuracy was then calculated over all the folds and was used to optimize the tunning of the classification model parameters.These hyper-parameters, which are presented in Table 3, were determined by an automatic hyper-parameter optimization using two methods: Bayesian [34,35] and random research [36] optimization.The final validation accuracy gave a good estimate of the predictive accuracy of the classifier, which was used in the next stage with the full training subset, excluding any data reserved for the testing subset.

3.
Prediction stage: The trained models obtained during the previous stage were then applied to the testing image patches, and the overall test accuracy of the classifier was determined.The testing subset here was independent of the training subset.
Based on these stages, Table 4 presents the top ten classifiers that were tested with different dimensions of the feature subspace obtained by PCA (96, 64, and 48) and with the two HP optimization methods.The first column of this table gives the name of the tested classifier; the second one indicates the dimension of the feature subspace; the third column gives the name of the used HP optimization method.The goal of the optimization algorithm is to find a combination of HP values that minimizes an objective function, here the classification error rate.To find this combination, the iteration number of the used algorithm was fixed to 120.Table 4 also describes the determined optimized hyperparameters in the fourth column.This column is divided into several cells, whose number depends on the classification method.For each tested classifier, the validation accuracy computed with the training subset and the test accuracy computed with the testing subset appear in the fifth and sixth columns, respectively.Accuracy is given as the percentage of patches (training or testing) that were correctly classified.Although the validation accuracy was not the highest for this classifier (89.7 %), it was very close to the test accuracy despite the imbalanced classification problem.This result showed that the validation accuracy provided a good estimate for the model performance on new data compared to the training data.The top validation accuracy was obtained with Model5-PCA64-KNN, which uses the 64 principal components of PCA, but this classifier achieved a test accuracy of 87.2% and, therefore, gave a lower performance.
Figure 11 gives the test confusion matrix, which details the performances per class obtained for Model2-PCA48-KNN on the testing subset.Its rows correspond to the predicted class and its columns to the true class.The diagonal cells correspond to patches that were correctly classified, and the off-diagonal cells correspond to incorrectly classified patches.Both the percentage of patches and the number of patches (in brackets) are shown in each cell.This matrix shows that the testing patches corresponding to the four minority classes (PUR, POM, PS, and ABS) were well classified (100% of test accuracy) despite the imbalanced classification issue.The class of other materials that represents non-plastic objects corresponded also to 100% accuracy.This result proved that the classifier was able to predict whether an object was a plastic or not.The testing patches with the lowest accuracy belonged to the PP and PVC classes.There were 16.7% of PP samples assigned to the PET class and 8.3% to the HDPE class.There were 14.5% of PVC samples assigned to the class other.These misclassification rates could be explained by the presence of wet samples with distorted spectra, by the presence of transparent and black plastics with poor reflectance, and by the diversity of samples in the class Other.The HDPE and LDPE classes can be confused since 13.5% of the LDPE samples were assigned to the HDPE class.These two types of plastic are based on the same polymer.Finally, the accuracy obtained for each of the other class was greater than 90%, which is a very good performance with a classical classifier.These experimental results highlighted that the RHIS provided a very satisfactory wet and dry plastic recognition performance by using classical supervised machine learning methods such as the Model2-PCA48-KNN classifier.With more training data and more sophisticated classification approaches such as deep learning approaches, this performance can obviously be further improved for the detection and identification of plastic litter [22,30].
The RHIS is the first embedded hyperspectral imaging system that observes the aquatic environment in the near-field and automatically quantifies and qualifies polymers of floating plastic litter with accuracy.

Conclusions
This paper addressed the problem of plastic litter pollution in the aquatic environment resulting from human activity.The observation and quantification of this waste by remote sensing at different scales is crucial to determine its exact nature and fight against this pollution.Hyperspectral imaging is emerging as an appropriate technology to characterize, detect, and identify floated plastic waste in terms of shape and type of polymers.In this paper, a new remote hyperspectral imaging system, embedded on an unmanned aquatic drone for plastic detection and identification in coastal environment, was presented.
This new hyperspectral imaging system, named the ROV-ULCO, was designed around a hyperspectral camera that captures reflectance spectra in the NIR to SWIR range to discriminate different types of plastic.It works in the near-field for the observation of floating litter (plastic and non-plastic type).The first obtained results were very encouraging and proved the marine litter automatic recognition capability using a simple supervised machine learning method.Indeed, these results reached an overall accuracy close to 90% with a K-nearest neighbors classifier associated with a principal component analysis These experimental results highlighted that the RHIS provided a very satisfactory wet and dry plastic recognition performance by using classical supervised machine learning methods such as the Model2-PCA48-KNN classifier.With more training data and more sophisticated classification approaches such as deep learning approaches, this performance can obviously be further improved for the detection and identification of plastic litter [22,30].
The RHIS is the first embedded hyperspectral imaging system that observes the aquatic environment in the near-field and automatically quantifies and qualifies polymers of floating plastic litter with accuracy.

Conclusions
This paper addressed the problem of plastic litter pollution in the aquatic environment resulting from human activity.The observation and quantification of this waste by remote sensing at different scales is crucial to determine its exact nature and fight against this pollution.Hyperspectral imaging is emerging as an appropriate technology to characterize, detect, and identify floated plastic waste in terms of shape and type of polymers.In this paper, a new remote hyperspectral imaging system, embedded on an unmanned aquatic drone for plastic detection and identification in coastal environment, was presented.
This new hyperspectral imaging system, named the ROV-ULCO, was designed around a hyperspectral camera that captures reflectance spectra in the NIR to SWIR range to discriminate different types of plastic.It works in the near-field for the observation of floating litter (plastic and non-plastic type).The first obtained results were very encouraging and proved the marine litter automatic recognition capability using a simple supervised machine learning method.Indeed, these results reached an overall accuracy close to 90% with a K-nearest neighbors classifier associated with a principal component analysis for the classification of nine plastic types and their distinction with a tenth class of non-plastic objects.
This study showed that the new hyperspectral imaging system, the ROV-ULCO, is a promising approach to detect and identify plastic waste in aquatic environments.It can be improved by focusing on challenges such as transparent and black plastic waste or wet and submerged plastic waste, which are more difficult to recognize [39].From our perspective, the databases will be enlarged to add these plastic types with more representative samples, and classification approaches based on artificial intelligence will be applied in order to improve the performance of this original system.In addition, our prototype can be equipped with other optical or radar sensors to meet these challenges, but also to make it autonomous so that it can automatically navigate to areas where plastic waste is present.

Figure 1 .
Figure 1.Conceptual framework for remote marine litter detection from[22] with different remote devices (satellite, airborne, drone, etc.) and with the new system (ROV-ULCO) proposed in this paper.

Figure 2 .
Figure 2. ROV-ULCO system: RHIS embedded on a UAD, namely the Jellyfishbot, which can be connected with a net to collect macroplastics or microplastics in different water surfaces.

•
Two inflatable boat floaters; • Two batteries (each inside a removable waterproof case); • An illumination device of halogen lamps; • Protections against solar illumination; • A long-range WiFi antenna to communicate remotely; • A waterproof box that contains the following inside components (Figure 3b): o A line-scan hyperspectral camera (Resonon PIKA-NIR-320) with a 12 mm focal length objective lens; o An optical mirror system; o An industrial Central Processing Unit (CPU) as the on-board computer; o An Arduino unit that controls two temperature sensors and a water velocity sensor via an integrated board.

Figure 2 .
Figure 2. ROV-ULCO system: RHIS embedded on a UAD, namely the Jellyfishbot, which can be connected with a net to collect macroplastics or microplastics in different water surfaces.

Figure 4 .
Figure 4. Experimental setup of the new remote hyperspectral imaging system.(a) Hyperspectral camera setup in controlled laboratory environment with WiFi antenna; (b) black container under the RHIS on the linear translation stage.

Figure 4 .
Figure 4. Experimental setup of the new remote hyperspectral imaging system.(a) Hyperspectral camera setup in controlled laboratory environment with WiFi antenna; (b) black container under the RHIS on the linear translation stage.

26 Figure 5 .
Figure 5. Examples of hyperspectral reflectance computed from five ROIs (Spec. 1 to 5) of the LDPE plastic object "Blue Toothpaste Tube".(a) Extraction of a region of interest from a hyperspectral image represented in false color; (b) non-normalized mean hyperspectral reflectance of different ROIs (in orange, the mean spectral reflectance over all pixels of the selected ROI).

Figure 5 .
Figure 5. Examples of hyperspectral reflectance computed from five ROIs (Spec. 1 to 5) of the LDPE plastic object "Blue Toothpaste Tube".(a) Extraction of a region of interest from a hyperspectral image represented in false color; (b) non-normalized mean hyperspectral reflectance of different ROIs (in orange, the mean spectral reflectance over all pixels of the selected ROI).

Figure 6 .
Figure 6.Examples of the reference mean hyperspectral reflectance of the tube and the cap of the LDPE plastic object "Blue Toothpaste Tube".

Figure 7 .
Figure 7. Examples of the reference mean hyperspectral reflectance of different HDPE plastic objects.

Figure 6 .
Figure 6.Examples of the reference mean hyperspectral reflectance of the tube and the cap of the LDPE plastic object "Blue Toothpaste Tube".

Figure 6 .
Figure 6.Examples of the reference mean hyperspectral reflectance of the tube and the cap of the LDPE plastic object "Blue Toothpaste Tube".

Figure 7 .
Figure 7. Examples of the reference mean hyperspectral reflectance of different HDPE plastic objects.

Figure 7 .
Figure 7. Examples of the reference mean hyperspectral reflectance of different HDPE plastic objects.

1 -Figure 8 .
Figure 8. Mean hyperspectral reflectance of plastic objects (HDPE, LDPE, and PET) depending on whether they were dry or wet and their corresponding absorption features.Figure 8. Mean hyperspectral reflectance of plastic objects (HDPE, LDPE, and PET) depending on whether they were dry or wet and their corresponding absorption features.

Figure 8 .
Figure 8. Mean hyperspectral reflectance of plastic objects (HDPE, LDPE, and PET) depending on whether they were dry or wet and their corresponding absorption features.Figure 8. Mean hyperspectral reflectance of plastic objects (HDPE, LDPE, and PET) depending on whether they were dry or wet and their corresponding absorption features.

4 -Figure 9 .
Figure 9. Mean hyperspectral reflectance of plastic objects (PP, PVC, and PS) depending on whether they are dry or wet and their corresponding absorption features.

Figure 9 .
Figure 9. Mean hyperspectral reflectance of plastic objects (PP, PVC, and PS) depending on whether they are dry or wet and their corresponding absorption features.

Figure 10 .
Figure 10.Examples of RGB and hyperspectral images of plastic objects.(a) RGB image of plastic objects; (b) false color image using three spectral bands of wavelength values of 1575.4 nm, 1257.1 nm, and 1100.0 nm; (c) selection of the image patches.

Figure 10 .
Figure 10.Examples of RGB and hyperspectral images of plastic objects.(a) RGB image of plastic objects; (b) false color image using three spectral bands of wavelength values of 1575.4 nm, 1257.1 nm, and 1100.0 nm; (c) selection of the image patches.

Table 1 .
Examples of waste samples for each plastic type.

Table 2 .
Plastic type and "other" classes with the number of used patches for the training and testing image patch subsets.

Table 4
first shows that the KNN classification model outperformed the SVM and ANN ones in terms of validation accuracy and test accuracy.This table also shows that, for the KNN model, PCA drastically increased the test accuracy.The KNN model with the highest test accuracy (89.1 %) was Model2-PCA48-KNN, which uses the 48 principal components of PCA.The optimized HP were determined with a random search method.