Next Article in Journal
Chemical Vapor Transport Synthesis of Fibrous Red Phosphorus Crystal as Anodes for Lithium-Ion Batteries
Next Article in Special Issue
Synthesis of Ni-Doped Tremolite Fibers to Help Clarify the Aetiology of the Cytotoxic Outcome of Asbestos
Previous Article in Journal
Single-Walled vs. Multi-Walled Carbon Nanotubes: Influence of Physico-Chemical Properties on Toxicogenomics Responses in Mouse Lungs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning to Predict the Adsorption Capacity of Microplastics

by
Gonzalo Astray
1,
Anton Soria-Lopez
1,
Enrique Barreiro
2,
Juan Carlos Mejuto
1 and
Antonio Cid-Samamed
1,*
1
Universidade de Vigo, Departamento de Química Física, Facultade de Ciencias, 32004 Ourense, Spain
2
Universidade de Vigo, Departamento de Informática, Escola Superior de Enxeñaría Informática, 32004 Ourense, Spain
*
Author to whom correspondence should be addressed.
Nanomaterials 2023, 13(6), 1061; https://doi.org/10.3390/nano13061061
Submission received: 14 February 2023 / Revised: 10 March 2023 / Accepted: 11 March 2023 / Published: 15 March 2023

Abstract

:
Nowadays, there is an extensive production and use of plastic materials for different industrial activities. These plastics, either from their primary production sources or through their own degradation processes, can contaminate ecosystems with micro- and nanoplastics. Once in the aquatic environment, these microplastics can be the basis for the adsorption of chemical pollutants, favoring that these chemical pollutants disperse more quickly in the environment and can affect living beings. Due to the lack of information on adsorption, three machine learning models (random forest, support vector machine, and artificial neural network) were developed to predict different microplastic/water partition coefficients (log Kd) using two different approximations (based on the number of input variables). The best-selected machine learning models present, in general, correlation coefficients above 0.92 in the query phase, which indicates that these types of models could be used for the rapid estimation of the absorption of organic contaminants on microplastics.

1. Introduction

Since the appearance of plastics, their production has grown exponentially in recent decades, and due to their versatility, they are used in different fields, such as packaging, building, or electronic industries, among others [1]. The use of microbeads or nanobeads based on plastic polymers (e.g., Bisphenol-A diglycidyl ether, polyetheramine, and Polyvinyl alcohol) to functionalize and decorate CNTs has been reported to improve their shielding against electromagnetic interference, and the electrical and mechanical properties of elastic, functional composites can also be improved by interlacing the beaded and coated fibers into a smart tissue [2,3]. The transformation of plastics into microplastics (MPs) and then into nanoplastics (NPs) through fragmentation makes the presence of micro- and nanoplastics (MNPs) in both water sources and our planet’s agroecosystems a worldwide concern [4,5,6,7]. In this sense, and as reported by Matthews et al. (2021), microplastics are plastic fragments of less than 5 mm, and the most commonly accepted size for nanoplastics is that falling within the range of 1–1000 nm [8].
The cycle from the production of plastics to their entry into the environment includes different stages, as reported by Woods et al. (2021) [9]: production for textile manufacturing and use, tires use, or packaging production, among others [10]. Additionally, there are different pollution sources by MPs; due to this, they can be differentiated between primary and secondary sources [11]. Regarding the primary MPs, they are commonly generated during the manufacturing of different products or the fabrication of microbeads or microfibers; while the secondary ones are formed from larger plastic litters due to degradation processes by external factors (chemical, physical or biological) [11].
Once released into the aquatic environment, it has been reported that microplastics can be vectors for persistent organic pollutants (POPs) [12,13]. Chemical contaminants can be captured on the surfaces of microplastics and nanoplastics due to, for example, their surface charge [14].
Currently, micro- and nanoplastics can be found in terrestrial and aquatic environments, are able to affect a large number of organisms [8], and can be considered Trojan horses, as reported by Katsumiti et al. (2021) and Hu et al. (2022) [15,16].
The adsorption capacity of organic polluting substances among microplastics and water could be understood as an equilibrium partitioning coefficient (Kd) [17]. Since, according to the authors [17], the absorption data are nowadays limited, it would be interesting to dispose a method to predict the Kd values under various conditions. For this reason, the application of the quantitative structure–property relationship (QSPR) together with machine learning (ML) techniques would be a possible alternative for the determination of these values.
Machine learning is one of the subsets that includes artificial intelligence (AI) which consists of training machines that are capable of imitating the intellectual ability of the human being based on the knowledge and experience learned [18]. Most existing machine-learning methods involve the use of a set of training cases, where each case contains input and output labels [19]. Then, the method tries to predict the output labels using the inputs from the training data [19]. Due to these characteristics, this type of supervised methods are suitable for the predictive modeling required in this research. The three supervised learning models used in this research will be briefly presented below. They were chosen because they have the common objective of building models that can generalize patterns from input data.

1.1. Random Forest

The first of the selected ML models is a random forest (RF) model, which is composed of decision trees that can be applied for regression and classification purposes [20]. These decision trees can be considered as among the main methods for solving real problems [21]. Random forests are an ensemble machine learning method that has been proposed by Breiman (2001) [22] that can overcome the instability and overfitting problems that arise when only using a single decision tree [23]. When working on regression and classification mode, the random forest uses the bootstrap samples of the original training data to develop and train each decision tree, generating a series of decision trees [24]. Therefore, the random forest involves the development of different decision trees using random subsets of the original training data [23]. Each decision tree in RF starts at a root node and recursively splits into two subnodes based on minimizing the mean square error [25]. The predictions of each individual tree are aggregated to form the final prediction. For regression tasks with quantitative data, the predictions are averaged, whereas for classification tasks with qualitative data, a voting process is carried out [20,21,26]. RF is generally considered a robust method that can achieve good results compared to different regression algorithms [27].
Random forest models can be used in different research fields, such as: cloud computing to develop a DDoS-attack-detection method [20], chemistry to determine molecular electronic transitions [25], or in farming to predict regional and local-scale wheat yield [28], among others.

1.2. Support Vector Machine

The second of the ML models developed in this research is the support vector machine (SVM). According to Geppert et al. (2010) [29], this method became popular during the 1990s based on the work carried out by Cortes and Vapnik (1995) [30] and, as reported by Rodríguez-Pérez et al. (2017) [31], has become more and more popular. A support vector machine is based on statistical learning theory [32] and can be used for regression and classification tasks [32,33]. According to Houssein et al. [33], the support vector machine can work with linear and nonlinear problems. When working in classification mode, the main aim of an SVM is to identify a hyperplane within an N-dimensional space to clearly classify the data points [34]. As Sareminia (2022) reports, different hyperplanes could be chosen for two different data classes, but the aim of the SVM is to locate the plane with the largest margin, which is the maximum gap between both classes’ data points [34]. Whether the problem is linear or nonlinear, the support vector machine separates the data into two classes by mapping the information into spaces with dimensions greater than two [35]. On the other hand, Rodríguez-Pérez et al. (2017) [31] report that SVM can also be used in the regression model (support vector regression, SVR) to predict numerical property values [36,37]. In this type of SVM model, a different function is derived from the training data for predicting numerical values [31].
Support vector machines can be applied in different fields such as chemistry to identify the polar liquids [38], energy storage to self-discharge prediction in batteries of lithium-ion [39], or in medicine to diagnose breast cancer [40], inter alia.

1.3. Artificial Neural Networks

The last ML models carried out in this research were models based on artificial neural networks (ANNs). An artificial neural network is a well-documented/known artificial intelligence model [41] that can be defined as a mathematical model which is inspired by the behavior of biological neurons [21,41]. As reported by Paturi et al. (2022) [42], McCulloch and Pitts (1943) [43] were the first who could explain the logical relationship that exists between the neural events of the nervous system. This imitation of the behavior of biological neurons can be learned through a process of backpropagation [44].
ANN is a powerful tool to find relationships between data, in this case, input and output data [41], and can be used to solve complex problems in optimization, clustering, or prediction, among others [45]. ANN is formed by units (neurons) that are organized into different layers [44]. A neuron performs two functions: collecting the inputs and producing an output [45]. An ANN architecture is usually made up of three elements, a first layer (named the input layer), a second layer (known as the hidden layer), and a final layer (named the output layer) [46]. One of the existing neural network types, the multi-layer perceptron network (MLP), possesses one or more hidden layers [42,46], and, in principle, and according to Saikia et al. (2020) [47], it is possible to approximate any continuous function with only one hidden layer [48]. According to [47], an artificial neural network is a popular ML procedure due to its capacity for complex nonlinear function modeling. The neurons number located in the intermediate layer can be established by trial and error [46,49].
To find the relationship between the input and the output data, the artificial neural network must necessarily be subjected to a training process using the database containing both input and output data [41]. Following Niazkar and Niazkar (2020) [41], the first layer contains neurons associated with the input vector; a hidden layer connects the input neurons and the output neuron/neurons and turns the input data into the correspondent output data. Finally, the output layer presents the neuron/neurons associated with the output vector. Each processing neuron is typically modeled as a computational unit that takes the input value and multiplies it by the learned importance of the connection, also called weight, and the result and bias are processed by an activation function and provide an output in the neuron [45]. In fact, there are different activation functions, such as sigmoid or Gaussian, among others [46].
Finally, artificial neural networks can be used in engineering to predict the building construction time and cost [50], om water management to model and predict the amount of salt removed by the capacitive deionization method [51], or in biotechnology to optimize the parameters in Ganoderma lucidum residue aerobic composting process [52].
Therefore, this research aimed to apply machine learning models (RF, SVM, and ANN) to predict the adsorption capacity of MPs ((polyethylene (PE), polystyrene (PS), polypropylene (PP)) in different waters using different configurations of input variables (molecular mass (Mw), n-octanol/water distribution coefficient under special pH condition (log D), and other quantum chemical descriptors) obtained from the literature [17]. These computational models will enable the quick prediction of the adsorption capacity of organic pollutants onto these three types of microplastics in water environments.

2. Materials and Methods

2.1. Experimental Data Used

The data used for the development of the different ML models were extracted from the work developed by Li et al. (2020) [17]. Li et al. (2020) [17] also used different articles reported in the literature to obtain data. These articles can be consulted in Table 2 of the research paper of Li et al. (2020).
In their study and accompanying supplementary material, Li et al. (2020) provided: (i) the n-octanol/water distribution coefficient under special pH conditions (log D); (ii) the molecular mass (Mw); and (iii) six different quantum chemical descriptors that allow the modeling of the microplastic/water partition coefficients (log Kd) for diverse organics between and polyethylene/seawater–freshwater–pure water, polystyrene/seawater, and polypropylene/seawater [17]. The quantum chemical descriptors calculated by Li et al. (2020) were: (i) molecular volume (V′); (ii) the most negative atomic charge (q); (iii) the most positive atomic charge on the H atom (qH+); (iv) the ratio of average molecular polarizability and molecular volume (π) and the covalent; (v) basicity (εβ); and (vi) acidity (εα).
In the present research work, two approximations were carried out. The first is using the same variables that the researchers used to develop their models (Type 1) [17]. On the other hand, due to the authors’ data having 8 different input variables, models that included the maximum number of these variables (Type 2) were developed to improve the previous models (Type 1). Table 1 shows the variables selected for each selected model.
The database was divided into three datasets. In this sense, the cases used by Li et al. (2020) to develop the models were used to generate two groups, a training group to elaborate different ML models and another group, the validation group, to select the best model (considering to the RMSE value in the validation phase). The query group (the cases reported by Li et al. (2020) as test cases in their Table 2) were used to check the adjustments provided by the different ML models.

2.2. Models Implemented

2.2.1. Random Forest Models

Random forest model have been successfully applicated in fields related to this research, for example, to identify and monitor different microplastics in environmental samples [53]. Hufnagl et al. (2019) developed a methodology to discriminate five different polymers (polyethylene, poly(methyl methacrylate), polypropylene, polystyrene, and polyacrylonitrile) and determine their abundance and size distribution [53]. Later, some previous authors extended the previous research to develop a model capable of differentiating more than 20 types of polymers [54].
The RF models (Figure 1A) were carried out using different parameter combinations. The following parameters were studied: the number of trees (1 to 100 using 99 steps in linear scale), maximum depth (1 to 100 using 99 steps in linear scale), and prepruning (false or true). All models were developed using the least square criterion.

2.2.2. Support Vector Machine Models

Support vector machine models have also been successfully used in related fields. An example of this is the research carried out by Yan et al. (2022) [59]; the aim was to develop an ensemble machine learning method capable of classifying and identifying MPs by attenuated total reflection–Fourier transform infrared spectroscopy (ATR-FTIR) data. On the other hand, Bifano et al. (2022) [60] developed a method based on a support vector machine to detect polypropylene and polyolefin in water using electrical impedance spectroscopy.
In the research presented in this article, the SVM models (Figure 1B) were implemented using the LibSVM learner developed by [61,62]. The following parameters combinations were studied: the SVM type (ε-SVR or ν-SVR); γ was studied between ≈2−20 and 28 using 28 steps in linear or logarithmic scale; and C between ≈ 2−10 and 220 using 30 steps in linear or logarithmic scale (SVM and SVMlog). These values are an extension of the proposed values of Hsu et al. (2016) [63]. In addition to using the database in their real-scale, they were also normalized in the interval [−1,1] (first just normalizing the input variables (SVMn and SVMn log) and then normalizing the input and the output variables (SVMn2 and SVMn2 log). The normalization was applied to the training input data, and later applied to the other phases. After the model selection, the output data were de-normalized to allow the real-scale comparison between all developed models

2.2.3. Artificial Neural Network Models

Artificial neural network models have been used to categorize microplastic contamination in the soil using infrared spectroscopy [64]. On the other hand, ANN has also been used successfully to determine the sorption capacity of heavy metal ions onto microplastics [65]. In this sense, Guo and Wang (2021) developed an ANN model using data from the literature and were able to determine the sorption capacity of different heavy metal ions onto microplastics in global environments with correlation coefficients greater than 0.92.
In the research presented in this article, the ANNs (Figure 1C) were developed with one single hidden layer. The hidden neurons were analyzed in the range between 1 and 2n+1, where n is the input neurons number. The training cycles were studied between 1 and 131072 using 17 steps in linear or logarithmic scale (ANNlin and ANNlog). In addition, the decay was studied in mode true or false. The neural net operator to develop the ANN models scaled the values between −1 and 1 [66].

2.2.4. Statistics Used to Analyze the Models

Different statistical parameters were used to evaluate the ML models implemented in this research. In this sense, the correlation coefficient (r), the root mean square error (RMSE), and the mean absolute percentage error (MAPE, expressed in %) were calculated (for training, validation, and query phases).
The best model for each ML approach was chosen considering the root mean square error for the validation phase. Once each best ML model was chosen, they were compared using the query data.

2.2.5. Equipment and Software Used for the Development of the Models

The developed ML models were implemented in two computers; the first, an Intel® Core™ i9-10900 at 2.80 GHz with 64GB RAM and Windows 10 Pro 21H1, and the second, an AMD Ryzen 7 3700X 8-Core at 3.60 GHz with 32 GB RAM and Windows 11 Pro 21H2.
The data used in this research were collected from Li et al. (2020) [17] using Microsoft Excel 2016 from Microsoft Office Professional Plus 2016. The ML models (RF, SVM, and ANN) were developed using two versions of RapidMiner Studio 9.10.001 and 9.10.011 software (Educational and a free). Figures were drawn with Microsoft PowerPoint 2016 from Microsoft Office Professional Plus 2016 and SigmaPlot v. 13.0 from Systat Software, Inc. (Palo Alto, Santa Clara, CA, USA).

3. Results and Discussion

The following sections analyzed the results obtained by the different machine learning methods for each of the analyzed assumptions.

3.1. ML Models Using Input Variables Type 1

Table 2 shows the adjustments obtained for the selected machine learning models to predict log Kd, developed with the same variable combination used by Li et al. (2020) [17].
The first models (PE/seawater) correspond with ML models to predict the adsorption capacity for polyethylene in seawater. In this case, the three best-selected models (each according to their RMSE value for the validation phase) can be seen. The model with the best adjustments is the artificial neural network (ANNlog) model (0.236), followed by the support vector machine (SVMn2 log) model (0.248), and finally, the random forest model (0.380). As can be seen, the three models present very high correlation coefficients for the validation phase, equal to or greater than 0.988; in addition, the mean absolute percentage error remains low, between 4.42% and 7.48%.
The good adjustments shown in the validation phase can also be observed in the training phase, where the values of RMSE remain similar to those of the validation phase, except for the random forest model, where the RMSE and MAPE values grow to 0.525 and 18.67%, respectively. As can be seen in the case of the query phase, the model that provided the best result in the validation and training phases, the ANN model, presents the worst results in terms of RMSE and MAPE (0.561 and 23.33%, respectively), despite it maintaining a high coefficient of correlation (0.979). The other two models, the support vector machine and the random forest model present slightly higher errors in terms of RMSE than those presented in the validation phase (0.357 and 0.523, respectively).
Given these results (Table 2), it can be said that the three models generally show a good performance, although, for the query phase, the errors slightly increase. Despite this, the errors, in terms of RMSE, remain below the test error reported by Li et al. (2020) (0.752) for the model developed with these three input variables (log D, εα, and εβ).
The second group of models (PE/freshwater) corresponds to machine learning models that predict the adsorption capacity for polyethylene in freshwater. In this case, it can be seen, in the case of the validation phase, that the errors made in terms of RMSE are closer to each other compared to the model’s behavior in the previous block. In this case, it can be seen that the worst model corresponds to the artificial neural network (ANNlin) model that presents an RMSE of 0.865, followed by the support vector machine (SVMn log) model with a value of 0.770, with the best model being the random forest, which has a root mean square error of 0.744. In this case, it can be seen that the mean absolute percentage errors exceed those obtained by the ML models of the first block, varying between 11.14% and 13.67%.
Regarding the training phase, it can be seen that the validation phase adjustments are improved in a significant way, presenting RMSE values falling between 0.489 and 0.549. Regarding the query phase, it can be seen that the root mean square error remains at acceptable levels, corresponding to mean absolute percentage errors between 7.23% and 10.46%. The best model for the validation phase (RF with RMSE of 0.744) presents the worst results for the query phase (RMSE of 0.565) and vice versa; the best model of the query phase (ANN with RMSE of 0.464) is the worst model in the validation phase (RMSE of 0.865).
Despite these behaviors, the three selected models have suitable adjustments for all phases (Table 2). If these models are compared with the model developed by Li et al. (2020), it can be seen that all of them improve the adjustments in terms of the RMSE value in the test phase (0.661 vs. 0.464, 0.475, and 0.565) for the model developed with this input variable (log D).
The two following groups (PE/pure water—1; and PE/pure water—2) correspond to the machine learning models developed to determine the adsorption capacity for polyethylene in pure water. In this case, two blocks were developed because Li et al. (2020) presented two different approaches, one using two input variables (PE/pure water—1 with log D and Mw) and the other one using only one input variable (PE/pure water—2 with log D).
In our research, for the model development with two input variables (PE/pure water—1), the case of 17α-ethinyl estradiol was not considered because the authors did not report the experimental log Kd value, so this model lacks this case. As expected, the models offer different results depending on the input variables. When two input variables are used, the model that presents the best results for the validation phase is the support vector machine (SVMn log) model, while when only one input variable is used, the best model is the random forest. It can be seen that the use of two input variables improves the adjustments in the training and validation phases (except for the RF model). For the query phase, the adjustments remain practically unchanged, except for the case of the ANN (ANNlin) model, where the error, in terms of RMSE, drops from 0.729 to 0.431. As can be seen, the models developed with two input variables present low mean absolute percentage errors between 2.06% and 3.92% for the validation phase. This behavior worsens slightly for the training phase, passing to 4.92% and 5.93% for the ANN and SVM models, respectively, and 11.28% for the RF model. On the other hand, in the case of the query phase, the MAPE values are between 6.90% and 12.21%. Despite the increase in both the RMSE and the MAPE values, these models developed with two variables seem to behave adequately to predict log Kd.
The models developed to predict the adsorption capacity for polyethylene in pure water (PE/pure water—2) present, in general, slightly lower adjustments than those obtained by PE/pure water—1). In this case, the best model, considering the value of the root mean square error in the validation phase, is the random forest model, which presents an RMSE of 0.132. This model presents, in the query phase, an increase in its RMSE value (0.526). The other two models, namely the SVM (SVMn2 log) model and the ANN (ANNlin) model, present an RMSE value of 0.439 and 0.431 for this phase, slightly improving the results of the RF model for this phase.
According to these results (Table 2), it can be said that the SVM and ANN models for PE/pure water—2 show good performances in terms of RMSE and improve the adjustment of RMSE value for the test phase (0.471) provided by the model developed by Li et al. (2020) using only one input variable (log D).
Before continuing, it is necessary to emphasize that all the machine learning models developed to predict the adsorption capacity for PE in the different water samples present, in terms of mean absolute percentage error for the query phase, adequate values, generally, below 10%. In other cases, the value is slightly higher (SVM for PE/freshwater and ANN for PE/pure water—1), and in others, the difference is more significant, for example, for the models intended to predict log Kd in seawater, which present errors between 13.24% and 23.33%.
The following models (PP/seawater) correspond to the models developed to predict the adsorption capacity of polypropylene in seawater. Based on the results provided in the validation phase, it can be said that the best model corresponds to the random forest model (0.199), followed by the SVM (SVMlog) model with an RMSE of 0.244 and, finally, the artificial neural network (ANNlin) model (0.270). The other statistics parameters of the validation phase show favorable behavior with MAPE values below 9% and with correlation coefficients above 0.980. For the training phase, the adjustments are similar to the validation phase, although an increase in the MAPE value of the random forest model is observed; even so, it remains below 10%.
For the query phase, an inconsistent behavior can be observed. Thus, in the case of the RF model and the ANN models, it can be observed that the statistics remain close to the values of the training and the validation phase, while the SVM model suffers an increase in terms of RMSE that makes this statistic parameter reach a value of 0.779, lowering its correlation coefficient to 0.817.
Given these results (Table 2), it can be said that the RF and ANN models can perform prediction tasks correctly. These two models present lower RMSE values (0.298 and 0.307) than the model proposed by Li et al. (2020) in the test phase (0.369), which was developed with two input variables (log D and εβ). The SVM model presents high generalization errors, which imply that it should not be used for prediction tasks. It should be noted that this SVM model, which is the one with the lowest error for the validation phase among all the SVM models developed, is the one with the highest error for the query phase. Other SVM models with close RMSE values in the validation phase (0.255 and 0.262) subsequently showed a better result in the query phase (0.287 and 0.266, respectively).
Finally, the last group of models (PS/seawater) developed corresponds to the machine learning models aiming to predict the adsorption capacity for polystyrene in seawater. Based on the results shown in Table 2, and taking into account the value of RMSE for the validation phase, it can be stated that the model presenting the best behavior in this phase is the support vector machine (SVMn2 log) model (0.524), followed by the artificial neural (ANNlin) network (0.643) and the random forest model (0.794). Based on the results presented by the mean absolute percentage error, it can be affirmed that these models destined to predict the adsorption for PS in seawater are the models that present the worst adjustments for the validation phase, varying between 14.61% and 21.69%. Despite this, the correlation coefficients remain high, with values greater than 0.960, except for the random forest model, whose correlation coefficient falls to 0.883. For the query phase, the values in terms of RMSE remain close, except for the random forest model, keeping the MAPE values above 15.1%.
Taking into account the results shown in Table 2, it can be concluded that the models predicting the adsorption capacity for PS in seawater do not present, in general, good results, except for the SVM model, which improves the RMSE value for the test phase (0.714) of the model developed by Li et al. (2020) with two input variables (log D and π).
Taking into account the fact that the results obtained by the machine learning models used the same variables as Li et al. (2020), it can be said that, in general, the ML models improve the results obtained by Li et al. (2020). However, these types of ML models often need a large number of experimental cases and input variables to correlate the desired variable. Therefore, in this research, in addition to developing ML models with the variables used by Li et al. (2020), other ML models have been developed with more input variables. This is possible because Li et al. (2020) reported eight different input variables; therefore, the results obtained by the models with the input variables selection Type 2 are shown below (Table 3).

3.2. ML Models Using Input Variables Type 2

Table 3 shows the adjustments obtained for the machine learning models developed with the input variables combination Type 2 using all the available input variables (except for the cases in which the variable qH+ is not possible).
The first models (PE/seawater) correspond with to ML models to for predicting the adsorption of polyethylene in seawater. Unlike the Type 1 models for PE/seawater where three input variables, log D, εα and εβ were used, in this new PE/seawater model, seven input variables were used (log D, Mw, εα, εβ, q, V′, π). It can be observed (Table 3), based on the RMSE value for the validation phase, that the best-developed machine learning model is the SVM (SVMn log) model, which has a value of 0.243 followed by the ANN (ANNlin) model (0.306), which is the random forest model and the model with the highest RMSE value for this phase (0.373). It is clear that, for this phase, the three selected models present suitable adjustments. In addition, these models also present high values of the correlation coefficient, all greater than 0.990. These promising results are also obtained for the training phase, although the random forest model presents an important increase regarding RMSE (from 0.373 to 0.824).
For the query phase, the RMSE values obtained by the model show an increase that, in the same way, happened for the models with the input variables shown in selection Type 1. In addition, looking at the data for the query phase of Table 2 and Table 3, it can be seen that the incorporation of the five variables concerning the input variables’ selection Type 1 destabilizes the models’ prediction, causing in all of them an increase in the RMSE value for this phase.
Despite this, the random forest and support vector machine models improve the results of the three-variable model proposed by Li et al. (2020) (0.693, 0.443 vs. 0.752, respectively, in terms of RMSE values for the test phase). The artificial neural network model developed with seven input variables presents an RMSE value close to the value of the Li et al. (2020) model for the query phase (0.762 vs. 0.752). Only the SVM model developed using the input variables selection Type 2 has improved the results over the ML models that used the input variables’ selection Type 1.
The second group of models (PE/freshwater) corresponds to machine learning models aimed at predicting the adsorption capacity of polyethylene in freshwater using eight input variables (log D, Mw, εα, εβ, qH+, q, V′, π). In this case, the best model, based on the RMSE value for the validation phase, corresponds to the ANN (ANNlog) model (0.446), followed by the SVM (SVMn) model (0.473) and the RF model (0.697). These reasonable adjustments are reflected in the high correlation coefficients all greater than 0.960. This behavior is improved in all statistical parameters for the training phase, except for the mean absolute percentage error of the random forest model. In the case of the query phase, these new models present RMSE values between 0.210 and 0.392, maintaining high correlation coefficients, all higher than 0.980. Comparing the ML models developed using the input variables selection Type 2 with the previously developed models using the input variables selection Type 1, it can be said that the ML models developed with eight variables improve the models developed with only one variable; the improvement is appreciable in all the parameters except three MAPE values.
Because of the results reported in Table 3, it can be concluded that the RF, SVM, and ANN models developed using eight input variables improve the model developed by Li et al. (2020) (0.392, 0.210, and 0.272 vs. 0.661, respectively, in terms of RMSE values for test phase).
The next group of models (PE/pure water) corresponds with ML models to predict the adsorption of polyethylene in pure water. In this case, these models were developed using the eight input variables (log D, Mw, εα, εβ, qH+, q, V′, π) instead of the two or one which were used by Li et al. (2020) and that was also used in the development of the previous ML models (Table 2). In this case, the optimization process carried out by the RF model involved the elimination of the variable V′ in the trees of the forest.
It can be seen in Table 3 that the best-selected model, according to the RMSE value for the validation phase, is the SVM (SVMlog) model, which presents a value of 0.154, followed by the RF model (0.204) and the ANN (ANNlog) model (0.403). As in the previous models developed using the input variables selection Type 2, the correlation coefficients are high, all greater than 0.930. This good behavior for the validation phase is also observed in the training phase, although a small increase in the errors made by the models can be seen. For the query phase, the different models present RMSE values between 0.433 and 0.551, keeping the MAPE value at approximately 10% and correlation coefficients greater than 0.920.
Comparing the ML models Type 2 with the previously developed models Type 1, it can be said that, for the query phase, the random forest and support vector machine models present similar adjustments in terms of RMSE to those presented by the Type 1 models. Despite this, only the support vector machine model improves the results of the best model proposed by Li et al. (2020) (0.433 vs. 0.471, respectively, in terms of RMSE values for the test phase).
The next models (PP/seawater) correspond to the models developed to predict the adsorption of polypropylene in seawater using seven input variables (log D, Mw, εα, εβ, q, V′, π).
Based on the results provided by the root mean square error in the validation phase, it can be said that the best model is the support vector machine (SVMlog) model (0.229), followed by the random forest model (0.245), and finally, the artificial neural network (ANNlin) model, which presents a higher error than the other two models (0.419). The correlation coefficients of the three models are greater than 0.975. This good behavior in the validation phase is also observed in the training phase, for both the random forest model and the support vector machine model; however, it should be noted that the artificial neural network model presents an error of 0.029 in the training phase. The three models present RMSEs for the query phase between 0.215 and 0.494, with the support vector machine model offering the best results, as was the case in the validation phase.
If the results obtained by the models developed using the input variables selection Type 2 are compared with Type 1, it can be said that the increase in the number of variables has led to a significant decrease in the RMSE values obtained in the query phase for the RF and the SVM models. This can be seen in the support vector machine model, which goes from an RMSE of 0.779 to 0.240.
Given the results reported in Table 3, it can be concluded that the RF and the SVM models developed using seven input variables improve the model developed by Li et al. (2020) with two variables (0.215 and 0.240 vs. 0.369, respectively, in terms of RMSE values for test phase). In addition, these models also improve the machine learning models developed using the input variables selection Type 1 except for the ANN model, which is slightly worse.
Finally, the last group of models (PS/seawater) corresponds to the ML models to predict the adsorption for polystyrene in seawater using seven input variables (log D, Mw, εα, εβ, q, V′, π). In these new models, a significant improvement can be seen in the validation and query phase adjustment parameters. In fact, for the validation phase, the RMSE values are between 0.290 and 0.475 for the SVM (SVMn2 log) model and the RF model, respectively, while in the Type 1 models, the RMSE values were included between 0.524 and 0.794. Similar behavior is observed for the query phase, with the RMSE values between 0.385 and 0.873. As can be seen in Table 3, the best model on this occasion is the support vector machine model, which also offers the best adjustment parameters for the query phase (0.385).
Given the results, it can be said that the SVM and the ANN (ANNlog) models developed using seven input variables improve the model developed by Li et al. (2020) with two input variables (0.385 and 0.407 vs. 0.714, respectively, in terms of RMSE values for the test phase).
Figure 2 represents the experimental and predicted values of log Kd for the best machine learning models according to RMSE in the validation phase of each block shown in Table 3.
Each graph shows that the adjustments of the training, validation, and query cases are conveniently fitted to the line of slope 1, although some deviation can be observed as it happens in a query case for the PE/seawater model or the PE/pure water model. In general, it can be seen that all the best models consistently predict the log Kd values.
Given the results shown in Table 1 and Table 2, key points can be drawn about the results obtained for the different machine learning models developed.
  • Regardless of the input variables chosen, there is always some machine learning model that improves (in terms of RMSE for the query phase) the good adjustments provided by the models developed by Li et al. (2020).
  • Including additional variables to develop the ML models does not always improve the variable selection carried out by Li et al. (2020). This is especially evident in the ML models destined to predict PE/seawater, where no model developed using the input variables selection Type 2 improves the Type 1 models. In this sense, it can be said that the selection of variables carried out by Li et al. (2020) is a good and reliable selection for the model’s development.
  • Both models developed by Li et al. (2020) and the models developed in this research are models that could be used to determine the log Kd values.
  • To the best of the authors’ knowledge, increasing the number of experimental cases for each microplastic/water group used to develop the models would be appropriate. Presumably, this increase would help the models present better adjustments.

4. Conclusions

In this research, various prediction models based on machine learning were developed using different variables to determine the adsorption capacity for PE, PP, and PS towards organic pollutants in various specific water environments.
Given the results, it can be concluded, regardless of the variables chosen for the development of the model, that there is always some machine learning model that provides good results.
On the other hand, the increase in input variables does not necessarily mean an improvement in the results of the models. This can be seen in the models intended to be used in PE/seawater, where no model developed using the variables selection Type 2 improves the Type 1 models.
To the best of the authors’ knowledge, it would be necessary to improve all models using: i) more experimental cases for each microplastic/water group; and ii) different datasets for training, validation, and query, or the means of different configuration parameters, among others.

Author Contributions

Conceptualization, A.C.-S., J.C.M. and G.A.; methodology, J.C.M. and G.A.; validation, G.A. and E.B.; formal analysis, G.A. and E.B.; investigation, A.C.-S., J.C.M. and G.A.; data curation, A.C.-S., G.A. and A.S.-L.; writing—original draft preparation, A.C.-S., G.A. and A.S.-L.; writing—review and editing, A.C.-S., J.C.M. and G.A.; visualization, A.S.-L., A.C.-S. and G.A.; supervision, J.C.M. and A.C.-S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used were taken from the research paper of Li et al. (2020).

Acknowledgments

The authors thank Li et al. (2020) [17] for publishing their experimental data in Scientific Reports under Attribution 4.0 International (CC BY 4.0) (https://creativecommons.org/licenses/by/4.0/), which has allowed to be carried out this study. The authors thank RapidMiner Inc. for the Educational and the free license of RapidMiner Studio 9.10.001 and 9.10.011 software. A.C.-S. thanks the action financed by the Ministry of Universities under application 33.50.460A.752 and by the European Union NextGenerationEU/PRTR through a María Zambrano contract from the University of Vigo, belonging to the launch of a European Recovery Instrument (“Next Generation EU”), aimed at requalifying the Spanish university system, specifically for teachers and attracting international talent. A. Soria-Lopez is grateful for the FPU grant (FPU2020/06140).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, H.; Shim, J.E.; Park, I.H.; Choo, K.S.; Yeo, M.-K. Physical and Biomimetic Treatment Methods to Reduce Microplastic Waste Accumulation. Mol. Cell. Toxicol. 2022, 19, 13–25. [Google Scholar] [CrossRef] [PubMed]
  2. Ghosh, S.; Greenfeld, I.; Daniel Wagner, H. CNT Coating and Anchoring Beads Enhance Interfacial Adhesion in Fiber Composites. Compos. Part A Appl. Sci. Manuf. 2023, 167, 107427. [Google Scholar] [CrossRef]
  3. Ghosh, S.; Mondal, S.; Ganguly, S.; Remanan, S.; Singha, N.; Das, N.C. Carbon Nanostructures Based Mechanically Robust Conducting Cotton Fabric for Improved Electromagnetic Interference Shielding. Fibers Polym. 2018, 19, 1064–1073. [Google Scholar] [CrossRef]
  4. Jaiswal, K.K.; Dutta, S.; Banerjee, I.; Pohrmen, C.B.; Singh, R.K.; Das, H.T.; Dubey, S.; Kumar, V. Impact of Aquatic Microplastics and Nanoplastics Pollution on Ecological Systems and Sustainable Remediation Strategies of Biodegradation and Photodegradation. Sci. Total Environ. 2022, 806, 151358. [Google Scholar] [CrossRef]
  5. Singh, S.; Kumar Naik, T.S.S.; Anil, A.G.; Dhiman, J.; Kumar, V.; Dhanjal, D.S.; Aguilar-Marcelino, L.; Singh, J.; Ramamurthy, P.C. Micro (Nano) Plastics in Wastewater: A Critical Review on Toxicity Risk Assessment, Behaviour, Environmental Impact and Challenges. Chemosphere 2022, 290, 133169. [Google Scholar] [CrossRef]
  6. Ng, E.-L.; Huerta Lwanga, E.; Eldridge, S.M.; Johnston, P.; Hu, H.-W.; Geissen, V.; Chen, D. An Overview of Microplastic and Nanoplastic Pollution in Agroecosystems. Sci. Total Environ. 2018, 627, 1377–1388. [Google Scholar] [CrossRef]
  7. Vivekanand, A.C.; Mohapatra, S.; Tyagi, V.K. Microplastics in Aquatic Environment: Challenges and Perspectives. Chemosphere 2021, 282, 131151. [Google Scholar] [CrossRef]
  8. Matthews, S.; Mai, L.; Jeong, C.-B.; Lee, J.-S.; Zeng, E.Y.; Xu, E.G. Key Mechanisms of Micro- and Nanoplastic (MNP) Toxicity across Taxonomic Groups. Comp. Biochem. Physiol. Part C Toxicol. Pharmacol. 2021, 247, 109056. [Google Scholar] [CrossRef]
  9. Woods, J.S.; Verones, F.; Jolliet, O.; Vázquez-Rowe, I.; Boulay, A.-M. A Framework for the Assessment of Marine Litter Impacts in Life Cycle Impact Assessment. Ecol. Indic. 2021, 129, 107918. [Google Scholar] [CrossRef]
  10. Peano, L.; Kounina, A.; Magaud, V.; Chalumeau, S.; Zgola, M.; Boucher, J. Plastic Leak Project, Methodological Guidelines. 2020. Available online: https://quantis.com/report/the-plastic-leak-project-guidelines/ (accessed on 13 February 2023).
  11. Ramachandraiah, K.; Ameer, K.; Jiang, G.; Hong, G.-P. Micro- and Nanoplastic Contamination in Livestock Production: Entry Pathways, Potential Effects and Analytical Challenges. Sci. Total Environ. 2022, 844, 157234. [Google Scholar] [CrossRef]
  12. Abihssira-García, I.S.; Kögel, T.; Gomiero, A.; Kristensen, T.; Krogstad, M.; Olsvik, P.A. Distinct Polymer-Dependent Sorption of Persistent Pollutants Associated with Atlantic Salmon Farming to Microplastics. Mar. Pollut. Bull. 2022, 180, 113794. [Google Scholar] [CrossRef]
  13. Gouin, T. Addressing the Importance of Microplastic Particles as Vectors for Long-Range Transport of Chemical Contaminants: Perspective in Relation to Prioritizing Research and Regulatory Actions. Microplastics Nanoplastics 2021, 1, 14. [Google Scholar] [CrossRef]
  14. Ali, I.; Tan, X.; Li, J.; Peng, C.; Naz, I.; Duan, Z.; Ruan, Y. Interaction of Microplastics and Nanoplastics with Natural Organic Matter (NOM) and the Impact of NOM on the Sorption Behavior of Anthropogenic Contaminants—A Critical Review. J. Clean. Prod. 2022, 376, 134314. [Google Scholar] [CrossRef]
  15. Katsumiti, A.; Losada-Carrillo, M.P.; Barros, M.; Cajaraville, M.P. Polystyrene Nanoplastics and Microplastics Can Act as Trojan Horse Carriers of Benzo(a)Pyrene to Mussel Hemocytes In Vitro. Sci. Rep. 2021, 11, 22396. [Google Scholar] [CrossRef]
  16. Hu, L.; Zhao, Y.; Xu, H. Trojan Horse in the Intestine: A Review on the Biotoxicity of Microplastics Combined Environmental Contaminants. J. Hazard. Mater. 2022, 439, 129652. [Google Scholar] [CrossRef] [PubMed]
  17. Li, M.; Yu, H.; Wang, Y.; Li, J.; Ma, G.; Wei, X. QSPR Models for Predicting the Adsorption Capacity for Microplastics of Polyethylene, Polypropylene and Polystyrene. Sci. Rep. 2020, 10, 14597. [Google Scholar] [CrossRef]
  18. Kathuria, C.; Mehrotra, D.; Misra, N.K. A Novel Random Forest Approach to Predict Phase Transition. Int. J. Syst. Assur. Eng. Manag. 2022, 13, 494–503. [Google Scholar] [CrossRef]
  19. Varnek, A.; Baskin, I. Machine Learning Methods for Property Prediction in Chemoinformatics: Quo Vadis? J. Chem. Inf. Model. 2012, 52, 1413–1437. [Google Scholar] [CrossRef]
  20. Alduailij, M.; Khan, Q.W.; Tahir, M.; Sardaraz, M.; Alduailij, M.; Malik, F. Machine-Learning-Based DDoS Attack Detection Using Mutual Information and Random Forest Feature Importance Method. Symmetry 2022, 14, 1095. [Google Scholar] [CrossRef]
  21. Taoufik, N.; Boumya, W.; Achak, M.; Chennouk, H.; Dewil, R.; Barka, N. The State of Art on the Prediction of Efficiency and Modeling of the Processes of Pollutants Removal Based on Machine Learning. Sci. Total Environ. 2022, 807, 150554. [Google Scholar] [CrossRef]
  22. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  23. He, S.; Wu, J.; Wang, D.; He, X. Predictive Modeling of Groundwater Nitrate Pollution and Evaluating Its Main Impact Factors Using Random Forest. Chemosphere 2022, 290, 133388. [Google Scholar] [CrossRef]
  24. Saglam, C.; Cetin, N. Prediction of Pistachio (Pistacia vera L.) Mass Based on Shape and Size Attributes by Using Machine Learning Algorithms. Food Anal. Methods 2022, 15, 739–750. [Google Scholar] [CrossRef]
  25. Kang, B.; Seok, C.; Lee, J. Prediction of Molecular Electronic Transitions Using Random Forests. J. Chem. Inf. Model. 2020, 60, 5984–5994. [Google Scholar] [CrossRef]
  26. Svetnik, V.; Liaw, A.; Tong, C.; Culberson, J.C.; Sheridan, R.P.; Feuston, B.P. Random Forest:  A Classification and Regression Tool for Compound Classification and QSAR Modeling. J. Chem. Inf. Comput. Sci. 2003, 43, 1947–1958. [Google Scholar] [CrossRef] [PubMed]
  27. Bienefeld, C.; Kirchner, E.; Vogt, A.; Kacmar, M. On the Importance of Temporal Information for Remaining Useful Life Prediction of Rolling Bearings Using a Random Forest Regressor. Lubricants 2022, 10, 67. [Google Scholar] [CrossRef]
  28. Pang, A.; Chang, M.W.L.; Chen, Y. Evaluation of Random Forests (RF) for Regional and Local-Scale Wheat Yield Prediction in Southeast Australia. Sensors 2022, 22, 717. [Google Scholar] [CrossRef]
  29. Geppert, H.; Vogt, M.; Bajorath, J. Current Trends in Ligand-Based Virtual Screening: Molecular Representations, Data Mining Methods, New Application Areas, and Performance Evaluation. J. Chem. Inf. Model. 2010, 50, 205–216. [Google Scholar] [CrossRef]
  30. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  31. Rodríguez-Pérez, R.; Vogt, M.; Bajorath, J. Support Vector Machine Classification and Regression Prioritize Different Structural Features for Binary Compound Activity and Potency Value Prediction. ACS Omega 2017, 2, 6371–6379. [Google Scholar] [CrossRef] [Green Version]
  32. Liu, G.; Zhu, H. Displacement Estimation of Six-Pole Hybrid Magnetic Bearing Using Modified Particle Swarm Optimization Support Vector Machine. Energies 2022, 15, 1610. [Google Scholar] [CrossRef]
  33. Houssein, E.H.; Hosney, M.E.; Oliva, D. A Hybrid Seagull Optimization Algorithm for Chemical Descriptors Classification. In Proceedings of the 2021 International Mobile, Intelligent, and Ubiquitous Computing Conference (MIUCC), Cairo, Egypt, 26–27 May 2021; pp. 381–386. [Google Scholar]
  34. Sareminia, S. A Support Vector Based Hybrid Forecasting Model for Chaotic Time Series: Spare Part Consumption Prediction. Neural Process. Lett. 2022, 1–17. [Google Scholar] [CrossRef]
  35. Orgeira-Crespo, P.; Míguez-Álvarez, C.; Cuevas-Alonso, M.; Doval-Ruiz, M.I. Decision Algorithm for the Automatic Determination of the Use of Non-Inclusive Terms in Academic Texts. Publications 2020, 8, 41. [Google Scholar] [CrossRef]
  36. Drucker, H.; Surges, C.J.C.; Kaufman, L.; Smola, A.; Vapnik, V. Support Vector Regression Machines. In Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA, 1–6 December 1997; pp. 155–161. [Google Scholar]
  37. Smola, A.J.; Schölkopf, B. A Tutorial on Support Vector Regression. Stat. Comput. 2004, 14, 199–222. [Google Scholar] [CrossRef] [Green Version]
  38. Prasanna, T.H.; Shantha, M.; Pradeep, A.; Mohanan, P. Identification of Polar Liquids Using Support Vector Machine Based Classification Model. IAES Int. J. Artif. Intell. 2022, 11, 1507–1516. [Google Scholar] [CrossRef]
  39. Liu, Z.; He, H.; Xie, J.; Wang, K.; Huang, W. Self-Discharge Prediction Method for Lithium-Ion Batteries Based on Improved Support Vector Machine. J. Energy Storage 2022, 55, 105571. [Google Scholar] [CrossRef]
  40. Elkorany, A.S.; Marey, M.; Almustafa, K.M.; Elsharkawy, Z.F. Breast Cancer Diagnosis Using Support Vector Machines Optimized by Whale Optimization and Dragonfly Algorithms. IEEE Access 2022, 10, 69688–69699. [Google Scholar] [CrossRef]
  41. Niazkar, H.R.; Niazkar, M. Application of Artificial Neural Networks to Predict the COVID-19 Outbreak. Glob. Health Res. Policy 2020, 5, 50. [Google Scholar] [CrossRef]
  42. Paturi, U.M.R.; Cheruku, S.; Reddy, N.S. The Role of Artificial Neural Networks in Prediction of Mechanical and Tribological Properties of Composites—A Comprehensive Review. Arch. Comput. Methods Eng. 2022, 29, 3109–3149. [Google Scholar] [CrossRef]
  43. McCulloch, W.S.; Pitts, W. A Logical Calculus of the Ideas Immanent in Nervous Activity. Bull. Math. Biophys. 1943, 5, 115–133. [Google Scholar] [CrossRef]
  44. Khan, M.T.; Kaushik, A.C.; Ji, L.; Malik, S.I.; Ali, S.; Wei, D.-Q. Artificial Neural Networks for Prediction of Tuberculosis Disease. Front. Microbiol. 2019, 10, 395. [Google Scholar] [CrossRef] [PubMed]
  45. Mohamed, Z.E. Using the Artificial Neural Networks for Prediction and Validating Solar Radiation. J. Egypt. Math. Soc. 2019, 27, 47. [Google Scholar] [CrossRef] [Green Version]
  46. Dikshit, A.; Pradhan, B.; Santosh, M. Artificial Neural Networks in Drought Prediction in the 21st Century–A Scientometric Analysis. Appl. Soft Comput. 2022, 114, 108080. [Google Scholar] [CrossRef]
  47. Saikia, P.; Baruah, R.D.; Singh, S.K.; Chaudhuri, P.K. Artificial Neural Networks in the Domain of Reservoir Characterization: A Review from Shallow to Deep Models. Comput. Geosci. 2020, 135, 104357. [Google Scholar] [CrossRef]
  48. Hornik, K.; Stinchcombe, M.; White, H. Multilayer Feedforward Networks Are Universal Approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  49. Shin-ike, K. A Two Phase Method for Determining the Number of Neurons in the Hidden Layer of a 3-Layer Neural Network. In Proceedings of the Proceedings of SICE Annual Conference 2010, Taipei, Taiwan, 18–21 August 2010; pp. 238–242. [Google Scholar]
  50. Ujong, J.A.; Mbadike, E.M.; Alaneme, G.U. Prediction of Cost and Duration of Building Construction Using Artificial Neural Network. Asian J. Civ. Eng. 2022, 23, 1117–1139. [Google Scholar] [CrossRef]
  51. Salari, K.; Zarafshan, P.; Khashehchi, M.; Pipelzadeh, E.; Chegini, G. Modeling and Predicting of Water Production by Capacitive Deionization Method Using Artificial Neural Networks. Desalination 2022, 540, 115992. [Google Scholar] [CrossRef]
  52. Shi, C.-F.; Yang, H.-T.; Chen, T.-T.; Guo, L.-P.; Leng, X.-Y.; Deng, P.-B.; Bi, J.; Pan, J.-G.; Wang, Y.-M. Artificial Neural Network-Genetic Algorithm-Based Optimization of Aerobic Composting Process Parameters of Ganoderma Lucidum Residue. Bioresour. Technol. 2022, 357, 127248. [Google Scholar] [CrossRef]
  53. Hufnagl, B.; Steiner, D.; Renner, E.; Löder, M.G.J.; Laforsch, C.; Lohninger, H. A Methodology for the Fast Identification and Monitoring of Microplastics in Environmental Samples Using Random Decision Forest Classifiers. Anal. Methods 2019, 11, 2277–2285. [Google Scholar] [CrossRef] [Green Version]
  54. Hufnagl, B.; Stibi, M.; Martirosyan, H.; Wilczek, U.; Möller, J.N.; Löder, M.G.J.; Laforsch, C.; Lohninger, H. Computer-Assisted Analysis of Microplastics in Environmental Samples Based on ΜFTIR Imaging in Combination with Machine Learning. Environ. Sci. Technol. Lett. 2022, 9, 90–95. [Google Scholar] [CrossRef]
  55. Yang, J.; Gong, J.; Tang, W.; Shen, Y.; Liu, C.; Gao, J. Delineation of Urban Growth Boundaries Using a Patch-Based Cellular Automata Model under Multiple Spatial and Socio-Economic Scenarios. Sustainability 2019, 11, 6159. [Google Scholar] [CrossRef] [Green Version]
  56. Sarraf Shirazi, A.; Frigaard, I. SlurryNet: Predicting Critical Velocities and Frictional Pressure Drops in Oilfield Suspension Flows. Energies 2021, 14, 1263. [Google Scholar] [CrossRef]
  57. Moldes, Ó.A.; Morales, J.; Cid, A.; Astray, G.; Montoya, I.A.; Mejuto, J.C. Electrical Percolation of AOT-Based Microemulsions with n-Alcohols. J. Mol. Liq. 2016, 215, 18–23. [Google Scholar] [CrossRef]
  58. Zou, Z.-M.; Chang, D.-H.; Liu, H.; Xiao, Y.-D. Current Updates in Machine Learning in the Prediction of Therapeutic Outcome of Hepatocellular Carcinoma: What Should We Know? Insights Imaging 2021, 12, 31. [Google Scholar] [CrossRef] [PubMed]
  59. Yan, X.; Cao, Z.; Murphy, A.; Qiao, Y. An Ensemble Machine Learning Method for Microplastics Identification with FTIR Spectrum. J. Environ. Chem. Eng. 2022, 10, 108130. [Google Scholar] [CrossRef]
  60. Bifano, L.; Meiler, V.; Peter, R.; Fischerauer, G. Detection of Microplastics in Water Using Electrical Impedance Spectroscopy and Support Vector Machines. In Proceedings of the Sensors and Measuring Systems; 21th ITG/GMA-Symposium, Nuremberg, Germany, 10–11 May 2022; pp. 356–359. [Google Scholar]
  61. Chang, C.-C.; Lin, C.-J. LIBSVM: A Library for Support Vector Machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27. [Google Scholar] [CrossRef]
  62. Chang, C.C.; Lin, C.J. LIBSVM—A Library for Support Vector Machines. Available online: https://www.csie.ntu.edu.tw/~cjlin/libsvm/ (accessed on 17 October 2022).
  63. Hsu, C.-W.; Chang, C.-C.; Lin, C.-J. A Practical Guide to Support Vector Classification. Available online: https://www.csie.ntu.edu.tw/~cjlin/papers/guide/guide.pdf (accessed on 17 October 2022).
  64. Ng, W.; Minasny, B.; McBratney, A. Convolutional Neural Network for Soil Microplastic Contamination Screening Using Infrared Spectroscopy. Sci. Total Environ. 2020, 702, 134723. [Google Scholar] [CrossRef]
  65. Guo, X.; Wang, J. Projecting the Sorption Capacity of Heavy Metal Ions onto Microplastics in Global Aquatic Environments Using Artificial Neural Networks. J. Hazard. Mater. 2021, 402, 123709. [Google Scholar] [CrossRef]
  66. RapidMiner Documentation. Neural Net. Available online: https://docs.rapidminer.com/latest/studio/operators/modeling/predictive/neural_nets/neural_net.html (accessed on 17 October 2022).
Figure 1. Schemes of the different ML models developed in this research: (A) RF model (in regression mode)—inspired by the figure of Yang et al. (2019) [55]; (B) SVM model—inspired by the figure of Sarraf Shirazi and Frigaard (2021) [56]; and (C) ANN model—inspired by the figures of Moldes et al. (2016) and Zou et al. (2021) [57,58].
Figure 1. Schemes of the different ML models developed in this research: (A) RF model (in regression mode)—inspired by the figure of Yang et al. (2019) [55]; (B) SVM model—inspired by the figure of Sarraf Shirazi and Frigaard (2021) [56]; and (C) ANN model—inspired by the figures of Moldes et al. (2016) and Zou et al. (2021) [57,58].
Nanomaterials 13 01061 g001
Figure 2. Scatter plots for the experimental and predicted values of log Kd for the selected ML models developed using the input variables selection Type 2. The dashed line corresponds to the line with slope 1.
Figure 2. Scatter plots for the experimental and predicted values of log Kd for the selected ML models developed using the input variables selection Type 2. The dashed line corresponds to the line with slope 1.
Nanomaterials 13 01061 g002
Table 1. Input variables, marked in purple, are used according to the input variable selection to predict log Kd. Type 1 and type 1 * are the configurations used by Li et al. (2020) [17], and Type 2 is the configuration used in this research. Polyethylene (PE), polystyrene (PS), polypropylene (PP), and the eight variables reported used by Li et al. (2020) [17]: (i) n-octanol/water distribution coefficient at special pH condition (log D); (ii) molecular mass (Mw); covalent; (iii) acidity (εα); and (iv) basicity (εβ); (v) most positive atomic charge on H atom (qH+); (vi) most negative atomic charge (q); (vii) molecular volume (V′); and (viii) molecular volume (π).
Table 1. Input variables, marked in purple, are used according to the input variable selection to predict log Kd. Type 1 and type 1 * are the configurations used by Li et al. (2020) [17], and Type 2 is the configuration used in this research. Polyethylene (PE), polystyrene (PS), polypropylene (PP), and the eight variables reported used by Li et al. (2020) [17]: (i) n-octanol/water distribution coefficient at special pH condition (log D); (ii) molecular mass (Mw); covalent; (iii) acidity (εα); and (iv) basicity (εβ); (v) most positive atomic charge on H atom (qH+); (vi) most negative atomic charge (q); (vii) molecular volume (V′); and (viii) molecular volume (π).
ModelInput VariablesModelInput Variables
log DM’wεαεβqH+qVπ log DM’wεαεβqH+qVπ
PE/seawaterPP/seawater
Type 1 Type 1
Type 2 Type 2
PE/freshwaterPS/seawater
Type 1 Type 1
Type 2 Type 2
PE/pure water
Type 1
Type 1 *
Type 2
Table 2. Adjustments for the different machine learning models developed using the input variables selection Type 1. RMSE is the root mean square error; MAPE corresponds to the mean absolute percentage error; and r is the correlation coefficient. RF is the random forest model; SVM is the support vector machine model; and ANN corresponds to the artificial neural network model. T, V, and Q are the training, validation, and query phases, respectively. The best models (regarding RMSE for the validation phase) are in bold.
Table 2. Adjustments for the different machine learning models developed using the input variables selection Type 1. RMSE is the root mean square error; MAPE corresponds to the mean absolute percentage error; and r is the correlation coefficient. RF is the random forest model; SVM is the support vector machine model; and ANN corresponds to the artificial neural network model. T, V, and Q are the training, validation, and query phases, respectively. The best models (regarding RMSE for the validation phase) are in bold.
TVZ
ModelRMSEMAPErRMSEMAPErRMSEMAPEr
PE/seawater
RF0.52518.670.9830.3807.480.9880.52313.380.979
SVM0.2872.830.9930.2484.610.9930.35713.240.990
ANN0.2573.130.9940.2364.420.9940.56123.330.979
PE/freshwater
RF0.5498.080.9730.74413.670.9440.5657.230.963
SVM0.5368.930.9760.77011.140.9450.47510.460.978
ANN0.4896.790.9780.86513.200.9320.4648.590.974
PE/pure water—1
RF0.47111.280.9680.1763.310.9920.5319.480.929
SVM0.3565.930.9740.1322.060.9930.4116.900.958
ANN0.3094.920.9810.2253.920.9820.72912.210.937
PE/pure water—2
RF0.4107.790.9670.1322.250.9930.5268.590.936
SVM0.4669.510.9550.2053.470.9830.4398.100.953
ANN0.4096.450.9650.2314.230.9810.4317.720.955
PP/seawater
RF0.2559.950.9900.1996.690.9940.2984.970.968
SVM0.2605.120.9890.2446.920.9880.7797.320.817
ANN0.1603.190.9960.2708.940.9880.3074.210.956
PS/seawater
RF0.2215.280.9960.79414.610.8831.00315.110.820
SVM0.55423.100.9690.52421.690.9650.43612.850.988
ANN0.3379.210.9880.64315.690.9720.77315.070.956
Table 3. Adjustments for the different machine learning models developed using the input variables selection Type 2. RMSE is the root mean square error, MAPE corresponds to the mean absolute percentage error, and r is the correlation coefficient. RF is the random forest model, SVM is the support vector machine model, and ANN corresponds to the artificial neural network model. T, V, and Q are the training, validation, and query phases, respectively. The best models (regarding RMSE for the validation phase) are in bold.
Table 3. Adjustments for the different machine learning models developed using the input variables selection Type 2. RMSE is the root mean square error, MAPE corresponds to the mean absolute percentage error, and r is the correlation coefficient. RF is the random forest model, SVM is the support vector machine model, and ANN corresponds to the artificial neural network model. T, V, and Q are the training, validation, and query phases, respectively. The best models (regarding RMSE for the validation phase) are in bold.
TVZ
ModelRMSEMAPErRMSEMAPErRMSEMAPEr
PE/seawater
RF0.82438.890.9540.3737.690.9880.69326.800.970
SVM0.3365.520.9910.2435.220.9940.44316.380.984
ANN0.0400.561.0000.3065.460.9890.76215.280.946
PE/freshwater
RF0.42416.780.9910.6978.780.9620.39211.860.986
SVM0.3206.870.9910.4737.050.9900.2108.180.999
ANN0.2894.940.9920.4467.100.9910.27210.400.997
PE/pure water
RF0.47310.770.9550.2043.310.9830.54210.370.929
SVM0.3065.340.9810.1542.560.9900.4337.250.956
ANN0.63414.700.9160.4037.900.9370.55111.570.926
PP/seawater
RF0.2956.440.9880.2459.420.9940.2153.360.983
SVM0.2224.740.9920.2296.980.9900.2403.660.974
ANN0.0290.541.0000.41912.200.9790.4948.200.938
PS/seawater
RF0.48611.070.9800.47515.160.9700.87323.010.882
SVM0.2484.720.9940.2908.500.9860.38512.050.976
ANN0.3097.010.9900.4459.740.9840.40712.430.973
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Astray, G.; Soria-Lopez, A.; Barreiro, E.; Mejuto, J.C.; Cid-Samamed, A. Machine Learning to Predict the Adsorption Capacity of Microplastics. Nanomaterials 2023, 13, 1061. https://doi.org/10.3390/nano13061061

AMA Style

Astray G, Soria-Lopez A, Barreiro E, Mejuto JC, Cid-Samamed A. Machine Learning to Predict the Adsorption Capacity of Microplastics. Nanomaterials. 2023; 13(6):1061. https://doi.org/10.3390/nano13061061

Chicago/Turabian Style

Astray, Gonzalo, Anton Soria-Lopez, Enrique Barreiro, Juan Carlos Mejuto, and Antonio Cid-Samamed. 2023. "Machine Learning to Predict the Adsorption Capacity of Microplastics" Nanomaterials 13, no. 6: 1061. https://doi.org/10.3390/nano13061061

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop