Abstract
Objectives: Determining individuals’ addiction levels plays a crucial role in facilitating more effective smoking cessation. For this purpose, the Fagerstrom Test for Nicotine Dependence (FTND) is used all over the World as a traditional testing method. It can be subjective and may influence the evaluation results. This study’s key innovation is the use of physiological signals to provide an objective classification of addiction levels, addressing the limitations of the inherently subjective Fagerström Test for Nicotine Dependence (FTND). Methods: Physiological parameters were recorded from 123 voluntary participants (both male and female) aged between 18 and 60 for 120 s using the Masimo Rad-G pulse oximeter and the Hartman–Veroval blood pressure monitor. All participants were categorized into four addiction groups: healthy, lightly addicted, moderately addicted, or heavily addicted with the help of FTND. The recorded data were classified using Decision Tree, KNN, and SVM methods. SMOTE and class-weighting techniques were used to eliminate class imbalance. Also, the PCA technique was applied for dimensionality reduction, and the k-fold cross-validation method was employed to enhance the reliability of the machine learning algorithms. Results: Machine learning methods, when evaluated using the SMOTE with a (7380×7) sample of physiological signals recorded every 2 s from 123 participants, showed a high recall of 98.74%, specificity of 99.58%, precision of 98.79%, F-score of 98.74%, and accuracy of 98.75%. Also, it is extracted that there is a direct relationship between physiological parameters and smoking addiction levels. Conclusions: The study’s core novelty lies in leveraging non-invasive physiological signals to objectively classify addiction levels, addressing the subjectivity of the Fagerström Test for Nicotine Dependence (FTND). This study provides a proof-of-concept for the feasibility of using machine learning and physiological signals to assess addiction levels. The results indicate that the approach is promising.
Keywords:
smoking addiction; physiological parameters; FTND; SMOTE; class-weighting; k-fold; PCA; machine learning 1. Introduction
Tobacco addiction represents the most prevalent form of dependency worldwide. According to the World Health Organization (WHO), over 1.7 billion people, corresponding to approximately 22.3% of the global population, were reported as smokers in 2024, comprising 36.7% of men and 7.8% of women. Annually, more than 8 million people die from tobacco-related causes, including 1.3 million non-smokers and passive smokers [,]. This equates to an average of 19,100 tobacco-related deaths daily, underscoring its status as a critical global public health issue that demands urgent attention through early diagnosis and effective intervention strategies [,,,].
Traditionally, research in this field has heavily relied on self-report questionnaires like the Fagerström Test for Nicotine Dependence (FTND). Numerous studies have drawn conclusions based on FTND results. For instance, Marakoğlu et al. and Yazıcı and Ak used surveys and the Beck Depression Inventory to establish a significant link between smoking and higher depression scores among medical students [,]. Zincir et al. utilized FTND alongside other inventories to find that lower FTND scores were correlated with higher success rates in smoking cessation []. Çelepkolu et al. applied FTND to investigate demographic correlations, finding higher addiction rates in men and those who started smoking younger []. While practical, self-report tools are susceptible to bias, leading researchers to seek validation through objective biochemical markers like cotinine and exhaled carbon monoxide (CO) []. Although correlations between FTND scores and these biomarkers exist, they can be variable, highlighting a critical area for improvement in study design [].
To overcome the limitations of subjective reporting, recent studies have turned to objective neuroimaging and electrophysiological methods. Notable works include Pariyadath et al. and Wetherill et al., who used SVM-based classification on resting-state fMRI data to identify brain regions associated with nicotine addiction with high accuracy [,]. Similarly, Eken et al. and Çay et al. achieved over 90% accuracy in classifying addiction levels using machine learning algorithms on ECG, PPG, and EEG data from participants assessed with FTND [,]. However, while powerful, technologies like fMRI and EEG are often high-cost, require controlled lab environments, and involve operator dependency, limiting their scalability for widespread screening or long-term monitoring.
In this context, the proliferation of wearable technology presents a significant opportunity for the objective and continuous monitoring of addictive behaviors. Wearables like smartwatches can provide passive physiological data that serve as digital biomarkers, offering valuable insights into substance use and craving symptoms. For example, studies have successfully used wearable sensor data to detect opioid use [] and nicotine craving []. Similarly, other research demonstrates the analysis of cardiovascular and respiratory signals from wearables using machine learning [,].
Our study is founded on this promising paradigm. In contrast to complex and costly systems, the approach proposed here relies on low-cost, non-invasive physiological signals, such as SpO2, PR, RRp, Pi, and PVi, which can be easily recorded via wearable devices. This methodology not only offers scalability and accessibility but also holds the potential for continuous monitoring in real-world conditions outside the laboratory.
Given the limitations of subjective questionnaires, our study classifies participants’ dependency levels based on the FTND as a practical approach. However, a key future improvement will be to validate our model’s predictions against a biochemical “gold standard,” such as cotinine or CO measurements, as seen in recent studies [,]. This would significantly enhance the reliability of our class labels and the clinical validity of our model.
Therefore, this study should be regarded as a proof-of-concept investigation rather than a clinically validated diagnostic system. It aims not only to provide an alternative to subjective questionnaires but also to propose a portable, applicable, and reliable digital biomarker framework that overcomes the practical constraints of existing objective methods.
2. Materials and Methods
2.1. Dataset
The dataset used in the study was collected from 123 volunteer participants (58 women, 65 men) aged 18–60, and no known health problems. Masimo Rad-G pulse oximeter [] and Hartmann–Veroval [] digital blood pressure devices were used for data recording. The corresponding dataset was recorded for 120 s for each participant and recorded every 2 s, including seven different physiological parameters: oxygen saturation (SpO2), pulse rate (PR), respiratory rate (RRp), perfusion index (Pi), pleth variability index (PVi), and blood pressure values (systolic-diastolic, mmHg). The data obtained for each person is every 2 s, as shown in Figure 1, and 60 different data points were obtained for each variable.
Figure 1.
Data acquisition flow.
The PPG signals obtained from the participants were recorded at any time during the day for the non-smoking group, and the addicted group was recorded immediately after cigarette consumption (within 1–2 min of extinguishing the cigarette). This specific timing was chosen to capture the acute physiological response and regulation period following nicotine intake, which we hypothesized might correlate with dependency levels.
Further, 2400 different data points were recorded for the non-smoking group of 40 people (25 women, 15 men), and 4980 different data points were recorded for the addicted group (33 women, 50 men), with a total dataset of (7380×7). The systolic and diastolic blood pressure values obtained for each person were spread over the 120 s dataset belonging to the participant. Figure 2 shows the frequency distributions of the physical parameters.
Figure 2.
Frequency distribution of dataset on physiological parameters.
As shown in Figure 2, the Spo2 data form a left-skewed histogram. This is an expected situation because the normal oxygen saturation value should be between 95 and 100%. The PR graph, which expresses the pulse value, exhibits an ordinary histogram. It is normal for this value to be between 60 and 100/min. Similarly, it can be said that the RRp graph, which expresses the respiratory rate, also exhibits an ordinary histogram. The average respiratory rate in a normal individual is between 10 and 20/min. PVi is known as the pleth variability index and is the parameter that allows the automatic monitoring of dynamic changes that occur during the respiratory cycle. It is normal for these parameter data to exhibit a left-skewed histogram because the normal values of the parameter should be between 0 and 100%. Similarly, the perfusion index Pi parameter, which expresses the circulation rate, also exhibits a left-skewed histogram. The range value of this parameter is known as 0.02–20% and its average value is 1.4%. In the last graph, it was normal for systolic (green) and diastolic (red) blood pressure change histograms to form an ordinary histogram because normal blood pressure values should be 80–120 mmHg. As a result, all parameters of forming the dataset showed frequencies within the expected limits.
Upon examination of Figure 3, distinct dynamic behavior in pulse rate values is observed across all classes. The PR values generally fluctuate within a range of 60–120 beats per minute, which falls within clinically accepted normal limits. The variations observed between different classes can be considered potentially related to addiction levels.
Figure 3.
Distribution of pulse rates and respiratory rates: (a) PR, (b) RRp.
The analysis of respiratory rate values shows that they generally range between 10 and 30 breaths per minute across all classes. These values largely align with the normal respiratory rate range for adults (12–20 breaths per minute). The prominent periodic fluctuations visible in the graph reflect the complex regulatory mechanisms of the respiratory system. The differences between classes may suggest information about pathophysiological mechanisms underlying variations in respiratory patterns.
Additionally, Table 1 provides a more detailed statistical analysis of the changes in basic vital parameters recorded from the participants.
Table 1.
Basic vital parameters.
An examination of Table 1 reveals a consistent decrease in oxygen saturation (SpO2) from Class 1 (98.1%) to Class 4 (87.1%). This trend indicates a decline in respiratory efficiency and suggests progressively unstable oxygenation with increasing dependency levels. In contrast, pulse rate (PR) demonstrates a corresponding increase (79.8 → 118.3 bpm). This pattern suggests a tendency for heart rate to rise with higher levels of addiction. Similarly, a progressive decrease is observed in PR SDNN (36.4 → 20.1 ms), a measure of short-term pulse rate variability. This finding indicates reduced heart rate stability associated with increased dependency. Furthermore, a marked increase is seen in both respiratory rate period (RRp) and its coefficient of variation (CV%) (16.8 → 27.9 s; 16.8% → 31.1%). This suggests that higher addiction levels are associated with more irregular and rapid breathing patterns. The overall pattern indicates a transition from efficient, stable respiration toward inefficient, labored breathing characterized by increased respiratory effort.
Table 2 displays the statistical variations in heart rate variability (HRV) analysis. The heart rate variability parameters show a gradual decline from Class 1 to Class 4. SDNN and RMSSD, which reflect overall and short-term HRV, respectively, demonstrate a steady decrease (SDNN: 35.8 → 19.3 ms; RMSSD: 30.2 → 13.1 ms). This pattern may indicate reduced autonomic adaptation and diminished parasympathetic activity in the higher dependency classes. The LF/HF ratio increased significantly (1.58 → 3.52), suggesting a shift in autonomic balance. Similarly, pNN50 and the HRV Index followed comparable declining trends, indicating a reduction in beat-to-beat variability. Overall, these results reveal a class-dependent decrease in HRV metrics, which can be interpreted as progressive autonomic imbalance and potential cardiovascular strain at higher levels of dependency.
Table 2.
Heart rate variability analysis.
Table 3 presents the analysis of respiratory dynamics, revealing a gradual shift toward rapid and unstable breathing patterns with increasing dependency levels. The respiratory rate demonstrates a marked increase from Class 1 to Class 4 (16.8 → 27.9 breaths/min). Furthermore, the respiratory rate coefficient of variation and breath interval standard deviation increased significantly (%16.8 → %31.1; 1.9 → 5.8 sec), indicating heightened respiratory irregularity. The Rapid Shallow Breathing Index (RSBI) shows a strong upward trend (37.3 → 85.8), potentially reflecting the onset of inefficient respiratory effort. Collectively, these results indicate that respiratory instability and a rapid, shallow breathing pattern are hallmark characteristics of higher dependency levels. When interpreted together, Table 1, Table 2 and Table 3 provide evidence that tobacco dependency is associated with a coordinated autonomic dysregulation, concurrently affecting both the cardiovascular and respiratory systems.
Table 3.
Respiratory dynamics.
2.2. Method
Participants were categorized into four distinct groups based on their scores from the Fagerström Test for Nicotine Dependence (FTND). The FTND, originally developed by Karl-Olov Fagerström in 1978 [] and updated in 1991 [], is a widely used instrument. It is a subjective assessment that evaluates nicotine dependence based on smoking habits and behavioral patterns, rather than definitive medical criteria. The test comprises six questions, with item-level scores ranging from 0 to 3 or 0 to 1. The total score ranges from 0 to 10, where a higher score indicates a greater degree of dependence. In this study, each participant completed the FTND. Based on their total scores, participants were classified as follows (see Figure 4):
Figure 4.
Evaluation and class scale.
The labeled dataset was classified with the help of machine learning algorithms, as shown in the methodology protocol given in Figure 5, and the classified outputs were evaluated by comparing with the FNTD results.
Figure 5.
The methodology protocol of study.
2.3. Principal Component Analysis (PCA)
Principal Component Analysis (PCA) is a method applied to reduce the size of a large dataset with multiple variables. After applying the PCA method, the calculated variables are expressed as the principal components of the first variables of the dataset. In the PCA method, as the dataset gets closer to the normal distribution, PCA reaches the best performance value [,,].
As presented in the histogram graph in Figure 6, the dataset consists of seven physiological parameters: SpO2, Pr, RRp, PVi, Pi, systolic, and diastolic pressure. PCA is employed to reduce the dimensionality of the features by analyzing the variances within the PCA space. It was determined that 5 principal components retain 99% of the total variance for all physiological parameters, as illustrated in Figure 5. Consequently, the original seven-dimensional dataset can be effectively represented using only five dimensions, preventing classifiers from processing redundant information. Based on the variance ratios, the dataset, originally sized at (7380×7), is reduced to (7380×5) within the PCA space while preserving 99% of the information.
Figure 6.
Physiological parameters on PCA space and their variances.
2.4. Synthetic Minority Over-Sampling Technique (SMOTE)
The Synthetic Minority Over-sampling Technique (SMOTE) is a pioneering algorithm designed to mitigate the class imbalance problem in machine learning datasets. Unlike naive random oversampling, which simply duplicates minority class instances and can lead to overfitting, SMOTE generates synthetic examples to construct a more robust and generalized decision region for the minority class []. The core idea is to operate directly in the feature space, creating new, artificial instances that lie between existing minority class examples, thereby effectively enlarging the minority class cluster.
The algorithm’s mechanism is rooted in a controlled interpolation strategy between a seed instance and its neighbors. For a given minority class instance , the process begins by identifying its k-nearest neighbors belonging to the same class, typically using the Euclidean distance metric. Subsequently, one of these neighbors, denoted as , is randomly selected. A synthetic data point is then generated along the line segment connecting and in the feature space. This fundamental operation is mathematically defined by the following convex combination formula:
Here, represents the feature vector of the new synthetic instance; is the feature vector of the i-th seed instance from the minority class; and is the feature vector of the randomly chosen j-th nearest neighbor from the k neighbors of and is a random number drawn from a uniform distribution between 0 and 1. This scalar acts as an interpolation coefficient, determining the precise location of the new instance between and .
The vector difference () defines the direction and magnitude of the line segment connecting the two original instances. Multiplying this vector by yields a random fraction of the segment. Finally, adding this scaled vector to the original point creates a new point that is a linear interpolation of the two. This procedure is applied iteratively and independently to each feature, generating a plausible synthetic instance within the convex hull defined by the minority class examples.
The primary advantage of this approach is that it forces the decision region for the minority class to become more general, as the classifier is exposed to a broader, more continuous distribution of minority samples rather than just the specific, and potentially sparse, original points. This helps in reducing the overfitting risk associated with mere replication and aids in constructing a more effective and generalized classifier []. The number of synthetic samples to be generated for each is controlled by an oversampling rate n, allowing for precise control over the final class distribution.
2.5. Class-Weighting Technique for Imbalanced Data
Class-weighting, also known as cost-sensitive learning, is a fundamental and widely adopted technique for addressing class imbalance. It operates at the algorithm level by assigning a higher misclassification cost to the minority class. This approach directly modifies the objective function of a learning algorithm, incentivizing the model to pay more attention to correctly classifying the minority class instances [,].
The core principle is to treat the classification errors unequally. Misclassifying a minority class instance is considered more detrimental than misclassifying a majority class instance. This is formally integrated into the model training process by assigning a weight to each class, which is typically inversely proportional to its class frequency.
The class weights are incorporated into the model’s loss function. For a general loss function L, the weighted loss L_weighted over a dataset of N samples is computed as follows:
where is the loss for the prediction compared to the true label for the i-th sample. is the weight assigned to the true class of the i-th sample. N is the total number of samples.
The most common strategy for determining the class weights for a class j is to use the inverse of the class frequency:
where N is the total number of training samples. C is the total number of classes. is the number of samples in class j.
This formula ensures that the total weight assigned to each class is balanced, preventing the model from being biased toward the majority class.
2.6. Machine Learning
Powerful methods commonly used in classification and pattern recognition tasks, such as Decision Trees, KNN, and SVM, have been implemented in this study. To improve the reliability of their accuracy assessments, the 5-fold cross-validation technique has been applied.
2.6.1. k-Fold Cross-Validation Technique
In the k-fold cross-validation technique, the dataset is divided into multiple subsets (folds). In each iteration, one fold is designated as the test set, while the remaining folds are used for training the model. This process is repeated k times, ensuring that each subset serves as the test set exactly once, leading to a more robust and reliable evaluation of the model’s performance. Ultimately, the overall performance of the model is determined by averaging the evaluation metrics obtained from all iterations, as illustrated in Figure 7 []. The k-value was tested between 1 and 10, and the best classifier performances were obtained when k = 5.
Figure 7.
The process of 5-fold cross-validation.
The mathematical formulas for performance of k-fold are shown in Equations (4) and (5):
here, TF represents the test fold used for classification, TP denotes true positive results, TN refers to true negative results, FP indicates false positive results, and FN corresponds to false negative results.
Additionally, performance evaluation metrics play a critical role in assessing the effectiveness and reliability of the classifiers []. These metrics are defined by Equations (6)–(9).
Recall represents the proportion of true positive cases correctly identified by the model. Precision measures the accuracy of the model’s positive predictions. Specificity reflects the model’s effectiveness in correctly identifying negative classes. The F-Score, calculated as the harmonic mean of precision and recall, provides a balanced assessment of both the model’s accuracy and sensitivity.
2.6.2. Decission Tree
The Decision Tree method introduced by Breiman is one of the supervised classification algorithms []. Root, node and leaves are the basic structures of a decision tree. It is based on the principle that each decision node in the tree is divided into two branches. Branching is continued until a new node is created [].
The Gini Algorithm is used in the Decision Tree method. This algorithm provides classification by dividing the data into four categories: healthy, lightly addicted, moderately addicted, and heavily addicted. The relevant attribute values are grouped into four. The class values corresponding to the groupings obtained in this way are distributed. The Gini values for each attribute are calculated as in the Formulas (10)–(13);
here, k is the total number of classes in the data, T is the number of samples in a node, is the number of samples taken from the control group, is the number of samples taken from lightly addicted participants, is the number of samples taken from moderately addicted participants, is the number of samples taken from heavily addicted participants, L is the number of samples in the control group, K is the number of samples in the lightly addicted group, R is the number of samples in the moderately addicted group, and P is the number of samples in the heavily addicted group.
The classification formula is calculated as in Equation (14), where the relevant j value is the number of rows in the training set and n is the number of rows in the training set:
The smallest value calculated for the relevant attribute j is selected, thus the division is performed on this attribute. Finally, the process is continued by returning to the first step [].
2.6.3. K-Nearest Neighbor (KNN)
The K-Nearest Neighbors (KNN) algorithm is widely utilized in machine learning due to its simplicity and effectiveness. It classifies a new data point by analyzing the class labels of its nearest neighbors, thereby performing the classification task [,]. The fundamental principle of KNN involves identifying the K closest neighbors of a given data point and determining its class based on the majority class of these neighbors. In this process, the distances between data points are commonly measured using the Euclidean distance, as defined in Equation (15) [].
here, xi and xj represent the two data points being compared, xik and xjk denote the k-th feature values of the respective points, and n is the number of features.
The number of neighbors (k was systematically varied across the set k = {1, 3, 5, 7, 9, 11, 15, 21} and the optimal value (k = 3) was determined using the elbow method, which balances model bias and variance []. To assess the impact of distance computation, Euclidean, Manhattan, Minkowski, and Cosine distance metrics were compared. Euclidean distance produced the best classification results [].
2.6.4. Support Vector Machine (SVM)
Support Vector Machine (SVM) is fundamentally based on the principle of identifying an optimal hyperplane that effectively separates the data []. The primary objective is to maximize the margin between different classes by constructing a hyperplane with the widest possible separation. The margin represents the distance between the hyperplane and the nearest data points, known as support vectors, which are crucial in determining the hyperplane’s position [,]. The SVM classification approach is mathematically formulated in Equations (16)–(19).
The training data (xi) belonging to two separated classes (yi),
represented with the optimal hyperplane,
An optimal hyperplane with the largest margin can be formulated as follows:
which is equivalent to
where is a function which maps the input space into a higher-dimensional space.
The hyperparameter space tested in the grid search was constructed as follows: [‘linear’, ‘RBF (radial basis function)’, ‘polynomial’, ‘gaussian’] values for the kernel function; [0.001, 0.01, 0.1, 1, 10, 100, 1000] values for the BoxConstraint (C) on a logarithmic scale; and [‘scale’, ‘auto’, 0.001, 0.01, 0.1] values for the kernel coefficient (gamma). This space allowed for exploring both linear and non-linear decision boundaries and optimizing the balance between model complexity and fit to the training data. The final hyperparameter values selected were [kernel = ‘gaussian’, C = 1, gamma = 1] [,].
3. Results
This study aimed to determine smoking addiction levels by analyzing physiological parameters recorded over a two-minute period from 123 volunteers. The analysis employed popular machine learning classifiers (Decision Tree, KNN and SVM) alongside techniques such as k-fold cross-validation and PCA. To address the significant class imbalance in the dataset, both the SMOTE and class-weighting were applied. The optimal results were achieved using 5-fold cross-validation, with the best performance observed when the KNN parameter was set to three neighbors in conjunction with the SMOTE. PCA indicated that the seven original physiological parameters could be reduced to five principal components while retaining 99% of the data variance. For each participant, 60 data points were recorded per parameter, resulting in a total dataset of 7380 instances (123 participants × 60 data points). The classifier performances obtained with 5-fold cross-validation are presented in Table 4.
Table 4.
Performance metrics of methods via 5-fold.
When the original dataset was used, all models demonstrated high classification performance, with mean accuracy rates above 97%. In this scenario, the Decision Tree model achieved the highest accuracy rate (97.89% ± 0.45). The KNN model similarly exhibited stable performance with an accuracy of 97.75% ± 0.33. Although slightly lower, the SVM model also produced reliable results overall (97.71% ± 0.22). These findings indicate that both distance-based and boundary-based methods can effectively capture the discriminative patterns in the original data distribution.
A significant performance increase was observed across all models following the application of the SMOTE. This demonstrates that mitigating class imbalance directly enhances classification success. The highest performance was achieved by the KNN model, which reached an accuracy of 98.75% ± 0.22, a recall (sensitivity) of 98.74% ± 0.22, and an F1-score of 98.74% ± 0.22. This result confirms that KNN is a highly stable and reliable model under balanced data conditions. Similarly, SVM attained a high accuracy rate (98.62% ± 0.16). The low standard deviation values across all metrics reveal that these models maintained stable performance throughout the 5-fold cross-validation.
In contrast, the results obtained with the class-weighting technique showed partial improvement over the original dataset but were not as effective as SMOTE. The KNN model again achieved the highest accuracy under these conditions (98.52% ± 0.41). However, the Decision Tree model showed a notable decline in accuracy (96.06% ± 1.27), which may be attributed to its higher sensitivity to weight adjustments. The SVM model demonstrated balanced performance with an accuracy of 97.78% ± 0.44. These findings suggest that while class-weighting partially alleviates the imbalance problem, synthetic sampling methods like SMOTE more effectively enhance model generalization.
In overall assessment, data balancing methods significantly improved classification performance. Specifically, the SMOTE + KNN combination provided the highest and most consistent results across all metrics. These results align with studies in the literature, which report that SMOTE-based methods achieve higher success than weighting methods on small to moderately imbalanced datasets [,,].
Figure 8 displays the classification performance of the Decision Tree model on the original dataset, presented through a confusion matrix and ROC curves. Analysis of the confusion matrix reveals that the correct classification rates for all classes are notably high, with misclassifications remaining at a low level. A distinct accuracy advantage is evident, particularly for Class 1, Class 3, and Class 4.
Figure 8.
The confusion matrix and ROC curves of Decision Tree on original data.
The ROC curves support this finding, showing that the AUC values for all classes exceed 0.98 (Class 1: 0.994, Class 2: 0.991, Class 3: 0.990, Class 4: 0.988). These results demonstrate that the Decision Tree model possesses high discriminatory power on the original data and exhibits strong non-linear separation performance between the classes.
Figure 9 presents the performance of the KNN model trained on the data balanced with the SMOTE. The confusion matrix reveals that the correct classification rates for all classes are exceptionally high, with misclassifications kept to a minimum. The pronounced dominance of the diagonal values across all classes indicates the model’s successful discrimination between categories. The ROC curves corroborate this finding, showing AUC values above 0.99 for all classes (Class 1: 0.994, Class 2: 0.997, Class 3: 0.987, Class 4: 0.999). These results demonstrate that the KNN model combined with SMOTE exhibits high sensitivity and strong generalization capability, confirming it as a highly effective approach for enhancing performance on imbalanced datasets.
Figure 9.
The confusion matrix and ROC curves of KNN on SMOTE.
Figure 10 presents the performance of the KNN model trained with the application of the class-weighting technique. The confusion matrix indicates that high correct classification rates were achieved for all classes, with the number of misclassifications remaining at a very low level. The distinct prominence of the diagonal elements demonstrates the model’s success in discriminating each class individually. Analysis of the ROC curves shows that the AUC values for all classes are above 0.99 (Class 1: 0.999, Class 2: 0.993, Class 3: 0.994, Class 4: 0.995). This result indicates that the weighted KNN model exhibits high discriminatory power by significantly mitigating biases arising from class imbalance. However, minor differences compared to the results obtained with the SMOTE method suggest that the synthetic sampling approach remains more effective in terms of overall accuracy.
Figure 10.
The confusion matrix and ROC Curves of KNN on class-weighting technique.
Table 5 presents the performance metrics of three different models (Decision Tree, KNN, SVM) following the application of PCA dimensionality reduction, which preserved 99% of the variance. The results demonstrate that applying PCA largely maintained and, in some cases, even enhanced classification accuracy.
Table 5.
Performance metrics of methods with PCA via 5-fold.
According to Table 5, on the PCA-reduced version of the original data, the SVM model achieved the highest accuracy rate (97.30% ± 0.20), followed by the KNN model (96.71% ± 0.39). The Decision Tree model showed lower performance in terms of accuracy (90.08% ± 0.72). This indicates that models sensitive to non-linear boundaries can generalize better within the feature space summarized by PCA.
When the SMOTE method was applied, a significant performance increase was observed across all models. The highest accuracy rate was achieved by the SVM model (98.58% ± 0.23), while the KNN model also demonstrated similarly high performance (98.22% ± 0.28). The Decision Tree model showed only limited improvement even after SMOTE application (92.81% ± 0.52). This result suggests that the reduced sample diversity after PCA may constrain the capacity of tree-based models to define decision boundaries.
When the class-weighting technique was employed, the KNN model again achieved the highest accuracy rate (97.93% ± 0.51). The SVM model similarly exhibited high performance (96.96% ± 0.83), while a substantial decline was observed in the Decision Tree model (86.07% ± 1.64). This decline suggests that the weighting method may cause overfitting or boundary ambiguity, particularly in decision trees.
Overall, on PCA-applied data, the KNN and SVM models stood out with high accuracy, specificity, and F-score values. The SMOTE + PCA combination proved to be the most effective approach for enhancing model generalization on imbalanced datasets. These results align with findings reported in the literature; feature reduction via PCA reduces noise while preserving meaningful variance and, when combined with balanced data, increases the learning stability of models [].
Figure 11 displays the classification performance of the SVM model on the PCA-reduced original dataset. It correctly predicted 2367 samples for Class 1, 1546 for Class 2, 2479 for Class 3, and 846 for Class 4. The correctly predicted classification rates for all classes are notably high, with misclassifications remaining at a low level. The ROC curves show AUC values between 0.996 and 0.998 for all classes. These results demonstrate that SVM effectively captured non-linear relationships and maintained robust generalization after PCA reduction.
Figure 11.
The confusion matrix and ROC curves of SVM on PCA with original data.
Figure 12 presents the confusion matrix and ROC curves of the SVM model on SMOTE. An examination of the confusion matrix indicates that the model achieved very high correct classification rates. Specifically, Class 2 achieved an almost perfect performance with 2516 correctly predicted samples, while Class 1 and Class 4 also demonstrated strong performance with 2491 and 2505 correct predictions, respectively. Even for Class 3, which appeared to be the most challenging, the model correctly classified 2445 samples, indicating consistently high accuracy. The ROC curves show that the AUC values for all classes exceed 0.96 (Class 1: 0.999, Class 2: 0.999, Class 3: 0.996, Class 4: 0.999). These results demonstrate that SVM is a highly effective approach for enhancing performance on imbalanced datasets.
Figure 12.
The confusion matrix and ROC curves of SVM on PCA with SMOTE.
Figure 13 presents the confusion matrix and ROC curves of the KNN with PCA and class-weighting. The highest absolute performance was achieved for Class 3, with 2470 correctly classified samples, while Class 4, despite having only 858 samples, achieved 858 correct predictions, indicating exceptionally high precision. Class 1 (2369 correct) and Class 2 (1527 correct) also maintained consistently strong performance. The ROC curves and corresponding AUC values further confirm the balanced performance of the KNN model with class-weighting. An AUC of 0.998 for Class 1 indicates near-perfect separability, while Classes 2, 3, and 4 all achieved AUC values around 0.993, reflecting the model’s consistent sensitivity and fairness across all classes. The KNN model achieved balanced and reliable classification across the four classes, primarily due to the effectiveness of the weighting strategy.
Figure 13.
The confusion matrix and ROC curves of KNN on PCA with class-weighting.
4. Discussion
Figure 14 and Figure 15 present the classification results obtained from the Decision Tree, KNN, and SVM evaluated using the original dataset and under two imbalance-handling strategies, SMOTE and class-weighting approaches. The experiments were conducted both before and after dimensionality reduction using PCA at a 99% variance retention level, and all results were validated using 5-fold cross-validation
Figure 14.
Performance results of classification methods on original data.
Figure 15.
Performance results of classification methods on PCA reduced data.
As seen in Figure 14, when trained on the original data, all models achieved high classification performance, with accuracies above 97%. The Decision Tree model slightly outperformed others (accuracy: 97.89%), followed by KNN (97.75%) and SVM (97.71%). After applying the SMOTE, a significant improvement was observed in all evaluation metrics, particularly for the KNN model, which achieved an accuracy of 98.75% and an F-score of 98.74%. The enhancement in recall and precision values suggests that SMOTE effectively mitigated class imbalance, providing more robust and generalized learning. This trend aligns with previous studies demonstrating SMOTE’s efficacy in improving minority class detection [,,]. Conversely, the class-weighting technique yielded more variable outcomes. While the KNN model remained strong (accuracy: 98.52%), the Decision Tree experienced a notable decline (accuracy: 96.06%), possibly due to its sensitivity to uneven class penalties. The SVM, on the other hand, maintained stable accuracy (97.78%), consistent with findings indicating its robustness to weighted penalty adjustments [].
As seen in Figure 15, following PCA-based dimensionality reduction (99% variance retention), the classification performance exhibited minor changes depending on the model type and balancing strategy. Using the original PCA-reduced dataset, the SVM model achieved the highest accuracy (97.3%), followed by KNN (96.71%) and Decision Tree (90.08%). The reduced performance of the Decision Tree indicates a potential sensitivity to feature compression and loss of non-linear feature dependencies. Upon applying SMOTE after PCA, the models achieved their best overall results, especially the KNN classifier, which reached an accuracy of 98.22% and a specificity of 99.42%. The SVM model also exhibited strong generalization (accuracy: 98.58%), confirming that combining SMOTE with PCA enhances class separability by generating synthetic samples in the transformed feature space [,]. The class-weighting technique, in contrast, showed more inconsistent behavior. Although KNN maintained competitive accuracy (97.93%), the Decision Tree suffered a substantial decrease (86.07%), highlighting its sensitivity to feature compression and weighting imbalance. SVM, however, remained relatively stable (96.96%), again demonstrating its resilience to class imbalance effects [].
Overall, both PCA and SMOTE contributed positively to model performance; however, SMOTE consistently outperformed algorithm-level weighting across all evaluation scenarios. The KNN classifier consistently achieved the best trade-off between recall, precision, and accuracy, both with and without PCA.
The correlation matrix and p-values between the parameters in the original dataset are shown in Figure 16. In the correlation matrix, a high positive correlation between the 6th (systolic) and 7th (diastolic) variables can be observed. This is normal because both variables reflect the blood pressure changes in the participants. Additionally, there is no relationship between 1–4 (SpO2-PVi) and 4–7 (PVi-diastolic). However, it can be said that there is a low positive or negative correlation between nearly all other variables. When the p-values are examined, it is noted that the p-values for the 1–4 and 4–7 variables are much greater than 0.05. This means that the correlation between these variables is not statistically significant. For all other variables, the obtained p-values are less than 0.05, indicating that the correlations between other variables are statistically significant. However, the values in the correlation table being relatively low in both the negative and positive directions imply that there is a weak, albeit present, relationship between the variables. Furthermore, when determining the main components of PCA, the relationships between variables are utilized. If there is no correlation or very low correlation between the variables, the dimensionality reduction effect of PCA may decrease [,]. This is clearly seen in Figure 15. The results in the PCA space provided lower performance for all machine learning algorithms. However, variables with low correlation provide different and independent information for classifiers. This helps the model to learn more general patterns, reduces the risk of overfitting, and makes the model more balanced [,,].
Figure 16.
Correlation matrix and p-values of original dataset.
When the recent studies in the literature, as summarized in Table 6, are examined, it is observed that significant research is being conducted in the field of digital health and addiction detection. In these studies, it is generally observed that smoking addiction is addressed at a binary classification level (smoker/non-smoker) using questionnaire data or physiological signals [,,,,].
Table 6.
Recent studies in the literature.
The main originality that distinguishes our study from these is its use of a four-level clinical addiction scale, verified by the clinically accepted Fagerström Test for Nicotine Dependence (FTND). In other words, in this study, subjective questionnaire results are verified by classifying them with high accuracy through objective physiological measurements. This approach demonstrates that the subjective addiction tests used in clinics can be made reliable, objective, and automated with the help of a portable and non-invasive PPG-based system. The main goal of our study is not to examine the acute cardiovascular effects of smoking but rather to present a proof-of-concept for objectively determining individuals’ levels of nicotine dependence with the help of machine learning. This aims to contribute to clinicians managing treatment processes more quickly and effectively.
The portable Massimo Rad-G pulse oximeter device used in our system has ideal features for such an application; besides being portable, it can directly provide the necessary signal dynamics effectively and quickly, rather than providing the raw time series signal.
The most promising practical implication of this study as a proof-of-concept is the high performance and relatively simple structure of the KNN algorithm. Such a model could be integrated into future clinical support systems. For example, a smartwatch or wearable monitoring device could passively collect pulse and respiration data from individuals during their daily lives. It could provide instant feedback that could be used to screen individuals at high risk of addiction or monitor the progress of existing addicts. Clinicians could use the reports from this system, along with traditional questionnaires (FTND), to conduct a more objective and continuous assessment of addiction severity and develop personalized intervention strategies (e.g., adjusting nicotine replacement therapy dosage or the frequency of behavioral therapy sessions). However, for such an application to be realized, the model first needs to be validated in larger and more diverse populations and integrated with a user-friendly interface.
5. Conclusions
This study focuses on the fine-grained, multi-level classification of nicotine dependence severity (via FTND scores) using non-invasive, wearable-grade physiological signals from a Massimo Rad-G pulse oximeter. The core novelty of the research lies in leveraging physiological signals to objectively classify addiction levels, thereby addressing the inherent subjectivity of the Fagerström Test for Nicotine Dependence (FTND).
In contrast to the subjective and traditional methods typically used to determine smoking addiction, this study developed an effective and innovative approach to predict addiction levels using physiological parameters (vital signs) and machine learning techniques. The participants’ recorded physiological data were classified by comparing them with their self-reported FTND results, which are based on behavioral and habitual conditions. It is important to note, however, that the FTND itself is a subjective, self-reported tool and does not represent a biochemical gold standard.
To enhance the robustness of the model, the k-fold cross-validation method was applied, and Principal Component Analysis (PCA) was utilized to reduce data redundancy. Furthermore, techniques like SMOTE and class-weighting were implemented to address class imbalance in the dataset. A comparison of performance metrics demonstrated that the KNN classifier consistently achieved the best trade-off between recall, precision, and accuracy, both with and without PCA.
Ultimately, this study provides a proof-of-concept for the feasibility of using machine learning and non-invasive physiological signals to assess nicotine addiction levels. It is believed that this work will offer a valuable contribution to future research in the field of digital health and behavioral diagnostics. However, this investigation should be regarded as a proof-of-concept rather than a clinically validated diagnostic system. Due to the subjective nature of the FTND and the limited dataset, future work should include biochemical validation using established markers such as exhaled CO or cotinine to further strengthen the findings.
Author Contributions
Conceptualization, M.S.B. and I.N.I.; methodology, M.S.B. and I.N.I.; software, M.S.B.; formal analysis, M.S.B.; investigation, M.S.B. and I.N.I.; data curation, M.S.B. and I.N.I.; writing—original draft preparation, M.S.B. and I.N.I.; writing—review and editing, M.S.B. and I.N.I.; visualization, M.S.B. All authors have read and agreed to the published version of the manuscript.
Funding
This work was funded by Selcuk University Research Projects Coordination Unit, grant number 23201103.
Institutional Review Board Statement
This study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of Faculty of Medicine of Selcuk University (protocol code 2023/433 and date of approval 26 September 2023).
Informed Consent Statement
Informed consent was obtained from all subjects involved in the study.
Data Availability Statement
The authors confirm that the data to support the findings of this study. are available upon request to the corresponding author.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- World Health Organization. Available online: https://www.who.int/news-room/fact-sheets/detail/tobacco?utm_source=chatgpt.com (accessed on 23 September 2025).
- Our World in Data. Available online: https://ourworldindata.org/grapher/number-of-deaths-from-tobacco-smoking#sources-and-processing (accessed on 23 September 2025).
- Dai, X.; Gakidou, E.; Lopez, A.D. Evolution of the global smoking epidemic over the past half century: Strengthening the evidence base for policy action. Tob. Control 2022, 31, 129–137. [Google Scholar] [CrossRef]
- De Silva, R.; Silva, D.; Piumika, L.; Abeysekera, I.; Jayathilaka, R.; Rajamanthri, L.; Wickramaarachchi, C. Impact of global smoking prevalence on mortality: A study across income groups. BMC Public Health 2024, 24, 1786. [Google Scholar] [CrossRef]
- Marakoğlu, K.; Çivi, S.; Şahsıvar, Ş.; Özdemir, S. The Relationship between Smoking Status and Depression Prevalence among First and Second Year Medical Students. J. Depend. 2006, 7, 129–134. [Google Scholar]
- Yazıcı, H.; Ak, İ. Depressive symptoms, personal health behaviors and smoking among university students. Anatol. J. Psychiatry 2008, 9, 224–231. [Google Scholar]
- Zincir, S.B.; Zincir, N.; Sünbül, E.A.; Kaynak, E. Relationship between nicotine dependence and temperament and character traits in adults with cigarette smoking. J. Mood Disord. 2012, 2, 160–166. [Google Scholar] [CrossRef]
- Çelepkolu, T.; Atli, A.; Palancı, Y.; Yılmaz, A.; Demir, S.; Ibiloglu, A.O.; Ekin, S. The relationship between nicotine dependence level and age-gender among the smokers: Diyarbakir sample. Dicle Med. J. 2014, 41, 712–716. [Google Scholar] [CrossRef]
- Kim, S. Overview of Cotinine Cutoff Values for Smoking Status Classification. Int. J. Environ. Res. Public Health 2016, 13, 1236. [Google Scholar] [CrossRef] [PubMed]
- Zettergren, A.; Sompa, S.; Palmberg, L.; Ljungman, P.; Pershagen, G.; Andersson, N.; Lindh, C.; Georgelis, A.; Kull, I.; Melen, E.; et al. Assessing tobacco use in Swedish young adults from self-report and urinary cotinine: A validation study using the BAMSE birth cohort. BMJ Open 2023, 13, e072582. [Google Scholar] [CrossRef]
- Pariyadath, V.; Stein, E.A.; Ross, T.J. Machine learning classification of resting state functional connectivity predicts smoking status. Front. Hum. Neurosci. 2014, 8, 425. [Google Scholar] [CrossRef] [PubMed]
- Wetherill, R.R.; Rao, H.; Hager, N.; Wang, J.; Franklin, T.R.; Fan, Y. Classifying and characterizing nicotine use disorder with high accuracy using machine learning and resting-state fMRI. Addict. Biol. 2019, 24, 811–821. [Google Scholar] [CrossRef]
- Eken, A.; Çalışkan, Ş.; Çivilibal, S.; Tosun, P.D. Prediction of smoking status by using multi-physiological measures and machine learning techniques. Dokuz Eylul Univ. Fac. Eng. J. Sci. Eng. 2021, 23, 55–69. [Google Scholar] [CrossRef]
- Çay, T.; Ölmez, E.; Tanık, N.; Altın, C. EEG based cigarette addiction detection with deep learning. Trait. Du Signal 2024, 41, 1183–1192. [Google Scholar] [CrossRef]
- Carreiro, S.; Ramanand, P.; Akram, W.; Stapp, J.; Chapman, B.; Smelson, D.; Indic, P. Developing a Wearable Sensor-Based Digital Biomarker of Opioid Dependence. Anesth. Analg. 2025, 141, 393–402. [Google Scholar] [CrossRef] [PubMed]
- Horvath, M.; Pittman, B.; O’Malley, S.S.; Grutman, A.; Khan, N.; Gueorguieva, R.; Brewer, J.A.; Garrison, K.A. Smartband-based smoking detection and real-time brief mindfulness intervention: Findings from a feasibility clinical trial. Ann. Med. 2024, 56, 2352803. [Google Scholar] [CrossRef] [PubMed]
- Gao, Z.; Xiang, L.; Fekete, G.; Baker, J.S.; Mao, Z.; Gu, Y. Data-Driven Approach for Fatigue Detection during Running Using Pedobarographic Measurements. Appl. Bionics Biomech. 2023, 2023, 7022513. [Google Scholar] [CrossRef]
- Mahoney, J.M.; Rhudy, M.B. Methodology and validation for identifying gait type using machine learning on IMU data. J. Med. Eng. Technol. 2019, 43, 25–32. [Google Scholar] [CrossRef] [PubMed]
- Kunutsor, S.K.; Spee, J.M.; Kieneker, L.M.; Gansevoort, R.T.; Dullaart, R.P.F.; Voerman, A.J.; Touw, D.J.; Bakker, S.J. Self-Reported Smoking, Urine Cotinine, and Risk of Cardiovascular Disease: Findings From the PREVEND (Prevention of Renal and Vascular End-Stage Disease) Prospective Cohort Study. J. Am. Heart Assoc. 2018, 7, e008726. [Google Scholar] [CrossRef]
- Murriky, A.; Allam, E.; Alotaibi, H.; Alnamasy, R.; Alnufiee, A.; AlAmro, A.; Al-Hammadi, A. The relationship between nicotine dependence and willingness to quit smoking: A cross-sectional study. Prev. Med. Rep. 2025, 53, 103066. [Google Scholar] [CrossRef]
- Masimo. Rad-G: Continuous Pulse Oximeter Opretors Manual. Available online: https://techdocs.masimo.com/globalassets/techdocs/pdf/lab-10076a_master.pdf (accessed on 23 September 2025).
- Hartmann Veroval Compact Upper Arm Blood Pressure Monitor. Available online: https://www.veroval.info/-/media/country/website/archive/diagnostics/veroval/pdfs/new-pdfs/instructions-for-use/veroval_compact_bpu_22.pdf?rev=7ab70eb24ab44ce5a52b69682dc797d1&sc_lang=en (accessed on 23 September 2025).
- Fagerström, K.O. Measuring degree of physical dependence to tobacco smoking with reference to individualization of treatment. Addict. Behav. 1978, 3, 235–241. [Google Scholar] [CrossRef]
- Heatherton, T.F.; Kozlowski, L.T.; Frecker, R.C.; Fagerstrom, K.O. The Fagerstrom Test for Nicotine Dependence: A revision of the Fagerstrom Tolerance Questionnaire. Br. J. Addict. 1991, 86, 1119–1127. [Google Scholar] [CrossRef]
- Jolliffe, I.T. Principal Component Analysis, 2nd ed.; Springer: New York, NY, USA, 2002. [Google Scholar] [CrossRef]
- Cao, L.J.; Chua, K.; Chong, W.K.; Lee, H.P.; Gu, O.M. A comparison of PCA, KPCA and ICA for dimensionality reduction in support vector machine. Neurocomputing 2003, 55, 321–336. [Google Scholar] [CrossRef]
- Howley, T.; Madden, M.G.; O’Connell, M.L.; Ryder, A.G. The effect of principal component analysis on machine learning accuracy with high-dimensional spectral data. Knowl.-Based Syst. 2006, 19, 363–370. [Google Scholar] [CrossRef]
- Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
- Fernandez, A.; Garcia, S.; Herrera, F.; Chawla, N.V. SMOTE for Learning from Imbalanced Data: Progress and Challenges, Marking the 15-year Anniversar. J. Artif. Intell. Res. 2018, 61, 863–905. [Google Scholar] [CrossRef]
- Benkendorf, D.J.; Schwartz, S.D.; Cutler, D.R.; Hawkins, C.P. Correcting for the effects of class imbalance improves the performance of machine-learning based species distribution models. Ecol. Model. 2023, 483, 110414. [Google Scholar] [CrossRef]
- Bakırarar, B.; Ehan, A.H. Class Weighting Technique to Deal with Imbalanced Class Problem in Machine Learning: Methodological Research. Turk. Klin. J. Biostat. 2023, 15, 19–29. [Google Scholar] [CrossRef]
- Bascil, M.S. A New Approach on HCI Extracting Conscious Jaw Movements Based on EEGSignals Using Machine Learnings. J. Med. Syst. 2018, 42, 169. [Google Scholar] [CrossRef]
- Ozer, I.; Karaca, A.C.; Ozer, C.K.; Gorur, K.; Kocak, I.; Cetin, O. The exploration of the transfer learning technique for Globotruncanita genus against the limited low-cost light microscope images. Signal Image Video Process. 2024, 18, 6363–6377. [Google Scholar] [CrossRef]
- Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
- Lee, S.; Lee, C.; Mun, K.G.; Kim, D. Decision Tree Algorithm Considering Distances Between Classes. IEEE Access 2022, 10, 69750–69756. [Google Scholar] [CrossRef]
- D’Ambrosio, A.; Tutore, V.A. Conditional Classification Trees by Weighting the Gini Impurity Measure. In Proceedings of the 7th Conference of the Classification and Data Analysis Group of the Italian Statistical Society, Catania, Italy, 9–11 September 2009. [Google Scholar] [CrossRef]
- Görür, K.; Bozkurt, M.R.; Bascil, M.S.; Temurtas, F. Tongue-Operated Biosignal over EEG and Processing with Decision Tree and kNN. Acad. Platf.-J. Eng. Sci. 2021, 9, 112–125. [Google Scholar] [CrossRef]
- Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
- Peterson, L.E. K-Nearest Neighbor. Scholarpedia 2009, 4, 1883. [Google Scholar] [CrossRef]
- Ali, N.; Neagu, D.; Trundle, P. Evaluation of k-nearest neighbour classifier performance for heterogeneous data sets. SN Appl. Sci. 2019, 1, 1559. [Google Scholar] [CrossRef]
- Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
- Ben-Hur, A.; Weston, J.A. User’s guide to support vector machines. Methods Mol. Biol. 2009, 609, 223–239. [Google Scholar] [CrossRef]
- Schölkopf, B.; Smola, A.J. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; The MIT Press: Cambridge, MA, USA, 2001. [Google Scholar] [CrossRef]
- Hidayaturrohman, Q.A.; Hanada, E. A Comparative Analysis of Hyper-Parameter Optimization Methods for Predicting Heart Failure Outcomes. Appl. Sci. 2025, 15, 3393. [Google Scholar] [CrossRef]
- Last, F.; Douzas, G.; Bacao, F. Oversampling for Imbalanced Learning Based on K-Means and SMOTE. Inf. Sci. 2018, 465, 1–20. [Google Scholar] [CrossRef]
- Elreedy, D.; Atiya, A.F.; Kamalov, F. A theoretical distribution analysis of synthetic minority oversampling technique (SMOTE) for imbalanced learning. Mach. Learn. 2024, 113, 4903–4923. [Google Scholar] [CrossRef]
- Jolliffe, I.T.; Cadima, J. Principal component analysis: A review and recent developments. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 2016, 374, 20150202. [Google Scholar] [CrossRef]
- Batista, G.; Prati, R.C.; Monard, M.C. A study of the behavior of several methods for balancing machine learning training data. ACM SIGKDD Explor. Newsl. 2004, 6, 20–29. [Google Scholar] [CrossRef]
- Tang, Y.; Zhang, Y.Q.; Chawla, N.V.; Krasser, S. SVMs modeling for highly imbalanced classification. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2009, 39, 281–288. [Google Scholar] [CrossRef] [PubMed]
- Guo, H.; Li, Y.; Shang, J.; Gu, M.; Huang, Y.; Gong, B. Learning from class-imbalanced data: Review of methods and applications. Expert Syst. Appl. 2017, 73, 220–239. [Google Scholar] [CrossRef]
- Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2009, 3, 1157–1182. [Google Scholar]
- Zheng, A.; Casari, A. Feature Engineering for Machine Learning: Principles and Techniques for Data Scientists, 1st ed.; O’Reilly Media: Sebastopol, CA, USA, 2018. [Google Scholar]
- Dormann, C.F.; Elith, J.; Bacher, S.; Buchmann, C.; Carl, G.; Carré, G.; Marquéz, J.R.G.; Gruber, B.; Lafourcade, B.; Leitão, P.J.; et al. Collinearity: A review of methods to deal with it and a simulation study evaluating their performance. Ecography 2013, 36, 27–46. [Google Scholar] [CrossRef]
- Kharabsheh, M.; Meqdadi, O.; Alabed, M.; Veeranki, S.; Abbadi, A.; Alzyoud, S. A machine learning approach for predicting nicotine dependence. Int. J. Adv. Comput. Sci. Appl. 2019, 10, 179–184. [Google Scholar] [CrossRef]
- Choi, J.; Jung, H.T.; Ferrell, A.; Woo, S.; Haddad, L. Machine learning-based nicotine addiction prediction models for youth e-cigarette and waterpipe (hookah) users. J. Clin. Med. 2021, 10, 972. [Google Scholar] [CrossRef]
- Issabakhsh, M.; Sánchez-Romero, L.M.; Le, T.; Liber, A.C.; Tan, J.; Li, Y.; Meza, R.; Mendez, D.; Levy, D.T. Machine learning application for predicting smoking cessation among US adults: An analysis of waves 1–3 of the PATH study. PLoS ONE 2023, 18, e0286883. [Google Scholar] [CrossRef] [PubMed]
- Aishwarya, S.; Siddalingaswamy, P.C.; Chadaga, K. Explainable artificial intelligence driven insights into smoking prediction using machine learning and clinical parameters. Sci. Rep. 2025, 15, 24069. [Google Scholar] [CrossRef] [PubMed]
- Qananwah, Q.; Quran, H.; Dagamseh, A.; Blazek, V.; Leonhardt, S. Investigating the correlation between smoking and blood pressure via photoplethysmography. BioMed. Eng. OnLine 2025, 24, 57. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).