Next Article in Journal
A Multi-Time-Scale Coordinated Scheduling Model for Multi-Energy Complementary Power Generation System Integrated with High Proportion of New Energy Including Electricity-to-Hydrogen System
Previous Article in Journal
Automated Dimension Recognition and BIM Modeling of Frame Structures Based on 3D Point Clouds
Previous Article in Special Issue
Modern Approaches to Software Vulnerability Detection: A Survey of Machine Learning, Deep Learning, and Large Language Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mouse Data Protection in Image-Based User Authentication Using Two-Dimensional Generative Adversarial Networks: Based on a WM_INPUT Message Approach

1
Interdisciplinary Program of Information & Protection, Mokpo National University, Muan 58554, Republic of Korea
2
School of Computer Science and Engineering, Information Security Major, Mokpo National University, Muan 58554, Republic of Korea
*
Author to whom correspondence should be addressed.
Electronics 2026, 15(2), 292; https://doi.org/10.3390/electronics15020292
Submission received: 19 November 2025 / Revised: 3 January 2026 / Accepted: 5 January 2026 / Published: 9 January 2026

Abstract

With the rapid evolution of computing technologies and the increased proliferation of online services, secure remote user authentication methods have become essential. Among these methods, password-based authentication remains dominant due to its straightforward implementation and ease of use. Nevertheless, password-based systems are particularly prone to credential theft from keylogging attacks, making user passwords easily compromised. To address these risks, image-based authentication methods were developed, allowing users to enter passwords through mouse clicks rather than keyboard input, thereby reducing vulnerabilities associated with conventional password entry. However, subsequent studies have shown that mouse movement and click information can still be obtained using APIs such as the GetCursorPos() function or WM_INPUT message, thus undermining the intended security benefits of image-based authentication. In response, various defense strategies have sought to inject artificial or random mouse data through functions such as SetCursorPos() or by utilizing the WM_INPUT message, in an effort to disguise authentic user input. Despite these defenses, recent machine learning-based attacks have demonstrated that such naïve bogus input can be distinguished from legitimate mouse data with up to 99% classification accuracy, resulting in substantial exposure of actual user actions. To address this, a technique leveraging Generative Adversarial Networks (GAN) was introduced to produce artificial mouse data closely mimicking genuine user input, which has been shown to reduce the attack success rate by roughly 37%, offering enhanced protection for mouse-driven authentication systems. This article seeks to advance GAN-based mouse data protection by integrating multiple adversarial generative models and conducting a comprehensive evaluation of their effectiveness with respect to data processing techniques, feature selection, generation intervals, and model-specific performance differences. Our experimental findings reveal that the enhanced approach reduces attack success rates by up to 48%, marking an 11% performance gain over previous mouse data protection approaches, and providing stronger empirical support that our method offers superior protection for user authentication data compared to prior techniques.

1. Introduction

With the advancement of computing technology and the widespread deployment of online services, Internet usage has increased significantly [1]. Users now routinely utilize web-based platforms, including e-commerce sites for purchasing goods and financial services for asset management, to accomplish a variety of tasks. Within these environments, secure remote user authentication methods are critical, with password-based authentication remaining the most prevalent technique. Although this method is straightforward to implement and has continued to be used extensively since the early days of the Internet, it exhibits a significant vulnerability: user passwords can be readily compromised through attacks such as keylogging, resulting in persistent security risks. To address these vulnerabilities, image-based authentication techniques have been developed.
Image-based authentication commonly uses mechanisms like virtual keypads, where users enter authentication information using mouse clicks instead of the keyboard. Since input is not directly provided via the keyboard, these schemes are intrinsically more resistant to keystroke logging attacks that intercept keyboard data [2]. Nevertheless, subsequent studies have revealed that mouse movement and click data may be recorded using APIs such as GetCursorPos() function or through WM_INPUT message, which enables attackers to reconstruct the user’s input and thereby undermines the security benefits of image-based authentication techniques [3,4].
To counter these risks, a number of mouse data protection mechanisms have been designed that utilize SetCursorPos() function or WM_INPUT message to safeguard against such attacks [5,6]. The fundamental principle of these strategies is to inject artificially generated bogus mouse data into the event stream so an attacker collecting mouse data via GetCursorPos() function or WM_INPUT message cannot reliably differentiate authentic user input from the entire set of recorded mouse data. Consequently, even if mouse data is compromised, the inability to separate genuine from bogus mouse data is intended to prevent the attacker from accurately extracting actual authentication credentials.
Faced with these enhanced defenses, attackers have started utilizing more advanced attack strategies. Recent studies indicate that, in contexts where such protection measures are deployed, machine learning–driven attack techniques can be trained to distinguish between synthetically generated bogus mouse data and actual mouse data [7]. These classification models have reached accuracy rates as high as 99%, thereby restoring the attacker’s ability to accurately identify genuine mouse data and undermining current mouse data protection strategies. To address this vulnerability, a defense mechanism based on GAN has been proposed [8]. This strategy creates bogus mouse data that closely imitates real user behavior, substantially impairing the success rate of machine learning–based classification attacks—experimentally reducing the attack success rate, defined as the classification accuracy for genuine input, from 99% to approximately 37%.
Building on these research developments, this article seeks to further fortify the resilience of GAN-based mouse data protection methodologies. Specifically, the study extends previous investigations by evaluating a range of GAN architectures, including those previously reported, and by systematically assessing how various model selections impact defense efficacy against classification-based mouse data inference attacks.
The core innovations and contributions of this article are presented below:
  • A WM_INPUT message–level mouse data protection method is put forward, using GAN-generated bogus mouse trajectories. In environments where attackers use machine learning to extract genuine mouse data, we conduct empirical comparisons across several GAN models to determine which achieves the greatest reduction in attack success rates. The outcomes indicate that our approach better secures user authentication data by refining the defensive generation of mouse event data.
  • We develop a two-dimensional GAN–based framework for generating plausible bogus mouse data, representing an advancement over earlier studies that focused predominantly on the CTGAN model. Notably, we explore the CopulaGAN model, the conditional GAN (CGAN) model, and the WGAN-GP model for their capability to produce highly realistic bogus trajectories, underscoring both methodological innovation and architectural diversity relative to previous work.
  • We show that the proposed defense method offers greater effectiveness compared to established mouse data obfuscation approaches. Whereas earlier GAN-based solutions achieved a reduction in attack success rates to around 37%, our experimental results reveal that the optimal GAN configuration lowers the attack success rate by 48% relative to the baseline, an improvement of 11% over the previous highest-performing defense. This demonstrates that our strategy provides a more resilient protection framework for mouse-driven user authentication data.
The remainder of this article is structured as follows. Section 2 reviews the background and examines related works concerning mouse data–based attacks and defense methods. Section 3 introduces the proposed GAN-based mouse data protection method and explains the experimental setup. Section 4 details and evaluates the experimental outcomes from various perspectives, including data processing strategies, feature definition, data generation intervals, and analysis of performance impact. Section 5 presents the conclusions of the article.

2. Background and Related Works

This section examines attack and defense approaches targeting mouse data in image-based user authentication methods. A key previous study on the WM_INPUT-based mouse data attack technique [4] showed that an attacker can capture mouse coordinates by intercepting WM_INPUT messages sent from the operating system to the application. As depicted in Figure 1, this method allows adversaries to reconstruct the sequence of mouse interactions during image-based authentication and eventually recover the user’s password on a legitimate service.
Under this attack scenario, a malicious program initially registers as a handler for WM_INPUT messages with the operating system to monitor all WM_INPUT traffic. After registration, the handler receives continuous streams of raw mouse input from the user’s mouse device. By recording all observed coordinates, the attacker can reconstruct the path of mouse actions and analyze user interaction behaviors on virtual keypads or graphical password systems. This process enables the attacker to deduce authentication information, such as image-based passwords, even without observing any keystrokes.
To mitigate such threats, multiple studies have introduced mouse data protection strategies that generate and insert random bogus mouse data to disrupt the precise identification of true mouse data [6]. In these methods, both authentic and artificially generated mouse coordinates are combined in the recorded event stream. Although all events remain accessible to the attacker, the objective is to make it difficult to accurately distinguish legitimate mouse input from decoy data, complicating the reconstruction of actual authentication inputs. Nonetheless, further studies have proposed machine learning–based attacks designed to compromise these defense techniques [7], with the experimental findings summarized in Table 1.
In this study, the attacker develops classifiers using datasets that comprise both genuine and bogus mouse data collected within attack scenarios. Table 1 presents the most effective models under several configurations, considering various bogus-data generation intervals and different combinations of features. The dataset was assembled by altering the injection periods of bogus mouse data and by selecting features such as elapsed time, the current X and Y coordinates, and the stepwise distances between consecutive X and Y coordinates. The findings indicate that utilizing all specified features typically results in optimal classification performance across nearly all generation intervals, with classification accuracy reaching up to 99%. This demonstrates that the machine learning–supported attack can nearly perfectly differentiate genuine mouse data from synthetically produced bogus mouse movements, thereby rendering the original random-injection defense ineffective.
To respond to this advanced attack strategy, a GAN-based mouse data protection method was introduced in [8]. Instead of injecting randomly generated bogus mouse movements, this method applies a GAN model to create bogus mouse trajectories that closely mimic genuine mouse activity. The outcomes of the performance evaluation from a defensive perspective are illustrated in Figure 2.
Figure 2 demonstrates the defensive efficacy of this technique by comparing classification outcomes for datasets produced at distinct generation intervals, such as 50 ms (minimal effect) and 500 ms (substantial effect). With the standard random generation method, all assessed machine learning models, except logistic regression, sustained high classification accuracy, reliably distinguishing genuine from bogus mouse data. Conversely, employing CTGAN, a specific GAN model, resulted in a notable decrease in several machine learning models’ classification capabilities. The attack success rate dropped by as much as 37%, signifying that, once enough genuine data points cannot be confidently identified, attackers face significant challenges reconstructing complete mouse trajectories and recovering the target password. These findings confirm the promise of GAN-based solutions as a more resilient measure for safeguarding mouse data in image-based authentication applications.
Ref. [9] extends prior acoustic side-channel attack studies that primarily focused on keyboard devices and keystroke data [10] by introducing mouse data attack techniques based on the potential leakage of sensitive information through subtle sounds generated during mouse operation. As countermeasures against such attacks, attack detection and defense techniques have been investigated by implementing and analyzing diverse side-channel-based attack scenarios, including optical, acoustic, electromagnetic, and vibration channels, with the goal of mitigating side-channel-based mouse data attacks. In Ref. [11], various IoT devices and equipment were utilized to collect sensor data in real time, followed by environment-specific noise reduction and feature extraction. The results indicate that, in optical environments, high detection performance can be achieved by employing machine learning models that classify normal versus abnormal signals using detection logic such as a sensor-box-based approach. Similar sensor-box-based detection frameworks may also enable effective detection in electromagnetic and acoustic environment. Nevertheless, practical limitations remain, including sensitivity to environmental noise and the need to account for false positives in real-time settings.
Behavior-based mouse data attack techniques have been primarily studied in the context of exploiting mouse data in web bots, along with corresponding countermeasures. Ref. [12] evaluated, through experiments, the effectiveness of deep reinforcement learning in circumventing existing behavior-based bot detection methods, with the objective of neutralizing a range of web bot detection techniques. Ref. [13] introduced an AI bot that generates human-like mouse movements and experimentally verified whether the generated mouse trajectories are nearly indistinguishable from those of real users. To counter such behavior-based mouse data attacks, ref. [14] proposed BeCAPTCHA-Mouse for human identification. This study collected real mouse movements as well as various types of bot trajectories and employed a self-constructed BeCAPTCHA-Mouse Benchmark database. The dataset includes mouse positions, click events, and timestamps. For analysis, multiple models, such as SVM, Random Forest, KNN, MLP, LSTM, and GRU, were employed and their performance was compared. Moreover, ref. [15] proposed a method to improve detection performance by visually transforming mouse movements and leveraging deep learning models. While prior studies largely relied on time-series analyses, this study is novel in that it distinguished bot and human behaviors using image-based representations. Mouse behavior data, such as webpage clicks, drags, and cursor movements, were collected from real users, and the approach was evaluated through experiments using CNN and RNN models.
For ease of reading, we give abbreviations of terms used in this paper, as shown in Table 2.

3. Proposed Mouse Data Protection Technique

In previous studies, the CTGAN model was utilized to generate synthetic mouse data that closely mimics authentic user trajectories. By integrating these realistic synthetic mouse data with actual mouse data, the defense method introduced ambiguity for attackers, and its efficacy was validated through empirical studies [16,17]. Expanding upon this approach, the current article seeks to further mitigate the risk of user authentication information exposure by generating even more lifelike synthetic mouse data. Accordingly, we examine several GAN models, including CTGAN, and develop a comprehensive mouse data protection method that exploits these models to enhance resilience against mouse data inference attacks. In this section, we outline the proposed defense methodology and describe the construction of datasets used for empirical evaluation.

3.1. Methodology of the Proposed Mouse Data Defense Technique

The mouse data protection approach we propose extends earlier CTGAN-based solutions by explicitly capturing the relationship between authentic and synthetic mouse trajectories. Consider A 1 ,   A 2 ,   ,   A n as the sequence of authentic mouse coordinates gathered from user interactions when the CTGAN-based system is deployed. From this genuine data, we train generative models to output synthetic mouse coordinates F 1 ,   F 2 ,   ,   F n designed to replicate the fundamental statistical and behavioral characteristics of the authentic mouse data. During the authentication process, an attacker monitoring mouse data encounters not just the authentic sequence { A i }, but a composite stream such as A 1 ,   F 1 ,   F 2 ,   A 2 ,   ,   A n ,   F n , where authentic and synthetic mouse data are interspersed. To extract the user’s authentication information, the attacker might employ ML-based classification algorithms in an effort to separate authentic from synthetic data; nevertheless, the adversarially trained generators significantly impair the efficacy of such classifiers, thereby lowering the attack’s probability of success. The central objective of this article is to further diminish attacker effectiveness through comprehensive evaluation of multiple GAN models to determine the configuration that achieves optimal defensive strength. Figure 3 illustrates an overview of the proposed defense methodology.
To implement this defense technique, authentic mouse movement data are initially gathered from users to train defensive generators. Each dataset contains the X and Y coordinates along with the associated elapsed time. Using these features, we utilize several GAN architectures, including CopulaGAN, CGAN, and WGAN-GP, as well as the baseline CTGAN, to produce high-fidelity bogus mouse trajectories. The generated bogus mouse data are subsequently injected into the system through WM_INPUT message events and are queued alongside genuine mouse data. During user interaction with an image-based authentication interface, any malware capturing mouse data via WM_INPUT messages inevitably acquires a composite dataset with both genuine { A i } and bogus { F i } samples. Since attackers lack prior knowledge of which events are GAN-generated and have no dependable side information for classification, it becomes considerably more challenging to identify the genuine mouse coordinates related to the authentication sequence. This mixed-stream obfuscation framework is the foundational concept of the proposed mouse data defense method.

3.2. Dataset Configuration for Experimental Evaluation

To validate the proposed defense approach, we set up our experimental procedure as a six-stage pipeline: (1) feature extraction, (2) bogus data generation, (3) data preprocessing, (4) dataset configuration, (5) model training, and (6) classification evaluation. For consistency with previous studies, we utilize the earlier collected mouse datasets under identical experimental settings and apply our GAN-based defense mechanism to these datasets.

3.2.1. Feature Extraction Step

During the feature extraction phase, we employ the same set of features as used in the previous machine learning–based mouse data attack studies, ensuring a fair comparison. In particular, we use elapsed time, current X coordinate, current Y coordinate, and the distances between previous and current X and Y coordinates. Based on these, three feature sets are defined:
F1: elapsed time, current X coordinate, current Y coordinate
F2: elapsed time, |ΔX|, |ΔY| (distances between consecutive X and Y coordinates)
F3: elapsed time, current X coordinate, current Y coordinate, |ΔX|, |ΔY|
where |ΔX| and |ΔY| represent the absolute differences between consecutive coordinates. For binary classification tasks, authentic mouse data are assigned the label 1, and bogus mouse data are assigned the label 0.

3.2.2. Bogus Data Generation Step

In this stage, we employ several GAN-based methods, including CopulaGAN [18,19], CGAN [20], and WGAN-GP [21], to synthesize bogus mouse data that mimic authentic mouse trajectories more closely than the uniformly random or heuristic noise used in prior defenses. The sample distributions generated by each GAN model are illustrated in Figure 4, Figure 5 and Figure 6.
For this study, a total of 70,314 mouse-trajectory samples were collected, and a single research operated the device according to typical, everyday usage patterns. In these figures, blue points indicate genuine mouse data while red points indicate bogus mouse data; the x-axis represents the X coordinate and the y-axis denotes elapsed time. By visual assessment, the bogus mouse data generated by GAN-based models demonstrate a high degree of similarity to the distributions of genuine mouse data. To ensure reproducibility, the experimental environment was established using the PyTorch (2.6.0) framework on a high-performance computing system equipped with an Intel Core i9-13900 CPU and an NVIDIA GeForce RTX 4090 GPU. As detailed in the hyperparameter settings, utilized the Adam optimizer with a learning rate of 0.00005, and the batch size was set to 32 for CGAN and 64 for WGAN-GP. Under these conditions, for each dataset configuration, we trained the GAN models across different epochs (50, 100, and 200) to investigate how training progression influences the authenticity and utility of the generated bogus data.
As illustrated in Figure 7, real-time WM_INPUT data capture and fake WM_INPUT event injection operate as follows. Based on the captured actual mouse coordinates ( R 1 , R 2 , , R n ), and generative adversarial networks produce bogus mouse coordinates ( F 1 , F 2 , …, F n ) that closely resemble actual mouse movements. These bogus coordinates are then injected into the system message queue by triggering corresponding WM_INPUT events. As a result, an attacker collects an interleaved sequence containing both actual and bogus data (e.g., R 1 , F 1 , F 2 , R 2 , , R n , F n ). The attacker subsequently attempts to steal the user’s actual mouse data by applying machine learning-based classification to the collected sequence. However, because the injected bogus mouse data is highly similar to actual data, the proposed method hinders reliable classification between the two classes, thereby reducing the attacker’s success rate.

3.2.3. Data Preprocessing Step

At the data preprocessing stage, normalization is performed on the features defined in Step 1, utilizing mouse data previously collected. To ensure stable and efficient model training, MinMaxScaler() [22] is employed to scale features within the [0, 1] range, as illustrated in Equation (1):
x = x   x m i n x m a x   x m i n
where xmin and xmax represent the minimum and maximum values for a feature, respectively. This normalization process reduces feature scale variance and enhances the convergence stability and accuracy of GAN-based algorithms.

3.2.4. Dataset Configuration Step

In the dataset construction phase, the outputs from Steps 1–3 are integrated to create datasets for training and evaluation in subsequent classification experiments outlined in Step 5 and Step 6. Four bogus mouse data generation intervals (50 ms, 100 ms, 250 ms, and 500 ms) are considered to produce multiple dataset variants. The comprehensive dataset configuration is outlined in Table 3.
Table 2 details that the count of genuine mouse data refers to authentic mouse events collected from user input devices, whereas the count of bogus mouse data refers to samples produced by GAN-based generation. All constructed datasets maintain a balanced 5:5 ratio between genuine and bogus samples. The dataset sample sizes are as follows: n − 1: 32,006; n − 2: 28,018; n − 3: 24,010; n − 4: 20,564; n − 5: 16,022; n − 6: 12,000; and n − 7: 8008.
For each dataset configuration, we conduct hyperparameter optimization of the classification models applied in the attack scenario to comprehensively assess the defense capability of the proposed method [23]. Table 4 summarizes the hyperparameter search space for each model.
Specifically:
K-Nearest Neighbors (KNN): metric is set to euclidean; n_neighbors ∈ {3, 5, 7, 10, 15, 20}
Logistic Regression: regularization parameter C ∈ {0.001, 0.1, 1, 10, 100, 1000}; max_iter∈{100, 200, 300}
Decision Tree: max_depth ∈ {3, 5, 7, 10, 15, 20}
Random Forest: min_samples_leaf = 5; min_samples_split = 5; max_depth ∈ {10, 20, 30}; n_estimators ∈ {100, 200, 300}
Gradient Boosting: learning_rate ∈ {0.01, 0.05, 0.1, 0.2, 0.3}; max_depth ∈ {3, 5, 7, 9}; n_estimators ∈ {100, 200, 300}
Multi-Layer Perceptron (MLP): alpha ∈ {0.0001, 0.001, 0.01, 0.1}; learning_rate = adaptive; max_iter ∈ {300, 500, 1000}
These configurations ensure an appropriate optimization of the attack models, leading to a conservative and rigorous assessment of our defense.

3.2.5. Model Training Step

At the training stage, an identical set of machine learning algorithms as used in prior research—KNN, Logistic Regression, Decision Tree, Random Forest, Gradient Boosting, and MLP—are trained on every constructed dataset. In all experiments, the dataset is split into training and test subsets following an 80:20 ratio. The test subsets are solely reserved for performance evaluation of the defense in Step 6, ensuring that the results accurately represent the generalization capability of the attack models.

3.2.6. Classification Evaluation

In this study, the attack success rate is defined in terms of the classification accuracy of the attacker’s model. To enable a meaningful comparison with prior defense approaches, we report the performance evaluation results as a percentage reduction, calculated as the relative decrease in accuracy with respect to the baseline performance.
In the concluding state, defensive performance is quantified by assessing the ability of the trained classifiers to differentiate between genuine mouse and bogus data within the held-out test data. We employ standard metrics—Accuracy, Precision, Recall, F1-Score, and AUC [24]—to evaluate classification outcomes. A greater reduction in these metrics, indicative of weaker attacker performance, signifies a stronger defense, as it implies increased complexity in classifying genuine mouse trajectories from composite mouse data. The model and parameterization that result in the greatest decline of attack classification capability are considered the most robust GAN-based mouse data protection strategy.

4. Experimental Results

This section presents an evaluation of the proposed mouse data protection technique from four analytical perspectives. First, the impact of varying data processing strategies on classification effectiveness is studied. Second, the influence of different feature definitions is explored. Third, the optimal dataset configurations identified from prior analyses are evaluated, with focus on different data generation interval settings. Lastly, variations in attack effectiveness are assessed relative to different GAN models. Across these analyses, the GAN model achieving the highest level of defensive performance is determined.

4.1. Performance Evaluation According to Different Data Processing Strategies

The first experiment explores the effect of data normalization on the efficacy of the defense technique presented. We directly compare the performance of GAN models trained with non-normalized (raw) data to those trained with normalized data. For each data generation interval, the classification accuracy of the attacker’s models is measured across all constructed datasets, and Figure 8 provides an overview of the experimental outcomes.
The findings demonstrate that training GAN models, including CopulaGAN, CGAN, and WGAN-GP, on normalized data substantially reduces the classification performance of machine learning–based attacks across every evaluated dataset. In contrast, when these models are trained on non-normalized data, most configurations experience a lesser decrease in attack performance, and in the CGAN case, normalization does not lead to a notable reduction in classifier effectiveness. This suggests that generating synthetic mouse data using normalized feature datasets is more successful at impairing the attacker’s classifiers. Therefore, we recommend normalization-based preprocessing as the most effective data processing strategy for GAN-based mouse data protection.

4.2. Performance Evaluation According to Feature Definitions

The second experiment assesses the influence of various feature definitions on the defense technique’s performance. We evaluate three different feature sets.
For each feature set (F1, F2, and F3), we train the GAN models and corresponding attacker classifiers, followed by a detailed analysis of the resulting degradation in classifier performance. The results are illustrated in Figure 9, Figure 10 and Figure 11.
The results demonstrate that feature set F2 results in the most significant decrease in attack performance in several scenarios. Nevertheless, F2 is constructed solely from elapsed time and inter-point distances, excluding the absolute X and Y coordinates. This lack of absolute positional data complicates the accurate modeling and realistic synthesis of mixed trajectories when both genuine and bogus mouse data are present, and also indicates that these features were initially designed to facilitate classification from the attacker’s perspective, rather than robust obfuscation from the defender’s perspective. As a result, F2 is not regarded as a suitable foundation for a principled defense strategy in our feature configuration, leading us to concentrate our detailed analysis on F1 and F3.
When F1 and F3 are utilized, the observed reduction in attack effectiveness is as follows. For CopulaGAN, the attack accuracy decreases by approximately 9% to 30% with F1, and by 12% to 41% with F3. In the case of CGAN, the reduction spans from 9% to 31% with F1 and 13% to 44% with F3. For WGAN-GP, the decline ranges from 9% to 29% with F1 and 12% to 39% with F3. In summary, the experimental results reveal that F3, which incorporates elapsed time, absolute X and Y coordinates, and differences between consecutive coordinates, consistently produces the greatest reduction in attack performance across all models. Therefore, in the subsequent experiments, F3 is chosen as the main feature configuration for assessing the impact of optimal dataset selection and model-specific performance.

4.3. Performance Evaluation According to Data Generation Intervals

The third evaluation assesses the effectiveness of data generation intervals determined through earlier analyses of data processing methods and feature selection. In this evaluation, (i) datasets produced using normalized features and (ii) the F3 feature set, which includes elapsed time, current X coordinate, current Y coordinate, and distances between sequential X and Y coordinates, are used. Utilizing these settings, we assess the average performance reduction observed for the attacker’s classifiers, with the outcomes displayed in Figure 12, Figure 13 and Figure 14.
The results demonstrate that, for all data generation intervals—50 ms, 100 ms, 250 ms, and 500 ms—the optimal configurations for each GAN model are achieved when CopulaGAN, CGAN, and WGAN-GP are each trained for 200 epochs. In the macro-averaged evaluation metrics, the 50 ms generation interval achieves the most substantial reduction in attack performance overall, while the 250 ms interval shows the least reduction. However, for the Gradient Boosting model, the decrease in classification accuracy is limited to about 10% across all tested configurations, indicating this model is comparatively less impacted by the addition of bogus data.
Among the GAN models evaluated, CGAN exhibits the highest defensive effectiveness. Notably, at a 50 ms generation interval, CGAN decreases the KNN model’s performance by about 44%, and at a 250 ms interval, by about 43%. These findings indicate that the use of normalized data, the F3 feature set, and well-tuned GAN training—particularly CGAN at 200 epochs with short generation intervals—offers an effective approach for reducing the success rate of machine learning–based mouse data inference attacks.

4.4. Performance Impact Analysis According to GAN Models

For the final evaluation, we assess the impact of each GAN model’s performance by comparing the relative changes in attack performance between this study and previous work. This assessment considers datasets created using (i) normalized features and (ii) the F3 feature set (elapsed time, current X coordinate, current Y coordinate, and distances between consecutive coordinates), and examines performance trends across different data generation intervals. For each GAN model, we analyze the datasets that yield the highest and lowest reductions in attack performance as observed in the 50 ms and 250 ms generation intervals, respectively. The results are summarized in Table 5, Table 6 and Table 7.
To measure the defensive impact, we calculate the percentage change in key evaluation metrics—Accuracy, Precision, Recall, F1-Score, and AUC—for each configuration by comparing the baseline attack performance from previous studies to the performance observed when applying our GAN-based defense technique. This comparison is conducted separately for CopulaGAN, CGAN, and WGAN-GP. In line with previous findings, macro-averaged metrics show that datasets generated using a 50 ms interval lead to the most substantial reduction in attacker performance, while datasets with a 250 ms interval yield the least reduction. For each GAN model, instead of assessing all possible datasets, we focus on representative datasets that produce the greatest reduction in attack performance.
For CopulaGAN, among the datasets generated with the most impactful interval (50 ms), dataset 1-1 is chosen. In this selected dataset, the highest performance degradation is observed for both the MLP and KNN classifiers: Accuracy is reduced by up to 44% (MLP), Precision by 42% (KNN and MLP), Recall by 53% (MLP), F1-Score by 48% (MLP), and AUC by 43% (MLP). For the interval with the lowest impact (250 ms), dataset 3-3 is chosen. Even here, significant degradation persists: Accuracy decreases by 46% (MLP), Precision by 47% (MLP), Recall by 42% (KNN), F1-Score by 44% (KNN and MLP), and AUC by 44% (MLP).
For CGAN, among the 50 ms datasets, dataset 1-4 is selected as the most impactful scenario. In this case, the most considerable reductions are again found for MLP and KNN: Accuracy is reduced by 48% (MLP), Precision by 49% (MLP), Recall by 48% (KNN), F1-Score by 48% (KNN), and AUC by 47% (MLP). At the 250 ms interval, dataset 3-1 is chosen. Here, KNN experiences the highest degradation, with reductions in Accuracy and Precision by 45%, Recall by 47%, F1-Score by 46%, and AUC by 43%.
For WGAN-GP, within the 50 ms interval, dataset 1-2 is identified as the representative for the highest-impact group. The MLP model shows the greatest reduction: both Accuracy and Precision fall by 48%, Recall drops by 62%, F1-Score decreases by 56%, and AUC by 49%. For the 250 ms interval, dataset 3-2 is selected. In this case, MLP and KNN both experience notable decreases: Accuracy and Precision decline by 48% (MLP), Recall by 44% (KNN), F1-Score by 45% (KNN and MLP), and AUC by 48% (MLP).

4.5. Statistical Analysis

Table 8 reports t-statistic, confidence interval, and p-value to validate performance differences across models and datasets, and Table 9 presents a statistical divergence analysis between actual and GAN-generated bogus mouse data (KL and JS divergence).
The results indicate that the WGAN-GP and CopulaGAN models generate bogus mouse data that is most similar to actual user mouse data. In particular, CopulaGAN achieves higher similarity in terms of spatial features, whereas WGAN-GP shows superior similarity with respect to temporal features. Although CGAN shows the largest divergence—suggesting relatively lower similarity to actual data—it, together with WGAN-GP, achieves a substantial decrease in attack success rate in the attack evaluation. Overall, the proposed approach can reduce the attack success rate by up to 48%, even when an attacker collects both actual and injected bogus mouse data and applies machine learning-based classification models. Moreover, this reduction is greater than that achieved by the baseline setting, demonstrating the improved effectiveness of our defense. This outcome is attributable to the fact that image-based authentication procedures typically generate a large volume of mouse movement data; consequently, the injected bogus mouse data introduces additional ambiguity into the attacker’s classification models, making it more difficult to reliably steal valid actual mouse data.

5. Conclusions

This article introduced an enhanced WM_INPUT message–based mouse data protection strategy for image-based user authentication, utilizing various GAN models to significantly diminish the effectiveness of machine learning–driven mouse data attacks. To thoroughly evaluate the defense’s effectiveness, we performed an extensive assessment across different data processing approaches, feature definitions, and data generation intervals, using standard classification metrics—Accuracy, Precision, Recall, F1-Score, and AUC—to quantify attacker performance.
Our experimental results indicate that the most effective configuration is obtained when bogus mouse data are generated using the CGAN model trained on normalized features derived from the F3 feature set, which comprises elapsed time, current X and Y coordinates, and the distances between consecutive coordinates. With this configuration, the injected bogus mouse trajectories result in the most significant reduction in the attacker’s classifier performance. In comparison with previous studies, where CTGAN-based defense lowered the attack accuracy by up to 37%, our CGAN- and WGAN-GP–based defenses achieve a reduction of up to 48%, representing an approximate 11% enhancement in defensive effectiveness. These results provide empirical evidence that meticulously designed GAN-based mouse data generation markedly enhances the robustness of image-based authentication against sophisticated machine learning–driven mouse data attacks.
In summary, the findings demonstrate that using CGAN and WGAN-GP to generate realistic bogus mouse data through the WM_INPUT message offers a viable and robust method for protecting image-based authentication information. For future research, we intend to explore three-dimensional GAN-based models that capture spatiotemporal mouse dynamics and more extensive behavioral context, aiming to further reduce the distinction between genuine and bogus mouse trajectories and to bolster resistance against increasingly advanced attack models. Additionally, we plan to address practical implementation challenges, including the optimization of data injection intervals, and computational considerations to facilitate real-world deployment.

Author Contributions

Conceptualization, J.K. and K.L.; methodology, J.K. and K.L.; software, J.K.; validation, J.K. and K.L.; data curation, J.K.; writing—original draft, J.K. and K.L.; writing—review and editing, K.L.; supervision, K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was mainly supported by the Technology Innovation Program (RS-2025-02317682, Shipyard-Partner Digital Production Cooperation Cloud Server Construction and Platform Development) funded By the Ministry of Trade, Industry & Energy (MOTIE, Korea) (50%). This research was also supported by the Regional Innovation System & Education (RISE) program through the Jeollanamdo RISE center, funded by the Ministry of Education (MOE) and the Jeollanamdo, Republic of Korea. (2025-RISE-14-001) (50%).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

This study was conducted as a research project commissioned by the Korean Association of Cybersecurity Studies (Project title: ‘2025 Cyber Security Papers Contest’, Project period: August 2025~September 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. National Information Society Agency. Survey on the Internet Usage. Available online: https://nia.or.kr/site/nia_kor/ex/bbs/View.do?cbIdx=99870&bcIdx=27869&parentSeq=27869&ref=blog.stibee.com (accessed on 11 March 2025).
  2. Ayo, F.; Awotunde, J.; Olalekan, O.; Imoize, A.; Li, C.; Lee, C. CBFISKD: A Combinatorial-Based Fuzzy Inference System for Keylogger Detection. Mathematics 2023, 11, 1899. [Google Scholar] [CrossRef]
  3. MSDN. GetCursorPos Function (winuser.h). Available online: https://learn.microsoft.com/ko-kr/windows/win32/api/winuser/nf-winuser-getcursorpos (accessed on 18 March 2025).
  4. Lee, K.; Yim, K.; Lee, S. Vulnerability analysis on the image-based authentication: Through the WM_INPUT message. Concurr. Comput. Pract. Exp. 2019, 32, e5596. [Google Scholar]
  5. MSDN. SetCursorPos Function (winuser.h). Available online: https://learn.microsoft.com/en-us/windows/win32/api/winuser/nf-winuser-setcursorpos (accessed on 18 March 2025).
  6. Lee, K.; Yim, K. A Protection Technique for Screen Image-based Authentication Utilizing the WM_INPUT message. In Proceedings of the Korean Society of Computer Information Conference, Busan, Republic of Korea, 11–13 January 2018; Volume 26, pp. 177–178. [Google Scholar]
  7. Jung, W.; Hong, S.; Lee, K. Mouse Data Attack Technique Using Machine Learning in Image-Based User Authentication: Based on a Defense Technique Using the WM_INPUT Message. Electronics 2024, 13, 710. [Google Scholar]
  8. Jung, W. A Mouse Data Protection Method using Generative Adversarial Networks for Image-Based Authentication. Master’s Thesis, Graduate School of Mokpo National University, Muan-gun, Republic of Korea, 2023; pp. 1–70. [Google Scholar]
  9. Conti, M.; Duroyon, M.; Orazi, G.; Tsudik, G. Acoustic Side-Channel Attacks on a Computer Mouse. Detect. Intrusions Malware Vulnerability Assess. 2024, 14828, 44–63. [Google Scholar]
  10. Cecconello, S.; Compagno, A.; Conti, M.; Lain, D.; Tsudik, G. Skype & Type: Keyboard Eavesdropping in Voice-over-IP. ACM Trans. Priv. Secur. 2019, 22, 1–34. [Google Scholar]
  11. Yu, J. Research on Airgap Attack and Defense Techniques in Closed Network Environment. Ph.D. Thesis, Sejong University, Seoul, Republic of Korea, 2023. [Google Scholar]
  12. Iliou, C.; Kostoulas, T.; Tsikrika, T.; Katos, V.; Vrochidis, S.; Kompatsiaris, I. Web Bot Detection Evasion Using Deep Reinforcement Learning. In Proceedings of the 17th International Conference on Availability, Reliability and Security, Vienna, Austria, 23–26 August 2022; pp. 1–10. [Google Scholar]
  13. Antal, M.; Buza, K.; Fejer, N. SapiAgent: A Bot Based on Deep Learning to Generate Human-Like Mouse Trajectories. IEEE Access. 2021, 9, 124396–124408. [Google Scholar] [CrossRef]
  14. Acien, A.; Morales, A.; Fierrez, J.; Vera-Rodriguez, R. BeCAPTCHA-Mouse: Synthetic mouse trajectories and improved bot detection. Pattern Recognit. 2022, 127, 1–13. [Google Scholar] [CrossRef]
  15. Niu, H.; Wei, A.; Song, Y.; Cai, Z. Exploring visual representations of computer mouse movements for bot detection using deep learning approaches. Expert Syst. Appl. 2023, 229, 120225. [Google Scholar] [CrossRef]
  16. Xu, L.; Skoularidou, M.; Cuesta-Infante, A.; Veeramachaneni, K. Modeling tabular data using conditional GAN. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, NC, Canada, 8–14 December 2019; Volume 659, pp. 7335–7345. [Google Scholar]
  17. Seong, D.; Lim, Y. A Study on the Optimization of Data Augmentation Ratio using CTGAN. In Proceedings of the Annual Conference of KIPS, Busan, Republic of Korea, 2–4 November 2023; Volume 30, pp. 327–330. [Google Scholar]
  18. Hong, C.; Lee, J. VaR Estimation of Multivariate Distribution Using Copula Functions. Korean J. Appl. Stat. 2011, 24, 523–533. [Google Scholar] [CrossRef]
  19. Byeon, H. Predictive Modeling for Employee Turnover Intention in South Korea Using CopulaGAN and Isolation Forests. J. Korean Soc. Mech. Technol. 2024, 26, 1244–1253. [Google Scholar]
  20. Kim, H.; Lee, D. Image denoising with conditional generative adversarial networks (CGAN) in low dose chest images. Nucl. Instrum. Methods Phys. Res. Sect. A Accel. Spectrometers Detect. Assoc. Equip. 2020, 954, 161914. [Google Scholar] [CrossRef]
  21. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved training of wasserstein GANs. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5769–5779. [Google Scholar]
  22. Deepa, B.; Ramesh, K. Epileptic seizure detection using deep learning through min max scaler normalization. Int. J. Health Sci. 2022, 6, 10981–10996. [Google Scholar] [CrossRef]
  23. Won, J.; Shin, J.; Kim, J.; Lee, J. A Survey on Hyperparameter Optimization in Machine Learning. J. Korean Inst. Commun. Inf. Sci. 2023, 48, 733–747. [Google Scholar] [CrossRef]
  24. Vujovic, Z. Classification Model Evaluation Metrics. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 599–606. [Google Scholar] [CrossRef]
Figure 1. Mouse data attack scenario using WM_INPUT message.
Figure 1. Mouse data attack scenario using WM_INPUT message.
Electronics 15 00292 g001
Figure 2. Overview of performance evaluation of the GAN-enhanced mouse data defense technique.
Figure 2. Overview of performance evaluation of the GAN-enhanced mouse data defense technique.
Electronics 15 00292 g002
Figure 3. Mouse data defense technique methodology using GAN.
Figure 3. Mouse data defense technique methodology using GAN.
Electronics 15 00292 g003
Figure 4. Visualization of the distribution of bogus data generation in the X coordinate of the CopulaGAN model.
Figure 4. Visualization of the distribution of bogus data generation in the X coordinate of the CopulaGAN model.
Electronics 15 00292 g004
Figure 5. Visualization of the distribution of bogus data generation in the X coordinate of the CGAN model.
Figure 5. Visualization of the distribution of bogus data generation in the X coordinate of the CGAN model.
Electronics 15 00292 g005
Figure 6. Visualization of the distribution of bogus data generation in the X coordinate of the WGAN-GP model.
Figure 6. Visualization of the distribution of bogus data generation in the X coordinate of the WGAN-GP model.
Electronics 15 00292 g006
Figure 7. Overview of the real-time capture of WM_INPUT data and injection of fake WM_INPUT events.
Figure 7. Overview of the real-time capture of WM_INPUT data and injection of fake WM_INPUT events.
Electronics 15 00292 g007
Figure 8. Performance evaluation results under different data processing strategies. (ac) correspond to the CopulaGAN, CGAN, and WGAN-GP models, respectively. The x-axis denotes the generation cycle of mouse data, and the y-axis denotes classification performance.
Figure 8. Performance evaluation results under different data processing strategies. (ac) correspond to the CopulaGAN, CGAN, and WGAN-GP models, respectively. The x-axis denotes the generation cycle of mouse data, and the y-axis denotes classification performance.
Electronics 15 00292 g008
Figure 9. Performance evaluation according to feature definitions of the CopulaGAN model. The x-axis denotes the features corresponding to F1, F2, and F3, and the y-axis denotes classification performance.
Figure 9. Performance evaluation according to feature definitions of the CopulaGAN model. The x-axis denotes the features corresponding to F1, F2, and F3, and the y-axis denotes classification performance.
Electronics 15 00292 g009
Figure 10. Performance evaluation according to feature definitions of the CGAN model. The x-axis denotes the features corresponding to F1, F2, and F3, and the y-axis denotes classification performance.
Figure 10. Performance evaluation according to feature definitions of the CGAN model. The x-axis denotes the features corresponding to F1, F2, and F3, and the y-axis denotes classification performance.
Electronics 15 00292 g010
Figure 11. Performance evaluation according to feature definitions of the WGAN-GP model. The x-axis denotes the features corresponding to F1, F2, and F3, and the y-axis denotes classification performance.
Figure 11. Performance evaluation according to feature definitions of the WGAN-GP model. The x-axis denotes the features corresponding to F1, F2, and F3, and the y-axis denotes classification performance.
Electronics 15 00292 g011
Figure 12. Performance evaluation results according to data generation intervals of the CopulaGAN model. The x-axis denotes the generation interval of mouse data, and the y-axis denotes classification performance.
Figure 12. Performance evaluation results according to data generation intervals of the CopulaGAN model. The x-axis denotes the generation interval of mouse data, and the y-axis denotes classification performance.
Electronics 15 00292 g012
Figure 13. Performance evaluation results according to data generation intervals of the CGAN model. The x-axis denotes the generation interval of mouse data, and the y-axis denotes classification performance.
Figure 13. Performance evaluation results according to data generation intervals of the CGAN model. The x-axis denotes the generation interval of mouse data, and the y-axis denotes classification performance.
Electronics 15 00292 g013
Figure 14. Performance evaluation results according to data generation intervals of the WGAN-GP model. The x-axis denotes the generation interval of mouse data, and the y-axis denotes classification performance.
Figure 14. Performance evaluation results according to data generation intervals of the WGAN-GP model. The x-axis denotes the generation interval of mouse data, and the y-axis denotes classification performance.
Electronics 15 00292 g014
Table 1. Summary of experimental results of mouse data attack technique using machine learning models.
Table 1. Summary of experimental results of mouse data attack technique using machine learning models.
DatasetGeneration
Interval
FeatureModelAccuracy
(Attack Success Rate)
Dataset
4–6
500ALLGradient Boosting,
Random Forest
1.000
Dataset
1–6
50ALLKNN1.000
Dataset
1–7
50ALLGradient Boosting, Decision Tree1.000
Dataset
4–6
500F2Gradient Boosting1.000
Dataset
2–6
500F2Random Forest0.998
ALL: Elapsed time, current X and Y coordinates, distances between consecutive X and Y coordinates. F2: Elapsed time, distances between successive X and Y coordinates.
Table 2. List of abbreviations used in this paper.
Table 2. List of abbreviations used in this paper.
AbbreviationDescription
GANGenerative Adversarial Networks
CTGANConditional Tabular Generative Adversarial Networks
CGANConditional Generative Adversarial Networks
CopulaGANCopula Generative Adversarial Networks
WGAN-GPWasserstein Generative Adversarial Networks—Gradient Penalty
AUCArea Under Curve
KNNK-Nearest Neighbors
MLPMulti-Layer Perceptron
Table 3. Overall dataset configuration reconstructed using GAN model outputs.
Table 3. Overall dataset configuration reconstructed using GAN model outputs.
DatasetGeneration
Intervals
Aggregate Count of Mixed Mouse MovementsAggregate Count of Genuine Mouse MovementsAggregate Count of Bogus Mouse MovementsRatio
1–150 ms32,00616,00316,0035:5
1–228,01814,00914,009
1–324,01012,00512,005
1–420,56410,28210,282
1–516,02280118011
1–612,00060006000
1–7800840044004
2–1100 ms32,00616,00316,0035:5
2–228,01814,00914,009
2–324,01012,00512,005
2–420,56410,28210,282
2–516,02280118011
2–612,00060006000
2–7800840044004
3–1250 ms32,00616,00316,0035:5
3–228,01814,00914,009
3–324,01012,00512,005
3–420,56410,28210,282
3–516,02280118011
3–612,00060006000
3–7800840044004
4–1500 ms32,00616,00316,0035:5
4–228,01814,00914,009
4–324,01012,00512,005
4–420,56410,28210,282
4–516,02280118011
4–612,00060006000
4–7800840044004
Table 4. Hyperparameter search space of all classification models.
Table 4. Hyperparameter search space of all classification models.
ModelHyperparameter
KNNmetric: euclidean, n_neighbors: 3~20
Logistic RegressionC: 0.001~1000, max_iter: 100~300
Decision Treemax_depth: 3~20
Random Forestmax_depth: 10~30, min_samples_leaf: 5, min_samples_split: 5, n_estimators: 100~300
Gradient Boostinglearning_rate: 0.01~0.3, max_depth: 3~9,
n_estimators: 100~300
MLPalpha: 0.0001~0.1, learning_rate: adaptive,
max_iter: 300~1000
Table 5. Summary of performance evaluation results according to the relative performance change of the CopulaGAN model.
Table 5. Summary of performance evaluation results according to the relative performance change of the CopulaGAN model.
Da/GIMDAC±P±R±F±AU±
1–1/
50 ms
KNNRD0.994−43%0.990−42%0.999−44%0.994−43%0.998−42%
GAN0.5620.5680.5620.5650.583
LRRD0.845−33%0.841−32%0.856−38%0.849−35%0.930−41%
GAN0.5150.5230.4740.4970.522
DTRD0.993−21%0.994−22%0.993−19%0.993−21%0.993−18%
GAN0.7800.7710.8020.7860.818
RFRD0.994−24%0.993−25%0.995−22%0.994−23%1−16%
GAN0.7540.7450.7800.7620.837
GBRD0.998−11%0.998−13%0.998−8%0.998−10%1−5%
GAN0.8910.8720.9200.8950.953
MLPRD0.986−44%0.986−42%0.987−53%0.987−48%0.998−43%
GAN0.5500.5670.4600.5080.568
3–3/
250 ms
KNNRD0.998−44%0.997−45%0.999−42%0.998−44%0.999−42%
GAN0.5610.5500.5750.5620.578
LRRD0.740−22%0.757−25%0.721−13%0.738−19%0.825−30%
GAN0.5170.5060.5930.5460.521
DTRD0.996−20%0.996−21%0.996−18%0.996−20%0.995−16%
GAN0.8010.7850.8180.8010.831
RFRD0.996−21%0.995−24%0.998−15%0.997−20%1−13%
GAN0.7900.7530.8500.7980.872
GBRD0.998−10%0.999−14%0.997−6%0.998−10%0.999−4%
GAN0.8960.8590.9410.8980.958
MLPRD0.997−46%0.996−47%0.998−41%0.997−44%1−44%
GAN0.5400.5270.5900.5570.565
Da/GI: Dataset/Generation Interval, LR: Logistic Regression, DT: Decision Tree, RF: Random Forest, GB: Gradient Boosting, RD: Random Dataset, GAN: GAN dataset, AC: ACcuracy, P: Precision, R: Recall, F: F1-score, AU: AUC, ±: Increase/Decrease rate.
Table 6. Summary of performance evaluation results according to the relative performance change of the CGAN model.
Table 6. Summary of performance evaluation results according to the relative performance change of the CGAN model.
Da/GIMDAC±P±R±F±AU±
1–4/
50 ms
KNNRD0.994−47%0.990−48%0.999−48%0.994−48%0.998−46%
GAN0.5280.5130.5230.5180.542
LRRD0.845−36%0.841−36%0.8566%0.849−22%0.930−41%
GAN0.4880.4850.9140.6330.520
DTRD0.993−24%0.994−26%0.993−23%0.993−24%0.993−21%
GAN0.7550.7380.7680.7520.786
RFRD0.994−25%0.993−27%0.995−25%0.994−26%1−18%
GAN0.7410.7270.7430.7350.825
GBRD0.998−16%0.998−18%0.998−14%0.998−16%1−8%
GAN0.8400.8190.8610.8390.920
MLPRD0.986−48%0.986−49%0.987−28%0.987−41%0.998−47%
GAN0.5050.4920.7040.5800.528
3–1/
250 ms
KNNRD0.998−45%0.997−45%0.999−47%0.998−46%0.999−43%
GAN0.5490.5490.5290.5390.567
LRRD0.740−24%0.757−26%0.721−20%0.738−23%0.825−32%
GAN0.5000.4980.5240.5110.503
DTRD0.996−20%0.996−22%0.996−17%0.996−20%0.995−15%
GAN0.7970.7800.8240.8010.846
RFRD0.996−21%0.995−22%0.998−18%0.997−20%1−13%
GAN0.7910.7750.8150.7950.873
GBRD0.998−10%0.999−12%0.997−7%0.998−10%0.999−4%
GAN0.9010.8770.9310.9030.963
MLPRD0.997−43%0.996−43%0.998−41%0.997−42%1−41%
GAN0.5700.5650.5870.5760.595
Da/GI: Dataset/Generation Interval, LR: Logistic Regression, DT: Decision Tree, RF: Random Forest, GB: Gradient Boosting, RD: Random Dataset, GAN: GAN dataset, AC: ACcuracy, P: Precision, R: Recall, F: F1-score, AU: AUC, ±: Increase/Decrease rate.
Table 7. Summary of performance evaluation results according to the relative performance change of the WGAN-GP model.
Table 7. Summary of performance evaluation results according to the relative performance change of the WGAN-GP model.
Da/GIMDAC±P±R±F±AU±
1–2/
50 ms
KNNRD0.994−46%0.99−45%0.999−44%0.994−45%0.998−45%
GAN0.5380.5380.5550.5460.548
LRRD0.845−34%0.841−34%0.856−39%0.849−36%0.930−43%
GAN0.5020.5040.4690.4860.504
DTRD0.993−22%0.994−24%0.993−17%0.993−21%0.993−18%
GAN0.7730.7500.8220.7840.818
RFRD0.994−24%0.993−25%0.995−22%0.994−24%1−16%
GAN0.7510.7410.7740.7750.837
GBRD0.998−13%0.998−15%0.998−11%0.998−13%1−6%
GAN0.8650.8470.8920.8690.935
MLPRD0.986−48%0.986−48%0.987−62%0.987−56%0.998−49%
GAN0.5020.5050.370.4270.506
3–2/
250 ms
KNNRD0.998−46%0.997−45%0.999−44%0.998−45%0.999−45%
GAN0.5420.5450.5550.5500.548
LRRD0.740−23%0.757−24%0.721−27%0.738−25%0.825−31%
GAN0.5150.5220.4500.4840.515
DTRD0.996−19%0.996−22%0.996−13%0.996−18%0.995−13%
GAN0.8080.7760.8710.8210.861
RFRD0.996−20%0.995−23%0.998−15%0.997−19%1−13%
GAN0.7940.7690.8450.8050.873
GBRD0.998−11%0.999−14%0.997−6%0.998−10%0.999−4%
GAN0.8880.8560.9350.8940.958
MLPRD0.997−48%0.996−48%0.998−41%0.997−45%1−48%
GAN0.5200.5210.5850.5510.525
Da/GI: Dataset/Generation Interval, LR: Logistic Regression, DT: Decision Tree, RF: Random Forest, GB: Gradient Boosting, RD: Random Dataset, GAN: GAN dataset, AC: Accuracy, P: Precision, R: Recall, F: F1-score, AU: AUC, ±: Increase/Decrease rate.
Table 8. Summary of performance comparison to validate differences across models and datasets.
Table 8. Summary of performance comparison to validate differences across models and datasets.
CopulaGAN
Featuret-StatisticConfidence Interval (95%)p-Value (One-Tailed)
F110.6830.182~0.2986.222 × 10−5
F219.5620.378~0.4923.222 × 10−6
F35.8160.154~0.3981.060 × 10−3
CGAN
Featuret-statisticConfidence interval (95%)p-Value (one-tailed)
F16.9480.143~0.3104.744 × 10−4
F219.5860.389~0.5063.203 × 10−6
F35.7000.164~0.4331.160 × 10−3
WGAN-GP
Featuret-statisticConfidence interval (95%)p-Value (one-tailed)
F17.1090.129~0.2754.270 × 10−4
F220.9240.363~0.4642.309 × 10−6
F35.8770.151~0.3871.013 × 10−3
Table 9. Statistical divergence analysis between actual and GAN-generated bogus mouse data (KL and JS Divergence).
Table 9. Statistical divergence analysis between actual and GAN-generated bogus mouse data (KL and JS Divergence).
CopulaGAN
FeatureKL (Actual||GAN)KL (GAN||Actual)JS Divergence
TIME0.0208630.6365830.007110
X0.0543110.0515120.012997
Y0.0317780.0301150.007667
DIFFPOSX0.0085410.0033570.000876
DIFFPOSY0.0038360.0038280.000955
CGAN
FeatureKL (Actual||GAN)KL (GAN||Actual)JS divergence
TIME0.0032740.0032640.000817
X0.5171830.1560800.022026
Y0.1120070.2475690.026299
DIFFPOSX0.0108270.0056450.001435
DIFFPOSY0.0051410.0050430.001270
WGAN-GP
FeatureKL (Actual||GAN)KL (GAN||Actual)JS divergence
TIME0.0108060.0093620.002478
X0.0456960.0948370.011406
Y0.1976050.0597120.010226
DIFFPOSX0.0038930.0038140.000961
DIFFPOSY0.0040540.0075410.000999
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, J.; Lee, K. Mouse Data Protection in Image-Based User Authentication Using Two-Dimensional Generative Adversarial Networks: Based on a WM_INPUT Message Approach. Electronics 2026, 15, 292. https://doi.org/10.3390/electronics15020292

AMA Style

Kim J, Lee K. Mouse Data Protection in Image-Based User Authentication Using Two-Dimensional Generative Adversarial Networks: Based on a WM_INPUT Message Approach. Electronics. 2026; 15(2):292. https://doi.org/10.3390/electronics15020292

Chicago/Turabian Style

Kim, Jinwook, and Kyungroul Lee. 2026. "Mouse Data Protection in Image-Based User Authentication Using Two-Dimensional Generative Adversarial Networks: Based on a WM_INPUT Message Approach" Electronics 15, no. 2: 292. https://doi.org/10.3390/electronics15020292

APA Style

Kim, J., & Lee, K. (2026). Mouse Data Protection in Image-Based User Authentication Using Two-Dimensional Generative Adversarial Networks: Based on a WM_INPUT Message Approach. Electronics, 15(2), 292. https://doi.org/10.3390/electronics15020292

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop