You are currently viewing a new version of our website. To view the old version click .
Systems
  • Article
  • Open Access

19 July 2025

A Hybrid CNN–BiLSTM Framework Optimized with Bayesian Search for Robust Android Malware Detection

Department of Information Science, College of Humanities and Social Sciences, King Saud University, Riyadh P.O. Box 11451, Saudi Arabia
This article belongs to the Special Issue Cyber Security Challenges in Complex Systems

Abstract

With the rapid proliferation of Android smartphones, mobile malware threats have escalated significantly, underscoring the need for more accurate and adaptive detection solutions. This work proposes an innovative deep learning hybrid model that combines Convolutional Neural Networks (CNNs) with Bidirectional Long Short-Term Memory (BiLSTM) networks for learning both local features and sequential behavior in Android applications. To improve the relevance and clarity of the input data, Mutual Information is applied for feature selection, while Bayesian Optimization is adopted to efficiently optimize the model’s parameters. The designed system is tested on standard Android malware datasets and achieves an impressive detection accuracy of 99.3%, clearly outperforming classical approaches such as Support Vector Machines (SVMs), Random Forest, CNN, and Naive Bayes. Moreover, it delivers strong outcomes across critical evaluation metrics like F1-score and ROC-AUC. These findings confirm the framework’s high efficiency, adaptability, and practical applicability, making it a compelling solution for Android malware detection in today’s evolving threat landscape.

1. Introduction

The proliferation of Android devices—driven by their affordability, open-source nature, and widespread market presence—has made the platform a prime target for mobile malware. The availability of alternative app distribution platforms and the fundamental openness of the Android ecosystem further shape an environment favorable to the spread of malicious applications. Reports indicate a sharp rise in Android-targeted malware, with earlier estimates suggesting that a new malicious app emerged every seven seconds as early as 2018 [1,2,3]. According to AV-TEST, there were approximately 6.06 billion recorded malware attacks worldwide in 2023, marking a 10% rise compared to the previous year. As of 2024, the situation continues to deteriorate, with daily detections exceeding 450,000 new malware variants and potentially harmful applications (PUAs) by cybersecurity experts. This ongoing surge in threats reflects the increasing sophistication of Android malware and highlights the urgent need for more intelligent and resilient detection strategies [4].
Android malware appears in multiple forms, including trojans, ransomware, spyware, adware, and riskware, with each variant engineered to take advantage of specific flaws in the operating system or in user behavior [5,6,7]. According to Kaspersky, adware represented 46% of all mobile threats in Q1 2024, while RiskTool-type apps (categorized as riskware) accounted for 21.27% of detections [8,9]. The continued dominance of these threats underscores the persistent challenges in mobile security and the adaptive nature of malware authors. This diversity in malware functionality necessitates detection systems capable of analyzing both behavioral and structural patterns.
To mitigate these growing threats, researchers have developed a variety of detection strategies, including techniques such as static and dynamic program inspection, along with conventional machine learning and modern deep learning methodologies [10,11,12]. Although traditional detection systems have proven useful, they often struggle to counter advanced evasion strategies employed by modern and highly sophisticated malware variants [11,13]. In contrast, recent research highlights the potential of deep learning models, which are particularly effective at identifying intricate behavioral patterns and irregularities that static and rule-based systems frequently miss [10,14,15]. Among these, hybrid deep learning architectures have demonstrated particular promise, as they integrate mechanisms to extract both localized and sequential characteristics of application behavior, resulting in notable gains in detection accuracy and model resilience.
In light of this, the current research proposes a hybrid framework for Android malware detection that combines Convolutional Neural Networks (CNNs) with Bidirectional Long Short-Term Memory (BiLSTM) layers. To optimize the model, Mutual Information (MI) was chosen for feature selection due to its ability to capture non-linear dependencies between features and the target variable, its model-independence, and its efficiency in large feature spaces. Unlike wrapper-based methods such as Recursive Feature Elimination [10] or filter-based methods like chi-square, MI avoids classifier-specific biases and reduces computational overhead—critical in high-dimensional Android malware datasets [11]. Bayesian Optimization is then employed to systematically adjust key hyperparameters and maximize model efficiency. This architecture was selected due to its synergistic advantages: CNNs are well-suited for extracting localized behavioral signatures from app activity, and BiLSTMs effectively capture temporal patterns in both forward and backward sequences—essential for identifying complex malware behavior. When compared with earlier configurations, including those using LSTM–GRU chains or tuning methods based on the Earthworm Optimization Algorithm, this model exhibits superior accuracy, achieving a detection rate of 99.3%. The system has been rigorously validated using publicly available Android malware datasets and shows strong adaptability across a wide range of attack types and evaluation benchmarks.
The remainder of this paper is structured as follows: Section 2 provides a review of the related literature. Section 3 outlines the methodology employed in this study. Section 4 describes the experimental setup and presents the simulation results. Section 5 concludes the study and suggests potential avenues for future work.

3. Proposed Methodology

This section outlines the architecture and operational pipeline of the proposed Android malware detection framework, which integrates CNN and BiLSTM networks in a hybrid model. The overall goal is to advance detection accuracy, generalizability, and computational effectiveness. The process is structured into four key stages: acquiring and preparing the dataset, identifying relevant features through Mutual Information analysis, tuning hyperparameters via Bayesian Optimization, and ultimately classifying malware using the CNN–BiLSTM model.

3.1. Data Collection and Preprocessing

Let D = { X 1 , X 2 , , X n } represent the collected dataset, where each X i denotes a feature vector for the i t h Android application sample, composed of mmm behavioral or permission-based features extracted from static and/or dynamic sources. Inconsistencies such as absent entries and uneven feature scaling are often present in the original dataset, necessitating preprocessing to enhance learning stability and model performance.
Handling Missing Values
Let M ( X i ) represent the subset of features with missing entries in the sample X i . Missing data is addressed either by discarding incomplete samples or by applying imputation strategies (e.g., mean or median imputation):
X i = I m p u t e X i , for   all   i   where   M ( X i )
The resulting cleaned dataset is denoted as D = { X i M ( X i ) = } .
Feature Normalization
To ensure consistent feature scaling, Z-score normalization is applied across all features. Prior to normalization, missing values in continuous features were imputed using the mean, while categorical feature gaps (if any) were filled using the mode. For the j t h feature of the i t h sample X i j , the normalized value X * i j is computed as follows:
X * i j = X i j μ j 1 σ j
where μ j i   σ j 1 denote the mean and standard deviation of the j t h feature across the entire dataset:
μ j i = 1 n X i j ,   σ j 1 = 1 n ( X i j μ j 1 ) 2
The normalized dataset is thus defined as follows:
D = { X * i 1 X * i 1 = ( X * i 1 , X * i 2 , , X * i m ) } , i = 1 , , n
This normalized dataset D is then passed to the feature selection stage, during which unnecessary or duplicate features are removed to improve the model’s clarity and effectiveness.

3.2. Feature Selection Using Mutual Information

After data normalization, the process of selecting features is carried out to isolate the most relevant variables within the refined dataset D . This step enhances model performance by reducing dimensionality, eliminating redundant features, and improving interpretability. In this study, Mutual Information (MI) is used as the feature selection method due to its effectiveness in measuring the dependency between individual features and the target labels.
Mutual Information Computation
Let X R n × m represent a matrix of input features, where n refers to the number of observations and m to the number of attributes per observation. Let y { 0,1 } n denote the binary labels assigned to each instance. The Mutual Information I j for each feature j is computed as follows:
I j = M I ( X j ; y ) , j = 1,2 , , m
where X j is the j t h feature column in M I ( X j ; y ) and denotes the Mutual Information between feature X j and the label vector y . The MI values reflect how much information a feature shares with the output class.
Feature Ranking and Selection
All features are ranked in descending order of their MI scores. Let C F i j be the permutation of feature indices such that
C F i j = c o u n t ( i , j ) N
The top k features are selected based on their relevance rankings. Let X t o p denote the refined feature matrix that includes only the k most informative features:
X t o p = X [ : , Φ ( 1 : k ) ]
In this study, we set k = 10 , selecting the 10 most informative features:
X t o p = X [ : , Φ ( 1 : 10 ) ]
This choice balances model performance with computational efficiency, as smaller feature subsets reduce training time while retaining high discriminative power. Similar settings have been used effectively in malware detection studies [29]. The refined feature subset, denoted as X t o p , is subsequently utilized for training the model and tuning its hyperparameters through Bayesian Optimization.

3.3. Hyperparameter Tuning Using Bayesian Optimization

To enhance the effectiveness of the proposed CNN–BiLSTM model, this phase concentrates on tuning its hyperparameters through Bayesian Optimization—a probabilistic, globally focused optimization method designed to efficiently identify optimal settings within complex, high-dimensional search spaces. Unlike evolutionary or population-based strategies, Bayesian Optimization builds a surrogate probabilistic model, specifically a Gaussian Process, to approximate the objective function. It then applies an acquisition function, the Expected Improvement (EI), to strategically balance the trade-off between exploring new configurations and exploiting known promising regions. This setup—using GP as the surrogate and EI as the acquisition function—was chosen due to their strong performance in high-dimensional and non-convex deep learning problems, as documented in prior studies [21,29,34].
The number of iterations for Bayesian Optimization was set to 30, which provided a balance between computational feasibility and thorough exploration of the hyperparameter space. This configuration was selected for its proven ability to handle non-convex objective functions and to converge rapidly towards optimal hyperparameter settings in deep learning applications.
Hyperparameter Search Space
The search space for Bayesian Optimization was defined using a hybrid approach that combines empirical tuning and prior research in malware detection and deep learning frameworks [24,28]. Specifically, the learning rate was constrained between 0.0001 and 0.01, dropout rates ranged from 0.1 to 0.5, and the number of BiLSTM units was set between 64 and 256. These bounds reflect common practice in cybersecurity-focused deep learning studies and were validated during preliminary experiments to ensure coverage of the most relevant parameter regions. This design balances flexibility with computational feasibility, enabling Bayesian Optimization to effectively navigate toward optimal configurations.
Optimization Objective
Let Θ R d denote the vector of d hyperparameters to be optimized (e.g., learning rate, batch size, number of filters, units in BiLSTM, dropout rate). The objective is to find the optimal configuration Θ * that minimizes the loss function L ( Θ ) on the validation set:
Θ * = a r g   m i n Θ H L ( Θ )
where H represents the feasible hyperparameter space, and L ( Θ ) measures the validation loss of the model trained using Θ .
Bayesian Optimization Process
The process begins by evaluating the objective function at a small set of initial points sampled from H . A surrogate model f : H R is then trained to approximate L ( Θ ) . At each iteration t , BO selects the next candidate i t by optimizing an acquisition function α ( Θ ; f ) , that manages a trade-off between investigating less certain regions and leveraging areas already identified as potentially optimal:
Θ t = a r g m a x Θ H α ( Θ ; f )
The true objective L ( Θ t ) is then evaluated, and the surrogate model is refined by incorporating this newly obtained data point ( Θ t ,   L ( Θ t ) ) .
Termination Criteria
The optimization proceeds until a predefined condition is satisfied, such as reaching the maximum allowed iterations T max or a convergence threshold ϵ on the improvement:
| L ( Θ t ) L ( Θ t 1 ) | < ϵ   o r   t > T max
Upon termination, the best-performing hyperparameter vector Θ * is selected:
Θ * = a r g m i n _ { Θ 1 , , Θ t } L ( Θ )
Model Training with Optimized Parameters
The optimized hyperparameters Θ * are subsequently employed to train the final CNN–BiLSTM model using the refined feature subset X top identified in Section 3.2:
T r a i n e d   M o d e l = T r a i n ( X top , y , Θ * )
The final trained model is subsequently evaluated on the test dataset to assess generalization and classification performance.

3.4. Prediction Using the Proposed Model

The CNN–BiLSTM hybrid model, at this stage, is trained and assessed using the best-performing set of hyperparameters, denoted as Θ * , obtained through Bayesian Optimization. Let X top   R nxk represent the optimized input matrix resulting from Stage 2, where n corresponds to the total number of samples and k refers to the dimensionality of the features selected.
CNN Layer
The model begins with two one-dimensional convolutional layers designed to extract local feature patterns from the input sequence. The first convolutional layer employs 32 filters with a kernel size of 3 × 3 and a stride of 1, followed by a ReLU activation function to introduce nonlinearity. A max-pooling layer (pool size = 2 × 2) follows to downsample the feature maps.
The second convolutional layer uses 64 filters with the same kernel size and activation settings, again followed by a max-pooling layer. This design helps reduce the dimensionality of intermediate representations while preserving important spatial features.
Let x t denote the input vector at time step t . A convolution operation produces a feature map f t as follows:
f t = σ ( W conv x t + b conv )
where
denotes the convolution operation;
W conv and b conv represent the convolutional filter’s associated weights and bias terms;
σ is an activation function (e.g., ReLU).
The output F = [ f 1 , f 2 , , f T ] is then passed to the BiLSTM layer.
BiLSTM Layer
The Bidirectional LSTM (BiLSTM) consists of one layer with 128 hidden units, configured to process the output from the CNN stack in both forward and backward directions. This configuration enables the model to capture temporal dependencies in Android app behavior from both past and future contexts.
At each time step t, the forward and backward hidden states are computed as follows:
h t = L S T M f w ( f t ) , h t = L S T M b w ( f t )
The final hidden state h t is the concatenation of both:
h t = h t ; h t
The last BiLSTM output h T is used for classification.
Prediction Layer
The last BiLSTM output hT is fed into a dense output layer, which employs a softmax activation to produce class probability distributions:
y ^ = s o f t m a x ( W o u t h T + b o u t )
where
W o u t is the weight matrix;
b o u t is the bias vector;
y ^ is the predicted class probability distribution.
Model Training
To improve generalization and prevent overfitting, the training process integrates several regularization techniques beyond dropout, including the following:
  • Batch normalization after each convolutional and BiLSTM layer to stabilize learning and reduce internal covariate shift;
  • Early stopping, monitoring validation loss with a patience of 2 epochs to halt training if no improvement occurs;
  • L2 weight regularization (λ = 0.0001) applied to all dense layers to penalize large weights.
The model is trained to minimize the categorical cross-entropy loss function L using the optimized hyperparameters Θ * :
L ( y , y ^ ) = c = 1 1 C 1 y c l o g ( y ^ c )
where
C is the number of output classes;
y c is the true label, and y ^ c is the predicted probability for class c .
Final Prediction
Once trained, the model is used to classify new input data X new :
y ^ new = P r e d i c t ( X new , Θ * )
The classification decision for each input is made by selecting the category associated with the highest probability score. This method enables the model to assign the most probable label based on its internal representations. This strategy is widely adopted in classification problems that generate probability distributions via a softmax function:
C l a s s = a r g m a x ( y ^ new )

4. Results and Discussion

4.1. System Specification

The proposed CNN–BiLSTM malware detection framework was implemented and evaluated in a robust computational environment using Python 3.11.5, managed via Anaconda. For deep learning development, we employed PyTorch 2.4v [40], a widely adopted framework known for its flexibility and performance. Key data preprocessing tasks, including cleaning and normalization, were performed using Pandas and NumPy [41]. Feature selection through Mutual Information and classification metric evaluations leveraged tools from the scikit-learn library [40,41].
For hyperparameter tuning, we integrated Bayesian Optimization using the Optuna library [42], replacing traditional heuristic optimizers to ensure a more efficient and probabilistically guided search of the hyperparameter space. This setup provided sufficient computational resources for effectively training and evaluating the deep learning framework.

4.2. Dataset Representation

The dataset used in this study was obtained from Kaggle, which provides a publicly available benchmark for Android malware detection research. It comprises a total of 10,000 Android application samples, evenly split between 4950 malware and 5050 benign applications. These samples include diverse categories of Android apps collected from official and third-party sources, ensuring a broad representation of real-world behaviors and potential attack vectors.
Each application is described using 34 distinct features, encompassing both system-level attributes (e.g., memory usage, CPU time, file descriptors) and behavior-specific characteristics (e.g., access frequency, execution patterns, API call statistics). Notable attributes include utime, static_prio, free_area_cache, and vm_truncate_count, which have been identified as critical indicators of malware activity.
As shown in Figure 1, the dataset exhibits a nearly uniform class distribution (50.5% benign vs. 49.5% malware), which minimizes the risk of model bias toward any single class during training. This natural balance eliminated the need for synthetic rebalancing or data augmentation techniques.
Figure 1. Class distribution of Android applications (malware vs. benign).
For model development, the dataset was randomly partitioned into 80% for training (8000 samples) and 20% for testing (2000 samples). The training set was used to optimize the model parameters and hyperparameters, while the test set served exclusively for evaluating the final performance metrics, ensuring unbiased validation.

4.3. Performance of Feature Selection

To enhance the detection performance and computational efficiency of our CNN–BiLSTM framework, we applied the Mutual Information (MI) technique for feature selection. Unlike ensemble-based methods such as Random Forest, MI offers a model-independent, information-theoretic approach that measures the dependency between each feature and the target label. This makes it particularly effective in identifying features that carry the highest discriminative power, regardless of linearity assumptions.
The MI scores were computed for all 34 original features in the dataset, and the top 10 features with the highest MI values were selected for training the model. These top-ranked features provide the greatest amount of information gain regarding the target classes (malware vs. benign). As illustrated in Figure 2, features such as static_prio, utime, and free_area_cache showed significantly high MI scores, underscoring their critical role in differentiating between harmful and safe Android applications.
Figure 2. Top 10 features selected via Mutual Information for Android malware detection.
This dimensionality reduction process not only helps mitigate overfitting but also improves the training speed and inference efficiency of the model. By prioritizing the most relevant features, the detection framework gains improved robustness and clarity, which is especially beneficial for practical applications in mobile cybersecurity.
To further investigate inter-feature relationships among the selected features, a correlation matrix is plotted in Figure 3. This visualization aids in understanding the degree of linear association between each feature pair and the classification label. Correlation values close to +1 or −1 indicate strong relationships, whereas those near 0 propose minimal interaction. Such insight supports a more informed model design, especially in the context of deep learning, where feature interactions can impact network convergence and performance.
Figure 3. Correlation matrix of selected features and classification label.
Further analysis of the top 10 features revealed clear correlations with the classification target, offering insight into their predictive value. Notably, features such as nvcsw (−0.36), utime (−0.34), and static_prio (−0.31) demonstrated strong negative associations with the malware label. These trends indicate that lower values of these features are more frequently linked to malicious applications, whereas higher values are more indicative of benign behavior. In contrast, prio exhibited a mild positive correlation (approximately +0.12), suggesting a limited but potentially informative contribution to the classification process.
Understanding these correlation patterns is essential not only for interpreting feature relevance but also for detecting multicollinearity among input variables—an important consideration in deep learning models. Excessive multicollinearity can lead to redundant representations, which may impact the convergence and generalization performance of the model. The correlation matrix in Figure 3 helped identify such dependencies, ensuring that the selected features contribute uniquely and effectively to the model’s learning process.
To gain further insight into how these features influence classification outcomes, we visualized the distribution of the top 10 features across both target classes: benign and malware. As shown in Figure 4, the density plots clearly demonstrate differentiable patterns between the two categories. Features such as utime and free_area_cache display non-overlapping density ranges, indicating strong class-separability. This visualization confirms the relevance of the selected features and validates their inclusion in the proposed CNN–BiLSTM model for effective Android malware detection.
Figure 4. Comparative box plot of top 10 discriminative features between malware and benign Android applications.
When examined across the two target classes, the density plots further highlight how individual features contribute to discriminating between benign and malicious applications. Notably, features such as static_prio, nvcsw, and utime exhibit clear distributional separation, with benign samples concentrated within specific value ranges that differ markedly from those of malware instances. This class-specific variation underscores their strong predictive utility within the model.
Other features, including free_area_cache and vm_truncate_count, although displaying some degree of distributional overlap, still convey informative signals that enhance the model’s decision-making ability when combined with more distinctive attributes. Their inclusion contributes to a richer and more nuanced feature representation, which is particularly beneficial in capturing subtle behavioral patterns often exhibited by sophisticated Android malware.
These visualizations not only complement the correlation matrix presented in Figure 3 but also validate the Mutual Information-based selection process employed in our study. Features with higher MI scores—corresponding to those with more distinctive class distributions—demonstrate their effectiveness both statistically and visually. Ultimately, this ensures that our CNN–BiLSTM framework is trained on the most discriminative subset of features, leading to enhanced classification accuracy and generalization performance in Android malware detection.

4.4. Performance of Bayesian Optimization

To improve the detection performance of our CNN–BiLSTM framework, we utilized Bayesian Optimization for hyperparameter tuning, moving beyond traditional grid search or manual trial-and-error methods. This approach was chosen for its efficiency in modeling the objective function and strategically exploring the hyperparameter space, focusing on the most promising areas to enhance model accuracy while minimizing computational cost. The Bayesian Optimization process employed a Gaussian Process (GP) surrogate model and Expected Improvement (EI) as the acquisition function. This combination is widely adopted in deep learning hyperparameter tuning due to its ability to manage non-linear, noisy, and expensive-to-evaluate objective functions [43].
Key hyperparameters adjusted during the optimization process encompassed the learning rate, dropout percentage, mini-batch size, and the count of BiLSTM units, each significantly influencing model convergence behavior, generalization capacity, and training stability. Bayesian Optimization enhanced the tuning process by constructing a probabilistic representation of the objective function, using previously observed outcomes to efficiently explore the most promising parameter combinations.
As demonstrated in Figure 5, Figure 6, Figure 7 and Figure 8, the hyperparameter optimization process—driven by the Bayesian Optimization algorithm—exhibited efficient convergence behavior and effective performance tuning for the proposed CNN–BiLSTM model.
Figure 5. Execution time per iteration during EOA hyperparameter optimization.
Figure 6. Population diversity across EOA iterations.
Figure 7. Exploration vs. exploitation trade-off in EOA process.
Figure 8. Convergence of global best fitness over iterations.
Figure 5 illustrates the runtime per iteration, which remained within an acceptable computational range (approximately 407 to 439 s). This consistency underscores the scalability and feasibility of applying Bayesian Optimization to complex deep learning frameworks without incurring excessive resource consumption. Moreover, we estimated the theoretical computational cost of the proposed CNN–BiLSTM model in terms of floating point operations (FLOPs). The model requires approximately 1.8 GFLOPs per inference, which is competitive compared to other lightweight hybrid architectures reported in the literature. While Bayesian Optimization and BiLSTM layers introduce moderate computational overhead, their inclusion significantly improves detection accuracy and robustness, justifying the trade-off. Compared to traditional models such as SVM and XGBoost, which require substantially fewer FLOPs (~0.05–0.1 GFLOPs), our framework achieves a balance between computational complexity and detection performance, making it suitable for deployment in environments with moderate hardware capabilities.
Figure 6 presents the diversity metric across iterations, indicating that the optimization algorithm preserved a high level of population diversity throughout the search process. Maintaining this diversity is essential for exploring the global search space and avoiding premature convergence to local optima.
Figure 7 visualizes the exploration-to-exploitation ratio maintained during the optimization process, with exploration consistently exceeding 90%. This strategic balance enabled a broad search of the hyperparameter space while progressively refining regions with high performance potential.
Figure 8 shows the trajectory of the global best fitness score, which improved significantly during the early iterations and stabilized by the fifth iteration. This trend reflects the rapid convergence behavior of Bayesian Optimization and its success in identifying high-performing configurations early in the tuning cycle. To ensure that such rapid convergence did not result in overfitting, additional regularization techniques were applied alongside dropout, including early stopping and batch normalization. These measures enhanced the model’s generalization ability, enabling it to achieve a classification accuracy of 99.3% while maintaining robustness across unseen data. The resulting configuration enabled the model to achieve a classification accuracy of 99.3%, significantly outperforming all baseline models evaluated in this study. Beyond accuracy, the model’s performance was further assessed using additional evaluation metrics including precision (99.1%), Recall (98.9%), F1-score (99.0%), and ROC-AUC (99.5%), as summarized in Table 2. These metrics provide a more comprehensive perspective on the model’s ability to minimize false positives and false negatives, which is critical for malware detection systems. To evaluate the statistical significance of the observed improvements over baseline models, paired t-tests were conducted on accuracy scores from repeated runs. The CNN–BiLSTM model outperformed all baseline models with statistical significance (p < 0.01), indicating that the performance gains are unlikely to be due to random chance.
While Bayesian Optimization proved highly effective in identifying optimal settings with relatively few evaluations, it may face scalability limitations in high-dimensional hyperparameter spaces or when applied to very large datasets. In such scenarios, scalable alternatives like Tree-structured Parzen Estimators (TPEs) or ensemble-based surrogate models could be explored in future work.
Nevertheless, within the scope of our Android malware detection task, Bayesian Optimization offered a statistically robust, computationally efficient, and accuracy-enhancing mechanism for fine-tuning the CNN–BiLSTM model. Its application was instrumental in achieving high precision, low loss, and strong generalization—reinforcing the proposed framework’s suitability for real-world cybersecurity deployments.

4.5. Performance of CNN–BiLSTM Model

After applying Bayesian Optimization to fine-tune the hyperparameters, we proceeded to train the CNN–BiLSTM model to assess its performance in identifying Android malware. The hybrid structure was intentionally built to combine the strengths of CNN for spatial pattern extraction and BiLSTM for capturing sequential behavioral dependencies. Training was conducted across five epochs using the optimized configuration, with training and validation metrics—namely accuracy and loss—monitored throughout.
As illustrated in Figure 9, there was a consistent decline in both training and validation loss across the five epochs, indicating that the model was effectively learning and generalizing from the data. At the outset, the training loss stood at 0.165, with a validation loss of 0.125. These values decreased progressively over subsequent epochs, culminating in a training loss of 0.025 and a validation loss of 0.010 by epoch five.
Figure 9. Training and validation loss and accuracy trends over epochs for the proposed CNN–BiLSTM model.
The observed high accuracy after only five epochs can be attributed to the combined use of regularization strategies such as dropout, batch normalization, early stopping, and L2 regularization, which enhanced the model’s ability to generalize and avoid overfitting.
This consistent decrease in loss across training and validation sets demonstrates that the CNN–BiLSTM model, with its optimized configuration, effectively minimizes prediction errors and avoids overfitting—establishing its suitability for robust malware detection in real-world Android environments.
Figure 9 illustrates the substantial improvement in the CNN–BiLSTM model’s accuracy over successive training epochs. The training accuracy progressed from 92.8% to 98.9%, while the validation accuracy climbed from 93.6% to 99.7% by the fourth epoch. Although only five epochs were needed for convergence, this was supported by the combination of advanced optimization (Bayesian search) and regularization techniques, ensuring robust learning without overfitting.
To further assess the model’s classification capability, a confusion matrix was generated (Figure 10) using the predictions of the trained CNN–BiLSTM model. This matrix provides detailed insights into classification outcomes, which are essential for evaluating the trustworthiness of the detection mechanism. The model accurately detected 9992 malware instances (true positives) and 9947 benign instances (true negatives). The number of incorrect predictions was very low, with just 23 malware instances wrongly labeled as benign (false negatives) and 38 benign instances incorrectly flagged as malicious (false positives).
Figure 10. Confusion matrix of the CNN–BiLSTM model on the Android malware detection task.
These findings clearly highlight the model’s strong ability to correctly identify both benign and malicious instances, effectively reducing the occurrence of incorrect classifications in both directions.
Statistical validation of these results was performed by running the experiments across 10 independent trials with randomized data splits. The average performance metrics were calculated with their corresponding standard deviations (e.g., F1-score: 99.0% ± 0.3%). Confidence intervals of 95% confirmed the stability of results, and paired t-tests (p < 0.01) verified statistical significance compared to baseline models.
This comprehensive evaluation approach ensures that the proposed CNN–BiLSTM framework delivers not only high accuracy but also consistent and statistically validated performance, making it highly reliable for real-world Android malware detection.

4.6. Comparative Analysis

To place the effectiveness of the proposed CNN–BiLSTM framework into perspective, we carried out a comparative analysis involving several commonly used classical and baseline models in malware detection. These included CNN, LSTM, RNN, XGBoost, Support Vector Machine (SVM), and a basic feedforward neural network. The comparison centered on two key performance indicators—accuracy (Figure 11) and loss (Figure 12)—monitored across training epochs.
Figure 11. Accuracy trends across baseline and proposed models.
Figure 12. Loss trends across baseline and proposed models.
The CNN model (Figure 11 and Figure 12) demonstrated a consistent decrease in loss and increase in accuracy over time. However, it plateaued slightly below the performance of the proposed CNN–BiLSTM model, highlighting its inability to capture sequential dependencies critical to Android malware behavior, as also noted in prior studies [14,15].
Similarly, the LSTM model (Figure 11 and Figure 12) exhibited a steep learning curve during the initial epochs, benefiting from its ability to model temporal sequences [44,45]. Yet, its accuracy ultimately stabilized below the CNN–BiLSTM model, indicating the added value of incorporating spatial features through CNN layers—an insight supported by recent hybrid deep learning research in cybersecurity [46,47].
The RNN model (Figure 11 and Figure 12) lagged behind in both learning rate and final performance. Although it showed gradual loss reduction, its limited memory structure constrained its ability to model complex patterns, reinforcing known limitations of shallow recurrent networks in high-dimensional cybersecurity data [5,15].
XGBoost (Figure 11 and Figure 12), while efficient and fast-converging, plateaued early with moderate accuracy and minimal further improvement. This result reflects its limitations in modeling non-linear temporal interactions, which are often essential in behavioral malware classification.
The Simple Neural Network (Figure 11 and Figure 12) showed slightly better performance than XGBoost, with a slow but steady rise in accuracy and corresponding decline in loss. However, it still fell short of capturing deeper behavioral relationships that models like LSTM and CNN–BiLSTM can learn.
SVM (Figure 11 and Figure 12), although initially competitive, exhibited stagnation in accuracy improvement and higher loss values compared to deep learning models. Its performance limitations stem from its inability to handle large feature spaces and nonlinearity [48,49], particularly in contexts requiring layered, contextual interpretation of app behavior.
In contrast, our CNN–BiLSTM model consistently outperformed all benchmarks, achieving both the highest accuracy and the lowest loss values. This performance validates the architectural synergy between spatial feature extraction and bidirectional temporal modeling, further strengthened by Mutual Information-based feature selection and Bayesian hyperparameter optimization [42,50,51]. Together, these enhancements make the proposed framework a strong candidate for real-world Android malware detection scenarios, where both accuracy and generalization are critical.
To further assess the contribution of individual components within the proposed framework, we conducted an ablation study, the results of which are summarized in Table 2. This analysis examined four configurations: (1) CNN only, (2) BiLSTM only, (3) CNN–BiLSTM without Mutual Information and Bayesian Optimization, and (4) CNN–BiLSTM with full integration of Mutual Information and Bayesian Optimization. The full configuration achieved a 99.3% accuracy and an F1-score of 99.0%, while removing Mutual Information or Bayesian Optimization resulted in noticeable drops in accuracy (to 96.7% and 95.8%, respectively). Models using CNN only and BiLSTM only performed comparably but were outperformed by the hybrid architecture, highlighting the importance of combining spatial and temporal feature learning. These findings demonstrate that each component—CNN, BiLSTM, Mutual Information, and Bayesian Optimization—provides a measurable contribution to the overall performance.
Table 2. Ablation study of proposed CNN–BiLSTM framework and component contributions.
Table 2. Ablation study of proposed CNN–BiLSTM framework and component contributions.
Model ConfigurationAccuracy (%)F1-Score (%)Precision (%)Recall (%)
CNN only94.293.893.594.1
BiLSTM only95.194.794.395
CNN–BiLSTM (w/o MI and BO)9695.695.495.8
CNN–BiLSTM (w/o MI)96.796.396.196.6
CNN–BiLSTM (w/o BO)95.895.39595.6
CNN–BiLSTM + MI + BO (proposed)99.39999.198.9
To provide a deeper comparative assessment, we visualized the confusion matrices of all baseline models, as shown in Figure 13. These visualizations clearly highlight disparities in false-positive and false-negative rates across the evaluated models. When compared with the confusion matrix of our proposed CNN–BiLSTM framework (Figure 10), it is evident that our model achieves superior classification accuracy with significantly fewer misclassifications. This performance advantage underscores the robustness and reliability of the proposed approach in accurately detecting Android malware, even under challenging conditions such as code obfuscation or previously unseen (zero-day) threats.
Figure 13. Confusion matrices of baseline models for Android malware classification.
These quantitative comparisons clearly establish that our CNN–BiLSTM model—enhanced with Mutual Information for feature selection and Bayesian Optimization for hyperparameter tuning—delivers superior classification performance [52,53,54]. The cascading architecture effectively captures both spatial and temporal malware behavior, while the optimization steps ensure efficient and robust training. This confirms the architectural and methodological decisions that underpin the model’s exceptional ability to detect Android malware with minimal error and high generalization.

5. Conclusions

This study introduces TRIM-SEC, a compact yet powerful deep learning solution for detecting Android malware. It incorporates sophisticated feature preparation techniques along with a hybrid neural network design. The framework utilizes Convolutional Neural Networks (CNNs) to identify spatial characteristics and employs Bidirectional Long Short-Term Memory (BiLSTM) layers to model time-dependent behavior in Android applications. To enhance the model’s transparency and generalization capabilities, feature selection was conducted using Mutual Information, and Bayesian Optimization was applied to refine key hyperparameters for improved performance.
The proposed framework achieved 99.3% classification accuracy, outperforming traditional and deep learning baselines—including SVM, XGBoost, CNN, CNN, and Naive Bayes, across multiple evaluation metrics. Its superiority was further validated through comparative and ablation analyses, highlighting the effectiveness of each stage, from denoising and dimensionality reduction to sequence modeling.
A notable strength of TRIM-SEC lies in its computational efficiency. By reducing the input dimensionality and optimizing hyperparameters systematically, the model maintains a low-complexity profile without sacrificing accuracy. This makes it highly adaptable for practical deployment in real-world scenarios, particularly where scalability and responsiveness are critical.
Preliminary inference tests indicate that TRIM-SEC achieves an average prediction time of approximately 20 ms per sample on a standard desktop CPU. This suggests that the framework holds promise for deployment on higher-end mobile devices or edge computing platforms, enabling near real-time detection capabilities.
However, some limitations still exist. While the framework is lighter than many deep learning models, the inclusion of the BiLSTM layer adds a moderate level of computational complexity, potentially restricting its use on devices with limited resources, such as entry-level smartphones or embedded platforms. Moreover, the dataset used in this study maintained a near-balanced distribution between benign and malicious samples, whereas real-world datasets are often highly imbalanced. This imbalance could impact performance metrics like precision and recall, and future evaluations should assess the model’s robustness under such conditions. Another limitation lies in the model comparison scope. Although TRIM-SEC outperformed classical ML baselines, it was not evaluated against recent hybrid deep learning architectures such as CNN–GRU combinations or Transformer-based models. Future work should include such comparisons to further validate its relative performance.
Another limitation is the exclusion of comparisons with more recent hybrid deep learning architectures, such as CNN–GRU models or Transformer-based networks. These emerging architectures have shown promise in sequence modeling and could provide valuable benchmarks for contextualizing TRIM-SEC’s performance.
Future work will explore enhancements such as reducing model size through pruning, applying quantization techniques, and leveraging knowledge distillation methods to optimize computational efficiency. Additionally, incorporating advanced hybrid architectures into the evaluation pipeline is planned to provide a more comprehensive performance comparison.
Future work will explore enhancements like reducing model size through pruning, applying quantization techniques, and leveraging knowledge distillation methods to optimize the framework’s computational efficiency. Additionally, integrating imbalance mitigation strategies—such as cost-sensitive learning or resampling—will further strengthen TRIM-SEC’s applicability to real-world mobile security challenges. These improvements are intended to support the deployment of TRIM-SEC in edge computing and on-device environments, ensuring scalable, fast, and reliable Android malware detection amid growing mobile cybersecurity threats. Future studies may also investigate the integration of LLMs to complement deep learning architectures like CNN–BiLSTM. LLMs could provide richer contextual understanding of application behavior, especially for detecting previously unseen or obfuscated malware variants.

Funding

This research was funded by the Ongoing Research Funding program (ORF-2025-233), King Saud University, Riyadh, Saudi Arabia.

Data Availability Statement

Data can be made available upon request, to ensure that privacy restrictions are upheld.

Acknowledgments

The authors would like to extend their sincere appreciation to the Ongoing Research Funding program (ORF-2025-233), King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Nikkhah, H.; Ghazizadeh-Ahsaee, M.; Mirvaziri, H. A New Android Malware Detection Approach Using Multilayer Stacking. In Proceedings of the 2024 10th International Conference on Web Research (ICWR), Tehran, Iran, 24–25 April 2024; pp. 321–328. [Google Scholar] [CrossRef]
  2. Kurniawan, F.; Stiawan, D.; Antoni, D.; Heriyanto, A.; Idris, M.Y.; Budiarto, R. Hybrid Machine Learning Model for Anticipating Cyber Crime Malware in Android: Work on Progress. In Proceedings of the 2024 11th International Conference on Electrical Engineering, Computer Science and Informatics (EECSI), Yogyakarta, Indonesia, 26–27 September 2024; pp. 499–505. [Google Scholar] [CrossRef]
  3. Mutambik, I. Sustainable IoT-Enabled Parking Management: A Multiagent Simulation Framework for Smart Urban Mobility. Sustainability 2025, 17, 6382. [Google Scholar] [CrossRef]
  4. Baghirov, E. Malware detection based on opcode frequency. Probl. Inf. Technol. 2023, 14, 3–7. [Google Scholar] [CrossRef]
  5. Lakshmanarao, A.; Shashi, M. Android Malware Detection with Deep Learning Using RNN from Opcode Sequences. Int. J. Interact. Mob. Technol. 2022, 16, 145–157. [Google Scholar] [CrossRef]
  6. Aurangzeb, S.; Anwar, H.; Naeem, M.A.; Aleem, M. BigRC-EML: Big-Data Based Ransomware Classification Using Ensemble Machine Learning. Clust. Comput. 2022, 25, 3405–3422. [Google Scholar] [CrossRef]
  7. Mutambik, I. Enhancing IoT Security Using GA-HDLAD: A Hybrid Deep Learning Approach for Anomaly Detection. Appl. Sci. 2024, 14, 9848. [Google Scholar] [CrossRef]
  8. Mutambik, I.; Almuqrin, A. Balancing Efficiency and Efficacy: A Contextual Bandit-Driven Framework for Multi-Tier Cyber Threat Detection. Appl. Sci. 2025, 15, 6362. [Google Scholar] [CrossRef]
  9. Kotzias, P.; Caballero, J.; Bilge, L. How Did That Get In My Phone? Unwanted App Distribution on Android Devices. In Proceedings of the 2021 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 24–27 May 2021; pp. 53–69. [Google Scholar] [CrossRef]
  10. Almuqrin, A.; Mutambik, I.; Alomran, A.; Zhang, J.Z. Enforcing Information System Security: Policies and Procedures for Employee Compliance. Int. J. Semant. Web Inf. Syst. 2023, 19, 1–17. [Google Scholar] [CrossRef]
  11. Salem, A.H.; Azzam, S.M.; Emam, O.E.; Abohany, A.A. Advancing Cybersecurity: A Comprehensive Review of AI-Driven Detection Techniques. J. Big Data 2024, 11, 105. [Google Scholar] [CrossRef]
  12. Mutambik, I. IoT-Enabled Adaptive Traffic Management: A Multiagent Framework for Urban Mobility Optimisation. Sensors 2025, 25, 4126. [Google Scholar] [CrossRef] [PubMed]
  13. Alzakari, S.A.; Aljebreen, M.; Ahmad, N.; Alhashmi, A.A.; Alahmari, S.; Alrusaini, O.; Al-Sharafi, A.M.; Almukadi, W.S. An Intelligent Ransomware Based Cyberthreat Detection Model Using Multi Head Attention-Based Recurrent Neural Networks with Optimization Algorithm in IoT Environment. Sci. Rep. 2025, 15, 8259. [Google Scholar] [CrossRef] [PubMed]
  14. Almakayeel, N. Deep Learning-Based Improved Transformer Model on Android Malware Detection and Classification in Internet of Vehicles. Sci. Rep. 2024, 14, 25175. [Google Scholar] [CrossRef] [PubMed]
  15. Panigrahi, R.; Bele, N.; Panigrahi, P.K.; Gupta, B.B. Features Level Sentiment Mining in Enterprise Systems from Informal Text Corpus Using Machine Learning Techniques. Enterp. Inf. Syst. 2024, 18, 2328186. [Google Scholar] [CrossRef]
  16. Yu, Y.; Cai, B.; Aziz, K.; Wang, X.; Luo, J.; Iqbal, M.S.; Chakrabarti, P.; Chakrabarti, T. Semantic Lossless Encoded Image Representation for Malware Classification. Sci. Rep. 2025, 15, 7997. [Google Scholar] [CrossRef] [PubMed]
  17. Alamro, H.; Mtouaa, W.; Aljameel, S.; Salama, A.S.; Hamza, M.A.; Othman, A.Y. Automated Android Malware Detection Using Optimal Ensemble Learning Approach for Cybersecurity. IEEE Access 2023, 11, 72509–72517. [Google Scholar] [CrossRef]
  18. Almuqrin, A.; Mutambik, I.; Zhang, J.Z.; Farahat, H.; Alharbi, Z.H. Cracking the Code: A Comprehensive Analysis of Information Security Strategies Among Professionals. J. Organ. End User Comput. 2024, 36, 1–21. [Google Scholar] [CrossRef]
  19. Smmarwar, S.K.; Gupta, G.P.; Kumar, S. Android Malware Detection and Identification Frameworks by Leveraging the Machine and Deep Learning Techniques: A Comprehensive Review. Telemat. Inform. Rep. 2024, 14, 100130. [Google Scholar] [CrossRef]
  20. Shaukat, K.; Luo, S.; Varadharajan, V. A Novel Machine Learning Approach for Detecting First-Time-Appeared Malware. Eng. Appl. Artif. Intell. 2024, 131, 107801. [Google Scholar] [CrossRef]
  21. Dai, H.; Zhen, Z.; Wang, F.; Lin, Y.; Xu, F.; Duić, N. A Short-Term PV Power Forecasting Method Based on Weather Type Credibility Prediction and Multi-Model Dynamic Combination. Energy Convers. Manag. 2025, 326, 119501. [Google Scholar] [CrossRef]
  22. Zhu, H.-J.; You, Z.-H.; Zhu, Z.-X.; Shi, W.-L.; Chen, X.; Cheng, L. DroidDet: Effective and Robust Detection of Android Malware Using Static Analysis along with Rotation Forest Model. Neurocomputing 2018, 272, 638–646. [Google Scholar] [CrossRef]
  23. Mohaisen, A.; Alrawi, O.; Mohaisen, M. AMAL: High-Fidelity, Behavior-Based Automated Malware Analysis and Classification. Comput. Secur. 2015, 52, 251–266. [Google Scholar] [CrossRef]
  24. Alasmary, H.; Khormali, A.; Anwar, A.; Park, J.; Choi, J.; Abusnaina, A.; Awad, A.; Nyang, D.; Mohaisen, A. Analyzing and Detecting Emerging Internet of Things Malware: A Graph-Based Approach. IEEE Internet Things J. 2019, 6, 8977–8988. [Google Scholar] [CrossRef]
  25. Abusnaina, A.; Abuhamad, M.; Alasmary, H.; Anwar, A.; Jang, R.; Salem, S.; Nyang, D.; Mohaisen, D. DL-FHMC: Deep Learning-Based Fine-Grained Hierarchical Learning Approach for Robust Malware Classification. IEEE Trans. Dependable Secur. Comput. 2022, 19, 3432–3447. [Google Scholar] [CrossRef]
  26. Alasmary, H.; Abusnaina, A.; Jang, R.; Abuhamad, M.; Anwar, A.; Nyang, D.; Mohaisen, D. Soteria: Detecting Adversarial Examples in Control Flow Graph-Based Malware Classifiers. In Proceedings of the 2020 IEEE 40th International Conference on Distributed Computing Systems (ICDCS), Singapore, 29 November–1 December 2020; pp. 888–898. [Google Scholar] [CrossRef]
  27. Wang, J.; Ni, T.; Lee, W.-B.; Zhao, Q. A Contemporary Survey of Large Language Model Assisted Program Analysis. Trans. Artif. Intell. 2025, 1, 6. [Google Scholar] [CrossRef]
  28. Papachristou, I.; Dimitroulakos, G.; Vassilakis, C. Automated Test Generation and Marking Using LLMs. Electronics 2025, 14, 2835. [Google Scholar] [CrossRef]
  29. Ali, S.; Wang, J.; Leung, V.C.M.; Bashir, F.; Bhatti, U.A.; Wadho, S.A.; Humayun, M. CLDM-MMNNs: Cross-Layer Defense Mechanisms through Multi-Modal Neural Networks Fusion for End-to-End Cybersecurity—Issues, Challenges, and Future Directions. Inf. Fusion. 2025, 122, 103222. [Google Scholar] [CrossRef]
  30. Shen, F.; Del Vecchio, J.; Mohaisen, A.; Ko, S.Y.; Ziarek, L. Android Malware Detection Using Complex-Flows. IEEE Trans. Mob. Comput. 2019, 18, 1231–1245. [Google Scholar] [CrossRef]
  31. İbrahim, M.; Issa, B.; Jasser, M.B. A Method for Automatic Android Malware Detection Based on Static Analysis and Deep Learning. IEEE Access 2022, 10, 117334–117352. [Google Scholar] [CrossRef]
  32. Maray, M.; Maashi, M.; Alshahrani, H.M.; Aljameel, S.S.; Abdelbagi, S.; Salama, A.S. Intelligent Pattern Recognition Using Equilibrium Optimizer with Deep Learning Model for Android Malware Detection. IEEE Access 2024, 12, 24516–24524. [Google Scholar] [CrossRef]
  33. Ji, X.; Tang, L.; Liu, T.; Guo, H. HeteroNet: A Heterogeneous Encoder–Decoder Network for Sea–Land Segmentation of Remote Sensing Images. J. Electron. Imaging 2023, 32, 053016. [Google Scholar] [CrossRef]
  34. Shu, L.; Dong, S. Enhanced Unknown Android Malware Detection Using LG-PN: A Local–Global Fusion Approach in Prototypical Networks. J. Inf. Secur. Appl. 2025, 91, 104062. [Google Scholar] [CrossRef]
  35. Pinhero, A.; Anupama, M.L.; Vinod, P.; Visaggio, C.A.; Aneesh, N.; Abhijith, S.; AnanthaKrishnan, S. Malware Detection Employed by Visualization and Deep Neural Network. Comput. Secur. 2021, 105, 102247. [Google Scholar] [CrossRef]
  36. Sinha, P.; Sahu, D.; Prakash, S.; Yang, T.; Rathore, R.S.; Pandey, V.K. A High Performance Hybrid LSTM CNN Secure Architecture for IoT Environments Using Deep Learning. Sci. Rep. 2025, 15, 9684. [Google Scholar] [CrossRef] [PubMed]
  37. Surendran, R.; Uddin, M.; Thomas, T.; Pradeep, G. Android Malware Detection Based on Informative Syscall Subsequences. IEEE Access 2024, 12, 189180–189190. [Google Scholar] [CrossRef]
  38. Isohanni, J. Customised ResNet Architecture for Subtle Color Classification. Int. J. Comput. Appl. 2025, 47, 341–355. [Google Scholar] [CrossRef]
  39. Das, T.; Roy, R.; Mandal, K.K. Solving the Cost Minimization Problem of Optimal Reactive Power Dispatch in a Renewable Energy Integrated Distribution System Using Rock Hyraxes Swarm Optimization. Electr. Eng. 2025, 107, 741–773. [Google Scholar] [CrossRef]
  40. Imambi, S.; Prakash, K.B.; Kanagachidambaresan, G.R. PyTorch. In Programming with Tensorflow; Springer Nature: Berlin/Heidelberg, Germany, 2021; pp. 87–104. [Google Scholar] [CrossRef]
  41. Lemenkova, P. Processing oceanographic data by Python libraries Numpy, Scipy and Pandas. Aquat. Res. 2019, 73–91. [Google Scholar] [CrossRef]
  42. Shahriari, B.; Swersky, K.; Wang, Z.; Adams, R.P.; de Freitas, N. Taking the Human Out of the Loop: A Review of Bayesian Optimization. Proc. IEEE 2016, 104, 148–175. [Google Scholar] [CrossRef]
  43. Mutambik, I. An Entropy-Based Clustering Algorithm for Real-Time High-Dimensional IoT Data Streams. Sensors 2024, 24, 7412. [Google Scholar] [CrossRef] [PubMed]
  44. Zha, W.; Liu, Y.; Wan, Y.; Luo, R.; Li, D.; Yang, S.; Xu, Y. Forecasting Monthly Gas Field Production Based on the CNN-LSTM Model. Energy 2022, 260, 124889. [Google Scholar] [CrossRef]
  45. Mutambik, I. An Efficient Flow-Based Anomaly Detection System for Enhanced Security in IoT Networks. Sensors 2024, 24, 7408. [Google Scholar] [CrossRef] [PubMed]
  46. Lakhwani, K.; Sharma, G.; Sandhu, R.; Nagwani, N.K.; Bhargava, S.; Arya, V.; Almomani, A. Adaptive and Convex Optimization-Inspired Workflow Scheduling for Cloud Environment. Int. J. Cloud Appl. Comput. 2023, 13, 1–25. [Google Scholar] [CrossRef]
  47. Tomáš, P. A Few Remarks on the (Im)Perfection of the Term Securities: A Theoretical Study. Jurid. Trib. 2021, 11, 135–149. [Google Scholar] [CrossRef]
  48. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  49. Joachims, T. Text Categorization with Support Vector Machines: Learning with Many Relevant Features; Springer: Berlin/Heidelberg, Germany, 1998; pp. 137–142. [Google Scholar] [CrossRef]
  50. Srebalová, M.; Peráček, T. Effective Public Administration as a Tool for Building Smart Cities: The Experience of the Slovak Republic. Laws 2022, 11, 67. [Google Scholar] [CrossRef]
  51. Potvin, M.J.; Leblanc, S.P. Detecting Malicious Anomalies in Heavy-Duty Vehicular Networks Using Long Short-Term Memory Models. Sensors 2025, 25, 4430. [Google Scholar] [CrossRef]
  52. Rhanoui, M.; Mikram, M.; Yousfi, S.; Barzali, S. A CNN-BiLSTM Model for Document-Level Sentiment Analysis. Mach. Learn. Knowl. Extr. 2019, 1, 832–847. [Google Scholar] [CrossRef]
  53. Shi, H.; Miao, K.; Ren, X. Short-term Load Forecasting Based on CNN-BiLSTM with Bayesian Optimization and Attention Mechanism. Concurr. Comput. 2023, 35, e6676. [Google Scholar] [CrossRef]
  54. Peráček, T.; Kaššaj, M. A Critical Analysis of the Rights and Obligations of the Manager of a Limited Liability Company: Managerial Legislative Basis. Laws 2023, 12, 56. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.