Next Article in Journal
Reivew of Light Field Image Super-Resolution
Next Article in Special Issue
Business Email Compromise Phishing Detection Based on Machine Learning: A Systematic Literature Review
Previous Article in Journal
A Decision-Support Informatics Platform for Minimally Invasive Aortic Valve Replacement
Previous Article in Special Issue
A Novel Android Botnet Detection System Using Image-Based and Manifest File Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Digital Forensics Classification Based on a Hybrid Neural Network and the Salp Swarm Algorithm

Faculty of Artificial Intelligence, Al-Balqa Applied University, Amman 1705, Jordan
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(12), 1903; https://doi.org/10.3390/electronics11121903
Submission received: 30 April 2022 / Revised: 4 June 2022 / Accepted: 5 June 2022 / Published: 17 June 2022
(This article belongs to the Special Issue High Accuracy Detection of Mobile Malware Using Machine Learning)

Abstract

:
In recent times, cybercrime has increased significantly and dramatically. This made the need for Digital Forensics (DF) urgent. The main objective of DF is to keep proof in its original state by identifying, collecting, analyzing, and evaluating digital data to rebuild past acts. The proof of cybercrime can be found inside a computer’s system files. This paper investigates the viability of Multilayer perceptron (MLP) in DF application. The proposed method relies on analyzing the file system in a computer to determine if it is tampered by a specific computer program. A dataset describes a set of features of file system activities in a given period. These data are used to train the MLP and build a training model for classification purposes. Identifying the optimal set of MLP parameters (weights and biases) is a challenging matter in training MLPs. Using traditional training algorithms causes stagnation in local minima and slow convergence. This paper proposes a Salp Swarm Algorithm (SSA) as a trainer for MLP using an optimized set of MLP parameters. SSA has proved its applicability in different applications and obtained promising optimization results. This motivated us to apply SSA in the context of DF to train MLP as it was never used for this purpose before. The results are validated by comparisons with other meta-heuristic algorithms. The SSAMLP-DF is the best algorithm because it achieves the highest accuracy results, minimum error rate, and best convergence scale.

1. Introduction

Great technological development has led to the use of the Internet in many areas of life [1]. Many companies have taken advantage of the Internet to provide many services using e-commerce without having to submit to market restrictions. This reflected positively in the country’s economy by increasing competitiveness and achieving great returns. This has caused a huge positive shift in the number of customers who use the Internet to buy, sell, and transfer large amounts of money [2]. These large sums are tempting for many hackers and scammers to engage in many activities that violate privacy. These put a lot of people and companies at risk through the web and cause huge financial losses [3]. Other violations that may occur online include impersonation, loss of privacy, brand theft, and loss of customers’ trust in institutions. Hence, the suitability of the Internet for carrying out business and banking operations is called into question.
DF is a direct result of cybercrime, which is typically applied in computer-related crimes [4,5,6]. This includes Intellectual property infringement, use of unauthorized privileges to deal with the computer systems, privacy infringement, a security breach of confidential data repositories, carrying out terrorist operations via the Internet, misuse of electronic data, etc. DF is defined as an organized process that uses scientific techniques to collect, document, and analyze electronically stored data. This helps utilize the computer equipment and storage media to provide evidence to detect the abnormal events [7].
Suspicious events and unsafe activities may lead organizations to lose a lot of money, in addition to losing their prestige. Therefore, such events may lead to serious repercussions for both individuals and institutions. According to the statistics shown by the published reports, the number of computer violations has reached more than five million violations so far [8]. Many cybercrimes go unrecorded because their victims do not report them to officials. Victims of cybercrime feel confused and humiliated or believe that the authorities are not taking the necessary measures to punish the attackers. In addition to the lack of competencies and expertise from the workforce, the government needs to mitigate computer crimes [9].
Crime scene cyber investigations include conducting forensic analysis of all types of storage and digital media increasing the volume of data obtained. The value of data depends on the extent to which it is used in decision-making. The process of analyzing digital data during forensic practice is traditionally manual in most cases. Investigators perform reports using some statistical tools to give a picture of the collected data [10]. However, the cost of digital investigations increases with increasing data dimensions. Furthermore, forensic cases increase the data analysis complexity, which leads to the deterioration of the manual process quickly [11]. It is necessary to use more advanced methods with potential beyond the capacity of the conventional manual analysis in dealing with big data. Machine learning (ML) is an efficient and more sophisticated method that facilitates the production of useful knowledge for decision-makers. This is carried out by analyzing data sets from different perspectives and fixing them into meaningful forms [12].
The primary goal of DF analysis is to identify who is responsible for these cyber security crimes. This is performed by the selection, classification, and reordering of the sequence of events of the digital crime. Acts reorganization analyzes DF and prepares a schedule of these cyber events. It determines digital evidence, which is information with true value gained from trusted sources that admit or do not admit to committing cybercrime. Over there, there are many sources from which reliable information can be collected to rebuild cybercrime events, such as web browsers, history files, cookies messages, temporary files, log files, and system files [13].
Some of the valuable sources of information that can aid the DF are the system files and their metadata. These files are normally modified by users and the usual use of computer machines. The reconstruction process for digital events can benefit from the system files affected by cyber-crime. However, these files may be modified by normal programs [14]. For this reason, the reconstruction of digital events and the determination of the timeline sequence of events and activities are essential. This helps to recognize whether the type of application that manipulates the file system is reliable or malicious.
The main problem to be solved in this paper is the classification of files that have been accessed, manipulated, changed, and deleted by application programs. This depends on using some features that are represented by footprints. Identifying the affected files by specious acts facilitates the process of event reconstruction in the file system.
This paper presents one of the Neural Network techniques for rebuilding the acts of a digital crime. Multi-layer perceptron (MLP) is one of the artificial neural networks (ANNs) that imitates the neural human system [15]. It has been used in many applications as a supervised classifier. The main advantage of MLP is that it can learn and tackle many complex problems with promising results in science and engineering. It can be adopted efficiently for either supervised or unsupervised learning. In addition, it has a large ability to tackle parallelism, fault tolerance, and robustness to noise. MLP has a great capacity to generalize as well.
The most common problems of gradient-based MLP are stagnation in local minima, a tendency to the starting values of its parameters, and slow convergence. Due to the local minima problem, several studies in the literature proposed different approaches to train MLP. Swarm-based algorithms have been widely used to train MLPs. These algorithms simulate the natural survival of animals such as the Gradient-Based Optimizer (GBO) [16], Slime mould algorithm (SMA) [17], Heap-based optimizer (HBO) [18], and Harris hawks optimization (HHO) [19]. However, the local minima problem still exists. Furthermore, based on the no-free-lunch theorem (NFL), there is no preference among swarm algorithms. The superiority of one swarm algorithm over another does not imply superiority in all applications. This means that there is no guarantee that it will perform similarly when applied to other optimization problems. These reasons motivated us to propose a new method to train MLPs using the SSA algorithm in the context of DF.
SSA  [20] is one effective meta-heuristic algorithm that belongs to a swarm-based family. It has a set of characteristics that motivated us to select it. First, it has a single parameter that decreases in an adaptive manner relative to the increasing iterations. Second, it performs an extensive exploration in the initial iterations then it adaptively switches to exploit the most promising areas of the search space. Third, SSA preserves the best-found solution so that it never loses the optimal solution. Fourth, follower salps change their locations adaptively following other members of the population, so it has the power to alleviate the local minima problem.
SSA has been implemented to solve different optimization problems. In [21], Yang proposed a memetic SSA (MSSA) with multiple independent salp chains. He aimed to make more exploration and exploitation. Another improvement was using a population-based regroup operation for coordination between different salp chains. The MSSA is applied for an efficient maximum power point tracking (MPPT) of PV systems. The results showed the output energy generated by MSSA in Spring in Hong Kong is 118.57%, which is greater than the other algorithms. In [22], the author integrated the SSA with the Ant Lion Optimization algorithm (ALO) for intrusion detection in IoT networks. The true positive rates reached 99.9% and with a minimal delay. In [20], the authors proposed a new phishing detection system using SSA. They aimed to increase the classification performance and reduce the number of features of the phishing system. Different transfer function (TF) families: S-TFs, V-TFs, X-TFs, U-TFs and Z-TFs were used to generate a binary version of the SSA. The results showed that BSSA with X-TFs obtained the best results. In [23], the target of the proposed approach was to reduce the number of symmetrical features and obtain a high accuracy after implementing three algorithms: particle swarm optimization (PSO), the genetic algorithm (GA), and SSA. The used datasets are the Koirala dataset. The proposed COVID-19 fake news detection model obtained a high accuracy (75.4%) and reduced the number of features to 303. In [24], the authors introduced four optimizations algorithms integrated with an MLP neural network, namely artificial bee colony (ABC-MLP), grasshopper optimization algorithm (GOA-MLP), shuffled frog leaping algorithm (SFLA-MLP), and SSA-MLP to approximate the concrete compressive strength (CSC) in civil engineering. The results show that SSA-MLP and GOA-MLP can be promising alternatives to laboratory and traditional CSC evaluative methods.
The main contributions can be summarized as follows:
  • The integration of SSA with MLP assists it since it can professionally avoid local minima;
  • The high speed of convergence of SSA helps reach the optimal MLP structure quickly;
  • The proposed method achieves better results compared with other algorithms in terms of accuracy, error rate, and convergence curve;
  • The proposed method can deal with different scenarios with varying complexity.
This paper has the following sections: Section 2 reviews some of the previous studies that investigate DF. Section 3 provides a description for MLP. Section 4 discusses the details of the new MLPSSA-DF method. Section 5 discusses the dataset, experiments, and results. Finally, Section 6 summarizes the findings of the whole paper.

2. Forensic’s Background

The initiation of the DF investigations depends on the presence of some indications on the computer system. There are several indications that a computer system has experienced suspicious incidents and is a victim of cybercrime [25]. Table 1 shows these indications.
The goal of conducting DF investigations is to answer a range of questions as quickly as possible. These include:
  • What are the motives for cybercrime?
  • Where is the cybercrime site?
  • When did the suspicious act happen?
  • What suspicious acts occurred on the computer that warrants the action Investigation?
  • What is the application program or tool that was used in the execution of the cybercrime?
DF investigations rely heavily on system files because they provide accurate and important information about the sequence of events involved in cybercrimes [26]. At the end of the DF investigations comes the question about how the cybercrime occurred, and this question depends on the answer to all the previous questions.
In the past, a general rule was devised regarding potential digital evidence based on documentary evidence considered admissible in court [27]. These steps include earning, identification, assessing, and admission. Years later, a specific methodology was developed for cybercrime research [28]. This methodology is an approach to all the proposed models that have been followed so far and includes the following steps:
  • Define;
  • Preserve;
  • Gather;
  • Test;
  • Analyze;
  • Present;
  • Check.
In [29], the authors expanded the model in [28] by including two additional steps, which are set up and approach strategies. The new model was called the abstract DF model. The additional two steps were carried out after the define step and before the preserving step. The setup step involves the tools that are used to deal with the suspicious acts. On the other hand, the approach step involves a formal strategy for investigating the effect on the technology and viewers. The problem with this model is that it is quite generic, and there is no clear method to test it.
Through the models that have been proposed in [27,28,29], we can say that the steps of the DF are compatible with the steps of machine learning. In general, a DF investigation relies heavily on reconstructing events and preparing accurate evidence.
In [30], the authors devised a new method that initially requires the collection of evidence to reconstruct the acts, followed by the development of initial hypotheses. These hypotheses are studied, analyzed, and examined. From here begins the actual process of DF investigations. It ends with finding the outcomes of the electronic case. A new technique has been developed in [31] to reconstruct the latest acts. It is based primarily on defining the condemnation of data objects and the relationships between them.
In [32], the authors used an ant system to take the place on Grid hosts. The Ant-based Replication and Mapping Protocol (ARMAP) is used to spread resource information with a decentralized technique, and its performance is assessed using an entropy index. The authors in [33] developed a new method to reconstruct acts. It uses a finite state machine. It shows the transition from one state to another based on some conditions put by the evidence. In [34], the authors proposed ontology for reconstructing acts by representing the acts as entities. The events are represented by developing a time model to show the instance change in the state instead of using a period. The shellbag-based technique was proposed in [35], which preserves information in the Windows Registry. This information is about the files, folders, and windows that appear, such as deleted, modified, and relocated files and folders.
Nowadays, many tools are used to collect acts and save them in a temporal repository. However, it can be difficult to analyze the raw data.
In [36], Hall investigated the potential of EXplainable artificial intelligence (XAI) to enhance the analysis of DF evidence using examples of the current state of the art as a starting point. Bhavsar [37] pointed out the challenges in this forensics standard to design the framework of the investigation process for criminal activities using the best digital forensic instruments. Dushyant [38] discussed the advantages and disadvantages of incorporating machine learning and deep learning into cybersecurity to protect systems from unwanted threats. In [39], Casino conducted a review of all the relevant reviews in the field of digital forensics. The main challenges are related to evidence acquisition and pre-processing, collecting data from devices, the cloud, etc.
Unlike previous studies in the literature, where the proposed algorithm was trained on general datasets or limited applications, the newly developed MLP-SSA is proposed for the first time in the context of DF. Different evaluations are performed to test the viability of the MLPSSA to be used as a robust classifier for DF.

3. DF Using MLP

The MLP neural network can be used for the classification of files that have been accessed, manipulated, changed, and deleted by application programs. Once act depends on using some features that are represented by footprints. Identifying the affected files by specious acts facilitates the process of event reconstruction in the file system.

3.1. Dataset

The dataset used in the experiments is collected from three resources: the audit log, file system, and registry. It ensures that if some features are missed in one resource, they may be found in another resource. The dataset represents a database for the collected features of the system file or the metadata. The record contains the values of the files’ features that have been affected by a specific program. These are the footprints that are specific for each application program. It describes the system events or acts or metadata. As in supervised classification, the dataset has one column for class. In this dataset, the last column is the application program.
The collection of the features related to the system file using an application program is carried out using some programs such as the .Net program. This paper runs a program called VMware, which is used for collecting the features and building a training dataset. The main advantage of VMware is that it can get red from the useless programs and reduce their effects. The operating system used in this experiment is Windows 7 as it is a commonly used platform. The application programs used in the experiments are Internet Explorer, Acrobat Reader, MS-Word, MS-Excel (Microsoft Office version 2007), and VLC media player.
These applications are performed in different ways. First, the applications are loaded separately. Thus, if one application is completely loaded and then closed, the other program is loaded after it. The main issue in this execution of programs is that they do not do anything in the file system except loading and closing. Second, the applications also are loaded and executed and then closed one by one. In this case, the first three applications perform one act, which is saving a file. The last application visit (www.msn.com accessed on 20 April 2022) website. Third, as the first and second execution, the applications are loaded separately. However, different acts are executed by each application.
The acts of the first three applications include saving files, opening files, and creating new files. However, the last application performs a set of acts that include visiting secured/unsecured websites and sending/receiving emails with/without attachments. Fourth, in this execution, the applications are executed at the same time as opposed to previous executions. As in execution three, different acts take place. The number of examples of records in a database is 23,887. Table 2 shows the features on the dataset that is used to investigate the digital forensics.

3.2. Preparing a Dataset

The performance of a machine learning model is greatly affected by the dataset. Therefore, preparing a dataset is an important stage for developing efficient and reliable models. It enhances the generalizing process. Different issues have been applied to a dataset to preprocess it before using it in the learning algorithm. First, restructuring of the dataset by distributing the values into sets in such a way the feature’s states reduce.
Another important concern in this regard is that most machine learning algorithms deal with numerical values instead of text values. Therefore, in the used dataset, there is a need to assign numerical values to some feature values using the word2vect tool. Using this tool, an index is assigned to each word. Second, cleaning the dataset and git-rid of missing and outlier values. These have a positive impact on enhancing the generalization and generating less biased models. Third, normalizing the dataset by scaling the values of the features into a predefined range using the min-max method. This helps to deal with all features equally instead of making one feature overwhelm the others.

3.3. Salp Swarm Algorithm (SSA)

The inspiration for the SSA algorithm is from the salp aquatic animals. They have a specialized technique for obtaining food. The first salp in the swarm leads the other members in the sequence. This implies that other swarm members change their positions dynamically concerning the leader. Figure 1 shows a single salp on the right side and a swarm of salps on the left side.
Algorithm 1 (SSA) is an evolutionary algorithm that was developed by Mirjalili [20]. The swarm S of n salps can be represented in Equation (1), where F o o is the source of food. The population of salps is represented by a matrix. Each row is a salp or solution. The length of the salp is the number of features in a dataset (d). The number of salps (n) is the swarm’s size. The first row in the matrix is the leader salp, and the other rows are for the follower salps.
S i = s 1 1 s 2 1 s d 1 s 1 2 s 2 2 s d 2 s 1 n s 2 n s d n
Equation (2) illustrates the location of the first salp
S j 1 = F o o j + c p 1 ( ( u p _ b o u n d j l o w _ b o u n d j ) c p 2 + l o w _ b o u n d j ) , c p 3 0.5 F o o j c p 1 ( ( u p _ b o u n d j l o w _ b o u n d j ) c p 2 + l o w _ b o u n d j ) , c p 3 < 0.5
where s j 1 and F o o j are the locations of leader salp and the source of food in the j t h dimension, respectively. In Equation (3), c p 1 gradually decreases and changes its value across cycles, and c u r r and l a s t are the current and the last cycles, respectively. The other parameters c p 2 and c 3 in Equation (2) are randomly chosen from [0, 1]. c p 2 and c p 3 direct the next location in the j t h dimension to + or − and determine the step size. The  u p _ b o u n d j and l o w _ b o u n d j are the limits of the j t h dimension.
c p 1 = 2 e ( 4 c u r r l a s t ) 2
S j i = 1 2 ( s j i + s j i 1 )
In Equation (4), i 2 , and s j i is the location of the i t h salp at the j t h dimension.
Algorithm 1 SSA
In_variables: n is the size_swarm, d is the no_dimensions
Out_variables: The best salp (Foo)
Initial_variables s_i (i = 1, 2, . . . , n), up_bound and low_bound
while (last cycle is not reached) do
    Compute each salp’s fitness value
    Determine Foo as the optimal salp
    Update c 1 by Equation (3)
    for (each salp s i ) do
        if l == 1 then
           Change the location of the first salp by Equation (2)
        else
           Change the locations of the other salps by Equation (4)
        end if
        Change the locations of the salps using the lower and upper bound
    end for
end while
Return (F)

3.4. Multi-Layer Perceptron Neural Networks (MLP)

MLP is a type of feed-forward NN, which is used for training the data and discovering the hidden patterns in the training data. The pattern is then applied to the hidden instances of the dataset to obtain the results. Three layers are in the architecture of MLP: the first, the middle, and the last layers. Each layer consists of a set of computational nodes that simulate the human neurons. The MLP’s complexity increases by adding more middle layers. The standard MLP contains a single hidden layer. Figure 2 shows a standard MLP, with a first layer that has n nodes, a single middle layer that has m nodes, and a last layer that has k nodes.
The MLP can be visualized as a directed graph that connects the first layer with the middle layer and the middle layer with the last layer. Each middle node is connected with the first layer with n weights and with r weights with the last layer. In addition, there are m biases. The main processes that take place in the hidden nodes are the summation as in Equation (5) and the activation as in Equation (6). The output of Equation (5) of node m is performed using Equation (5). After computing the sum, a transfer function is applied to the input as in Equation (6).
S u m F u n m = n = 1 n d n m P n + c m
where d n m is the weight between the first node P n and the middle node h m , and c m is the bias m that enters the middle node m.
o m = S i g ( S u m F u n m )
where o m is the output node m; m = 1 , 2 , , s ; S i g is the sigmoid function as in Equation (7)
S i g ( S u m F u n m ) = 1 1 + e S u m F u n m
After collecting the results from all the m nodes, the final result O m can be generated as shown in Equations (8) and (9).
S u m F u n m = n = 1 s d n m o m + c m
where d n m is the weight between the node n in the middle layer and the node m in the last layer, and c m is the bias m that enters to the output node m.
O m = f ( S u m F u n m )
where O m is the final result m; m = 1 , 2 , , r ; S i g is the function applied in Equation (7).

4. SSAMLP-DF Model

MLP is a type of NN that uses the feed-forward NN architecture and backpropagation method to propagate data from input nodes to hidden nodes to output nodes. The following list shows the stages of the SSAMLP-DF model:
  • Collecting data;
  • Data pre-processing;
  • MLP construction;
  • MLP training phase;
  • MLP testing phase.
There are many areas in which the MLP has been implemented. The main features of MLP that helped it to be commonly used include the nonlinear structure, adaptive update of its parameters, and the ability to generalize well compared with other algorithms. Initially, the dataset is divided into three parts: 70% to train the model, 15% to validate the model, and 15% to test the model and mitigate the overfitting problem. This is called the “Hold-Out” validation method. Overfitting is a popular machine learning problem in which the error rate in the training phase is too small, but it increases in the testing phase. Of course, this is not what is required.
The main target of MLP is to build an accurate model with a minimum error achieved in the testing phase. By applying the MLP to the training part of a dataset, the initial MLP structure is constructed. These include the MLP’s layers and the number of nodes in each middle layer. The weights of the MLP are set to nonzero values. MLP trains a model until a specific condition is satisfied, such as reaching the maximum number of cycles or achieving a threshold error rate. If the trainer achieves the stopping condition, then the weight parameters of the generated model are kept. If the stopping condition is not met, then the MLP structure is updated by changing the number of nodes in the middle layer.
The MLP structure has a large number of nodes in the middle layer and is then progressively resourced across cycles until an acceptable performance is achieved. This method is called the punning method, and it is used to determine the number of nodes in the middle layer that is suitable to generate the desired performance. After the model is generated, the testing dataset is applied to this model, and the error rate is approximated. The basic measurement is the accuracy, which is calculated by dividing the correctly classified instances by the total number of instances in the testing part of the dataset. It is computed as follows:
A C C - D F = T P - D F + T N - D F T P - D F + T N - D F + F P - D F + F N - D F
where ( T P - D F ) is the number of files that are predicted to have tampered with a specific computer program, and it is tampered by a specific computer program. ( F N - D F ) is the number of files that are tampered by a specific computer program and incorrectly predicted that they are not tampered by a specific computer program. ( F P - D F ) is the number of files that are not tampered by a specific program and incorrectly predicted to be tampered by a specific program. ( T N - D F ) is the number of files that are not tampered by a specific program and incorrectly predicted to be not tampered by a specific program. The generated DF-model is trained using MATLAB.
This section presents the proposed DF model by integrating the SSA and MLP algorithms. In this model, the SSA is used to train the MLP based on one hidden layer. Two important issues must be taken into account: the representation of the solution in the SSA and the fitness function. Each solution is represented as a one-dimensional array and its values represent a candidate MLP structure. The solution in the SSA-MLP model is partitioned into three parts: the weight parameters that connect the input nodes with the hidden nodes, the weight parameters that connect the hidden nodes with the output nodes, and the biases. The solution’s length equals the number of connection weights in addition to the number of biases. Equation (11)
S o l u t i o n s l e n g t h = ( P × M ) + ( 2 × M ) + 1
where P is the number of input nodes, and M is the the number of the hidden nodes.
The fitness value in the SSA-MLP model is the mean square error (MSE). This is calculated by subtracting the predicted value from the actual value by the generated solutions (MLPs) for the training part instances of the dataset. MSE is shown in Equation (12), where R is the actual value, R ^ is the value generated from prediction, and N is the number of examples in the training dataset.
M S E = i = 1 N ( R R ^ ) 2
The steps of the proposed SSA-MLP can be summarized as follows:
  • The candidate individuals of the SSA are initialized randomly. These represent the possible solutions of the MLP.
  • Evaluating the fitness of solutions. In this step, randomly generated values are assigned to the bits of solutions that represent the possible values of the connection weights and biases. These solutions are then assessed by the MLP. MSE is a popular function that is selected to evaluate the evolutionary-based MLP models. The main purpose of the training stage is to obtain the best network structure of the MLP that generates the minimum error rate or MSE value when implemented on a training part of the dataset
  • Update each solution by changing the position of search agents in the SSA algorithm.
  • Steps 2 and 3 are repeated until the stopping condition is satisfied. The last stage of the proposed model is generating the optimal MLP structure with the best weight parameters and biases. This MLP structure is tested on the testing and validation parts of the dataset. Therefore, the obtained error rate by MSE is the minimum. Figure 3 shows the assignment of a salp to MLP. Figure 4 shows the general steps of the proposed DF-SSA-MLP model.

5. Experiments and Results

This section presents the classification results of the proposed SSAMLP versus other metaheuristic algorithms in terms of accuracy. Twenty-four experiments were established to evaluate the different algorithms in terms of the accuracy of results using different MLP structures. Furthermore, an analysis of their convergence curves is illustrated. The proposed approach is compared against seven metaheuristic algorithms integrated with MLP using the same experiment specifications. These algorithms are: Particle Swarm Optimization (PSO) [40], Ant Colony Optimization (ACO) [41], Genetics Algorithm (GA) [42], Differential Evolution algorithm (DE) [43], and BackPropagation [44].
These experiments performed the proposed methods on 14 benchmark mathematical functions used for minimization problems. Table 3 shows the values of the parameters for the applied algorithms that will be used to validate the proposed SSAMLP model.
Table 4 shows the specification of the performed experiments. These include the structure of the MLP related to the number of layers and the number of hidden nodes. The target is to study the effects of different MLP structures on the performance of the algorithms and determine the best MLP structure for all the studied algorithms.
Table 5 and Figure 5 show the accuracy results of the proposed SSAMLP model versus other algorithms that are integrated into the MLP algorithm.
As can be seen in Table 5, the best MLP structure is achieved when the number of hidden nodes is four and the number of layers is one. The results show that in the first three experiments, the accuracy of all hybrid models increases dramatically by decreasing the number of hidden nodes in the MLP’s structures. This can be explained because reducing the complexity of computations enhances the performance of DF prediction. The accuracy results increase until a specified limit, which is when the number of layers is one and the number of hidden nodes is four. This structure represents the best MLP structure for all algorithms in which the best DF prediction accuracy is achieved. After that, the accuracy started to decrease linearly by decreasing the number of nodes in the middle layer. The reason is that decreasing the number of nodes in the middle layer makes the model become simple, and the computations are not sufficient to produce an efficient model. Hence, it is not recommended to decrease the number of hidden nodes to less than four or increase it to more than four.
Choosing a medium number of computational nodes in the middle layer can produce the best classification model and achieve the best DF prediction performance. In the remaining experiments, the number of layers in the MLP structure does not benefit the prediction performance. Conversely, increasing the number of layers inversely affects the classification performance. This can increase the complexity of the classification model and cause a major machine learning problem, which is overfitting. This problem occurs because complex models that are unable to produce a general prediction pattern in the learning phase are generated. The produced model is complex and passes by a large number of learning instances. This reflects badly on the testing phase and causes degradation in the prediction performance. Figure 6 shows the convergence curves of the proposed SSAMLP versus other algorithms. The Y-axes is the error rate in terms of MSE. The X-axis is the number of iterations. It can be seen that SSAMLP achieved the best convergence curves, as it obtains the least error rates in the final iterations.
Overall, the experiments confirm that MLP can be applied as a reliable prediction model for DF. Furthermore, it can be revealed that the number of features is sufficient to investigate the DF and identify the application programs that have affected the file system. The best accuracy achieved is 95.84%, which is somehow high. This indicates that there is about a 4.16 error rate. Although there is no recommended error rate threshold commonly reported for the DF models, it can be explained that this small error rate comes from the application programs that access the same files in the file system. This makes the extracted features of the application programs overlap for several files.

6. Conclusions and Future Works

Recently, cybercrime has increased significantly and tremendously. This made the need for DF urgent. The target of this paper is to propose the SSA in training MLPs. The few parameters, fast convergence, and strong ability to avoid local minima motivated our attempts to use SSA to train MLPs. This is a minimization problem in which the main objective is to select the optimal structure of MLP (best connection weight and bias parameters) to achieve the minimum MSE. The purpose is to apply the optimal MLP in determining the growing evidence by checking the historical acts on the file system to identify how the application programs affected these files.
The used dataset in the experiments has been collected based on applying for five different application programs and checking the footprint on the system files that are the results of system actions and log entries. There are four scenarios for applying the application programs to the system files. These represent simple, medium, and complex scenarios to access the file system. The dataset is used to train the hybrid MLP-SSA model and determine the best structure of MLP that produces the minimum MSE. The results of the experiments show that the proposed SSAMLP outperformed others compared with hybrid meta-heuristic algorithms in terms of accuracy, error rate, and convergence scale. Furthermore, SSAMLP proved its suitability to be used as a reliable model to investigate DF, with an accuracy of 95.84% when the number of middle layers is one and the number of hidden nodes is four.
To verify the proposed method, a set of meta-heuristic algorithms were applied to the same dataset, and their results were compared with the SSAMLP. The comparison results show the out-performance of the SSAMLP in the majority of cases.
For future works, it is worthy to train other types of MLPs using the SSA. The applications of the multiobjective SSA to train MLP in the context of DF are recommended as well.

Author Contributions

Conceptualization, R.A.K. and M.A.; methodology, R.A.K. and M.A.; software, A.A.; validation, M.A., A.A. and M.W.; formal analysis, R.A.K.; investigation, M.A.; data curation, M.W.; writing—original draft preparation, R.A.K.; writing—review and editing, R.A.K.; visualization, A.A.; supervision, M.A.; project administration, M.A.; funding acquisition, M.A. All authors have read and agreed to the published version of the manuscript.

Funding

The research reported in this publication was funded by the Deanship of Scientific Research and Innovation at Al-balqa Applied University, Al-Salt, Jordan. (Grant Number: DSR-2021#388).

Conflicts of Interest

The author declares that no conflict of interest.

References

  1. Purnaye, P.; Kulkarni, V. A comprehensive study of cloud forensics. Arch. Comput. Methods Eng. 2022, 29, 33–46. [Google Scholar] [CrossRef]
  2. Wu, H.; Zhou, J.; Tian, J.; Liu, J.; Qiao, Y. Robust Image Forgery Detection against Transmission over Online Social Networks. IEEE Trans. Inf. Forensics Secur. 2022, 17, 443–456. [Google Scholar] [CrossRef]
  3. Li, B.; Ma, S.; Deng, R.; Choo, K.K.R.; Yang, J. Federated Anomaly Detection on System Logs for the Internet of Things: A Customizable and Communication-Efficient Approach. IEEE Trans. Netw. Serv. Manag. 2022, 4, 104–111. [Google Scholar] [CrossRef]
  4. Alazab, M.; Alazab, M.; Shalaginov, A.; Mesleh, A.; Awajan, A. Intelligent mobile malware detection using permission requests and API calls. Future Gener. Comput. Syst. 2020, 107, 509–521. [Google Scholar] [CrossRef]
  5. Alazab, M.; Venkatraman, S.; Watters, P.; Alazab, M.; Alazab, A. Cybercrime: The case of obfuscated malware. In Global Security, Safety and Sustainability & e-DEMOCRACY; Springer: Berlin/Heidelberg, Germany, 2011; pp. 204–211. [Google Scholar]
  6. Alazab, M.; Alhyari, S.; Awajan, A.; Abdallah, A.B. Blockchain technology in supply chain management: An empirical study of the factors affecting user adoption/acceptance. Clust. Comput. 2021, 24, 83–101. [Google Scholar] [CrossRef]
  7. Padma, K.; Don, K. Artificial Neural Network Applications in Analysis of Forensic Science. Cyber Secur. Digit. Forensics 2022, 2, 59–72. [Google Scholar]
  8. Flatley, J. Crime in England and Wales: Year Ending December 2015; Office for National Statistics: London, UK, 2016. [Google Scholar]
  9. Jin, Z.; Huang, J.; Wang, W.; Xiong, A.; Tan, X. Estimating Human Weight from A Single Image. IEEE Trans. Multimed. 2022, 1, 12–23. [Google Scholar] [CrossRef]
  10. Mohammad, R.M. A neural network based digital forensics classification. In Proceedings of the 2018 IEEE/ACS 15th International Conference on Computer Systems and Applications (AICCSA), Aqaba, Jordan, 28 October–1 November 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–7. [Google Scholar]
  11. Jolfaei, A.; Usman, M.; Roveri, M.; Sheng, M.; Palaniswami, M.; Kant, K. Guest Editorial: Computational Intelligence for Human-in-the-Loop Cyber Physical Systems. IEEE Trans. Emerg. Top. Comput. Intell. 2022, 6, 2–5. [Google Scholar] [CrossRef]
  12. Han, J.; Pei, J.; Kamber, M. Data Mining: Concepts and Techniques; Elsevier: Amsterdam, The Netherlands, 2011. [Google Scholar]
  13. Dong, W.; Zeng, H.; Peng, Y.; Gao, X.; Peng, A. A deep learning approach with data augmentation for median filtering forensics. Multimed. Tools Appl. 2022, 2, 1–19. [Google Scholar] [CrossRef]
  14. Carrier, B.; Spafford, E. An event-based digital forensic investigation framework. Digit. Investig. 2004, 1, 5–14. [Google Scholar]
  15. Alhmoud, L.; Abu Khurma, R.; Al-Zoubi, A.; Aljarah, I. A Real-Time Electrical Load Forecasting in Jordan Using an Enhanced Evolutionary Feedforward Neural Network. Sensors 2021, 21, 6240. [Google Scholar] [CrossRef] [PubMed]
  16. Pasti, R.; de Castro, L.N. Bio-inspired and gradient-based algorithms to train MLPs: The influence of diversity. Inf. Sci. 2009, 179, 1441–1453. [Google Scholar] [CrossRef]
  17. Izci, D. An enhanced slime mould algorithm for function optimization. In Proceedings of the 2021 3rd International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA), Online, 11–13 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–5. [Google Scholar]
  18. Askari, Q.; Saeed, M.; Younas, I. Heap-based optimizer inspired by corporate rank hierarchy for global optimization. Expert Syst. Appl. 2020, 161, 113702. [Google Scholar] [CrossRef]
  19. Khurma, R.A.; Awadallah, M.A.; Aljarah, I. Binary Harris Hawks Optimisation Filter Based Approach for Feature Selection. In Proceedings of the 2021 Palestinian International Conference on Information and Communication Technology (PICICT), Gaza, Palestine, 28–29 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 59–64. [Google Scholar]
  20. Khurma, R.A.; Sabri, K.E.; Castillo, P.A.; Aljarah, I. Salp Swarm Optimization Search Based Feature Selection for Enhanced Phishing Websites Detection. In Proceedings of the EvoApplications, Virtual Event, 7–9 April 2021; pp. 146–161. [Google Scholar]
  21. Yang, B.; Zhong, L.; Zhang, X.; Shu, H.; Yu, T.; Li, H.; Jiang, L.; Sun, L. Novel bio-inspired memetic salp swarm algorithm and application to MPPT for PV systems considering partial shading condition. J. Clean. Prod. 2019, 215, 1203–1222. [Google Scholar] [CrossRef]
  22. Abu Khurma, R.; Almomani, I.; Aljarah, I. IoT Botnet Detection Using Salp Swarm and Ant Lion Hybrid Optimization Model. Symmetry 2021, 13, 1377. [Google Scholar] [CrossRef]
  23. Al-Ahmad, B.; Al-Zoubi, A.; Abu Khurma, R.; Aljarah, I. An Evolutionary Fake News Detection Method for COVID-19 Pandemic Information. Symmetry 2021, 13, 1091. [Google Scholar] [CrossRef]
  24. Ma, X.; Foong, L.K.; Morasaei, A.; Ghabussi, A.; Lyu, Z. Swarm-based hybridizations of neural network for predicting the concrete strength. Smart Struct. Syst. Int. J. 2020, 26, 241–251. [Google Scholar]
  25. Czap, H. Self-Organization and Autonomic Informatics (I); IOS Press: Amsterdam, The Netherlands, 2005; Volume 1. [Google Scholar]
  26. Cho, G.S.; Rogers, M.K. Finding forensic information on creating a folder in logfile of ntfs. In Proceedings of the International Conference on Digital Forensics and Cyber Crime, Dublin, Ireland, 26–28 October 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 211–225. [Google Scholar]
  27. Pollitt, M. Computer forensics: An approach to evidence in cyberspace. In Proceedings of the National Information Systems Security Conference, Baltimore, MD, USA, 10–13 October 1995; Volume 2, pp. 487–491. [Google Scholar]
  28. Palmer, G. A Road Map for Digital Forensic Research/DFRWS. 2001. Available online: https://dfrws.org/wp-content/uploads/2019/06/2001_USA_a_road_map_for_digital_forensic_research.pdf (accessed on 22 April 2022).
  29. Reith, M.; Carr, C.; Gunsch, G. An examination of digital forensic models. Int. J. Digit. Evid. 2002, 1, 1–12. [Google Scholar]
  30. Lee, H.C.; Palmbach, T.; Miller, M.T. Henry Lee’s Crime Scene Handbook; Academic Press: Cambridge, MA, USA, 2001. [Google Scholar]
  31. Rynearson, J.M. Evidence and Crime Scene Reconstruction; National Crime Investigation and Training: Glynco, Georgia, 2002. [Google Scholar]
  32. Forestiero, A.; Mastroianni, C.; Spezzano, G. Building a peer-to-peer information system in grids via self-organizing agents. J. Grid Comput. 2008, 6, 125–140. [Google Scholar] [CrossRef]
  33. Chabot, Y.; Bertaux, A.; Nicolle, C.; Kechadi, M.T. A complete formalized knowledge representation model for advanced digital forensics timeline analysis. Digit. Investig. 2014, 11, S95–S105. [Google Scholar] [CrossRef]
  34. Schatz, B.; Mohay, G.; Clark, A. Rich event representation for computer forensics. In Proceedings of the Fifth Asia-Pacific Industrial Engineering and Management Systems Conference (APIEMS 2004); Queensland University of Technology Publications: Brisbane, Australia, 2004; Volume 2, pp. 1–16. [Google Scholar]
  35. Zhu, Y.; Gladyshev, P.; James, J. Using shellbag information to reconstruct user activities. Digit. Investig. 2009, 6, S69–S77. [Google Scholar] [CrossRef]
  36. Hall, S.W.; Sakzad, A.; Choo, K.K.R. Explainable artificial intelligence for digital forensics. Wiley Interdiscip. Rev. Forensic Sci. 2022, 4, e1434. [Google Scholar] [CrossRef]
  37. Bhavsar, K.; Patel, A.; Parikh, S. Approaches to Digital Forensics in the Age of Big Data. In Proceedings of the 2022 9th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 23–25 March 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 449–453. [Google Scholar]
  38. Dushyant, K.; Muskan, G.; Gupta, A.; Pramanik, S. Utilizing Machine Learning and Deep Learning in Cybesecurity: An Innovative Approach. Cyber Secur. Digit. Forensics 2022, 2, 271–293. [Google Scholar]
  39. Casino, F.; Dasaklis, T.K.; Spathoulas, G.; Anagnostopoulos, M.; Ghosal, A.; Borocz, I.; Solanas, A.; Conti, M.; Patsakis, C. Research trends, challenges, and emerging topics in digital forensics: A review of reviews. IEEE Access 2022, 1, 3–9. [Google Scholar] [CrossRef]
  40. Marini, F.; Walczak, B. Particle swarm optimization (PSO). A tutorial. Chemom. Intell. Lab. Syst. 2015, 149, 153–165. [Google Scholar] [CrossRef]
  41. Katiyar, S.; Ibraheem, N.; Ansari, A.Q. Ant colony optimization: A tutorial review. In Proceedings of the National Conference on Advances in Power and Control, Hong Kong, China, 8–12 November 2015; pp. 99–110. [Google Scholar]
  42. Mirjalili, S. Genetic algorithm. In Evolutionary Algorithms and Neural Networks; Springer: Berlin/Heidelberg, Germany, 2019; pp. 43–55. [Google Scholar]
  43. Deng, W.; Shang, S.; Cai, X.; Zhao, H.; Song, Y.; Xu, J. An improved differential evolution algorithm and its application in optimization problem. Soft Comput. 2021, 25, 5277–5298. [Google Scholar] [CrossRef]
  44. Leung, H.; Haykin, S. The complex backpropagation algorithm. IEEE Trans. Signal Process. 1991, 39, 2101–2104. [Google Scholar] [CrossRef]
Figure 1. Single salp (A) and the salps chain (B).
Figure 1. Single salp (A) and the salps chain (B).
Electronics 11 01903 g001
Figure 2. MLP components.
Figure 2. MLP components.
Electronics 11 01903 g002
Figure 3. Assigning a salp to MLP.
Figure 3. Assigning a salp to MLP.
Electronics 11 01903 g003
Figure 4. DF-SSA-MLP flowchart.
Figure 4. DF-SSA-MLP flowchart.
Electronics 11 01903 g004
Figure 5. Boxplot representation for the proposed SSAMLP and other algorithms in terms of classification accuracy.
Figure 5. Boxplot representation for the proposed SSAMLP and other algorithms in terms of classification accuracy.
Electronics 11 01903 g005
Figure 6. Convergence curves of the proposed SSAMLP and other algorithms.
Figure 6. Convergence curves of the proposed SSAMLP and other algorithms.
Electronics 11 01903 g006
Table 1. DF indications.
Table 1. DF indications.
Computer System PartIndication
BootingSlow
Turn-offSlow
ResponseSlow
DataDeleted or destroyed
ProcessesUnknown processes run
Disk spaceLow
Pop-up windowsMany unexpected
BatteryDrains quickly
Wi-FiUnstable connection
Blue ScreenDisplayed many times
Table 2. The features in a dataset.
Table 2. The features in a dataset.
Feature NumberFeature NameFeature Value Example
1Filename-length255 characters
2Filenamespace_user/
3Object Id O b j I d
4Original-Volume-Id 94 F 8 9 C 08
5Domain-Id d d s : : d o m a i n : : D o m a i n P a r t i c i p a n t ( D O M A I N _ I D , q o s _ p r o v i d e r . p a r t i c i p a n t q o s ( ) )
6Original-Object-Id40dff02fc9b4d4118f120090273fa9fc
Table 3. The parameters’ values for the applied algorithms.
Table 3. The parameters’ values for the applied algorithms.
AlgoParameterVal
SSAc1[0–1]
c2[0–1]
GACrossProb0.9
Mut-prob0.1
SelecMethRoulette wheel
PSOAccCons[2.1, 2.1]
InerWe[0.9, 0.6]
DECrossProb0.9
DiffWe0.5
ACOIniPh ( τ )0.000001
PhUpCons (Q)20
PhCons (q)1
GPhDec-rate ( p g )0.9
LPhDecRate ( p t )0.5
PhSen ( α )1
VisSen ( β )5
ES λ 10
σ 1
PBILLeRate0.05
GPopMem1
BadPoMem0
EPara1
MutProb0.1
Table 4. Experiments specifications.
Table 4. Experiments specifications.
Experiment NumberNumber of LayersNumber of Hidden Nodes
118
217
316
415
514
613
712
811
1027,1
1126,2
1225,3
1324,4
1426,1
1525,2
1624,3
1725,1
1824,2
1923,3
2023,2
2124,1
2223,1
2322,2
2422,1
2521,1
Table 5. Accuracy results of the SSA-MLP against other algorithms based on different experiment specifications.
Table 5. Accuracy results of the SSA-MLP against other algorithms based on different experiment specifications.
Exp #SSAMLPBPMLPGAMLPPSOMLPACOMLPDEMLPESMLPPRILMLP
191.3487.3390.5690.2184.2883.2080.5882.79
292.8887.9891.3491.6685.0683.7880.9984.39
393.4188.0591.9786.0083.9081.4581.7784.67
494.2288.4793.1192.7185.1383.1081.9384.80
595.8489.5094.6193.9886.0984.3282.7785.39
695.7689.4892.7892.6785.0084.2282.5585.23
795.0088.0992.7192.3684.5384.0182.0785.14
894.7884.6790.8991.4582.7783.9981.2084.33
994.3882.6688.6091.5582.3083.9080.9983.80
1093.8881.0182.4590.6681.8681.7780.4382.26
1193.7681.0082.1990.6181.2380.9878.5180.87
1292.1780.9080.7788.5080.9079.7677.5680.28
1391.8980.7480.3687.9480.8979.3276.1179.90
1490.0480.0380.1284.4480.1377.6575.8978.94
1590.0180.0079.4584.2180.1076.3076.2677.85
1689.9979.9979.3983.9780.0176.1075.9777.48
1789.8779.4379.3483.5679.5675.9875.3677.31
1889.6079.2078.2282.3079.3674.3475.1277.22
1988.4579.1077.8382.2879.2074.2974.8474.09
2088.3478.3377.5182.1478.2073.2972.9173.67
2188.2378.1175.2381.9077.2872.1172.3771.88
2287.8877.2074.3381.3677.0272.0072.2171.51
2387.2176.3774.0781.2376.2671.9871.0071.47
2487.1175.3073.4481.1176.0971.8670.6671.22
2587.0275.1973.1981.0276.0070.9670.3470.18
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alazab, M.; Abu Khurma, R.; Awajan, A.; Wedyan, M. Digital Forensics Classification Based on a Hybrid Neural Network and the Salp Swarm Algorithm. Electronics 2022, 11, 1903. https://doi.org/10.3390/electronics11121903

AMA Style

Alazab M, Abu Khurma R, Awajan A, Wedyan M. Digital Forensics Classification Based on a Hybrid Neural Network and the Salp Swarm Algorithm. Electronics. 2022; 11(12):1903. https://doi.org/10.3390/electronics11121903

Chicago/Turabian Style

Alazab, Moutaz, Ruba Abu Khurma, Albara Awajan, and Mohammad Wedyan. 2022. "Digital Forensics Classification Based on a Hybrid Neural Network and the Salp Swarm Algorithm" Electronics 11, no. 12: 1903. https://doi.org/10.3390/electronics11121903

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop