Next Article in Journal
Special Issue of the Manufacturing Engineering Society—2021 (SIMES-2021)
Next Article in Special Issue
Surface Defect Detection Model for Aero-Engine Components Based on Improved YOLOv5
Previous Article in Journal
Tissue Mechanics and Tissue Engineering
Previous Article in Special Issue
Development and Optimization of Deep Learning Models for Weapon Detection in Surveillance Videos
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Transfer Learning Enabled Intelligent Object Detection for Crowd Density Analysis on Video Surveillance Systems

1
Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, Riyadh 11671, Saudi Arabia
2
Department of Information Systems, College of Computing and Information System, Umm Al-Qura University, Mecca 24382, Saudi Arabia
3
Department of Computer Science, College of Science & Art at Mahayil, King Khalid University, Abha 62529, Saudi Arabia
4
Faculty of Science, Mathematics and Computer Science Department, Menoufia University, Shebeen El-Kom 32511, Egypt
5
Department of Information Systems, College of Computer and Information Science, Majmaah University, Al-Majmaah 11952, Saudi Arabia
6
Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, Riyadh 11671, Saudi Arabia
7
Department of Digital Media, Faculty of Computers and Information Technology, Future University in Egypt, New Cairo 11845, Egypt
8
Department of Computer Science, College of Sciences and Humanities-Aflaj, Prince Sattam bin Abdulaziz University, Al-Kharj 16278, Saudi Arabia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(13), 6665; https://doi.org/10.3390/app12136665
Submission received: 10 May 2022 / Revised: 23 June 2022 / Accepted: 25 June 2022 / Published: 30 June 2022
(This article belongs to the Special Issue Deep Learning in Object Detection and Tracking)

Abstract

:
Object detection is a computer vision based technique which is used to detect instances of semantic objects of a particular class in digital images and videos. Crowd density analysis is one of the commonly utilized applications of object detection. Since crowd density classification techniques face challenges like non-uniform density, occlusion, inter-scene, and intra-scene deviations, convolutional neural network (CNN) models are useful. This paper presents a Metaheuristics with Deep Transfer Learning Enabled Intelligent Crowd Density Detection and Classification (MDTL-ICDDC) model for video surveillance systems. The proposed MDTL-ICDDC technique mostly concentrates on the effective identification and classification of crowd density on video surveillance systems. In order to achieve this, the MDTL-ICDDC model primarily leverages a Salp Swarm Algorithm (SSA) with NASNetLarge model as a feature extraction in which the hyperparameter tuning process is performed by the SSA. Furthermore, a weighted extreme learning machine (WELM) method was utilized for crowd density and classification process. Finally, the krill swarm algorithm (KSA) is applied for an effective parameter optimization process and thereby improves the classification results. The experimental validation of the MDTL-ICDDC approach was carried out with a benchmark dataset, and the outcomes are examined under several aspects. The experimental values indicated that the MDTL-ICDDC system has accomplished enhanced performance over other models such as Gabor, BoW-SRP, Bow-LBP, GLCM-SVM, GoogleNet, and VGGNet.

1. Introduction

Object detection is a computer technology related to computer vision and image processing that aims to determine and detect many target objects from still images or video data [1]. It widely comprises different important techniques, namely image processing, pattern recognition, artificial intelligence (AI), and machine learning (ML). It finds applicability in different domains such as road traffic accident prevention, theft detection, traffic management, etc. [2,3]. Intelligent video surveillance is a vintage subject in the domain of image processing and computer vision that has recently become well known. It has numerous significant benefits, such as accurate data processing, low human resource cost, and effective information gathering organization [4]. Crowd density estimation is regarded as a significant application in visual surveillance, and it plays an important role in crowd management and monitoring. Specifically for service providers in public areas, a crowd density estimation system could show how many consumers are currently waiting and therefore provide an appropriate reference to send the bounded sources reasonably and effectively [5].
Crowd density estimation refers to the assessment of crowd dispersal and the particular number of people [6]. Crowd analysis has attracted substantial interest among researchers in recent years because of various reasons. The massive increase in the global population and in urbanization has resulted in an increase in such events as public demonstrations, sporting events, political rallies, and so on. Similar to other computer vision issues, crowd analysis faces numerous difficulties, such as inter-scene variations in appearance, occlusions, uneven distribution of people, high clutter, intra-scene and scale issues, non-uniform illumination, and an unclear viewpoint; these problems are immensely difficult to solve [7,8]. The perplexity of the issue along with the extensive array of applications for crowd analysis has led to an increased focus among research scholars in recent years.
Convolutional neural network (CNN) models [9,10] have reached successful outcomes in image processing and in the prediction of crowd density. Recent crowd density estimation methodologies are primarily dependent on regression or identification. Detection approaches can be implemented in cases that have a minimum number of persons and no occlusion-like detectors depending on closer frames [11,12]. Other methodologies that depend on regression are of two categories. The first includes those that identify handmade features in the image, namely texture feature and edge feature; next, regression function is selected for estimating aggregate person numbers [13]. Another one relies on deep neural networks and density map regression, and this technique is considered the best for estimating crowd density.
This paper presents a Metaheuristics with Deep Transfer Learning Enabled Intelligent Crowd Density Detection and Classification (MDTL-ICDDC) model on video surveillance systems. The proposed MDTL-ICDDC technique leverages a Salp Swarm Algorithm (SSA) with NASNetLarge model as a feature extraction in which the hyperparameter tuning process is performed by the SSA. Furthermore, a weighted extreme learning machine (WELM) technique was employed for crowd density and classification process. Finally, the krill swarm algorithm (KSA) is applied for effectual parameter optimization process and thereby improves the classification results. The experimental validation of the MDTL-ICDDC technique is carried out using benchmark dataset.
The rest of the paper is organized as follows. Section 2 offers a brief survey of recently developed crowd density estimation and classification models. Next, Section 3 provides the proposed MDTL-ICDDC technique for crowd classification on surveillance videos. Then, Section 4 validates the performance of the proposed model, and finally, Section 5 concludes with the major key findings of the work.

2. Related Works

This section presents a detailed literature review of the existing crowd density analysis models. Ding et al. [14] proposed a novel encoder-decoder CNN that combines the feature map in encoding and decoding subnetworks for estimating the number of people accurately. In addition, the authors present a new assessment methodology called the Patch Absolute Error (PAE) that is more applicable for measuring the accuracy of density maps. Zhu et al. [15] resolve crowd density evaluation problems for dense and sparse conditions. Consequently, it generates two contributions: (1) a network called Patch Scale Discriminant Regression Network (PSDR). Considering an input crowd image, it splits the images as two patches and sends them into a regression network, which then yields a density map. It fuses the two patch density maps in order to predict the whole density map as the output. (2) A person classification activation map (CAM) technique is the other contribution.
In [16], the authors developed a Wi-Fi monitoring detection system that could capture smart phone passive Wi-Fi signal data involving a received signal strength indicator and MAC address. Next, the authors present a positioning model based on a dynamic fingerprint management strategy and smart phone passive Wi-Fi probe. In real time social activities, an individual might possess zero, one, two, or many smart phones with different Wi-Fi signals. Thus, it can be designed as a methodology for calculating the possibility of users generating one Wi-Fi signal to recognize the total population of people. Last, the authors developed a crowd density evaluation method based on a Wi-Fi packet positioning model.
In [17], the current research progression on density estimation and crowd counting has been comprehensively analyzed. First, the authors present the background of density estimation and crowd counting. Next, they summarized the traditional crowd counting method. Later, the authors focus on investigating the density estimation and crowd counting methodologies based on a CNN model. In [18], the authors proposed a crowd density estimation model, utilizing Hough circle transformation. Here, background and foreground datasets were segregated by ViBe technology and the segmentation of foreground datasets.
In Bouhlel et al. [19], a crowd density estimation method in an aerial image is proposed for examining a crowded region that shows an abnormal density. The presented technique comprises an inference and offline phase. The offline phase focused on generating a crowd model with a combination of handcrafted and relevant deep features designated by the use of the minimum-redundancy maximum-relevance (mRMR) method. During the inference phase, the previously generated models for classifying the aerial image patches yield the following four classifications: None, Sparse, Medium, and Dense. In [20], the network compression to the CNN-based crowd density estimation method is applied for reducing their computation and storage costs. In particular, the authors depend on l1-norm for selecting insignificant filters and physically pruning them. These models are trained to identify the insignificant filter and to increase the regression performance simultaneously.
The authors in [21] developed a new crowd density estimation model by the use of DL models for passenger flow recognition model in exhibition centers. At the initial stage, the difference amplitude feature and gray feature of the central pixel are derived to create the CLBP feature for obtaining more crowd-group description information. Moreover, the LR activation function is used for adding the non-linear factors to the CNN and exploiting dense blocks derived from crowd density estimation for calibrating the LR-CNN crowd density estimation model. Bhuiyan et al. [22] developed a fully convolutional neural network (FCNN) model for crowd density estimation on surveillance video captured by a camera at a distance. Li et al. [23] introduces a multi-scale feature fusion network (IA-MFFCN) depending upon reverse attention model that mapped the image into the crowd density map for counting purposes. Wang et al. [24] developed a lightweight CNN to estimate crowd density by the combination of the modified MobileNetv2 and the dilated convolution.

3. The Proposed Model

In this study, a new MDTL-ICDDC technique was established for effectual identification and classification of crowd density on a video surveillance system. The MDTL-ICDDC model initially presented an SSA with NASNetLarge model as a feature extraction model in which the hyperparameter tuning process is performed by the SSA. This was followed by the KSA-WELM model, which is employed for crowd-density and classification processes. Figure 1 illustrates the block diagram of MDTL-ICDDC technique.

3.1. Feature Extraction Module

At the initial stage, the SSA with NASNetLarge model functions as a feature extraction model in which the hyperparameter tuning process is performed by the SSA. For determining the optimal convolution architecture for the dataset, a search algorithm can be used. Neural architecture search (NAS) is the most important search technique that the authors deployed in this network. Child network is shown to accomplish some accuracy on a validation set; in other words, it is used for convergence [25]. The subsequent accuracy value is utilized to upgrade the controller that consecutively generates better architecture over time. The policy gradient takes place to update the controller weight.
A new searching space was designed, which permits the better architecture found on the CIFAR-10 dataset (available at http://www.cs.toronto.edu/~kriz/cifar.html, accessed on 12 Febuary 2022) that generalized for large, high-resolution image datasets from the range of computation environments. To adapt input of depth of filtering and spatial dimension, this cell is sequentially stacked. In this technique, the convolution net overall architecture is predefined manually. They are composed of convolution cells that possess a similar shape as the original but are weighted in a different way. Two kinds of convolution cells have been taking place for rapidly developing scalable architecture for images of any size: (1) convolution cells return a feature map with a 2-fold reduction in width and height, and (2) convolution cells produce a feature map with the similar dimension. Figure 2 depicts the framework of NasNetLarge.
These two kinds of convolution cells are represented as Normal Cell and Reduction Cell, correspondingly. The primary process used for the cell’s input gives a two-step stride to minimalize the cell’s width and height. The convolution cells support striding because they consider each operation. The Normal as well as Reduction Cells architecture that the controller RNN search for dissimilar to convolution net. The searching region is utilized for searching each cell shape. There are 2 hidden states (HS), namely h (i) and h (i −1), presented for every cell in the searching space. Initially, the HS is the outcome of 2 cells in the prior 2 lower layers or input image, correspondingly. The controller RNN make recursive prediction on the remaining convolution cells on the basis of 2 primary HSs. The controller prediction for all the cells is ordered into B blocks, with every block comprising five prediction steps implemented by five discrete SoftMax classifications representing different selections of block elements.
Step 1. Choose an HS from h (i), h (i −1), or the set of formerly generate HS.
Step 2. In the similar option as in Step1, choose the next HS.
Step 3. In Step 1, select the HS that the authors need to employ.
Step 4. After, choose an HS in Step2, choose an operation to employ.
Step 5. Define how the output from Steps 3 & 4 would be integrated to make a novel HS. It can be helpful to apply the recently generated HS as an input for the subsequent block.
For the optimal hyperparameter tuning process, the SSA is utilized. It was initially established as a swarm intelligence process [26]; it can be stimulated by the foraging performance of the salp swarm that forms a chain. Similar to other algorithms, it finds an optimum solution via the cooperation and division of salps. The population is divided into the following categories: the leader and followers. Initially, the leaders lead the direction of the population, and next the follower follows them; consequently, the population forms a chain. The population X includes N agents are established as a matrix using D columns and N rows. The target population from the searching region is a food source represented as F .
X = [ X 1 , 1 X 1 , 2 X 1 , p X 2 , 1 X 2 , 2 X 2 , p X N , 1 X N , 2 X N , D ] ,
whereas population size denotes N , D indicates dimension. The leader’s location is rehabilitated.
X 1 , j = { F j + c 1 × ( ( u b j 1 b j ) × c 2 + 1 b j ) c 3 0.5 F j c 1 × ( ( u b j 1 b j ) × c 2 + 1 b j ) c 3 0.5
Now X 1 , j and F j mean the j th parameter of the leader location and food source, respectively. c 1 signifies a control variable adoptively decreased through the iteration and calculated. The role of exploration and exploitation is to determine SSA. c 2 and c 3 are created within [ 0 ,   1 ] . u b j and l b j indicates the j th variable of upper and lower limits, respectively.
c 1 = 2 × e ( 4 × l L ) 2 .
Here, l and L denote the existing and maximal iterations, respectively. The follower location is rehabilitated as [27].
X i j = 1 2 × ( X i j + X i 1 j ) ,
where i 2 and X i j indicates the j t h parameter in the position of i - th follower.

3.2. Crowd Density Classification Module

Once the feature vectors are created, the next stage is to classify the crowd density with the WELM model. WELM is an enhanced version of ELM that manages imbalanced class distribution data [28]. The WELM approach presented the weighted matrix W similar to how the original ELM model works to balance the data distribution that weakens the majority class and strengthens the minority class. The W represents a matrix of misclassification values according to class distribution. The W in (5) signifies diagonal matrixes with N × N dimensional in which N indicates the overall dataset, and # ( t i ) denotes the overall amount of samples that belongings to class t i .
W i i = 1 # ( t i )
The WELM is based on the standardized ELM that minimizes the error vector ξ , as well as minimizes the output weight norm β to have improved generalization outcomes. In the following, the mathematical expression of the minimization problem is given.
m i n i m i z e : L P E M = 1 2 β 2 + 1 2 C W i = 1 N ξ i 2 s u b j e c t   t o : h ( x ) β = t i T ξ i T , i = 1 , , N
The solution of β is obtained from (6) on the basis of KKT condition into (7) if N is smaller and (8) if N is larger.
β = H T ( I C + W H H T ) 1 W T
β = ( I C + W H H T ) 1 H T W T
For multi-class classification with m class, all the labels are mapped into a vector of [−l, 1] through length of m ; for example, a dataset that is categorized into the 2nd classes from three classes are [ 1 , 1 , 1 ] . The output f ( x ) = H B in multi-class classification is a vector f ( x ) = [ f 1 ( x ) ,   f 2 ( x ) ,   f m ( x ) ] and the class label is evaluated by the following equation. Figure 3 showcases the framework of WELM.
l a b e l   o f   x = arg m a x ( x )

3.3. Parameter Tuning Process

For the optimal adjustment of the WELM parameters, the KSA is utilized in this study. The krill swarm algorithm (KSA) [29] is a commonly used intelligent optimization technique, due to its benefits of strong search diversity, few adjusted parameters, and simple operation. The KSA approach originated from mutual communication and krill foraging. In the presented approach, the location of every individual krill represents a potential solution. The location of each krill is commonly specified in the following:
  • Individual swimming due to population migration:
    N i n e w = N max α i + ω n N i o l d
    α i = α i l o c a 1 + α i t a r g e t
From the equation, the maximum induction speed can be represented as N max , and taken as 0.01 ( ms 1 ) ,  and ω n refers to the inertia weight of motion-induced range from [0, 1]. N i o l d represents prior movement, whereas α i t a r g e t and α i l o c a l indicate the target and the existing positions, respectively.
2.
Foraging behaviour:
F i = V f β i + ω f F i o l d
β i = β i f o o d + β i b e s t
Now, the individual direction of foraging for krill was denoted as β i . β i indicates the attractive direction of food, and β i b e s t indicates the direction of individual krill with the optimal fitness value. V f represents the foraging speed that takes as 0.02 ( ms 1 ) , and ω f represents the inertia weight that ranges from [0, 1]. F i o l d signifies the location change due to the preceding foraging motion of the i - t h individual krill; F i denotes the location change due to the existing foraging motion of the individual i - t h krill.
3.
Random diffusion of individual krill:
D i = D max δ
In Equation (14), D max indicates the maximal disturbance (diffusion) velocity, and δ denotes an arbitrary direction vector that ranges from [ 1 ,   1 ] ; D i signifies the location change due to arbitrary diffusion of i - th individual krill.
The swimming direction of each krill can be defined by the fusion of abovementioned factors that change in the direction with minimum fitness value. The foraging and induced motions have local and global searching functions. After the process is iteratively updated, two measures are simultaneously implemented, which make stable and powerful optimization algorithms.
The location vector from t to t + Δ t is formulated by:
X i ( t + Δ t ) = X t ( t ) + Δ t d X i d t
Here, Δ t indicates the factor for the step size.
Δ t = C t j = 1 N V ( U B j L B j )
Now, N V indicates the overall amount of parameters, whereas L B j and U B j denote the upper as well as lower bounds of the j - th parameter.
The KSA system grows a fitness function (FF) for attaining higher classifier performance. It resolves the positive integer for denoting the best efficiency of candidate results. In this work, the minimization of the classification error rate is measured FF, as shown in Equation (17).
f i t n e s s ( x i ) = C l a s s i f i e r E r r o r R a t e ( x i ) = n u m b e r   o f   m i s c l a s s i f i e d   s a m p l e s T o t a l   n u m b e r   o f   s a m p l e s 100

4. Results and Discussion

In this section, the experimental validation of the MDTL-ICDDC model is tested with a dataset comprising 1000 images under four class labels. The MDTL-ICDDC model is simulated using Python 3.6.5 tool on a PC i5-8600k, GeForce 1050Ti 4 GB, 16 GB RAM, 250GB SSD, and 1 TB HDD. The parameter settings are given as follows: learning rate: 0.01, dropout: 0.5, batch size: 5, epoch count: 50, and activation: ReLU. Few sample images are demonstrated in Figure 4. The details related to the dataset are given in Table 1.
Figure 5 demonstrates a set of confusion matrices formed by the MDTL-ICDDC model on distinct epoch counts. On epoch 200, the MDTL-ICDDC model has recognized 223, 220, 236, and 243 images under classes 0–3 respectively. Moreover, on epoch 600, the MDTL-ICDDC technique has recognized 206, 203, 234, and 241 images under classes 0–3 respectively. At the same time, on epoch 1000, the MDTL-ICDDC approach has recognized 197, 178, 233, and 234 images under classes 0–3 respectively. In line with, epoch 1200, the MDTL-ICDDC system has recognized 227, 178, 236, and 242 images under classes 0–3 respectively.
Table 2 and Figure 6 offer a detailed crowd density classification outcome of the MDTL-ICDDC model under distinct epochs and classes. The results indicated that the MDTL-ICDDC model has gained effectual outcomes under all classes and epochs. For instance, with 200 epochs, the MDTL-ICDDC model has provided average a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a s u r e of 96.10%, 92.20%, 92.20%, 92.18%, and 92.19% respectively. Likewise, with 600 epochs, the MDTL-ICDDC technique has obtainable average a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a s u r e of 94.20%, 88.38%, 88.40%, 88.31%, and 88.35% respectively. Similarly, with 1000 epochs, the MDTL-ICDDC system has provided average a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a s u r e of 92.10%, 84.06%, 84.20%, 83.94%, and 84.04% respectively. Eventually, with 1200 epochs, the MDTL-ICDDC methodology has offered average a c c u y , p r e c n , r e c a l , F s c o r e , and G m e a s u r e of 96.40%, 92.82%, 92.80%, 92.80%, and 92.80% respectively.
The training accuracy (TA) and validation accuracy (VA) attained by the MDTL-ICDDC method on test dataset are demonstrated in Figure 7. The experimental outcome implied that the MDTL-ICDDC model has gained maximal values of TA and VA. Specfically, the VA seemed superior to the TA.
The training loss (TL) and validation loss (VL) achieved by the MDTL-ICDDC model on test dataset are displayed in Figure 8. The experimental outcome exposed that the MDTL-ICDDC technique has been able least values of TL and VL. Specifically, the VL seemed lower than the TL.
A brief precision-recall examination of the MDTL-ICDDC method on test dataset is portrayed in Figure 9. An observation of the figure shows that the MDTL-ICDDC system has been able to achieve maximal precision-recall performance under all classes.
A detailed ROC investigation of the MDTL-ICDDC approach on test dataset is depicted in Figure 10. The results indicated that the MDTL-ICDDC model has exhibited its ability in categorizing four different classes 0–3 on the test dataset.
Table 3 and Figure 11 inspect a comparative p r e c n inspection of the MDTL-ICDDC model with other models under distinct classes [30,31]. The experimental results indicated that the Gabor and BoW-SRP models have shown lower classification results with least average p r e c n of 61.83% and 68.33% respectively. Likewise, the BoW-LBP and GLCM-SVM models have accomplished slightly improved average p r e c n values of 74.68% and 75.47% respectively. At the same time, the GoogleNet and VGGNet techniques have resulted in reasonable average p r e c n values of 82.98% and 86.14% respectively. However, the MDTL-ICDDC model has gained maximum average p r e c n of 92.90%.
Table 4 and Figure 12 demonstrate a comparative r e c a l analysis of the MDTL-ICDDC approach with other models under distinct classes. The experimental results indicated that the Gabor and BoW-SRP models have shown lower classification results with least average r e c a l of 62.30% and 67.85% respectively. Moreover, the BoW-LBP and GLCM-SVM models have accomplished somewhat enhanced average r e c a l values of 74.15% and 73.52% correspondingly. Simultaneously, the GoogleNet and VGGNet techniques have resulted in reasonable average r e c a l values of 85.26% and 82.78% respectively. But, the MDTL-ICDDC model has gained maximal average r e c a l of 92.90%.
Table 5 and Figure 13 depict a comparative a c c u y examination of the MDTL-ICDDC model with other algorithms under distinct classes. The experimental results indicated that the Gabor and BoW-SRP approaches have shown lower classification results with least average a c c u y of 71.83% and 80.40% respectively. Alongside these results, the BoW-LBP and GLCM-SVM methods have accomplished slightly improved average a c c u y values of 84.04% and 79.78% respectively. These results are followed by the GoogleNet and VGGNet approaches, which have resulted in reasonable average a c c u y values of 84.40 and 84.75% respectively. At last, the MDTL-ICDDC method has gained higher average a c c u y of 96.45%.
Table 6 and Figure 14 illustrate a comparative F s c o r e inspection of the MDTL-ICDDC algorithm with other techniques under distinct classes. The experimental results indicated that the Gabor and BoW-SRP models have exposed lesser classification results with minimal average F s c o r e of 61.98% and 67.88% respectively. In addition, the BoW-LBP and GLCM-SVM models have accomplished somewhat higher average F s c o r e values of 74.35% and 87.99% respectively.
Moreover, the GoogleNet and VGGNet approaches have resulted in reasonable average F s c o r e values of 81% and 84.95% respectively. Finally, the MDTL-ICDDC model has gained superior average F s c o r e of 92.87%. These results and discussion pointed out that the MDTL-ICDDC model has proficiently detected and classified the crowd density in enabled video surveillance systems.

5. Conclusions

In this study, a new MDTL-ICDDC method was established for effectual identification and classification of crowd density on video surveillance systems. The MDTL-ICDDC model initially presented an SSA with NASNetLarge model as a feature extraction model in which the hyperparameter tuning process is performed by the SSA. This was followed by the WELM model, which is employed for crowd density and classification processes. At last, the KSA is applied for the effectual parameter optimization process and thereby improves the classification results. The experimental validation of the MDTL-ICDDC system was carried out with a benchmark dataset and the outcomes are examined under several aspects. The experimental values indicated that the MDTL-ICDDC approach has accomplished enhanced performance over other models. The proposed model can be extended to crowd density estimation in public places such as railway stations, airports, sports stadiums, shopping malls, etc. In the future, the crowd density classification performance can be further boosted by the design of a hybrid metaheuristics algorithm.

Author Contributions

Conceptualization, F.A.; Data curation, F.A. and S.S.A.; Formal analysis, S.S.A.; Investigation, F.N.A.-W. and N.N.; Methodology, N.N. and R.A.; Project administration, R.A.; Software, R.M.; Supervision, R.M.; Validation, A.S.M. and M.A.D.; Visualization, A.S.M. and M.A.D.; Writing—original draft, F.A.; Writing—review & editing, F.N.A.-W. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through Large Groups Project under grant number (42/43). Princess Nourah Bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R77), Princess Nourah Bint Abdulrahman University, Riyadh, Saudi Arabia. The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code: (22UQU4210118DSR22). The authors would like to thank the Deanship of Scientific Research at Majmaah University for supporting this work under Project No. R-2022-xxx.

Institutional Review Board Statement

This article does not contain any studies with human participants performed by any of the authors.

Informed Consent Statement

Not Applicable.

Data Availability Statement

Data sharing was not applicable to this article as no datasets were generated during the current study.

Conflicts of Interest

The authors declare that they have no conflict of interest. The manuscript contains contributions from all authors. All authors have approved the final version of the manuscript.

References

  1. Liu, W.; Lis, K.; Salzmann, M.; Fua, P. Geometric and physical constraints for drone-based head plane crowd density estimation. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 244–249. [Google Scholar]
  2. Gao, G.; Gao, J.; Liu, Q.; Wang, Q.; Wang, Y. Cnn-based density estimation and crowd counting: A survey. arXiv 2020, arXiv:2003.12783. [Google Scholar]
  3. Fradi, H.; Dugelay, J.L. Towards crowd density-aware video surveillance applications. Inf. Fusion 2015, 24, 3–15. [Google Scholar] [CrossRef]
  4. Pai, A.K.; Karunakar, A.K.; Raghavendra, U. A novel crowd density estimation technique using local binary pattern and Gabor features. In Proceedings of the 2017 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Lecce, Italy, 29 August–1 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6. [Google Scholar]
  5. Zhou, B.; Song, B.; Hassan, M.M.; Alamri, A. Multilinear rank support tensor machine for crowd density estimation. Eng. Appl. Artif. Intell. 2018, 72, 382–392. [Google Scholar] [CrossRef]
  6. Sindagi, V.A.; Patel, V.M. A survey of recent advances in cnn-based single image crowd counting and density estimation. Pattern Recognit. Lett. 2018, 107, 3–16. [Google Scholar] [CrossRef] [Green Version]
  7. Weng, W.T.; Lin, D.T. Crowd density estimation based on a modified multicolumn convolutional neural network. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–7. [Google Scholar]
  8. Anwer, M.H.; Hadeel, A.; Fahd, N.A.; Mohamed, K.N.; Abdelwahed, M.; Anil, K.; Ishfaq, Y.; Abu Sarwar, Z. Fuzzy cognitive maps with bird swarm intelligence optimization-based remote sensing image classification. Comput. Intell. Neurosci. 2022, 2022, 4063354. [Google Scholar]
  9. Sreenu, G.; Durai, M.S. Intelligent video surveillance: A review through deep learning techniques for crowd analysis. J. Big Data 2019, 6, 48. [Google Scholar] [CrossRef]
  10. Abunadi, I.; Althobaiti, M.M.; Al-Wesabi, F.N.; Hilal, A.M.; Medani, M.; Hamza, M.A.; Rizwanullah, M.; Zamani, A.S. Federated learning with blockchain assisted image classification for clustered UAV networks. Comput. Mater. Contin. 2022, 72, 1195–1212. [Google Scholar] [CrossRef]
  11. Sirohi, P.; Al-Wesabi, F.N.; Alshahrani, H.M.; Maheshwari, P.; Agarwal, A.; Dewangan, B.K.; Hilal, A.M.; Choudhury, T. Energy-efficient cloud service selection and recommendation based on QoS for sustainable smart cities. Appl. Sci. 2021, 11, 9394. [Google Scholar] [CrossRef]
  12. PK, S.; Rabichith, S.N.S.; Borra, S. Crowd density estimation using image processing: A survey. Int. J. Appl. Eng. Res. 2018, 13, 6855–6864. [Google Scholar]
  13. Lamba, S.; Nain, N. A texture based mani-fold approach for crowd density estimation using Gaussian Markov Random Field. Multimed. Tools Appl. 2019, 78, 5645–5664. [Google Scholar] [CrossRef]
  14. Ding, X.; He, F.; Lin, Z.; Wang, Y.; Guo, H.; Huang, Y. Crowd density estimation using fusion of multi-layer features. IEEE Trans. Intell. Transp. Syst. 2020, 22, 4776–4787. [Google Scholar] [CrossRef]
  15. Zhu, L.; Li, C.; Yang, Z.; Yuan, K.; Wang, S. Crowd density estimation based on classification activation map and patch density level. Neural Comput. Appl. 2020, 32, 5105–5116. [Google Scholar] [CrossRef]
  16. Tang, X.; Xiao, B.; Li, K. Indoor crowd density estimation through mobile smartphone wi-fi probes. IEEE Trans. Syst. Man Cybern. Syst. 2018, 50, 2638–2649. [Google Scholar] [CrossRef]
  17. Fan, Z.; Zhang, H.; Zhang, Z.; Lu, G.; Zhang, Y.; Wang, Y. A survey of crowd counting and density estimation based on convolutional neural network. Neurocomputing 2022, 472, 224–251. [Google Scholar] [CrossRef]
  18. Purwar, R.K. Crowd Density Estimation Using Hough Circle Transform for Video Surveillance. In Proceedings of the 2019 6th International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 7–8 March 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 442–447.
  19. Bouhlel, F.; Mliki, H.; Hammami, M. Abnormal crowd density estimation in aerial images based on the deep and handcrafted features fusion. Expert Syst. Appl. 2021, 173, 114656. [Google Scholar] [CrossRef]
  20. Li, M.; Chen, T.; Li, Z.; Liu, H. An Efficient Crowd Density Estimation Algorithm Through Network Compression. In Traffic and Granular Flow 2019; Springer: Cham, Switzerland, 2020; pp. 165–173. [Google Scholar]
  21. Xiang, J.; Liu, N. Crowd Density Estimation Method Using Deep Learning for Passenger Flow Detection System in Exhibition Center. Sci. Program. 2022, 2022, 1990951. [Google Scholar] [CrossRef]
  22. Bhuiyan, M.R.; Abdullah, J.; Hashim, N.; Al Farid, F.; Haque, M.A.; Uddin, J.; Isa, W.N.M.; Husen, M.N.; Abdullah, N. A deep crowd density classification model for Hajj pilgrimage using fully convolutional neural network. PeerJ Comput. Sci. 2022, 8, e895. [Google Scholar] [CrossRef]
  23. Li, Y.C.; Jia, R.S.; Hu, Y.X.; Han, D.N.; Sun, H.M. Crowd density estimation based on multi scale features fusion network with reverse attention mechanism. Appl. Intell. 2022, 1–17. [Google Scholar] [CrossRef]
  24. Wang, S.; Pu, Z.; Li, Q.; Wang, Y. Estimating Crowd Density with Edge Intelligence Based on Lightweight Convolutional Neural Networks. Expert Syst. Appl. 2022, 206, 117823. [Google Scholar] [CrossRef]
  25. Zaman, K.; Sun, Z.; Shah, S.M.; Shoaib, M.; Pei, L.; Hussain, A. Driver Emotions Recognition Based on Improved Faster R-CNN and Neural Architectural Search Network. Symmetry 2022, 14, 687. [Google Scholar] [CrossRef]
  26. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp swarm algorithm: A bioinspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  27. Xia, J.; Zhang, H.; Li, R.; Wang, Z.; Cai, Z.; Gu, Z.; Chen, H.; Pan, Z. Adaptive Barebones Salp Swarm Algorithm with Quasi-oppositional Learning for Medical Diagnosis Systems: A Comprehensive Analysis. J. Bionic Eng. 2022, 19, 240–256. [Google Scholar] [CrossRef]
  28. Utomo, O.K.; Surantha, N.; Isa, S.M.; Soewito, B. Automatic sleep stage classification using weighted ELM and PSO on imbalanced data from single lead ECG. Procedia Comput. Sci. 2019, 157, 321–328. [Google Scholar] [CrossRef]
  29. Deng, Z.G.; Yang, J.H.; Dong, C.L.; Xiang, M.Q.; Qin, Y.; Sun, Y.S. Research on economic dispatch of integrated energy system based on improved krill swarm algorithm. Energy Rep. 2022, 8, 77–86. [Google Scholar] [CrossRef]
  30. Meynberg, O.; Cui, S.; Reinartz, P. Detection of high-density crowds in aerial images using texture classification. Remote Sens. 2016, 8, 470. [Google Scholar] [CrossRef] [Green Version]
  31. Pu, S.; Song, T.; Zhang, Y.; Xie, D. Estimation of crowd density in surveillance scenes based on deep convolutional neural network. Procedia Comput. Sci. 2017, 111, 154–159. [Google Scholar] [CrossRef]
Figure 1. Block diagram of MDTL-ICDDC technique.
Figure 1. Block diagram of MDTL-ICDDC technique.
Applsci 12 06665 g001
Figure 2. Architecture of NasNetLarge.
Figure 2. Architecture of NasNetLarge.
Applsci 12 06665 g002
Figure 3. Structure of Weighted ELM.
Figure 3. Structure of Weighted ELM.
Applsci 12 06665 g003
Figure 4. Sample Crowd Density Images. (a) Dense Crowd, (b) Medium Dense Crowd, (c) Sparse Crowd, (d) No Crowd.
Figure 4. Sample Crowd Density Images. (a) Dense Crowd, (b) Medium Dense Crowd, (c) Sparse Crowd, (d) No Crowd.
Applsci 12 06665 g004
Figure 5. Confusion matrices of MDTL-ICDDC technique (a) epoch 200, (b) 400 epoch, (c) epoch 600, (d) epoch 800, (e) epoch 1000, and (f) epoch 1200.
Figure 5. Confusion matrices of MDTL-ICDDC technique (a) epoch 200, (b) 400 epoch, (c) epoch 600, (d) epoch 800, (e) epoch 1000, and (f) epoch 1200.
Applsci 12 06665 g005
Figure 6. Average analysis of MDTL-ICDDC technique (a) epoch 200, (b) 400 epoch, (c) epoch 600, (d) epoch 800, (e) epoch 1000, and (f) epoch 1200.
Figure 6. Average analysis of MDTL-ICDDC technique (a) epoch 200, (b) 400 epoch, (c) epoch 600, (d) epoch 800, (e) epoch 1000, and (f) epoch 1200.
Applsci 12 06665 g006
Figure 7. TA and VA analysis of MDTL-ICDDC technique.
Figure 7. TA and VA analysis of MDTL-ICDDC technique.
Applsci 12 06665 g007
Figure 8. TL and VL analysis of MDTL-ICDDC technique.
Figure 8. TL and VL analysis of MDTL-ICDDC technique.
Applsci 12 06665 g008
Figure 9. Precision-recall curve analysis of MDTL-ICDDC technique.
Figure 9. Precision-recall curve analysis of MDTL-ICDDC technique.
Applsci 12 06665 g009
Figure 10. ROC curve analysis of MDTL-ICDDC technique.
Figure 10. ROC curve analysis of MDTL-ICDDC technique.
Applsci 12 06665 g010
Figure 11. Precision analysis of MDTL-ICDDC technique under various classes.
Figure 11. Precision analysis of MDTL-ICDDC technique under various classes.
Applsci 12 06665 g011
Figure 12. Recall analysis of MDTL-ICDDC technique under various classes.
Figure 12. Recall analysis of MDTL-ICDDC technique under various classes.
Applsci 12 06665 g012
Figure 13. Accuracy analysis of MDTL-ICDDC technique under various classes.
Figure 13. Accuracy analysis of MDTL-ICDDC technique under various classes.
Applsci 12 06665 g013
Figure 14. F-score analysis of MDTL-ICDDC technique under various classes.
Figure 14. F-score analysis of MDTL-ICDDC technique under various classes.
Applsci 12 06665 g014
Table 1. Dataset details.
Table 1. Dataset details.
LabelsClass NameNo. of Instances
Class-0Dense Crowd250
Class-1Medium Dense Crowd250
Class-2Sparse Crowd250
Class-3No Crowd250
Total No. of Instances1000
Table 2. Result analysis of MDTL-ICDDC technique with distinct measures and epochs.
Table 2. Result analysis of MDTL-ICDDC technique with distinct measures and epochs.
Class LabelsAccuracyPrecisionRecallF-ScoreG-Measure
Epoch-200
Class-095.6092.9289.2091.0291.04
Class-194.1088.3588.0088.1888.18
Class-296.4091.4794.4092.9192.92
Class-398.3096.0597.2096.6296.62
Average96.1092.2092.2092.1892.19
Epoch-400
Class-096.0094.4989.2091.7791.81
Class-194.9090.2889.2089.7489.74
Class-296.6091.8694.8093.3193.32
Class-398.3094.9898.4096.6696.68
Average96.4592.9092.9092.8792.89
Epoch-600
Class-093.3089.9682.4086.0186.10
Class-191.2083.2081.2082.1982.19
Class-295.0087.3193.6090.3590.40
Class-397.3093.0596.4094.7094.71
Average94.2088.3888.4088.3188.35
Epoch-800
Class-092.8089.3880.8084.8784.98
Class-190.8082.6480.0081.3081.31
Class-294.3085.8792.4089.0289.08
Class-396.1090.1194.8092.4092.43
Average93.5087.0087.0086.9086.95
Epoch-1000
Class-090.8083.4778.8081.0781.10
Class-188.7081.2871.2075.9176.07
Class-293.7083.5193.2088.0988.22
Class-395.2087.9793.6090.7090.74
Average92.1084.0684.2083.9484.04
Epoch-1200
Class-096.3094.1990.8092.4692.48
Class-194.3088.1489.2088.6788.67
Class-296.8092.9194.4093.6593.65
Class-398.2096.0396.8096.4196.42
Average96.4092.8292.8092.8092.80
Table 3. Precision analysis of MDTL-ICDDC technique with existing algorithms under various classes.
Table 3. Precision analysis of MDTL-ICDDC technique with existing algorithms under various classes.
Precision (%)
MethodsClass-0Class-1Class-2Class-3Average
Gabor57.8047.8056.5085.2061.83
BoW-SRP76.0049.6059.1088.6068.33
Bow-LBP75.8055.8072.2094.9074.68
GLCM-SVM72.7070.3378.4280.4175.47
GoogleNet74.0589.3878.4890.0182.98
VGGNet79.4282.1189.7893.2386.14
MDTL-ICDDC94.4990.2891.8694.9892.90
Table 4. Recall analysis of MDTL-ICDDC technique with existing algorithms under various classes.
Table 4. Recall analysis of MDTL-ICDDC technique with existing algorithms under various classes.
Recall (%)
MethodsClass-0Class-1Class-2Class-3Average
Gabor52.0047.4058.8091.0062.30
BoW-SRP63.4049.6067.0091.4067.85
Bow-LBP74.0061.8067.0093.8074.15
GLCM-SVM73.5969.5779.1771.7473.52
GoogleNet79.6083.9987.1890.2685.26
VGGNet76.0882.4689.8282.7582.78
MDTL-ICDDC89.2089.2094.8098.4092.90
Table 5. Accuracy analysis of MDTL-ICDDC technique with existing algorithms under various classes.
Table 5. Accuracy analysis of MDTL-ICDDC technique with existing algorithms under various classes.
Accuracy (%)
MethodsClass-0Class-1Class-2Class-3Average
Gabor55.1380.8365.6785.6771.83
BoW-SRP79.6586.9570.8684.1580.40
Bow-LBP91.8980.8674.8688.5584.04
GLCM-SVM72.4382.2170.9993.4879.78
GoogleNet93.0073.2392.8878.5084.40
VGGNet86.7382.7378.9590.5784.75
MDTL-ICDDC96.0094.9096.6098.3096.45
Table 6. F-score analysis of MDTL-ICDDC technique with existing algorithms under various classes.
Table 6. F-score analysis of MDTL-ICDDC technique with existing algorithms under various classes.
F-Score (%)
MethodsClass-0Class-1Class-2Class-3Average
Gabor54.7047.6057.6088.0061.98
BoW-SRP69.1049.6062.8090.0067.88
Bow-LBP74.9058.6069.5094.4074.35
GLCM-SVM89.4485.7891.6485.0987.99
GoogleNet83.5478.6775.9085.8981.00
VGGNet82.3787.9890.2879.1784.95
MDTL-ICDDC91.7789.7493.3196.6692.87
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alrowais, F.; Alotaibi, S.S.; Al-Wesabi, F.N.; Negm, N.; Alabdan, R.; Marzouk, R.; Mehanna, A.S.; Al Duhayyim, M. Deep Transfer Learning Enabled Intelligent Object Detection for Crowd Density Analysis on Video Surveillance Systems. Appl. Sci. 2022, 12, 6665. https://doi.org/10.3390/app12136665

AMA Style

Alrowais F, Alotaibi SS, Al-Wesabi FN, Negm N, Alabdan R, Marzouk R, Mehanna AS, Al Duhayyim M. Deep Transfer Learning Enabled Intelligent Object Detection for Crowd Density Analysis on Video Surveillance Systems. Applied Sciences. 2022; 12(13):6665. https://doi.org/10.3390/app12136665

Chicago/Turabian Style

Alrowais, Fadwa, Saud S. Alotaibi, Fahd N. Al-Wesabi, Noha Negm, Rana Alabdan, Radwa Marzouk, Amal S. Mehanna, and Mesfer Al Duhayyim. 2022. "Deep Transfer Learning Enabled Intelligent Object Detection for Crowd Density Analysis on Video Surveillance Systems" Applied Sciences 12, no. 13: 6665. https://doi.org/10.3390/app12136665

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop