Next Article in Journal
Deep Learning for Early Earthquake Detection: Application of Convolutional Neural Networks for P-Wave Detection
Previous Article in Journal
Optimized Hybrid Central Processing Unit–Graphics Processing Unit Workflow for Accelerating Advanced Encryption Standard Encryption: Performance Evaluation and Computational Modeling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DualCascadeTSF-MobileNetV2: A Lightweight Violence Behavior Recognition Model

1
Key Laboratory of CTC & IE (Engineering University of PAP), Ministry of Education, Engineering University of PAP, Xi’an 710086, China
2
Graduate Student Brigade, Engineering University of PAP, Xi’an 710086, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(7), 3862; https://doi.org/10.3390/app15073862
Submission received: 4 March 2025 / Revised: 25 March 2025 / Accepted: 27 March 2025 / Published: 1 April 2025

Abstract

:
This paper proposes a lightweight violent behavior recognition model, DualCascadeTSF-MobileNetV2, which is improved based on the temporal shift module and its subsequent research. By introducing the Dual Cascade Temporal Shift and Fusion module, the model further enhances the feature correlation ability in the time dimension and solves the problem of information sparsity caused by multiple temporal shifts. Meanwhile, the model incorporates the efficient lightweight structure of MobileNetV2, significantly reducing the number of parameters and computational complexity. Experiments were conducted on three public violent behavior datasets, Crowd Violence, RWF-2000, and Hockey Fights, to verify the performance of the model. The results show that it outperforms other classic models in terms of accuracy, computational speed, and memory size, especially among lightweight models. This research continues and expands on the previous achievements in the fields of TSM and lightweight network design, providing a new solution for real-time violent behavior recognition on edge devices.

1. Introduction

Social safety and stability are the desired living conditions for every citizen. To prevent and reduce the occurrence of violent behavior, a large number of security personnel monitor social security conditions in real-time through surveillance cameras or frontline duty terminal devices. This monitoring method has significantly improved the response speed to violent behavior. However, relying solely on human supervision has two main issues: on the one hand, it is difficult for surveillance personnel to maintain high concentration at all times, leading to potential misjudgments or missed detections of violent behavior; on the other hand, with the vast amount of surveillance data, it is difficult for personnel to quickly analyze complex situations and make effective judgments. Therefore, introducing intelligent monitoring systems to assist security personnel in detecting violent behavior is of great importance.
The core technology of intelligent monitoring systems lies in the intelligent recognition of violent behavior, which is an important direction in the field of computer vision for behavior recognition. In recent years, deep learning-based behavior recognition methods have made significant progress, with many new models achieving increasingly precise results on public datasets. However, considering the practical needs of violent behavior recognition tasks, models need not only higher accuracy but also faster recognition speed to reduce latency, fewer parameters, and smaller model size for deployment on edge devices.
In this context, we noticed the temporal shift module (TSM) designed by Lin et al. [1]. This model interacts with the action information at different times by shifting the time steps of different channels in the temporal dimension. This allows the extraction of action features across three consecutive time steps with only a single 2D convolution, effectively achieving 3D convolutional performance with 2D convolutional efficiency without sacrificing accuracy. This significantly improves computational speed and reduces recognition latency.
The design of TSM is very ingenious and still has potential for further exploration. For example, a longer-term information correlation can be achieved by fusing the features before and after the shift at the same time point; the number of parameters and the model volume can be further reduced by replacing the base model. Based on TSM and its subsequent research, this paper deeply explores its potential optimization space. The main contributions are as follows:
1.
Improved TSM based on previous studies and designed the Dual Cascade Temporal Shift and Fusion (DualCascadeTSF) module. This module cascades two temporal shift modules and enhances the model’s ability to extract long-term temporal information by adding feature fusion [2] after each temporal shift. It improves the robustness and accuracy of the model without increasing additional parameters.
2.
Combined the DualCascadeTSF module with the MobileNetV2. This not only improves the accuracy of MobileNetV2, but also makes it have fewer parameters and a faster operation speed compared to other classic models, making it more suitable for deployment on edge devices.

2. Related Work

2.1. Research on Lightweight Models

In the current development stage of behavior recognition tasks, researchers’ attention is no longer limited to continuously improving the recognition accuracy of models. Instead, they pay more attention to lightweight models with fewer parameters, a smaller size, and shorter inference times. Such models are crucial for effective deployment in environments with limited hardware resources, especially in scenarios that require edge computing, such as mobile devices, embedded systems, and miniaturized machines.
Currently, there are three main methods to obtain lightweight models: the first is to prune classic models; the second is to design lightweight models based on manual experience; and the third is to use the Neural Architecture Search (NAS) to generate efficient models.
Pruning is one of the most commonly used methods to obtain lightweight models. This method reduces the size and computational complexity of classic models by pruning weights, channels, or structures while maintaining high performance. VGGNet pruning [3] achieves channel pruning by applying L1 regularization to the scaling factor of the Batch Normalization (BN) layer during the network training process, which is widely used in image classification and object detection tasks. Inception pruning [4] evaluates the importance of each convolutional filter and removes less significant filters to reduce the size of the model and accelerate inference. BERT pruning [5] significantly reduces the number of model parameters and computational costs by removing redundant attention heads and layers, which is applicable to natural language processing tasks.
The design of lightweight models based on human experience is a traditional but still effective method for lightweight models. This method relies on experts’ knowledge and experience to design network structures and optimize components to achieve model lightweighting. The main methods include using lightweight convolutions, designing compact network structures, using lightweight activation functions, and optimizing pooling and normalization. MobileNetV1 [6] introduces depthwise separable convolutions, decomposing standard convolutions into depthwise and pointwise convolutions, significantly reducing computational load while using the ReLU6 activation function to limit output range and avoid numerical overflow. MobileNetV2 [7] further improves this by adopting inverted residual blocks and linear bottlenecks, expanding channels through expansion layers, processing feature maps with depthwise separable and pointwise convolutions, and then compressing back to the input channel number. ShuffleNet [8] uses group convolutions and channel shuffle techniques to enhance the information exchange between feature maps while reducing the amount of calculation and combines residual connections to improve model performance. SqueezeNet [9] significantly reduces the number of parameters through the fire module (including the squeeze layer and the expand layer) and global average pooling.
NAS is a model lightweighting method that automatically searches for the optimal neural network structure. Its implementation steps include setting goals, defining the search space, determining the search strategy, and evaluating performance. Google designed MobileNetV3 [10] through the NAS method based on reinforcement learning, optimizing the network structure, which is applicable to image classification, object detection, and real-time video processing tasks. EfficientNet-Lite [11], as a lightweight version of EfficientNet, further optimized the model structure through the NAS method and is widely used in image classification, object detection, and semantic segmentation tasks. BiX-Net [12] used a two-stage NAS algorithm to automatically search for an ultra-lightweight and effective bidirectional skip connection structure, suitable for medical image segmentation tasks. LightTrack [13] optimized the network structure for object-tracking tasks through the NAS method and is applicable to object tracking and video analysis tasks. OFA (Once-for-All) [14] generates a supernet through a one-time NAS search, from which multiple subnetworks are pruned, each optimized for different tasks and resource constraints, suitable for image classification, object detection, and semantic segmentation tasks.

2.2. TSM and Two-Cascade TSM

In behavior recognition tasks, different from static image input, the number of video stream parameters shows explosive growth, which poses a huge challenge to achieving high-accuracy and low-cost behavior recognition.
Traditional 2D convolutional neural networks (CNNs) have a low computational cost but cannot capture the unique temporal information of videos; while 3D CNNs can achieve good performance, their computational load is huge, and the deployment cost is too high. To solve this problem, the temporal shift module (TSM) came into being, which cleverly achieved the goal of approaching the performance of 3D CNNs with the computational cost of 2D CNNs.
Specifically, the working principle of TSM [1] is shown in Figure 1. Figure 1a represents the video stream before temporal shift, where C is the channel dimension, representing different channels of the video stream, and T is the temporal dimension, with each layer of color corresponding to a time step in the sequence. Figure 1b shows the video stream after temporal shifting, where some channels are shifted upward or downward by one time step in the temporal dimension, while the remaining parts stay unchanged. Through this approach, the task of extracting behavior features across any three consecutive time steps can be accomplished with a single convolution operation, meaning that 2D convolution alone can be used to extract spatio-temporal features for actions at a single time step.
Therefore, the operation of TSM can be divided into two steps: temporal shifting and convolution. Suppose that there is a 1 × 3 convolution kernel W = ( ω 1 , ω 2 , ω 3 ) and an infinitely long video stream X n × 3 k = ( X i k , j , X i , j , X i + k , j ) R n × 3 with three channels, where k represents the number of temporal shifts. After one temporal shift, the video stream becomes X n × 3 1 = ( X i 1 , j , X i 2 , X i + 1 , j ) R n × 3 , and then the convolution operation is performed:
y i = ω 1 x i 1 , 1 + ω 2 x i , 2 + ω 3 x i + 1 , 3
Y = j = 1 3 ( ω j x i 1 , j )
where y i represents the convolution operation process, and Y represents the calculation result. From the calculation process, it can be seen that the first step of the TSM operation does not generate computational costs, while the second step will generate a certain amount of calculation. However, compared to the reduction in computational cost from transforming the convolution operation from 3D to 2D, the slight increase in computation is acceptable.
Building on this foundation, Liang et al. [15] further proposed the Two-cascade TSM. This module extends the original TSM by cascading another TSM within the residual block, enhancing the model’s temporal correlation capability through the method of expanding the model’s receptive field, thus enabling it to capture longer-term information dependencies. However, while this improvement enhances model performance, the authors discovered a problem: when cascading more TSMs, the model’s performance deteriorates instead of improving. Analyzing the reason, this may be because too many temporal shift operations cause the feature maps that are originally adjacent in different channels to become sparse in the temporal dimension, thereby weakening the model’s performance. Based on this drawback of the Two-cascade TSM, this paper conducts research on how to address the issue of sparse temporal information connections after long-term shifting.

2.3. MobileNetV2

MobileNetV2 is a lightweight convolutional neural network developed based on MobileNetV1, both of which aim to provide efficient computing capabilities for mobile devices and resource-constrained environments. MobileNetV2 inherits the core idea of MobileNetV1; that is, replacing traditional standard convolutions with depthwise separable convolutions to significantly reduce the amount of calculation and the number of parameters.
As shown in Figure 2, depthwise separable convolution divides standard convolution into pointwise convolution and depthwise convolution. The main function of depthwise convolution is to extract features from each input channel using a single filter, while pointwise convolution applies a 1 × 1 convolution to combine the output of depthwise convolution. The computational cost generated by this method is K × K × C × H × W + C × N × H × W , where K × K represents the size of the convolution kernel, C is the number of channels, H × W is the size of the input video frame, and N is the number of input video frames. Compared to the computational cost K × K × H × W × C × N generated by standard convolution, which extracts and combines features simultaneously, the reduced computational cost is
K × K × C × H × W + C × N × H × W K × K × H × W × C × N = 1 N + 1 K 2
It can be seen that this algorithm indeed achieves the expected effect of reducing the amount of calculation and the number of parameters.
Meanwhile, MobileNetV2 introduces two important innovations based on MobileNetV1: inverted residuals and linear bottlenecks. The core idea of the linear bottleneck is to avoid using nonlinear activation functions in the low-dimensional bottleneck layer to ensure the integrity of the input information and prevent the loss of key information in the video frames. In contrast, non-linear activation functions are introduced in the high-dimensional expansion layer to enhance the model’s expressive power. The design of the inverted residuals is opposite to that of the traditional residual block. The shortcut connections are located between the low-dimensional bottleneck layers, while the intermediate layers perform feature extraction and nonlinear transformations using lightweight depthwise separable convolutions. This design not only significantly reduces the number of model parameters, but also further improves computational efficiency, making MobileNetV2 more suitable for deployment and operation in resource-constrained environments.

3. Methods

3.1. Conception

The Two-cascade TSM proposed in the literature [15] is an improvement on the traditional TSM, which significantly enhances the model’s ability to extract long-term information. However, when attempting to cascade more TSMs within the same residual block, the experimental results did not meet expectations. The authors speculate that this might be due to the excessive temporal shifting operations, which cause the originally adjacent feature maps in different channels to become sparse in the temporal dimension, thereby weakening the model’s performance.
Based on this observation, this paper proposes a new hypothesis: if a feature fusion step is added after each temporal shift to combine the feature maps at the same temporal position before and after the shift, could this further enhance the model’s ability to extract longer-term temporal information? Based on this, this paper designs the DualCascadeTSF module, which enhances the ability to extract long-term temporal information without adding extra parameters. It is hoped that through the method of feature fusion [2], the correlation between features in the temporal dimension can be strengthened without increasing the number of parameters, thereby compensating for the information sparsity caused by multiple shifts in the Two-cascade TSM design.
Moreover, to make the designed model more lightweight, this paper chooses to replace traditional ResNet with MobileNetV2 as the base model. MobileNetV2, with its unique inverted residual structure and linear bottleneck design, significantly reduces the number of parameters and computational complexity while maintaining performance. Therefore, combining the DualCascadeTSF module with MobileNetV2 can not only achieve higher accuracy but also effectively reduce the model size and improve the calculation speed, making it more suitable for deployment on resource-constrained edge devices.

3.2. DualCascadeTSF Module

Based on the concept mentioned above, this paper proposes a module, the Dual Cascade Temporal Shift and Fusion (DualCascadeTSF) module. As shown in Figure 3, this is the structural diagram of the Dual Cascade TSF module. The design of this module inherits the core idea of the TSM, which involves shifting the temporal positions of video frame channels to alter the manner of convolution operations, thereby improving computational efficiency. Building on this foundation, the DualCascadeTSF module further introduces a feature fusion mechanism that merges feature maps at the same temporal position during each temporal shift. This enhances the model’s ability to extract longer-term information.
As shown in Figure 4, the working principle of the DualCascadeTSF module is illustrated. Stage 1 represents the original input video stream, where each column corresponds to a channel, and each row represents a frame. In this figure, we assume that it has four frames and six channels. Stage 2 involves the first temporal shift, where one channel is shifted forward in time while another channel is shifted backward. Stage 3 shows the feature fusion of the original input video stream with the shifted video stream after the first temporal shift. Stage 4 involves a second temporal shift after the first feature fusion. Stage 5 shows the feature fusion of the original input video stream with the shifted video stream after the second temporal shift. At this point, the DualCascadeTSF module completes its entire operation. It can be seen from the figure that the working principle of the DualCascadeTSF module is very simple and does not introduce significant additional parameters.
After intuitively presenting the working principle of the module in the form of diagrams, this paper further explains the ingenuity of the module’s design from a mathematical perspective. Suppose that there are a 1 × 3 convolution kernel W = ( ω 1 , ω 2 , ω 3 ) and a 3-channel infinitely long video stream X n × 3 k = ( X i k , j , X i , j , X i + k , j ) R n × 3 , ( n = ) , where k represents the number of temporal shifts, i represents the position of the time step in the video stream, j represents the position of the channel in the video stream, and n represents the number of frames in the input video stream. After one temporal shift, the video stream becomes X n × 3 1 = ( X i 1 , j , X i 2 , X i + 1 , j ) R n × 3 , ( n = ) , and then the convolution operation is performed:
y i 1 = ω 1 x i 1 , 1 + ω 2 x i , 2 + ω 3 x i + 1 , 3
Y 1 = j = 1 3 ( ω j x i 1 , j , ω j x i , j , ω j x i + 1 , j )
where y i 1 represents the process of the first convolution operation, and Y 1 represents the calculation result. After the first temporal shift, feature fusion is performed:
X 1 = X 0 ± X 1 M
where X 1 represents the feature values of the infinitely long video stream after the first feature fusion, M is the 1 1 1 0 0 0 0 0 0 matrix, which is used to extract the columns that have undergone temporal shifting in X 1 , namely the feature values in columns 1 and 3. At this point, X 1 is
X 1 = ( x i , 1 ± ω 1 x i 1 , 1 , x i , 2 , x i , 3 ± ω 3 x i 1 , 3 ) n × 3
After completing the first temporal shift and feature fusion, the second temporal shift is then performed:
y i 2 = ω 1 ( x i 1 , 1 ± ω 1 x i 2 , 1 ) + ω 2 x i , 2 + ω 3 ( x i + 1 , 3 ± ω 3 x i + 2 , 3 )
Y 2 = j = 1 3 ω j ( x i 1 , j ± ω j x i 2 , j ) + ω j x i , j + ω j ( x i + 1 , j ± ω j x i + 2 , j )
Here, y i 2 represents the process of the first convolution operation, and Y 2 represents the calculation result. After the second temporal shift, feature fusion is performed:
X 2 = X 0 ± X 2 M
where X 2 represents the feature values of the infinitely long video stream after the second feature fusion, and at this point, its value is
X 2 = [ x i , 1 ± ω 1 ( x i , 1 ± ω 1 x i 1 , 1 ) , x i , 2 , x i , 3 ± ω 3 ( x i , 3 ± ω 3 x i + 1 , 3 ) ] n × 3
After simplification, it is
X 2 = ( x i , 1 ± ω 1 x i , 1 ± ω 1 2 x i 1 , 1 , x i , 2 , x i , 3 ± ω 3 x i , 3 ± ω 3 2 x i + 1 , 3 ) n × 3
At this point, it can be clearly seen that after two rounds of temporal shift and feature fusion, each shifted channel in every time step contains the features before the shift. When the model starts the convolution operation, each time step for convolution contains not only the original feature values of the unshifted channels (such as the second column in Equation (12)) but also the feature values of the channels shifted forward and backward in the temporal dimension (such as the first and third columns in Equation (12)). By this method, the correlation between feature maps and between channels can be improved, thus greatly enhancing the robustness of the model.

3.3. DualCascadeTSF-MobileNetV2

To meet the demand for lightweight and efficient algorithms on edge devices for intelligent violence recognition, this paper first compares the performance of classic lightweight convolutional structure models to select a baseline suitable for combination with the DualCascadeTSF module. By referring to the performance parameters of each model listed in Table 1, this paper finds that MobileNetV2 can achieve a good balance between the number of parameters and accuracy, reaching an ideal precision value with a relatively low number of parameters. Therefore, this paper combines the DualCascadeTSF module with MobileNetV2 and proposes a new model, DualCascadeTSF-MobileNetV2. The DualCascadeTSF module enhances the ability to extract temporal features through temporal shift, while MobileNetV2 is renowned for its efficient lightweight structure. The combination of the two provides an effective solution for achieving a model with high performance and low computational cost.
Table 2 briefly lists the core structural design of DualCascadeTSF-MobileNetV2. Here, “conv” denotes the convolutional layer, “bottleneck” denotes the bottleneck layer, and “Avgpool” denotes the average pooling layer. When designing the DualCascadeTSF-MobileNetV2 model, we comprehensively considered factors such as the demand for lightweight design, feature extraction ability, and computational efficiency. Therefore, the DualCascadeTSF module was only introduced at specific positions rather than including this temporal shift and feature fusion mechanism in every layer. The layers with DualCascadeTSF module are bottleneck layers, which enhance the feature extraction capability in the temporal dimension by embedding the DualCascadeTSF module. In contrast, the layers without DualCascadeTSF module are mainly responsible for initial feature extraction and final feature mapping. Since the primary task of these layers is to quickly extract and map features, the DualCascadeTSF module is not embedded to avoid unnecessary computational burden.
As shown in Figure 5, it is the overall structure of DualCascadeTSF-MobileNetV2. This figure demonstrates the specific deployment positions of the DualCascadeTSF module and the overall framework of the model. When designing the DualCascadeTSF-MobileNetV2, we comprehensively considered the requirements for model lightweight design, the performance of the baseline model, and the characteristics of the DualCascadeTSF module. Based on the experimental results, we decided to refer to the architectural pattern of MobileNet-TSM and, leveraging the ability of the DualCascadeTSF module to enhance the inter-channel information connection after multiple temporal shifts, increased the number of layers within a reasonable parameter range to achieve better performance. Therefore, from the perspective of the overall model architecture, the design using seven bottleneck layers comprehensively considers the relationship between model lightweight design and feature extraction, enabling the model to enhance its feature extraction ability while maintaining its lightweight nature. By setting multiple bottleneck layers, each bottleneck layer gradually increases the complexity of processing and the level of abstraction, from simple features to more complex patterns, which helps to capture high-level features in the input data and makes it more suitable for handling complex video analysis tasks. This strategy effectively improves the efficiency and effectiveness of the model and achieves a good balance between performance and resource utilization.
The bottleneck layer consists of a pointwise convolution for channel expansion, a depthwise convolution for feature extraction, and another pointwise convolution for channel compression. Regarding the specific deployment position of the DualCascadeTSF module, placing it between the first pointwise convolution and the depthwise convolution not only fully utilizes the rich channel information after the first pointwise convolution but also allows the depthwise convolution to effectively extract features processed by the DualCascadeTSF module. This design optimizes the temporal feature extraction process. It prevents information loss due to too few channels while avoiding increased computational burden from too many channels, thereby achieving a balance between performance and efficiency. As a result, the entire model runs more efficiently on resource-constrained devices while maintaining high accuracy.

4. Experiments

4.1. Datasets

To verify the performance of the DualCascadeTSF-MobileNetV2 model in violence recognition tasks, this paper selected three publicly available violence-related datasets—Crowd Violence (Figure 6) [26], RWF-2000 (Figure 7) [27], Hockey Fights (Figure 8) [28]—for experimentation.
The Crowd Violence dataset, also known as the Violent Flow dataset, mainly focuses on scenes of crowd violence. This dataset contains 246 video clips, with the video lengths ranging from 1 s to 6.5 s, and the average duration is approximately 3.6 s. Due to the low performance of the shooting equipment and the large scene, the overall video images are rather blurry. However, these videos cover a variety of violent behaviors, such as fights and brawls. And in order to provide a comparative analysis, the dataset also includes some video clips of non-violent behaviors, which are used to enhance the model’s ability to distinguish between different behaviors.
The RWF-2000 dataset is a newly proposed medium-sized violent behavior dataset, which contains 2000 surveillance video clips collected from YouTube. Among them, the training set includes 1600 video clips, and the validation set includes 400 video clips. Each video clip is 5 s long, with a total of 150 frames, and has not undergone any preprocessing. This dataset demonstrates violent behaviors between two people, multiple people, and crowds. The scenes are complex and diverse, and the recognition difficulty is relatively high, which is very close to the actual application requirements and has a high research value. In addition, the dataset not only contains videos of violent behaviors but also videos of nonviolent behaviors as a control group to help the model better learn the differences between the two types of behavior.
The Hockey Fights dataset contains 1000 video clips collected from ice hockey games. Among them, the training set has 800 video clips, and the validation set has 200 video clips. Each clip is 2 s long and contains 41 frames. This dataset records the fighting behaviors of ice hockey players during the games, as well as the physical contacts in normal games as a control group. This dataset is not only of great significance for the safety management of sports events but also provides valuable resources for research in fields such as human behavior recognition and abnormal behavior detection.

4.2. Parameter Settings

The environmental configuration used in the experiment is as follows: The operating system is Ubuntu 16.04, the deep learning framework is PyTorch 1.9.0, the CPU is Intel I9-10920X. At the same time, CUDA 11.1 is utilized to accelerate GPU computing, and two NVIDIA RTX 2080 Ti GPUs (with 11 GB of video memory per card) are equipped for parallel computing. During the experiment, the base model MobileNetV2 is initialized using the pre-trained parameters provided by PyTorch, and the training and testing are completed based on this. The Stochastic Gradient Descent (SGD) algorithm is selected as the optimization algorithm. The initial learning rate is set to 0.000375, and the learning rate adjustment strategy is to decay the learning rate to 80% of its original value every two epochs. The input modality is RGB images, the batch size is set to 8, and the entire training process runs for 100 epochs.
In order to verify whether the model designed in this paper will exhibit overfitting, we replicated the experiment in the literature [29] on the RWF-2000 dataset and compared the loss curve of the model in this paper with the results of the replicated experiment. The experimental results are shown in Figure 9. The light red line and the light green line represent the loss value change curves of DualCascadeTSF-MobileNetV2 and MobileNet-TSM, respectively. As can be seen from the figure, after 20 epochs, the loss values of the two curves gradually converge, and the loss value of DualCascadeTSF-MobileNetV2 is significantly lower than that of MobileNet-TSM. This result indicates that the model proposed in this paper not only has stronger generalization ability but also does not have the problem of overfitting.

4.3. Results

4.3.1. Ablation Experiments of Different Feature Fusion Methods

Firstly, this paper conducts an ablation experiment on the feature fusion methods, focusing on examining the impacts of three key factors on the model’s performance: the batch of temporal shift (the first shift, the second shift), the direction of temporal shift (forward, backward), and the feature fusion algorithm (enhancing the correlation between features through addition or eliminating redundant information through subtraction). By testing different combinations in the experiment, the aim is to determine the optimal feature fusion strategy so as to improve the overall performance of the model.
Since the naming of the feature fusion methods involved in the experiment is rather complicated, for the convenience of description, we use serial numbers to represent different feature fusion methods. The specific rules are as follows: From left to right, the first digit represents the operation mode of the first forward shift, the second digit represents the operation mode of the first backward shift, the third digit represents the operation mode of the second forward shift, and the fourth digit represents the operation mode of the second backward shift. Among them, “1” represents the addition operation, and “0” represents the subtraction operation. For example, the combination mode of “using the addition operation for the first forward shift, using the subtraction operation for the first backward shift, using the addition operation for the second forward shift, and using the subtraction operation for the second backward shift” corresponds to the serial number “1010”.
After 100 epochs of training and validation on the RWF-2000 dataset, the experimental results are shown in Figure 10. It can be clearly seen from the figure that method 1010 exhibits the highest accuracy throughout the entire training process, with its accuracy stably above 85%, which is significantly higher than the other three methods. This indicates that this method has an advantage in feature fusion, enabling more effective extraction and utilization of temporal features. In addition, the method numbered 1010 not only has a high accuracy but also has a relatively smooth accuracy curve with small fluctuations. This means that this method has good convergence and stability during the training process and is not easily affected by overfitting or underfitting. In contrast, the accuracy curves of the other methods have large fluctuations, which may affect the final performance of the model. Therefore, based on experimental verification, the feature fusion method of “using the addition operation for the first forward shift, using the subtraction operation for the first backward shift, using the addition operation for the second forward shift, and using the subtraction operation for the second backward shift” is the optimal solution, enabling the model to achieve the best performance. Subsequent experiments will adopt this method for feature fusion.

4.3.2. Ablation Study on Different Structures

To verify whether the proposed model suffers from performance degradation due to structural complexity, we simplified the model structure and compared the performance of different modules combined with MobileNetV2. Four module combinations were selected for the experiments: DualCascadeTSM-MobileNetV2, TSF-MobileNetV2, TSM-MobileNetV2, and DualCascadeTSF-MobileNetV2. These combinations were evaluated on three datasets: Crowd Violence, RWF-2000, and Hockey Fights. After 100 epochs of training, the experimental results are shown in Figure 11.
Figure 11a shows the accuracy curve of the model on the Crowd Violence dataset. It can be seen from the figure that the accuracy of DualCascadeTSF-MobileNetV2 and DualCascadeTSM-MobileNetV2 is significantly higher than that of TSF-MobileNetV2 and TSM-MobileNetV2. Figure 11b shows the accuracy curve of the model on the RWF-2000 dataset. It can be seen that the accuracy of DualCascadeTSF-MobileNetV2 is significantly better than the other three models. Figure 11c shows the accuracy curve of the model in the Hockey Fights dataset. The results indicate that DualCascadeTSF-MobileNetV2 not only achieves the highest accuracy but also maintains stability after convergence.
The ablation experiments demonstrate that the proposed DualCascadeTSF-MobileNetV2 does not suffer performance degradation due to the addition of extra modules. On the contrary, the model outperforms other module combinations on all three datasets, exhibiting superior stability and recognition capabilities. This indicates that the integration of the DualCascadeTSF module with MobileNetV2 not only enhances the model’s performance but also effectively avoids the negative impacts associated with increased structural complexity.

4.3.3. Comparative Experiments with Classic Models

In order to verify the advantages and disadvantages of the model proposed in this paper, we selected several classic behavior recognition models for comparative experiments. All the models were trained and tested on the three datasets mentioned above (Crowd Violence, RWF-2000, and Hockey Fights) under the same experimental environment. The comparative analysis was mainly carried out from four aspects: accuracy, the number of parameters, the occupied memory size, and the calculation speed. The results are shown in Table 3.
For the three indicators of the number of parameters, the occupied memory size, and the calculation speed, we used the third-party library torchsummary of PyTorch for calculation, and uniformly set the input parameters as a batch size of 8, the number of channels as 24, the height of the input video frame as 128, and the width as 128. The results are shown in Table 4.

5. Discussion

As can be seen in Table 3, the DualCascadeTSF-MobileNetV2 model proposed in this paper has achieved results that are close to or even better than those of other classic models in the three violent behavior datasets. On the Crowd Violence dataset, the accuracy of DualCascadeTSF-MobileNetV2 reaches 98.98%, which is higher than other classical models and represents an improvement of 0.83% compared to the previously best-performing model, SAM. On the RWF-2000 dataset, its accuracy was 88.5%, higher than eight models and slightly lower than four models, with only a 2.5% gap compared to the highest-accuracy model, P-TSM. On the Hockey Fights dataset, its accuracy was 98.0%, higher than nine models and slightly lower than four models, with only a 1.1% gap compared to the highest-accuracy model, SAM. Compared to MobileNet-TSM, which also uses MobileNetV2 as the base model, the accuracy of the model in this paper is 1.021%, 0.75%, and 0.5% higher on the three datasets, respectively. Compared with the lightweight model MiNet-3D, the accuracy of the model in this paper is 7.57%, 6.52%, and 3.29% higher on the three datasets, respectively. Overall, although there is still a certain gap between DualCascadeTSF-MobileNetV2 and the most advanced algorithms in terms of accuracy, its performance is better than that of most models, especially among lightweight models, and it is significantly better than MobileNet-TSM and MiNet-3D.
Moreover, it can be seen from Table 4 that in terms of the number of parameters, DualCascadeTSF-MobileNetV2 has a parameter size of 16.99 MB. In comparative experiments, although this value is slightly higher than that of MobileNet-TSM, it still significantly reduces the parameter scale compared to other classic models. This indicates that DualCascadeTSF-MobileNetV2 has adopted effective strategies in its design to reduce redundant parameters while ensuring that the model can capture key features in videos.
Secondly, regarding memory usage, DualCascadeTSF-MobileNetV2 occupies a memory size of 347.13 MB. Although slightly larger than MobileNet-TSM, considering its accuracy rate of 98.98% on the Crowd Violence dataset and its equally outstanding performance on the RWF-2000 and Hockey Fights datasets, this moderate increase in memory consumption is acceptable. This level of memory usage also means that the model can be deployed on edge devices, which is crucial for real-time applications.
Lastly, thanks to the superior structure and performance of MobileNetV2, DualCascadeTSF-MobileNetV2 not only excels in training efficiency, requiring only 35.92 min to complete training, which is far less than traditional models such as 3D-CNN (248.38 min) and LRCN (185.12 min). At the same time, DualCascadeTSF-MobileNetV2 has the same GFLOPs as MobileNet-TSM, at just 5.0 GFLOPs, indicating its low computational complexity. Compared to classic models such as 3D-CNN, LRCN, and TSN, DualCascadeTSF-MobileNetV2 requires significantly fewer computational resources. In particular, compared to TSM and its improved versions, the GFLOPs of DualCascadeTSF-MobileNetV2 are reduced by up to 14 times. This shows that DualCascadeTSF-MobileNetV2 is highly suitable for scenarios with limited computational resources, such as mobile devices or edge computing environments, and it offers faster speed and lower energy consumption. Therefore, whether in terms of training efficiency or performance during the inference stage, DualCascadeTSF-MobileNetV2 demonstrates strong advantages, providing powerful support for the rapid development and deployment of efficient video classification solutions.
To more intuitively demonstrate the performance of the models, this paper presents a comparison of DualCascadeTSF-MobileNetV2 with other classical models in terms of the number of parameters, memory size, and training speed in Figure 12. In the figure, the x-axis represents the number of parameters, the y-axis represents the training time, and the area of the circle indicates the memory size of the model. The colors are used solely to distinguish different models and do not carry any other special meanings. As shown in the figure, the proposed model has a parameter quantity of 16.99 MB, a training time of 35.92 min, and a memory usage of 347.13 MB. Except for being slightly larger in parameter quantity and memory usage compared to MobileNet-TSM, the proposed model outperforms the other classical models in these three metrics, successfully achieving the design goal of being “lighter and faster” and demonstrating feasibility for deployment on edge devices.

6. Conclusions

To meet the current demand for lightweight and efficient networks in violent behavior recognition tasks, based on the previous research results, this paper designs the DualCascadeTSF-MobileNetV2. The experimental results show that this model demonstrates excellent performance on the three datasets, namely Crowd Violence, Hockey Fights, and RWF-2000. Especially on the Crowd Violence dataset, its accuracy exceeds that of other classic models. In the other two datasets, although the accuracy does not reach the highest level, considering that this model has fewer parameters, a smaller size, and a faster operation speed, these advantages make its deployment on edge devices highly attractive. In practical applications, these performance optimizations can almost make up for the small accuracy gap. Therefore, the DualCascadeTSF-MobileNetV2 model has broad application prospects in future intelligent recognition edge devices.
Finally, this paper proposes a question worthy of further exploration: the literature [15] has pointed out that cascading more TSM modules within each residual block does not further enhance the model’s feature extraction capabilities. The likely reason is that excessive temporal shifting operations make the connections between originally adjacent feature maps sparse. In this paper, we have enhanced the temporal coupling between feature maps by introducing feature fusion of the feature maps before and after each shift. So, if we further cascade more temporal shifting and feature fusion modules within each linear bottleneck layer of this model, will it further improve the model’s feature extraction capabilities? We welcome readers to discuss this question with us!

Author Contributions

Conceptualization, Y.L. and Y.C.; methodology, Y.C.; software, Y.C. and S.L. (Shaohua Li); validation, Y.C., F.L. and S.L. (Shuhan Lv); formal analysis, Y.C.; investigation, F.L.; resources, Y.L.; data curation, S.L. (Shuhan Lv); writing—original draft preparation, Y.C.; writing—review and editing, Y.L.; visualization, S.L. (Shaohua Li); supervision, Y.L.; project administration, Y.L.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Science and Technology Innovation Project grant number No. ZZKY20222304.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are openly available in the dataset at https://aistudio.baidu.com/datasetdetail/99624, accessed on 25 March 2025.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DualCascadeTSFDual Cascade Temporal Shift and Fusion module
TSMtemporal shift module

References

  1. Lin, J.; Gan, C.; Han, S. TSM: Temporal Shift Module for Efficient Video Understanding. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–5 November 2019. [Google Scholar]
  2. Lian, Z.; Yin, Y.; Lu, J.; Xu, Q.; Zhi, M.; Hu, W.; Duan, W. A Survey: Feature Fusion Method for Object Detection Field. In Advanced Intelligent Computing Technology and Applications; Springer Nature: Singapore, 2024. [Google Scholar]
  3. Liu, Z.; Li, J.; Shen, Z.; Huang, G.; Yan, S.; Zhang, C. Learning Efficient Convolutional Networks through Network Slimming. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  4. Li, H.; Kadav, A.; Durdanovic, I.; Samet, H.; Graf, M.P. Pruning Filters for Efficient ConvNets. arXiv 2016, arXiv:1608.08710. [Google Scholar]
  5. Sanh, V.; Wolf, T.; Rush, A.M. Movement Pruning: Adaptive Sparsity by Fine-Tuning. arXiv 2020, arXiv:2005.07683. [Google Scholar]
  6. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Wey, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  7. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  8. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  9. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  10. Howard, A.; Sandler, M.; Chen, B.; Wang, W.; Chen, L.-C.; Tan, M.; Chu, G.; Vasudevan, V.; Zhu, Y.; Pang, R.; et al. Searching for MobileNetV3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  11. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; Volume 2. [Google Scholar]
  12. Wang, X.; Xiang, T.; Zhang, C.; Song, Y.; Liu, D.; Huang, H.; Cai, W. BiX-NAS: Searching Efficient Bi-directional Architecture for Medical Image Segmentation. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2021, Proceedings of the 24th International Conference, Strasbourg, France, 27 September–1 October 2021, Proceedings, Part I 24; Springer International Publishing: Cham, Switzerland, 2021; pp. 229–238. [Google Scholar]
  13. Yan, B.; Peng, H.; Wu, K.; Wang, D.; Fu, J.; Lu, H. LightTrack: Finding Lightweight Neural Networks for Object Tracking via One-Shot Architecture Search. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  14. Cai, H.; Gan, C.; Han, S. Once for All: Train One Network and Specialize it for Efficient Deployment. arXiv 2019, arXiv:1908.09791. [Google Scholar]
  15. Liang, Q.; Li, Y.; Chen, B.; Yang, K. Violence behavior recognition of two-cascade temporal shift module with attention mechanism. J. Electron. Imaging 2021, 30, 43009. [Google Scholar] [CrossRef]
  16. Gholami, A.; Kwon, K.; Wu, B.; Tai, Z.; Yue, X.; Jin, P.; Zhao, S.; Keutzer, K. SqueezeNext: Hardware-Aware Neural Network Design. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  17. Wu, B.; Wan, A.; Yue, X.; Keutzer, K. Shift: A Zero FLOP, Zero Parameter Alternative to Spatial Convolutions. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 9127–9135. [Google Scholar]
  18. Qin, D.; Leichner, C.; Delakis, M.; Fornoni, M.; Luo, S.; Yang, F.; Wang, W.; Banbury, C.; Ye, C.; Akin, B.; et al. MobileNetV4: Universal Models for the Mobile Ecosystem. In Computer Vision—ECCV 2024, Proceedings of the 18th European Conference, Milan, Italy, 29 September–4 October 2024, Proceedings, Part XL; Springer: Cham, Switzerland, 2024. [Google Scholar]
  19. Tan, M.; Le, Q.V. EfficientNetV2: Smaller Models and Faster Training. In Proceedings of the 38th International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022. [Google Scholar]
  20. Wang, A.; Chen, H.; Lin, Z.; Han, J.; Ding, G. Rep ViT: Revisiting Mobile CNN From ViT Perspective. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024. [Google Scholar]
  21. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. GhostNet: More Features From Cheap Operations. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  22. Tang, Y.; Han, K.; Guo, J. GhostNetV2: Enhance Cheap Operation with Long-Range Attention. In Proceedings of the Advances in Neural Information Processing Systems 35 (NeurIPS 2022), New Orleans, LA, USA, 28 November–9 December 2022. [Google Scholar]
  23. Han, K.; Wang, Y.; Xu, C.; Guo, J.; Xu, C.; Wu, E.; Tian, Q. GhostNets on Heterogeneous Devices via Cheap Operations. Int. J. Comput. Vis. 2022, 130, 1050–1069. [Google Scholar] [CrossRef]
  24. Ma, N.; Zhang, X.; Zheng, H.-T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. In Proceedings of the Computer Vision, 15th European Conference (ECCV 2018), Munich, Germany, 8–14 September 2018. [Google Scholar]
  25. Freeman, I.; Roese-Koerner, L.; Kummert, A. EffNet: An Efficient Structure for Convolutional Neural Networks. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018. [Google Scholar]
  26. Hassner, T.; Itcher, Y.; Kliper-Gross, O. Violent flows: Real-time detection of violent crowd behavior. In Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  27. Cheng, M.; Cai, K.; Li, M. RWF-2000: An Open Large Scale Video Database for Violence Detection. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021. [Google Scholar]
  28. Bermejo Nievas, E.; Deniz Suarez, O.; Bueno Garcia, G.; Sukthankar, R. Violence Detection in Video Using Computer Vision Techniques. In Proceedings of the International Conference on Computer Analysis of Images and Patterns, Seville, Spain, 29–31 August 2011. [Google Scholar]
  29. Zhang, Y.; Li, Y.; Guo, S. Lightweight mobile network for real-time violence recognition. PLoS ONE 2022, 17, e0276939. [Google Scholar]
  30. Shafiq, M.; Gu, Z. Deep Residual Learning for Image Recognition: A Survey. Appl. Sci. 2022, 12, 8972. [Google Scholar] [CrossRef]
  31. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 221–231. [Google Scholar] [PubMed]
  32. Donahue, J.; Hendricks, L.A.; Rohrbach, M.; Venugopalan, S.; Guadarrama, S.; Saenko, K.; Darrell, T. Long-Term Recurrent Convolutional Networks for Visual Recognition and Description. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 677–691. [Google Scholar] [CrossRef] [PubMed]
  33. Carreira, J.; Zisserman, A. Quo Vadis, action recognition? A new model and the kinetics dataset. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 22–25 July 2017. [Google Scholar]
  34. Wang, W.; Dong, S.; Zou, K.; Li, W. A Lightweight Network for Violence Detection. In Proceedings of the ICIGP 2022: 2022 the 5th International Conference on Image and Graphics Processing (ICIGP), Beijing, China, 7–9 January 2022. [Google Scholar]
  35. Meng, Y.; Lin, C.-C.; Panda, R.; Sattigeri, P.; Karlinsky, L.; Oliva, A.; Saenko, K.; Feris, R. AR-Net: Adaptive Frame Resolution for Efficient Action Recognition. In Computer Vision–ECCV 2020, Proceedings of the16th European Conference, Glasgow, UK, 23–28 August 2020, Proceedings, Part VII 16; Springer: Cham, Switzerland, 2020. [Google Scholar]
  36. Li, Y.; Ji, B.; Shi, X.; Zhang, J.; Kang, B.; Wang, L. TEA: Temporal Excitation and Aggregation for Action Recognition. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  37. Asad, M.; Jiang, H.; Yang, J.; Tu, E.; Malik, A.A. Multi-Level Two-Stream Fusion-Based Spatio-Temporal Attention Model for Violence Detection and Localization. Int. J. Pattern Recognit. Artif. Intell. 2022, 36, 2255002. [Google Scholar]
  38. Mohammadi, H.; Nazerfard, E. Video violence recognition and localization using a semi-supervised hard attention model. Expert Syst. Appl. 2023, 212, 118791. [Google Scholar]
  39. Zhang, Y.; Li, Y.; Guo, S.; Liang, Q. Not all temporal shift modules are profitable. J. Electron. Imaging 2022, 31, 043030. [Google Scholar]
  40. Wang, L.; Xiong, Y.; Wang, Z.; Qiao, Y.; Lin, D.; Tang, X.; Van Gool, L. Temporal Segment Networks: Towards Good Practices for Deep Action Recognition. In Computer Vision—ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016, Proceedings, Part VIII; Springer: Cham, Switzerland, 2016; Volume 9912, pp. 20–36. [Google Scholar]
Figure 1. The process of the video stream going through the temporal shift module operation. (a) The video stream before temporal shifting. (b) The video stream after temporal shifting.
Figure 1. The process of the video stream going through the temporal shift module operation. (a) The video stream before temporal shifting. (b) The video stream after temporal shifting.
Applsci 15 03862 g001
Figure 2. The two convolution forms of depthwise separable convolution. (a) Pointwise convolution and (b) depthwise convolution.
Figure 2. The two convolution forms of depthwise separable convolution. (a) Pointwise convolution and (b) depthwise convolution.
Applsci 15 03862 g002
Figure 3. The structure of the DualCascadeTSF module.
Figure 3. The structure of the DualCascadeTSF module.
Applsci 15 03862 g003
Figure 4. The working principle of the DualCascadeTSF module. Stage 1 represents the original state of the input video stream, Stage 2 involves the first temporal shift of the video stream, Stage 3 is the feature fusion of the video stream after the first temporal shift, Stage 4 involves the second temporal shift of the video stream, Stage 5 is the feature fusion of the video stream after the second temporal shift.
Figure 4. The working principle of the DualCascadeTSF module. Stage 1 represents the original state of the input video stream, Stage 2 involves the first temporal shift of the video stream, Stage 3 is the feature fusion of the video stream after the first temporal shift, Stage 4 involves the second temporal shift of the video stream, Stage 5 is the feature fusion of the video stream after the second temporal shift.
Applsci 15 03862 g004
Figure 5. Overall architecture of DualCascadeTSF-MobileNetV2. The left part is the entire structure of DualCascadeTSF-MobileNetV2. The middle part is the structure of one bottleneck in DualCascadeTSF-MobileNetV2. The right part is the structure of one DualCascadeTSF within the bottleneck.
Figure 5. Overall architecture of DualCascadeTSF-MobileNetV2. The left part is the entire structure of DualCascadeTSF-MobileNetV2. The middle part is the structure of one bottleneck in DualCascadeTSF-MobileNetV2. The right part is the structure of one DualCascadeTSF within the bottleneck.
Applsci 15 03862 g005
Figure 6. The Crowd Violence dataset.
Figure 6. The Crowd Violence dataset.
Applsci 15 03862 g006
Figure 7. The RWF-2000 dataset.
Figure 7. The RWF-2000 dataset.
Applsci 15 03862 g007
Figure 8. The Hockey Fights dataset.
Figure 8. The Hockey Fights dataset.
Applsci 15 03862 g008
Figure 9. Loss value curves of the two models.
Figure 9. Loss value curves of the two models.
Applsci 15 03862 g009
Figure 10. Accuracy curves of four different feature fusion methods on the RWF-2000 dataset.
Figure 10. Accuracy curves of four different feature fusion methods on the RWF-2000 dataset.
Applsci 15 03862 g010
Figure 11. Experimental accuracy curves: (a) accuracy curve on the Crowd Violence dataset; (b) accuracy curve on the RWF-2000 dataset; (c) accuracy curve on the Hockey Fights dataset.
Figure 11. Experimental accuracy curves: (a) accuracy curve on the Crowd Violence dataset; (b) accuracy curve on the RWF-2000 dataset; (c) accuracy curve on the Hockey Fights dataset.
Applsci 15 03862 g011
Figure 12. Comparison of DualCascadeTSF-MobileNetV2 with other classical models in terms of the number of parameters, memory size, and training speed.
Figure 12. Comparison of DualCascadeTSF-MobileNetV2 with other classical models in terms of the number of parameters, memory size, and training speed.
Applsci 15 03862 g012
Table 1. Comparison of the performance of classic convolutional structure models. The models we list are all versions without expansion or compression. In SqueezeNet_8bit, “8bit” refers to the data type. “1.0_SqueezeNext_23” indicates a version without expansion and composed of 23 modules. “ShiftNet_A” denotes its Type A. “MobileNetV4_Conv-M” represents the convolutional version and is the medium-sized model. “EfficientNet_B0” is the smallest base model, while “EfficientNetV2_M” is the medium-sized version. “RepViT_M1.0” indicates the medium-sized version without expansion, and “G_GhostNet_w/_mix” refers to the version with mix operations. “ShuffleNet_1(g = 8)” means that the channels are divided into eight groups.
Table 1. Comparison of the performance of classic convolutional structure models. The models we list are all versions without expansion or compression. In SqueezeNet_8bit, “8bit” refers to the data type. “1.0_SqueezeNext_23” indicates a version without expansion and composed of 23 modules. “ShiftNet_A” denotes its Type A. “MobileNetV4_Conv-M” represents the convolutional version and is the medium-sized model. “EfficientNet_B0” is the smallest base model, while “EfficientNetV2_M” is the medium-sized version. “RepViT_M1.0” indicates the medium-sized version without expansion, and “G_GhostNet_w/_mix” refers to the version with mix operations. “ShuffleNet_1(g = 8)” means that the channels are divided into eight groups.
ModelWeightsComputational ComplexityTop-1 Acc.Top-5 Acc.Dataset
SqueezeNet_8bit [9]0.6657.580.3ImageNet
1.0_SqueezeNext_23 [16]0.72282M MAC58.9882.33ImageNet
ShiftNet_A [17]4.170.1ImageNet
MobileNet_1 [6]4.2569M Mult-Add70.6ImageNet
MobileNetV2_1 [7]3.4300M Mult-Add72.0ImageNet
MobileNetV3_Large_1 [10]4.0155M Mult-Add73.3ImageNet
MobileNetV4_Conv_M [18]9.21.0G MAC79.9ImageNet
EfficientNet_B0 [11]5.30.39B FLOP77.193.3ImageNet
EfficientNetV2_M [19]5424B FLOP85.1ImageNet
RepViT_M1.0 [20]6.81.1G MAC80.0ImageNet
GhostNet_1 [21]5.2141M FLOP73.991.4ImageNet
GhostNetV2_1.0 [22]6.1167M FLOP75.392.4ImageNet
C_GhostNet_1 [23]5.2141M FLOP73.991.4ImageNet
G_GhostNet_w/_mix [23]14.62.3B FLOP73.191.2ImageNet
ShuffleNet_1(g=8) [8]140M FLOP67.6ImageNet
ShuffleNetV2_1 [24]146M FLOP69.4ImageNet
EffNet [25]0.1411.4M FLOP80.20CIFAR-10
Table 2. Structure of DualCascadeTSF-MobileNetV2.
Table 2. Structure of DualCascadeTSF-MobileNetV2.
LayerStructureDualCascadeTSF IncludedOutput SizeRepetition Times
conv13 × 3 × 32 112 × 112×1
bottleneck11 × 1 × 32, 3 × 3 × 32, 1 × 1 × 16Yes112 × 112×1
bottleneck21 × 1 × 96, 3 × 3 × 96, 1 × 1 × 24Yes56 × 56×2
bottleneck31 × 1 × 192, 3 × 3 × 192, 1 × 1 × 32Yes28 × 28×3
bottleneck41 × 1 × 384, 3 × 3 × 384, 1 × 1 × 64Yes14 × 14×4
bottleneck51 × 1 × 576, 3 × 3 × 576, 1 × 1 × 96Yes14 × 14×3
bottleneck61 × 1 × 960, 3 × 3 × 960, 1 × 1 × 160Yes7 × 7×3
bottleneck71 × 1 × 960, 3 × 3 × 960, 1 × 1 × 320Yes7 × 7×1
conv21 × 1 × 1280 7 × 7×1
AvgpoolPool 7 × 7 1 × 1×1
FCLinear 1280 → 2 1 × 1×1
Table 3. Comparison of accuracy rates of DualCascadeTSF-MobileNetV2 with other classical models on the three datasets.
Table 3. Comparison of accuracy rates of DualCascadeTSF-MobileNetV2 with other classical models on the three datasets.
ModelCrowd Violence (%)RWF-2000 (%)Hockey Fights (%)
ResNet-50 [30]93.8788495.5
3D-CNN [31]94.382.7594.4
LRCN [32]94.577797.1
I3D [33]88.8985.7597.5
MiNet-3D [34]91.4181.9894.71
AR-Net [35]95.91887.397.2
TSM [1]95.958897.5
TEA [36]96.93988.597.7
Two-cascade TSM [15]96.9398998.05
SAM [37]98.1589.199.1
SSHA [38]-90.498.7
P-TSM [39]96.9699198.5
MobileNet-TSM [29]97.95987.7597.5
DualCascadeTSM-MobileNetV2 (ours)98.9888.598.0
Table 4. Comparison of the number of parameters, memory size, and training speed between DualCascadeTSF-MobileNetV2 and other classic models.
Table 4. Comparison of the number of parameters, memory size, and training speed between DualCascadeTSF-MobileNetV2 and other classic models.
ModelParameters (MB)Training Speed (min)Memory Size (MB)GFLOPs
3D-CNN [31]297.83248.382647.7191.6
LRCN [32]237.83185.121212.93186.0
I3D [33]46.88123.81000.235.6
TSN [40]89.6946.68390.05125.3
TSM [1]89.69113.43397.7168.1
TEA [36]91.95190.8479.7869.9
Two-cascade TSM [15]89.69128.33397.7168.1
P-TSM [39]89.6987.63396.5768.1
MobileNet-TSM [29]8.4932.68175.865.0
DualCascadeTSM-MobileNetV2 (ours)16.9935.92347.135.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.; Li, Y.; Li, S.; Lv, S.; Lin, F. DualCascadeTSF-MobileNetV2: A Lightweight Violence Behavior Recognition Model. Appl. Sci. 2025, 15, 3862. https://doi.org/10.3390/app15073862

AMA Style

Chen Y, Li Y, Li S, Lv S, Lin F. DualCascadeTSF-MobileNetV2: A Lightweight Violence Behavior Recognition Model. Applied Sciences. 2025; 15(7):3862. https://doi.org/10.3390/app15073862

Chicago/Turabian Style

Chen, Yuang, Yong Li, Shaohua Li, Shuhan Lv, and Fang Lin. 2025. "DualCascadeTSF-MobileNetV2: A Lightweight Violence Behavior Recognition Model" Applied Sciences 15, no. 7: 3862. https://doi.org/10.3390/app15073862

APA Style

Chen, Y., Li, Y., Li, S., Lv, S., & Lin, F. (2025). DualCascadeTSF-MobileNetV2: A Lightweight Violence Behavior Recognition Model. Applied Sciences, 15(7), 3862. https://doi.org/10.3390/app15073862

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop