You are currently viewing a new version of our website. To view the old version click .
Big Data and Cognitive Computing
  • Article
  • Open Access

12 November 2025

Overcoming Domain Shift in Violence Detection with Contrastive Consistency Learning

,
and
1
School of Software, Northeastern University, Shenyang 110819, China
2
National Frontiers Science Center for Industrial Intelligence and Systems Optimization, Northeastern University, Shenyang 110819, China
3
Key Laboratory of Data Analytics and Optimization for Smart Industry, Ministry of Education, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.

Abstract

Automated violence detection in video surveillance is critical for public safety; however, existing methods frequently suffer notable performance degradation across diverse real-world scenarios due to domain shift. Substantial distributional discrepancies between source training data and target environments severely hinder model generalization, limiting practical deployment. To overcome this, we propose CoMT-VD, a new contrastive Mean Teacher-based violence detection model, engineered for enhanced adaptability in unseen target domains. CoMT-VD innovatively integrates a Mean Teacher architecture to adequately leverage unlabeled target domain data, fostering stable, domain-invariant feature representations by enforcing consistency regularization between student and teacher networks, crucial for bridging the domain gap. Furthermore, to mitigate supervisory noise from pseudo-labels and refine the feature space, CoMT-VD incorporates a dual-strategy contrastive learning module. DCL systematically refines features through intra-sample consistency, minimizing latent space distances for compact representations, and inter-sample consistency, maximizing feature dissimilarity across distinct categories to sharpen decision boundaries. This dual regularization purifies the learned feature space, boosting discriminativeness while mitigating noisy pseudo-labels. Broad evaluations on five benchmark datasets unequivocally demonstrate that CoMT-VD achieves the superior generalization performance (in the four integrated scenarios from five benchmark datasets, the improvements were 5.0∼12.0%, 6.0∼12.5%, 5.0∼11.2%, 5.0∼11.2%, and 6.3∼12.3%, respectively), marking a notable advancement towards robust and reliable real-world violence detection systems.

1. Introduction

Violence detection (VD), a subtask of human action recognition, addresses a research area of substantial practical importance [,,,,]. The development of robust violence detection systems is paramount for curbing the proliferation of harmful content on digital platforms, enabling the proactive monitoring of potential violent incidents in real-world environments, and facilitating preventive measures against such behaviors. Collectively, these applications notably contribute to safeguarding public safety and maintaining the integrity of both online and physical spaces.
Driven by notable advancements in artificial intelligence, methodologies for violent detection have undergone a paradigm shift, transitioning from traditional manual feature engineering approaches [,] to sophisticated deep learning-based architectures. Current research largely employs three main model categories: (a) hybrid frameworks combining two-dimensional convolutional neural networks (2D CNNs) with temporal modeling modules [,,] to capture both spatial and sequential dynamics; (b) end-to-end spatiotemporal representation learning via three-dimensional convolutional neural networks (3D CNNs) [,,]; and (c) more recently, attention-based architectures, particularly Transformer models [,], which excel at modeling long-range dependencies in video data. While these contemporary models often demonstrate strong performance in controlled experimental settings, their efficacy in practical, real-world deployments is frequently hindered by notable challenges, predominantly performance degradation when encountering unseen or dynamically changing environments.
Concretely, the reliable real-world deployment of these models is principally impeded by the domain shift [] phenomenon. This challenge stems from a fundamental distributional discrepancy between the data used for training and the data encountered during actual deployment [,,,,,,]. In violence detection, this manifests acutely due to the vast, unpredictable variability inherent in real-world scenarios, which datasets, by their nature, can only partially represent. For example, authentic violent incidents in the real world frequently occur under dynamically complex visual conditions such as occlusions (Figure 1a), where key elements may be obscured, or in low-illumination environments (Figure 1b), where visual clarity is compromised. These characteristics of target domain often differ qualitatively and in their range of variation from the training data subsets, which might include specific scenarios like actions against a pure-color background (Figure 1c) or with high motion blur (Figure 1d). Models trained on limited violence datasets may encounter unknown environmental changes when applied to real-world scenarios. This situation, where the data distribution used during training is inconsistent with the data encountered in practical applications, is a manifestation of domain shift [].
Figure 1. Examples of domain shift in violent detection. (a) Violent behavior with occlusion in real-world scenarios. (b) Violent behavior with low light in real-world scenarios. (c) Pure-color background. (d) High motion blur.
Such divergence is practically unavoidable, as real-world scenarios are inherently open-ended, encompassing countless variations in environmental factors, camera perspectives, and scene compositions that finite datasets cannot exhaustively capture [,,]. This inherent gap between training and real-world operational data critically exacerbates domain shift, profoundly undermining model generalization and leading to sharp performance declines in practical applications. This degradation carries severe security implications: systems may generate false alarms under benign environmental variations or, more critically, fail to detect genuine threats, potentially enabling exploitation by malicious actors. Addressing this pervasive domain shift is therefore paramount for developing reliable and robust VD technologies.
To mitigate this pervasive domain shift, we introduce CoMT-VD, a new domain adaptation framework specifically designed for violence detection. Our primary objective is to adequately transfer and adapt knowledge from labeled source domains to unlabeled target domains [,,], thereby notably enhancing generalization ability. CoMT-VD strategically leverages the Mean Teacher (MT) architecture [,], comprising the student θ S network and teacher one θ T with identical structures, where θ T is updated via Exponential Moving Average (EMA) of θ S . This design choice is fundamental because the EMA mechanism empowers the teacher network to generate highly stable and reliable pseudo-labels for sufficient unlabeled target data. These stable pseudo-labels, in conjunction with a consistency loss, are then critically utilized to guide θ S training, compelling it to align its predictions with θ T and thereby facilitating essential adaptation to diverse domain variations. While data augmentation [,,] further supports this adaptation by simulating real-world visual diversity, the inherent noise stemming from imperfect pseudo-labels generated by the MT setup necessitates a more robust feature refinement strategy. To address this crucial need and purify the learned feature space [,], we propose and integrate a dual-strategy contrastive learning (DCL) module. DCL meticulously refines feature representations through two complementary objectives. First, intra-sample consistency compels θ S to produce highly similar features for differently augmented views of the same input. This ensures that learned features are robust to superficial visual changes and capture the essential, invariant characteristics of violence. Second, inter-sample consistency enhances semantic coherence. It actively maps features of samples with similar violent patterns closer together while maximizing distance for dissimilar categories, thereby sharpening decision boundaries. This synergistic integrated approach allows for the Mean Teacher framework to adequately drive domain adaptation [,,], simultaneously ensuring that DCL purifies the learned feature space by enforcing robustness to nuisance visual variations and notably enhancing categorical discriminability. This process adequately mitigates the detrimental impact of potentially noisy pseudo-labels and ensures that the student network θ S learns highly discriminative yet domain-invariant features, ultimately leading to improved generalization in violence detection across shifting domains.
The main contributions of this paper are summarized as follows:
  • We present a pioneering investigation into the pervasive domain shift challenges in violence detection. To adequately address this, we propose CoMT-VD, a new contrastive Mean Teacher model specifically designed to enhance model adaptability and performance across diverse target domain distributions.
  • We introduce a new dual-strategy contrastive learning (DCL) module that integrates two distinct positive-negative pair matching strategies to compute complementary consistency losses. This promotes the learning of more effective discriminative features for violence detection under challenging domain shift conditions.
  • We conduct comprehensive evaluations of detecting violence under challenging domain shift scenarios, which unequivocally demonstrate consistent and notable performance improvements when CoMT-VD is integrated with various baseline models.
In the following sections, we will introduce current research on violence detection in the Related Work section, along with relevant studies involving domain adaptation, contrastive learning, knowledge distillation, and other related methodologies that this paper will address. In the Proposed Method section, we will provide a detailed explanation of the CoMT-VD training strategy we have developed. In the Experiments and Ablation Studies sections, we will demonstrate how the proposed CoMT-VD enhances model performance in dealing with domain shift issues. Finally, in the Conclusions section and Future Work sub-section, we will summarize the contributions of this paper and discuss the limitations of the current research.

2. Related Work

2.1. Violence Detection

Early efforts for recognizing violent behaviors predominantly relied on handcrafted feature engineering [,]. The rapid evolution of artificial intelligence has catalyzed a paradigm shift toward deep learning-based approaches. Recent violent detection architectures are primarily categorized into three technical streams: 2DCNN-based spatio-temporal methods: Extracts spatial features via 2D convolutional networks coupled with temporal modeling modules, e.g., LSTM or TSM [,,], for sequential reasoning. 3DCNN-based methods: Directly captures spatiotemporal correlations through 3D convolutions, e.g., C3D, I3D, or X3D [,,]. Transformer-based: As researchers extend Vision Transformer (ViT) architectures to video understanding, transformer-based methods are increasingly being explored for violent detection tasks. We note that 2DCNN+LSTM network excels in long-term temporal modeling, but struggles to capture short-term local continuous temporal variations. The 3DCNN-based network has limitations in detecting long-duration continuous actions. Transformer-based networks demonstrate superior capability in global relationship modeling for video context understanding, although challenges persist in balancing local feature sensitivity and reducing data dependency. Researchers have also explored avenues such as multi-stream networks to fuse richer multimodal information [,].
Despite notable advances, a fundamental and persistent challenge continues to hinder contemporary approaches: the domain shift between training and testing distributions. Models trained on specific, often curated datasets—such as cinematic depictions of violence or controlled surveillance footage—frequently suffer severe performance degradation when deployed in real-world settings. These environments typically exhibit substantial variations in lighting, occlusions, camera viewpoints, recording quality, and the subtle, unpredictable nature of violent behavior. This critical generalization gap, particularly evident in cross-domain scenarios, highlights an urgent need for models with greater adaptability. Addressing this challenge lies at the core of our CoMT-VD framework, which is specifically designed to enhance robustness against such distributional shifts and improve cross-domain generalization.

2.2. Knowledge Distillation and Mean Teacher

Knowledge Distillation (KD) presents a paradigm wherein a student model learns to emulate the decision-making processes of a larger teacher model []. This technique has found successful application in various domains, including human action recognition [], where student models have demonstrated notable performance enhancements by leveraging the distilled knowledge from more potent teacher architectures. One of the important evolutions of KD, particularly for semi-supervised and unsupervised learning scenarios, is the Mean Teacher (MT) framework []. In the MT paradigm, the teacher model’s parameters are not independently trained, but are instead maintained as an exponential moving average (EMA) of the student model. This “mean” teacher then provides stable and consistent pseudo-labels for unlabeled data, which in turn guide the student’s training process. This self-ensembling mechanism has exhibited distinct advantages, especially in contexts requiring adaptation to discrepancies in target domain data distributions and the enhancement of multi-scenario generalization capabilities [,,]. For instance, Kumar et al. [] introduces the stable Mean Teacher, incorporating an error recovery module to refine the quality of pseudo-labels, while Xiong et al. [] utilizes a teacher memory bank for the generation of pseudo-prototypes.
We observe that the systematic application of the Mean Teacher (MT) framework to address the nuanced challenges of violence detection—particularly under notable domain shift—remains underexplored. Unlike other action recognition tasks, violence detection often hinges on subtle visual and temporal cues, and misclassification can lead to far more severe consequences. Our work is the first to pioneer the adaptation of the MT framework specifically for violence detection. By leveraging its well-established stabilization properties, we generate reliable pseudo-labels from unlabeled target domains, which are typically characterized by real-world complexities such as partial occlusions and poor illumination. In doing so, we address a critical gap in the existing literature and advance the robustness of cross-domain violence detection.

2.3. Contrastive Learning (CL)

CL has rapidly emerged as a potent self-supervised learning paradigm, dedicated to acquiring rich and meaningful feature representations from data. It aims to learn representations by maximizing agreement between augmented views of the same sample (positives) and minimizing it with other samples (negatives) [,]. This approach has been successfully applied to various tasks, enabling models to learn domain-invariant features, even in the presence of limited or entirely unlabeled data, by compelling the model to focus on consistent, underlying patterns that transcend superficial domain variations [,,,]. Illustrative examples include work by [], who utilizes videos played at different speeds as a basis for temporal contrastive learning, and [], who augment positive pairs with samples from synchronous viewpoints and further refined the process by using classifier probabilities for hard negative mining. Similarly, Lorre et al. [] employed contrastive predictive coding (CPC) to learn long-term temporal dependencies that are robust to local noise and variations. Currently, methods based on contrastive learning have been widely adopted by scholars for video content analysis. In [], a novel form of multi-scale and cross-scale contrastive learning was instantiated to capture salient semantics between video moments. Dave et al. [] proposed, for the first time, a temporal contrastive loss by designing two types of temporal contrastive losses—local-local and global-local—to distinguish the temporal steps of feature maps in input clips, thereby enhancing the temporal diversity of the learned features and accomplishing video content understanding tasks. Building upon Dave et al. [], Altabrawee and Noor [] utilized sparse local clips with a local-local loss to eliminate the need for the second TCLR temporal loss (global-local loss). These studies collectively demonstrate the advantages of contrastive learning in video content analysis.
Nevertheless, contrastive learning (CL) approaches typically emphasize instance-level discrimination or pretext tasks, which may not adequately promote the semantic separability required for effective violence detection—especially when integrated with pseudo-labeling from a Mean Teacher framework. To address this limitation, our CoMT-VD introduces a dual-strategy contrastive learning (DCL) module, explicitly designed to synergize with the Mean Teacher paradigm. DCL enforces both augmentation consistency and feature-level coherence under the guidance of pseudo-labels, aiming to refine the feature space and enhance the discriminability of violent patterns in the presence of domain shift.

2.4. Domain Adaptation (DA)

DA encompasses a set of techniques designed to mitigate the distributional disparity, or “gap”, between a labeled source domain and an unlabeled or sparsely labeled target domain. Addressing this gap is of paramount importance for the successful and reliable deployment of machine learning models in real-world applications, where data heterogeneity is the norm rather than the exception []. DA methodologies have undergone notable evolution, progressively incorporating principles from adversarial learning, self-supervised learning, augmentation, and generative modeling, all aimed at achieving greater efficiency, robustness, and dynamic adaptability in bridging domain differences [,,,,]. Within the specific context of video-based action recognition, various DA methods have been proposed. For example, approaches like ACAN [] have employed adversarial training objectives to align pixel-wise correlations across domains, while other research [] has conceptualized unsupervised domain adaptation (UDA) as a problem of learning from noisy labels, strategically selecting small-loss samples from the target domain based on assessments of pseudo-label correctness. Sohn et al. [] addressed the challenge of recognizing unlabeled faces in videos based on still-image face recognition by employing an image-to-video feature-level domain adaptation method to learn discriminative video-frame representations. Kim et al. [] utilized multimodal information to align features using video domain adaptation, achieving cross-domain action recognition. Chen et al. [] proposed a Temporal Attention-aware Adversarial Adaptation Network (TA3N) to accomplish domain alignment. Aich et al. [] introduced a novel “Zero-shot Cross-domain Video Anomaly Detection (zxVAD)” framework that learns features by understanding how normal event video features differ from those in pseudo-anomaly examples. Such studies demonstrate the potential of DA techniques to improve cross-domain feature alignment and enhance the reliability of pseudo-labels in the face of disparate data distributions.
Many DA techniques, however, can be complex to implement or may not optimally address the subtle, discriminative cues defining violent acts. Our CoMT-VD framework offers an integrated DA strategy. The Mean Teacher inherently facilitates unsupervised adaptation. We bolster this with strategic cross-domain augmentation (simulating realistic variations like occlusion and low-light) and the DCL module (promoting feature alignment and discriminability). This holistic approach, embedded within the learning framework, aims for robust adaptation and superior generalization in violence detection scenarios affected by domain shifts.

3. Proposed Method

This section presents our proposed approach to violence detection, specifically designed to address key challenges such as domain shift. we introduce the Contrastive Mean Teacher Violence Detection (CoMT-VD) framework, detailing its overall architecture and elaborating on the motivation and design of its core components: the teacher-student paradigm, a dual-strategy contrastive learning mechanism, and the Mean Teacher optimization procedure.

3.1. Problem Definition

Let X denote the input space, representing video segments. Each video segment can be associated with a label from the label space Y = { 0 , 1 } , where y = 1 signifies the presence of violence and  y = 0 signifies its absence. The source domain D s o u , consists of | D s o u | samples and is formally represented as { ( x i s o u , y i s o u ) } i = 1 N s o u , where x i s o u X is the i-th video segment from the source domain and  y i s o u Y is its corresponding ground-truth label. In addition, we have access to a target domain, denoted as D t a r , with  | D t a r | samples, represented as D t a r = { x j t a r } j = 1 N t a r , where x j t a r X is the j-th video segment from the target domain. Crucially, the labels for the target domain data are often unavailable or only sparsely available, posing a notable challenge.
The core problem addressed is that of cross-domain violence detection. We aim to learn a model f : X Y , which generalizes to the target domain, even when there is a domain shift between the source D s o u and target distribution D t a r . The objective is to optimize f to learn domain-invariant features that are also highly discriminative for the violence detection task. Our approach seeks to achieve this by adequately leveraging any available unlabeled target data and by enhancing the model’s capacity to learn from augmented counterparts of the input data, thereby bridging the gap between domains.

3.2. Contrastive Mean Teacher Violence Detection

To achieve this, we propose the Contrastive Mean Teacher Violence Detection (CoMT-VD) model, which synergistically combines semi-supervised learning via the Mean Teacher paradigm with contrastive learning’s robust representation capabilities, further refined using a cross-domain augmentation strategy.

3.2.1. Method Overview

The CoMT-VD model, depicted in Figure 2, leverages a teacher–student framework [], where both teacher and student models, w.r.t. f T , f S , parameterized by θ T and θ S , shares the same network architecture. One critical aspect of our CoMT-VD is the use of cross-domain augmentation: the student model is fed data (from source and target domains) processed by the strong augmenter A s t r o n g , while the teacher model receives data processed by the weak augmenter A w e a k . As shown in Figure 2, labeled source domain D s o u data undergoes the strong augmenter A s t r o n g to { x ¯ k s o u } k = 1 K and is fed to the student network θ S for supervised learning via L sup . Unlabeled target domain D t a r data is processed through the weak augmenter A w e a k to { x ˜ k t a r } k = 1 K for the teacher θ T and the strong augmenter to { x ˜ k t a r } k = 1 K for the student. The teacher, updated via exponential moving average (EMA), provides pseudo-labels from { x ˜ k t a r } k = 1 K to guide θ S through a self-supervised consistency loss L self . The strong-augmented target samples { x ¯ k t a r } k = 1 K are also input to θ S to extract features for the dual-strategy contrastive learning (DCL) module, which applies a contrastive loss L con to learn invariant representations. The framework is optimized with a composite loss with L sup , L self , and  L con , enabling effective cross-domain violence detection.
Figure 2. Overview of the proposed CoMT-VD framework. To enable the model to tackle the domain shift problem, we fine-tuned the pre-trained model using the Mean Teacher framework. At the same time, to ensure that the student network can more effectively leverage the knowledge learned by the teacher network, we introduced a contrastive learning module during the fine-tuning process. Furthermore, to enhance the reliability of the knowledge acquired by the contrastive learning module, we designed a novel DCL contrastive learning approach. The blue line in the figure represents the data flow direction of the self-supervised branch, and the red part shows the data flow direction of the supervised branch.
The student model f S learns by minimizing a supervised loss on strongly augmented labeled source data. Furthermore, a consistency loss is enforced between the student’s predictions on strongly augmented data and the teacher’s predictions on weakly augmented data, applied to samples from both domains. This encourages the student to learn representations invariant to severe perturbations while being guided by more stable targets from the teacher f T . CoMT-VD integrates a dual-strategy contrastive learning (DCL) module that leverages student-extracted features (from strong augmentations) and teacher-extracted features (from weak augmentations) to foster domain-invariant yet discriminative representations. The teacher model updates via exponential moving average (EMA), ensuring training stability and continuous knowledge distillation. This unified framework effectively mitigates cross-domain distribution discrepancies, yielding a violence detection system robust to domain shifts.

3.2.2. Teacher–Student Framework

The student model f S ( · ; θ S ) learns through direct back-propagation using supervised loss of L s u p . In contrast, the parameter of the teacher model θ T are updated as EMA of the student θ T , creating a temporally ensembled, more stable version of the student. The key distinction of CoMT-VD is how data is presented to each model. The student model is trained on aggressively augmented data to learn robust features. In addition, the teacher model, receiving mildly augmented data, provides more consistent and reliable supervisory signals, especially for unlabeled target data and for regularizing the learning process of the student branch.
The student model learns from labeled source data x s o u via a supervised classification loss L s u p (e.g., binary cross-entropy), applied to strongly augmented source samples:
L sup = 1 | D s o u | x t a r D s o u y s o u log g f S ( A s t r o n g ( x s o u ) ) .
Here, x ¯ k , i s o u = A s t r o n g ( x i s o u ) is k-th augmentation from the strong augmenter A s t r o n g of i-th sample the source domain. g is the detection head.
Additionally, a consistency loss L c o n s (e.g., Mean Squared Error) is applied between the predictions of the student on strongly augmented inputs and the teacher on weakly augmented inputs. This loss is computed for all samples x i in a batch, which can include data D t a r :
L self = 1 | D t a r | x t a r D t a r f S A s t r o n g ( x t a r ) f T A w e a k ( x t a r ) 2 2 ,
where B is the batch size. This forces the student to produce consistent predictions, even under strong perturbations, aligning with the teacher’s more stable view of the data.

3.2.3. Cross-Domain Augmentation

Integral to the efficacy and robustness of our CoMT-VD framework is a carefully designed cross-domain augmentation pipeline. Far from being a mere data preprocessing step, it serves as a strategic component engineered to (1) fortify the student network θ S against challenging domain shifts, (2) provide diverse input views essential for dual-strategy contrastive learning to acquire invariant representations, and (3) ensure the stability of supervisory signals generated by the teacher network θ T . To achieve these objectives, we divide the augmentation process into two distinct modalities: a weak augmenter A w e a k , which preserves semantic content to support reliable pseudo-labeling by the teacher, and a strong augmenter A s t r o n g , which introduces substantial visual variation to enhance the student’s resilience and promote distinctive feature learning.
Weak Augmenter: This is denoted as the operator A w e a k . Its core tenet is the generation of K minimally perturbed view { x ˜ k } k = 1 K = A w e a k ( x ) , which retains high semantic fidelity to the original sample. This conservative approach is paramount because x ˜ serves as the input to the teacher network θ T , whose outputs (i.e., pseudo-labels) should be stable and reliable to guide the student’s optimization on unlabeled data. By minimizing aggressive counterparts, A w e a k ensures that the teacher’s predictions are not confounded by augmentation-induced artifacts, thereby fostering a more dependable knowledge transfer. We note that A w e a k concretely includes the following:
  • Minor Occlusion: Small, randomly positioned occluding patches are introduced. Given an input x and a binary mask m τ w defining these minor occlusions [,], the augmented sample is x ˜ = x ( 1 m τ w ) , where m τ w Bernoulli ( τ w ) , and  τ w is set 0.1 in our experiment.
  • Subtle Brightness Adjustment: Pixel intensities are moderately modulated, for example, x ˜ = γ w · x , where the brightness factor γ w is sampled from a narrow interval of [ 0.9 , 1.1 ] .
  • Gentle Frame Blending: Subtle temporal alterations or mild blending with non-violent frames x n v are performed using a minimal blending coefficient λ w = 0.1 , ensuring that the dominant pixel remains unaltered: x ˜ = ( 1 γ w ) x + γ w x n v .
Given the target domain sample x t a r , the corresponding weak-augmented results { x ˜ k t a r } k = 1 K can be obtained:
x ˜ k t a r = A w e a k ( x t a r ) , k = 1 , 2 , , K .
Then, { x ˜ k s o u } k = 1 K will be input into the teacher network θ T , to generate high-quality pseudo-labels, which is fundamental for the student’s robust adaptation to the target domain.
Strong Augmenter. In contrast, the strong augmenter A s t r o n g is designed to generate substantially diversified and challenging views of the input data. It serves two crucial purposes: (1) It is applied to labeled source samples x s o u D s o u to train a student θ S that is robust to various visual perturbations using self-supervised learning in Equation (2). (2) It is applied to unlabeled target samples x t a r D t a r to produce K augmentations. These strong-augmented target data are then processed by the student network θ S and are the primary inputs for DCL module, compelling it to learn features that are invariant to drastic appearance changes, yet retain semantic discriminability. Concretely, A s t r o n g is characterized by significant intensity, broader parameter ranges, as follows:
  • Significant Occlusion: Larger or more strategically disruptive occlusion masks m τ s are employed: x ¯ = x ( 1 m τ s ) , with  m τ s Bernoulli ( τ s ) and τ s = 0.3 .
  • Major Brightness and Contrast Shifts: Pixel intensities and contrast are altered dramatically. For instance, brightness [,] might be scaled by x ¯ = γ s · x , where γ s is sampled from a wider range [ 0.5 , 1.5 ] , and contrast adjustments are similarly intensified to simulate challenging real-world lighting conditions (i.e., very dark or overexposed scenes).
  • Aggressive Frame Blending and Temporal Manipulation: More profound temporal alterations, such as significant frame shuffling [], or aggressive blending [] with disparate scenes x n v (including non-violent content or noise) using a substantial blending factor λ s : x ¯ = ( 1 λ s ) x + λ s x n v , where λ s = 0.4 .
The rationale for the strong augmenter is to construct a challenging learning crucible for θ S . We note that the strong augmenter A s t r o n g receives the samples x t a r and x s o u from both target domain and the source domain, and yields the following conterparts { x ¯ t a r } k = 1 K and { x ¯ s o u } k = 1 K :
x ¯ k t a r = A s t r o n g ( x t a r ) , k = 1 , 2 , , K ,
x ¯ k s o u = A s t r o n g ( x s o u ) , k = 1 , 2 , , K .
By forcing the student to discern invariant characteristics across these radically different views of the same underlying instance (especially for target data in the DCL module) and across diverse source instances, we cultivate feature representations that are not only robust to superficial visual changes, but are also highly discriminative of the core semantic content pertaining to violence. This is indispensable for overcoming domain-specific idiosyncrasies and achieving superior generalization.
This strong-weak augmentation strategy [] establishes a symbiotic interaction within CoMT-VD, effectively addressing domain shift through reliable supervision from the Mean Teacher and DCL’s refinement of discriminative, domain-agnostic features. This synergy forms the foundation of our framework’s enhanced adaptability and performance across diverse violence detection scenarios.

3.2.4. Dual-Strategy Contrastive Learning

To align samples from different distributions, we propose a new contrastive learning mechanism with two key components: Intra-Sample Consistency: Treat a sample and its differently augmented versions as a positive pair, while considering all other samples as negative pairs. Inter-Sample Consistency: To identify positive pairs of the same class from different samples within a batch, we regard samples with similarity scores exceeding δ u p p e r as positive pairs. To better distinguish samples from different classes, samples with similarity scores below δ l o w e r within the same batch are treated as negative pairs. In this Figure, z ^ k i is the feature map obtained after sample x k i undergoes the strong augmenter and is fed into the encoder in the student network; z ¯ k i is the feature map obtained after sample x k i undergoes weak augmentation and is fed into the teacher encoder in the teacher network. Here, K represents the number of augmentations. Contrastive learning enables the unsupervised approach of pulling closer the distance between samples of the same category in the input space while repelling samples from different categories [,]. While this mechanism is commonly applied in few-shot learning, our work draws inspiration from the core principle of contrastive learning: pulling similar samples closer and pushing dissimilar ones apart in the feature space. As illustrated in Figure 2, we propose a dual positive-negative pair matching strategy to compute the NT-Xent loss (Normalized Temperature-scaled Cross Entropy) [,] for both positive and negative sample pairs. We call the module dual-strategy contrastive learning (DCL). The input contains 2 K augmented samples by the weak augmenter and strong augmenter. These samples first undergo positive-negative pair matching via the Pair Matcher in DCL, followed by the computation of the NT-Xent loss using these pairs to achieve cross-domain sample consistency alignment. The Pair Matcher employs two distinct positive-negative pair matching methods to perform consistency alignment:
  • Intra-Sample Consistency: To enable the model to maintain feature invariance of the same sample across different scenarios, we construct positive pairs by pairing the augmented features of a sample with those extracted from other augmentation methods applied to the same sample, while treating features from other samples as negative pairs. This ensures that the model preserves consistency in extracted features across varying environments while enhancing its robustness to noise and perturbations.
  • Inter-Sample Consistency: Normal samples consistency establishes positive pairs by matching a sample with its differently augmented views, while treating samples from other instances as negatives—a classic positive-negative pairing strategy. This encourages the network to preserve feature consistency across diverse transformations and real-world variations. However, relying solely on this approach may under-utilize valuable semantic relationships among distinct instances that share the same class label. To fully exploit the diverse information emerging from different views generated by the same sample within a batch, we design an intra-sample feature pairing strategy.
To address this limitation, we introduce Inter-Sample Consistency, which aligns feature representations of different samples belonging to the same class. This not only improves the utilization of intra-class diversity, but also promotes better alignment of heterogeneous yet semantically similar instances, thereby enhancing the compactness and discriminability of action-level features.
As show in Figure 3, we input K strong augmentations { x ¯ t a r } k = 1 K from target domain D t a r into the DCL, where each sample x ¯ k t a r is a strong augmentation of target domain using A s t r o n g . Also, K weak augmentations { x ˜ t a r } k = 1 K from A w e a k are input. Then, we encode each x ¯ k t a r by student network f S ( · ; θ S ) , and encode each x ˜ k t a r by teacher network f T ( · ; θ T ) to obtain the feature set Z ¯ = { z ¯ } k = 1 K and Z ˜ = { z ˜ } k = 1 K . We note that each feature z ¯ k , z ˜ k will be used to compute the similarity within the corresponding feature set. Taking a strong augmentation sample pair z ¯ m and z ¯ n as an example, the similarity is calculated by cosine similarity:
c o s _ s i m ( z ¯ m , z ¯ n ) = z ¯ m · z ¯ n z ¯ m z ¯ n .
If the c o s _ s i m ( z ¯ m i , z ¯ n j ) > δ u p p e r , we regard z ¯ m and z ¯ n as a positive pair. In contrast, if  c o s _ s i m ( z ¯ m , z ¯ n ) < δ l o w e r , the  z ¯ m and z ¯ n are treated as a negative pair. In our experiments, δ u p p e r = 0.7 is the positive critical threshold and  δ l o w e r = 0.3 is the negative critical threshold. In the dual-strategy contrastive learning module, for a batch of size K, the teacher network generates K weak augmented samples and the student network generates K strong augmented samples in the unsupervised branch. These 2K samples are then paired as positive and negative sample pairs. All other sample features are disregarded, and we set these similarities to 0. For an augmented feature z k z ¯ Z ˜ derived from a sample, its positive pair set is denoted as P ( z k ) . We use the NT-Xent loss [,] to compute the contrastive loss L con :
L con = 1 2 K k = 1 2 K log z p P ( z k ) exp c o s _ s i m ( z k , z p ) / τ z q P ( z k ) exp c o s _ s i m ( z k , z q ) / τ ,
where τ is the temperature scaling factor. We are consistent with the best performance in [], and set τ = 0.1. P ( z k ) represent positive pair sets for the z k , and  z p represents the positive feature for z k in the P ( z k ) , and  z q represents the negative sample of P ( z k ) , which is not in P and have a similarity δ l o w e r < 0.3 with z k .
Figure 3. Illustration of dual-strategy contrastive learning (DCL). To align samples from different distributions, we propose a new contrastive learning mechanism with two key components: Intra-Sample Consistency and Inter-Sample Consistency.

3.2.5. Mean Teacher Optimization

The Mean Teacher architecture enhances semi-supervised learning by leveraging the stability of the teacher model θ T to reduce noise in pseudo-labels []. To ensure that the teacher network acquires richer knowledge while maintaining stability, the weakly augmented samples x ˜ k t a r = A w e a k ( x t a r ) are fed into the teacher branch. This approach enables the model to learn sample features across diverse scenarios while preserving the teacher model’s stability. As a result, the teacher generates more reliable pseudo-labels to guide the parameters updating of the student network θ S . The parameters θ S of student network f S ( · ; θ S ) is updated using gradient descent:
θ S θ S α · θ S L CoMT ,
where α = 0.001 is the learning rate. In addition, L CoMT is the total loss function of the CoMT-VD framework, which is formulated as the combination of three components, i.e., the supervised loss L sup , self-supervised loss L self , and contrastive loss L con in Equation (7), expressed as
L CoMT = λ 1 L sup + λ 2 L self + λ 3 L con ,
where λ 1 = 1 , λ 2 = 0.5 , and  λ 3 = 0.3 are the weights of these three loss terms.
Meanwhile, the parameters θ T of the teacher branch are updated via θ S by Exponential Moving Average (EMA):
θ T η θ T + ( 1 η ) θ S .
Here, η is the momentum factor, which is recommended to set to 0.95 during deployment. The overall training procedure of the proposed CoMT-VD (Contrastive Mean Teacher for Violence Detection) model is outlined in Algorithm 1.
Algorithm 1 Training Procedure of CoMT-VD
  • Require: Labeled source data D s o u = { ( x i s o u , y i s o u ) } ; Unlabeled target data D t a r = { x t a r } ; Pre-trained model f θ ; Output: Optimized teacher f θ T and student model f θ S ;
1. Initialize the student network f θ S f θ and the teacher network f θ T f θ ;
2. For each x i s o u D s o u and x i t a r D t a r  do:
⊳ generate the strong augmentations
3.      { x ¯ k s o u } k = 1 K = A s t r o n g ( x s o u ) ;
4.      { x ¯ k t a r } k = 1 K = A s t r o n g ( x t a r ) ;
▹ generate the weak augmentations
5.      { x ˜ k t a r } k = 1 K = A w e a k ( x t a r ) ;
⊳ compute the feature of the strong augmented samples
6.     { z ¯ } k K = f θ T ( { x ¯ k t a r } k = 1 K ) and { z ˜ } k K = f θ T ( { x ˜ k t a r } k = 1 K ) ;
7.    Calculate L sup using Equation (1);
     L sup = 1 | D s o u | x t a r D s o u y s o u log g f S ( A s t r o n g ( x s o u ) ) ;
8.     Calculate L self via Equation (2);
9.     Calculate L con via Equation (7);
10.   Calculate L CoMT = L sup + L self + L con ;
11.   Update θ S using Equation (8);
12.   Update θ T using Equation (10);
13. End For

4. Experiments

We implement our proposed model based on TensorFlow 2.3.0, and all experiments are deployed on a machine with a 24 GB NVIDIA A10 GPU. In this section, we firstly introduce our experimental datasets, then we will present the experimental setup and results of the model. In this paper, all of the experiments adopted the classic frame sampling method in the violence detection [,,,], sampling 16 frames uniformly from each video. In this paper, all training runs were conducted with the random seed fixed at 42, and the best-performing model was selected after 200 epochs of training. Following this, we conduct a further ablation analysis, and, finally, we analyze our proposed model based on the experimental results.

4.1. Benchmark Datasets

In this paper, we verify the performance of our proposed model on five well-known datasets:
  • RLVS [] is a large-scale real-world captured dataset, containing 1000 violent and 1000 non-violent video clips. The video data in RLVS features more diverse scenarios including streets, classrooms, courtyards, corridors, and sports fields. Most footage is captured from a third-person perspective, predominantly showing fights involving multiple individuals, with a small number of group violence scenes.
  • RWF-2000 [] is designed for real-world violent behavior detection under surveillance cameras. The videos in this dataset are collected from raw surveillance footage on YouTube, segmented into clips of up to 5 s at 30 frames per second, with each clip labeled as either violent or non-violent behavior.
  • Hockey Fight [] comprises 1000 real video clips from ice hockey matches, evenly split between 500 fight scenes and 500 normal segments in the hockey games. However, a notable limitation of this dataset is the lack of scene diversity, as all videos are confined to ice hockey rinks during matches.
  • Movies [] is relatively small-scale, consisting of 200 video clips from action movies, with 100 clips depicting violent scenes and 100 non-violent scenes.
  • Violent Flow [] describes violent or non-violent group behaviors in real-world scenarios. The samples originates from surveillance cameras and monitoring devices, capturing large-scale crowd behaviors in public spaces such as stadiums, streets, and squares. The videos are divided into two categories, violent video and normal video, with a total of 246 video clips (123 violent and 123 non-violent clips).

4.2. Baselines

Our CoMT-VD framework can be integrated with any existing violence detection models. Therefore, to validate the generalization capability of our proposed CoMT-VD, we conduct experiments on 5 different models categorized into three groups based on their architectures: (1) 3DCNN-based: Song et al. [] extends the C3D network proposed in [] (originally designed for human action recognition) to the task of violent behavior recognition. Zhenhua et al. [] proposes full-temporal fusion violence detection model (SCTF). (2) 2DCNN-based: Soliman et al. [] uses 2DCNN to extract spatial features from each frame and employs an LSTM network for temporal feature fusion, called VGG16 + LSTM. Wang et al. [] proposes a plug-and-play module (ActionNet) for human action recognition networks, which is also used to the violence detection task. (3) Transformer-based: Arnab et al. [] introduces a pure Transformer encoder for video understanding, replacing conventional 3D convolutions in action recognition tasks.

4.3. Experimental Setup

To comprehensively evaluate the efficacy of our proposed CoMT-VD framework, we designed two distinct experimental setups, each addressing a critical aspect of real-world violence detection challenges: handling unseen environmental conditions and generalizing to variational datasets. In all experiments, inputs consist of 16 video frames uniformly segmented from the video, with each frame resized to a resolution of 224 × 224 .
(1) Setup-1: Performance under Simulated Variational Conditions: This setup is meticulously designed to assess CoMT-VD’s robustness and generalization abilities when confronted with data corrupted by various challenging, yet common, environmental factors not explicitly seen during standard training. Our objective is to simulate real-world unseen scenarios to validate the performance improvement of CoMT-VD on data exhibiting variational distributions. We introduce four challenging conditions, occlusion, fog, rain, and low-light environments, as shown in Figure 4.
Figure 4. Instances of domain shift under simulated real-world conditions: Occlusions, Rainy Scene, Foggy Scene, and low-illumination environments. These four examples are examples where the baseline model tends to yield detection errors. By combining CoMT VD, all models will obtain more accurate detection results.
  • Rain Simulation: Gaussian noise is strategically injected into the original images, with elongated noise points designed to simulate realistic raindrops. Raindrop parameters are precisely set to a length of 10 pixels and a count of 500 per frame, ensuring a consistent and reproducible simulation of rainy scenes [].
  • Fog Simulation: A fog-like color template [] (fixed at an RGB value of 200) is overlaid onto the original image at 50% intensity. This process emulates the reduced visibility and contrast of foggy conditions.
  • Low-Light Simulation: The brightness of each frame is systematically reduced to 40–70% of its original level. This simulates varying degrees of low-light conditions, a frequent challenge in surveillance and monitoring applications.
  • Occlusion Simulation: To mimic real-world occlusions [,], randomly positioned rectangular masks, with widths and heights sampled from the range of [ 30 , 70 ] pixels, are applied to each frame within the test videos. This simulates partial obstructions that can severely impact visual cues.
For this setup, five baseline models are evaluated across five benchmark datasets. For each dataset, 80% of the data is allocated for training and 20% for testing. Crucially, within our proposed CoMT-VD, the training set is further partitioned, with 20% designated as source domain data and 80% as target domain data. We conduct a rigorous comparative analysis of baseline model both before and after the integration of CoMT-VD to empirically validate its effectiveness in mitigating the impact of these unseen conditions.
(2) Setup-2: Generalization to Variational Datasets: This experimental setup focuses on evaluating the cross-scenario generalization efficacy. Here, the five baselines are initially trained on widely recognized real-world violence detection datasets, namely RLVS [] and RWF-2000 []. Subsequently, their performance is rigorously evaluated on distinct, scenario-specific benchmarks: Hockey Fight [] and Movies []. To ensure comparative validity and a fair assessment, baseline counterparts augmented with our proposed training strategy undergo identical evaluation protocols. This testing paradigm quantitatively assesses the detection accuracy improvements, thereby systematically validating the methodology’s ability to generalize adequately to entirely variational datasets, a critical requirement for real-world applicability.
In all experiments presented in this paper, the input consists of 16 video frames uniformly segmented from the video, with each frame resized to a resolution of 224 × 224 .

4.4. Experimental Results

4.4.1. Performance for Unseen Conditions

In real-world scenarios (such as the video obtained from security cameras without modification by multimedia technology), there are often uncertain dynamic environmental changes (e.g., occlusion, low illumination, rainy, foggy weather, etc.) that involve unseen data not encountered during training, leading to performance degradation of models. This experiment aims to validate the capability of our proposed CoMT in handling such unseen datasets under environmental variations. We train five baseline models on five benchmark datasets, respectively, using 80% of the data for training and 20% for testing. The test data simulates challenging scenarios including rainy, foggy, low-light, and occluded conditions to evaluate the models’ generalization capability on unseen situations potentially encountered in real-world deployments. To validate the effectiveness of the Contrastive Mean Teacher (CoMT), we integrate CoMT into each of the five pre-trained baseline models for fine-tuning, subsequently testing their adaptability and robustness in previously unseen extreme scenarios. Table 1 presents the performance comparisons between the baseline models and those integrated with CoMT-VD across five benchmark datasets for rainy scenarios, foggy scenarios, low-light scenarios, and occlusion scenarios, respectively. It can be observed that the incorporation of CoMT-VD leads to a notable improvement in the model’s performance on unseen scenarios. We conduct the experiment in the settings consistent with Setup-1. The results show that the introduction of CoMT-VD can enhance the generalization capability of the models, improving its detection performance on unseen data. The baseline models achieve improvements ranging from 5.0∼12.0%, 6.0∼12.5%, 5.0∼11.2%, 5.0∼11.2%, and 6.3∼12.3% across the five benchmark datasets in four different domain scenarios, respectively. We note that the results demonstrate that CoMT-VD consistently enhances the capability of various violence detection models for test samples with variational distributions.
Table 1. Volence detection performance (accuracy %) across five dataset under different domain shift scenarios: rainy, foggy, low-light, and occlusion scenes. The numbers in parentheses (blue color) indicate the absolute performance gains (percentage points) achieved by CoMT-VD over vanilla baseline models.

4.4.2. Performance for Cross Datasets

To validate CoMT’s ability to enhance model generalization and improve domain adaptation capability, we designed this experiment to demonstrate that models trained on real-world datasets (RLVS, RWF-2000) show improved domain adaptation and generalization on specific-scenario violence detection datasets (Hockey Fight, Movies). The experimental setup during training remained consistent with Section 4.4.1, using real-world datasets RLVS and RWF individually, as well as a mixed dataset combining RLVS and RWF-2000 for training. Testing was conducted on the Hockey Fight and Movies datasets. The Hockey Fight dataset contains violent behaviors occurring in ice hockey arenas, where all subjects wear protective gear. While the dataset exhibits notable variations in scene configurations and subject appearances, the characteristic patterns of violent behaviors remain consistent. The Movies dataset contains numerous exaggerated violent actions from a movie perspective, which differs from the distribution of daily violent behaviors in the training data. As shown in Table 2 and Table 3, baseline models incorporating the CoMT strategy achieve remarkable performance improvements. The results indicate that these baselines can enhance cross-domain detection capability when trained with CoMT-VD, specifically improving the models trained on real-world data by 5.1∼9.2% on the Hockey Fight test set and 5.0∼12.0% on the test set of the Movies dataset. Experimental results demonstrate that incorporating CoMT adequately enhances the model’s generalization performance in cross-domain detection under such specialized scenarios
Table 2. Performance (accuracy %) of models trained on real-world datasets test on the Hockey Fight dataset under Setup-1. We employ real-world datasets (RLVS and RWF-2000, both of them come from surveillance videos, which include indoor/outdoor fights and daily violent behaviors) as training data and evaluated on the Hockey Fight dataset to validate the model’s generalization capability for cross-domain detection. The numbers in parentheses (blue color) indicate the absolute performance gains (percentage points) achieved by CoMT-VD over vanilla baseline models.
Table 3. Performance of models trained on real-world datasets test on the Movies dataset under Setup-2. We employ real-world datasets (RLVS and RWF-2000, both of them come from surveillance videos, which include indoor/outdoor fights and daily violent behaviors) as training data and evaluated on the Movies dataset to validate the model’s generalization capability for cross-domain detection. The numbers in parentheses (blue color) indicate the absolute performance gains (percentage points) achieved by CoMT-VD over vanilla baseline models.

5. Ablation Studies

Finally, we investigate the impact of different components on the generalization capability of the CoMT-VD model. When we set the brightness γ s = [ 0.5 , 1.5 ]   γ w = [ 0.9 1.1 ] , occlusion τ w = 0.1   τ s = 0.3 , Frame Blending λ s = 0.4   λ w = 0.1 , and similarity threshold δ u p p e r = 0.7   δ l o w e r = 0.3 parameters, we fine-tuned the model to achieve the best performance. In this section, we introduce the impact of different parameter settings and modules on the model.

5.1. Different Augmentation Intensities

This study employs three distinct data augmentation strategies to imbue training data with multi-scenario noise perturbations, thereby enhancing the model’s adaptability to diverse environmental disturbances. The experiments investigate how varying augmentation intensities affect the model’s detection generalization capability: (1) Brightness Augmentation: We optimize brightness augmentation factors γ s and γ w through systematic strength calibration to identify optimal parameter values. (2) Occlusion Augmentation: By controlling occlusion ratios m τ s and m τ w , selecting diverse occlusion box dimensions, we enhance the model’s noise resistance against partial occlusions. The numbers of pixels with occlusion are calculated by m τ s x m τ s x , m τ w x m τ w x , respectively, where · represents rounding down to an integer. Ther weak augmenter randomly samples occlusion masks as integers within (0, [ m τ w x ], while the strong augmentation selects integers from [ m τ w x , m τ s x ]. (3) Temporal Frame Blending: A frame blending rate parameter governs the insertion of adversarial frames into video sequences. Given input dimensions 224 × 224 × 16 , the number of blended frames is calculated as λ n v × 16 , strengthening temporal feature extraction capabilities. In this section, all experiments maintain consistency with Setup-2 in terms of scenario configurations and data usage. During the testing phase, we amalgamate the four scenarios defined in Setup-2 into composite test data for experimental validation.
Impact of Brightness Adjustment Factor. As demonstrated in Table 4, we explored varying random sampling ranges for the brightness augmentation factor. Optimal results are achieved by configuring strong augmentation with a factor randomly sampled from [ 0.5 , 1.5 ] and weak augmentation from [ 0.9 , 1.1 ] . This reveals two key observations: A clear distinction exists between the ranges of strong and weak brightne augmentation factors. Weak augmentation operates within a narrower, more stable range, whereas strong augmentation notably enhances the model’s generalization capability when employing a broader yet appropriately bounded random sampling interval. The empirical observations demonstrate that both excessive or insufficient degradation levels adversely impact model generalization capability. The bold font indicates the best performance achieved by CoMT-VD across all setups.
Table 4. Performance (accuracy %) with different brightness augmentations. The numbers of red color represent the absolute performance degradation (percentage) lossed by CoMT-VD compared to the best performance. The first column specifies different value combinations of illumination γ s (strong augmentation) and γ w (weak augmentation) in the illumination-reduction augmentation.
Impact of Occlusion Augmentation Rate. As shown in Table 5, the model achieves optimal performance when the occlusion ratio is set to 0.3 for strong augmentation and 0.1 for weak augmentation. Excessively high occlusion ratios (i.e., τ s > 0.3 in strong augmentation) may excessively mask critical features during training, impairing feature extraction capabilities. Conversely, overly conservative ratios (e.g., lowering strong augmentation below τ s < 0.3 ) risk oversimplifying learned representations, thereby degrading generalization ability. Crucially, insufficient differentiation between strong and weak augmentation parameters (e.g., 0.2 for weak vs. 0.3 for strong and 0.1 for weak vs. 0.2 ) leads to dual failures: inadequate knowledge acquisition due to overlapping perturbation intensities and diminished pseudo-label reliability caused by weakened data credibility in weakly augmented samples.
Table 5. Performance (accuracy %) with different occlussion augmentation rates in the strong and weak augmenter. The numbers of red color represent the absolute performance degradation (percentage) lossed by CoMT-VD compared to the best performance. This experiment investigates the impact of occlusion size on model generalization capability. The numbers of pixels with occlusion are calculated by m τ s x m τ s x , b r e a k m τ w x m τ w x respectively, where · represents rounding down to an integer.
Impact of Temporal Frame Blending Rate: This experiment investigates the impact of introducing a cross-modal data mixing strategy within the Mean Teacher framework. In Table 6, by comparing the effects of inserting non-violent frames at different ratios in the strong/weak augmentation branches ( λ s = 0.3 for strong augmentation and λ w = 0.1 for weak augmentation), we observe that introducing 30% non-violent frames in the strong augmentation branch achieves optimal validation performance. This demonstrates the strategy’s effectiveness in enhancing the model’s discriminative capability against adversarial frames. The high-disturbance features generated by strong augmentation, combined with the semantic interference from non-violent frames, create an adversarial training paradigm that compels the network to prioritize temporal motion features over individual frame appearances. Meanwhile, maintaining a lower 10% mixing ratio in the weak augmentation branch preserves the reliability of pseudo-labels generated by the teacher network. It can be observed that excessive blending frames rate may lead to the replacement of key feature frames, which could degrade model performance. Insufficient distinction between strong and weak augmentations might result in inadequate knowledge acquisition by the model, while unstable weak augmentations could compromise pseudo-label quality and further impair model performance. In this study, we select the parameter combination of and λ s = 0.4 and λ w = 0.1 .
Table 6. Performance (%) with different frame blending rate. The numbers of red color represent the absolute performance degradation (percentage) lossed by CoMT-VD compared to the best performance. The number of blended frames is calculated as λ n v × 16 , strengthening temporal feature extraction capabilities.

5.2. The Impact of Other Factors on Performance

Impact of Mean Teacher and DCL in CoMT-VD: In this experiment, the encoder employs two baselines: the top-performing SCTF and the classic 3DCNN-based model C3D. We validate the generalization of CoMT-VD on unseen data across five benchmark datasets, specifically the models are tested on the Hockey Fight dataset, to demonstrate CoMT-VD’s capability. Following the experimental setups in Section 4.4.1, we evaluate the impact of CoMT-VD’s components by disabling the Mean Teacher strategy and DCL strategy separately across different baselines. Table 7, Table 8, Table 9 and Table 10 demonstrate that incorporating the Mean Teacher and dual-strategy contrastive learning (DCL) modules can enhance generalization by improving the effectiveness of training samples. While DCL alone—without the Mean Teacher framework—yields moderate gains in generalization, the combination of both components achieves the best performance, highlighting their complementary roles in facilitating cross-domain knowledge transfer. In this experiment, we aggregate unseen data from diverse categories across each dataset into the test set to evaluate the model’s out-of-distribution performance. Figure 5 illustrates the ROC curves of C3D and SCTF in cross-dataset violence detection after training on three source datasets with CoMT-VD. The results show that CoMT-VD improves the Area Under the ROC Curve (AUC), indicating enhanced model performance in balancing true positive and false positive rates when detecting violent behavior.
Table 7. Impact of accuracy by Mean Teacher and DCL for cross-dataset detection. To verify the contributions of both modules, we conduct experiments by selectively disabling each component through setting their corresponding loss terms to zero, thereby validating the effectiveness of both components in CoMT-VD. These experiments are performed by training on three datasets (RLVS, RWF-2000, and RLVS + RWF-2000) and testing on the Hockey Fight dataset, aiming to demonstrate CoMT-VD’s capability in addressing domain shift challenges for cross-dataset detection tasks. The numbers in parentheses (blue color) indicate the absolute performance gains (percentage points) achieved by CoMT-VD over vanilla baseline models. × indicates that the corresponding module has been included, while √ indicates that the corresponding module has not been included.
Table 8. Impact of F1 score by Mean Teacher and DCL for cross-dataset detection. The numbers in parentheses (blue color) indicate the absolute performance gains (percentage points) achieved by CoMT-VD over vanilla baseline models. × indicates that the corresponding module has been included, while √ indicates that the corresponding module has not been included.
Table 9. Validation of the Impact of accuracy by Mean Teacher and DCL on the Models for unseen scenarios dataset violent video detection. We conduct experiments by selectively disabling each component through setting their corresponding loss terms to zero, thereby validating the effectiveness of both components in CoMT-VD. These experiments are performed by training on five benchmark datasets (RLVS, RWF-2000, Hockey Fight, Movies and Violent) respectively, and test the performance on each simulated scenarios of test set to demonstrate CoMT-VD’s capability in addressing domain shift challenges for unseen conditions dataset detection tasks. The numbers in parentheses (blue color) indicate the absolute performance gains (percentage points) achieved by CoMT-VD over vanilla baseline models. × indicates that the corresponding module has been included, while √ indicates that the corresponding module has not been included.
Table 10. Validation of the Impact of F1 score by Mean Teacher and DCL on the Models for unseen scenarios dataset violent video detection. The numbers in parentheses (blue color) indicate the absolute performance gains (percentage points) achieved by CoMT-VD over vanilla baseline models. × indicates that the corresponding module has been included, while √ indicates that the corresponding module has not been included.
Figure 5. ROC curves of C3D and SCTF on cross-dataset voilence detection. The sub-figure (ac) show the changes in ROC curves of C3D and SCTF models tested on the Hockey Fight dataset before and after incorporating the proposed CoMT-VD framework. The results demonstrate that with CoMT-VD, the AUC values are improved for all models.
Impact of the threshold in DCL: Within the Inter-sample consistency in the proposed DCL, our pair-matching mechanism for positive/negative samples considers feature pairs with similarities above threshold δ u p p e r as positive matches (features from different samples but belonging to the same class), and those below threshold δ l o w e r as negative matches (features from different samples with distinct classes). This approach ensures that pseudo-labels generated in unlabeled self-supervised learning better approximate the true class distribution. The selection of these pair-matching thresholds becomes crucial. The results in Table 11 indicate that the optimal performance is achieved when δ u p p e r = 0.7 and δ l o w e r = 0.3. Excessively strict constraints on inter-sample similarity matching can lead to insufficient alignment capability learned from the data, hindering the model’s knowledge acquisition. Conversely, overly lenient constraints may introduce excessive invalid sample pairs, causing the model to learn irrelevant or even counterproductive knowledge.
Table 11. Performance (accuracy %) with different similarity thresholds in DCL. The numbers of red color represent the absolute performance degradation (percentage) lossed by CoMT-VD compared to the best performance. For the proposed DCL, the inter-sample consistency alignment controls the matching of positive and negative pairs through upper-bound δ u p p e r and lower-bound δ l o w e r . From the result, we observe the combination of 0.7 and 0.3 is the optimal values.
Impact of the weights in the loss function. Equation (9) represents the loss function of CoMT-VD. Since labeled data typically provides more reliable optimization for models in supervised settings, we set λ 1 = 1.0 . This experiment aims to explore the effects of varying λ 2 and λ 3 on model performance and identify the optimal combination of λ 2 and λ 3 for CoMT-VD. We integrate CoMT-VD into the pretrained SCTF model and trained it on RLVS, RWF-2000, and their combined dataset (RLVS + RWF-2000). By fine-tuning the model with different combinations and evaluating accuracy on the Hockey Fight and Movies datasets, we select the best parameter configuration via observing the accuracy trend with different λ 2 and λ 3 . Figure 6 demonstrate that the model achieves optimal performance when λ 2 = 0.5 and λ 3 = 0.3 .
Figure 6. Influence of different parameter combinations with loss function for the model. We fix λ 1 at 1.0 and explore the most suitable combination of λ 2 and λ 3 by varying their values within the range [ 0 , 1 ] with a step size of 0.1 . Using the pre-trained SCTF model fine-tuned with CoMT-VD on RLVS (a), RWF-2000 (b), and their mixed dataset (c), and test different weight combinations on Hockey Fight. For clearer visualization, the figure shows the Gaussian-smoothed accuracy variations of fine-tuned models across different weight configurations. The trend reveals that the model achieves peak performance when λ 2 = 0.5 and λ 3 = 0.3 .
Performance of other augmentation: In this ablation study, we evaluate classical image augmentation techniques, rotation and cropping, finding they slightly reduce model performance, as summarized in Table 12. For rotation, we implement 30 (strong) and 15 (weak) angles. For cropping, randomly selected dimensions are used with ratios β = 0.6 (strong) and β = 0.8 (weak). The performance decrease can be attributed to the inherent translation invariance of CNNs, where rotation fails to alter feature relationships adequately. Furthermore, unlike occlusion, cropping inadvertently obscures critical continuous features in sequential data, leading to inferior performance. This suggests that inappropriate augmentation causes models to learn ineffective features, indicating an upper bound to the knowledge gain from augmented data.
Table 12. Validation of the performance (accuracy %) of other augmentations. The numbers in parentheses (red color) represent the absolute performance degradation (percentage) loss by CoMT-VD compared to the best performance sets.

6. Conclusions

6.1. Discussion

In this paper, we address the critical challenge of domain shift in automated violence detection systems, which notably impedes their reliable deployment in diverse real-world surveillance scenarios. We introduce CoMT-VD, a new Contrastive Mean Teacher-based Violence Detection model, specifically engineered to enhance model adaptability and robustness across unseen target domains. The efficacy of CoMT-VD stems from its innovative integration of a Mean Teacher architecture, which adequately leverages unlabeled target domain data to foster stable, domain-invariant feature representations through consistency regularization. Crucially, to counteract the potential supervisory noise from pseudo-labels and further refine the learned feature space, we incorporate a sophisticated dual-strategy contrastive learning module. It meticulously refines features by enforcing both intra-sample consistency for robust representation learning and inter-sample consistency for sharpened categorical discriminability. Broad evaluations on five benchmark datasets unequivocally demonstrate the superior capacity to substantially improve generalization performance on unseen data and across challenging cross-dataset assessments, marking a notable advancement towards more robust and reliable real-world violence detection systems.

6.2. Limitation

This paper verifies the proposed training strategy for enhancing the model’s ability to handle domain shift problems. Due to the limitation of difficulty in collecting violent behavior datasets in specific scenarios, this study employs the addition of various types of noise to general open source datasets for violent behavior detection, simulating common real-world conditions such as low-light, rainy, foggy, and occluded scenes (as illustrated in Figure 4). However, whether the proposed training strategy can maintain consistent performance under corresponding real-world environmental changes requires further validation in subsequent research using datasets from actual specific scenarios. Additionally, this paper focuses on fine-tuning pre-trained models to improve model performance, but it has limitations in detecting continuous video sequences with fewer than 16 frames, necessitating special processing techniques (such as padding video frames or implementing looped playback). Furthermore, according to the ablation experiments, the model has limitations in detecting videos where the number of frames containing violent behavior is extremely low.

6.3. Future Work

For future work, we plan to explore several promising directions, including the integration of more advanced spatiotemporal backbone networks to further enhance feature extraction capabilities, investigating adaptive weighting mechanisms for dual consistency losses or other self-supervised learning forms tailored for adversarial domain shifts, extending CoMT-VD to a multi-modal framework incorporating audio cues for richer contextual understanding, and optimizing CoMT-VD for real-time inference and deployment on edge devices to address computational efficiency.

Author Contributions

Conceptualization, Z.X.; Methodology, Z.X., Z.T. and B.Z.; Validation, Z.X.; Formal analysis, Z.X.; Data curation, Z.X.; Writing—original draft, Z.X.; Writing—review & editing, Z.T. and B.Z.; Funding acquisition, Z.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant No. 61772125; the Fundamental Research Funds for the Central Universities, China under Grants No. N2317004.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data used in this article are all baseline public datasets for violence detection. The sources of each public dataset are introduced in the experimental section.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Clarin, C.; Dionisio, J.; Echavez, M.; Naval, P. DOVE: Detection of movie violence using motion intensity analysis on skin and blood. PCSC 2005, 6, 150–156. [Google Scholar]
  2. De Souza, F.D.; Chavez, G.C.; do Valle Jr, E.A.; Araújo, A.d.A. Violence detection in video using spatio-temporal features. In Proceedings of the 2010 23rd SIBGRAPI Conference on Graphics, Patterns and Images, Gramado, Brazil, 30 August–3 September 2010; pp. 224–230. [Google Scholar]
  3. Khan, H.; Yuan, X.; Qingge, L.; Roy, K. Violence Detection from Industrial Surveillance Videos Using Deep Learning. IEEE Access 2025, 13, 15363–15375. [Google Scholar] [CrossRef]
  4. Sultani, W.; Chen, C.; Shah, M. Real-world anomaly detection in surveillance videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6479–6488. [Google Scholar]
  5. Maqsood, R.; Bajwa, U.I.; Saleem, G.; Raza, R.H.; Anwar, M.W. Anomaly recognition from surveillance videos using 3D convolution neural network. Multimed. Tools Appl. 2021, 80, 18693–18716. [Google Scholar] [CrossRef]
  6. Wang, Z.; She, Q.; Smolic, A. Action-net: Multipath excitation for action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13214–13223. [Google Scholar]
  7. Soliman, M.M.; Kamal, M.H.; Nashed, M.A.E.M.; Mostafa, Y.M.; Chawky, B.S.; Khattab, D. Violence recognition from videos using deep learning techniques. In Proceedings of the 2019 Ninth International Conference on Intelligent Computing and Information Systems (ICICIS), Cairo, Egypt, 8–10 December 2019; pp. 80–85. [Google Scholar]
  8. Pandey, B.; Sinha, U.; Nagwanshi, K.K. A multi-stream framework using spatial–temporal collaboration learning networks for violence and non-violence classification in complex video environments. Int. J. Mach. Learn. Cybern. 2025, 16, 4737–4766. [Google Scholar] [CrossRef]
  9. Sun, S.; Gong, X. Multi-scale bottleneck transformer for weakly supervised multimodal violence detection. In Proceedings of the 2024 IEEE International Conference on Multimedia and Expo (ICME), Niagara Falls, ON, Canada, 15–19 July 2024; pp. 1–6. [Google Scholar]
  10. Rendón-Segador, F.J.; Álvarez-García, J.A.; Salazar-González, J.L.; Tommasi, T. Crimenet: Neural structured learning using vision transformer for violence detection. Neural Netw. 2023, 161, 318–329. [Google Scholar] [CrossRef]
  11. Pan, S.J.; Tsang, I.W.; Kwok, J.T.; Yang, Q. Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 2010, 22, 199–210. [Google Scholar] [CrossRef]
  12. Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning spatiotemporal features with 3D convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4489–4497. [Google Scholar]
  13. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  14. Xu, Y.; Cao, H.; Mao, K.; Chen, Z.; Xie, L.; Yang, J. Aligning correlation information for domain adaptation in action recognition. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 6767–6778. [Google Scholar] [CrossRef]
  15. Da Costa, V.G.T.; Zara, G.; Rota, P.; Oliveira-Santos, T.; Sebe, N.; Murino, V.; Ricci, E. Dual-head contrastive domain adaptation for video action recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 1181–1190. [Google Scholar]
  16. Li, J.; Xu, R.; Liu, X.; Ma, J.; Li, B.; Zou, Q.; Ma, J.; Yu, H. Domain adaptation based object detection for autonomous driving in foggy and rainy weather. arXiv 2023, arXiv:2307.09676. [Google Scholar] [CrossRef]
  17. Dasgupta, A.; Jawahar, C.; Alahari, K. Source-free video domain adaptation by learning from noisy labels. Pattern Recognit. 2025, 161, 111328. [Google Scholar] [CrossRef]
  18. Xu, Y.; Cao, H.; Xie, L.; Li, X.L.; Chen, Z.; Yang, J. Video unsupervised domain adaptation with deep learning: A comprehensive survey. ACM Comput. Surv. 2024, 56, 1–36. [Google Scholar] [CrossRef]
  19. Gao, Z.; Zhao, Y.; Zhang, H.; Chen, D.; Liu, A.A.; Chen, S. A novel multiple-view adversarial learning network for unsupervised domain adaptation action recognition. IEEE Trans. Cybern. 2021, 52, 13197–13211. [Google Scholar] [CrossRef]
  20. Huang, S.W.; Lin, C.T.; Chen, S.P.; Wu, Y.Y.; Hsu, P.H.; Lai, S.H. Auggan: Cross domain adaptation with gan-based data augmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 718–731. [Google Scholar]
  21. Choi, J.; Sharma, G.; Chandraker, M.; Huang, J.B. Unsupervised and semi-supervised domain adaptation for action recognition from drones. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 1717–1726. [Google Scholar]
  22. Cen, F.; Zhao, X.; Li, W.; Wang, G. Deep feature augmentation for occluded image classification. Pattern Recognit. 2021, 111, 107737. [Google Scholar] [CrossRef]
  23. Wang, X.; Zhang, R.; Shen, C.; Kong, T.; Li, L. Dense contrastive learning for self-supervised visual pre-training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 3024–3033. [Google Scholar]
  24. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  25. Guo, Y.; Ma, S.; Su, H.; Wang, Z.; Zhao, Y.; Zou, W.; Sun, S.; Zheng, Y. Dual mean-teacher: An unbiased semi-supervised framework for audio-visual source localization. Adv. Neural Inf. Process. Syst. 2023, 36, 48639–48661. [Google Scholar]
  26. Mahmoodi, J.; Nezamabadi-pour, H. A spatio-temporal model for violence detection based on spatial and temporal attention modules and 2D CNNs. Pattern Anal. Appl. 2024, 27, 46. [Google Scholar] [CrossRef]
  27. Mahmoodi, J.; Nezamabadi-Pour, H. Violence Detection in Video Using Statistical Features of the Optical Flow and 2D Convolutional Neural Network. Comput. Intell. 2025, 41, e70034. [Google Scholar] [CrossRef]
  28. Buciluǎ, C.; Caruana, R.; Niculescu-Mizil, A. Model compression. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Philadelphia, PA, USA, 20–23 August 2006; pp. 535–541. [Google Scholar]
  29. Wu, M.C.; Chiu, C.T.; Wu, K.H. Multi-teacher knowledge distillation for compressed video action recognition on deep neural networks. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 2202–2206. [Google Scholar]
  30. Kumar, A.; Mitra, S.; Rawat, Y.S. Stable mean teacher for semi-supervised video action detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 4419–4427. [Google Scholar]
  31. Wang, X.; Hu, J.F.; Lai, J.H.; Zhang, J.; Zheng, W.S. Progressive teacher-student learning for early action prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3556–3565. [Google Scholar]
  32. Xiong, B.; Yang, X.; Song, Y.; Wang, Y.; Xu, C. Modality-Collaborative Test-Time Adaptation for Action Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 26732–26741. [Google Scholar]
  33. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9729–9738. [Google Scholar]
  34. Singh, A.; Chakraborty, O.; Varshney, A.; Panda, R.; Feris, R.; Saenko, K.; Das, A. Semi-supervised action recognition with temporal contrastive learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10389–10399. [Google Scholar]
  35. Shah, K.; Shah, A.; Lau, C.P.; de Melo, C.M.; Chellappa, R. Multi-view action recognition using contrastive learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 3381–3391. [Google Scholar]
  36. Lorre, G.; Rabarisoa, J.; Orcesi, A.; Ainouz, S.; Canu, S. Temporal contrastive pretraining for video action recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 662–670. [Google Scholar]
  37. Zheng, S.; Chen, S.; Jin, Q. Few-shot action recognition with hierarchical matching and contrastive learning. In Computer Vision—ECCV 2022, Proceedings of the 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 297–313. [Google Scholar]
  38. Nguyen, T.T.; Bin, Y.; Wu, X.; Hu, Z.; Nguyen, C.D.T.; Ng, S.K.; Luu, A.T. Multi-scale contrastive learning for video temporal grounding. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 6227–6235. [Google Scholar]
  39. Dave, I.; Gupta, R.; Rizve, M.N.; Shah, M. Tclr: Temporal contrastive learning for video representation. Comput. Vis. Image Underst. 2022, 219, 103406. [Google Scholar] [CrossRef]
  40. Altabrawee, H.; Noor, M.H.M. STCLR: Sparse Temporal Contrastive Learning for Video Representation. Neurocomputing 2025, 630, 129694. [Google Scholar] [CrossRef]
  41. Sohn, K.; Liu, S.; Zhong, G.; Yu, X.; Yang, M.H.; Chandraker, M. Unsupervised domain adaptation for face recognition in unlabeled videos. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3210–3218. [Google Scholar]
  42. Kim, D.; Tsai, Y.H.; Zhuang, B.; Yu, X.; Sclaroff, S.; Saenko, K.; Chandraker, M. Learning cross-modal contrastive features for video domain adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 13618–13627. [Google Scholar]
  43. Chen, M.H.; Kira, Z.; AlRegib, G.; Yoo, J.; Chen, R.; Zheng, J. Temporal attentive alignment for large-scale video domain adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6321–6330. [Google Scholar]
  44. Aich, A.; Peng, K.C.; Roy-Chowdhury, A.K. Cross-domain video anomaly detection without target domain adaptation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 2579–2591. [Google Scholar]
  45. Valois, P.H.V.; Niinuma, K.; Fukui, K. Occlusion Sensitivity Analysis With Augmentation Subspace Perturbation in Deep Feature Space. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 1–6 January 2024; pp. 4829–4838. [Google Scholar]
  46. Wang, Z.; Jiang, J.x.; Zeng, S.; Zhou, L.; Li, Y.; Wang, Z. Multi-receptive field feature disentanglement with Distance-Aware Gaussian Brightness Augmentation for single-source domain generalization in medical image segmentation. Neurocomputing 2025, 638, 130120. [Google Scholar] [CrossRef]
  47. Kandel, I.; Castelli, M.; Manzoni, L. Brightness as an augmentation technique for image classification. Emerg. Sci. J. 2022, 6, 881–892. [Google Scholar] [CrossRef]
  48. Choi, J.; Sharma, G.; Schulter, S.; Huang, J.B. Shuffle and attend: Video domain adaptation. In Computer Vision–ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XII 16; Springer: Cham, Switzerland, 2020; pp. 678–695. [Google Scholar]
  49. Sahoo, A.; Shah, R.; Panda, R.; Saenko, K.; Das, A. Contrast and mix: Temporal contrastive video domain adaptation with background mixing. Adv. Neural Inf. Process. Syst. 2021, 34, 23386–23400. [Google Scholar]
  50. Yuan, J.; Liu, Y.; Shen, C.; Wang, Z.; Li, H. A simple baseline for semi-supervised semantic segmentation with strong data augmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 8229–8238. [Google Scholar]
  51. Sohn, K. Improved deep metric learning with multi-class n-pair loss objective. In Proceedings of the Advances in Neural Information Processing Systems 29 (NIPS 2016), Barcelona, Spain, 5–10 December 2016; Volume 29. [Google Scholar]
  52. Cheng, M.; Cai, K.; Li, M. RWF-2000: An open large scale video database for violence detection. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 4183–4190. [Google Scholar]
  53. Bermejo Nievas, E.; Deniz Suarez, O.; Bueno García, G.; Sukthankar, R. Violence detection in video using computer vision techniques. In Computer Analysis of Images and Patterns, Proceedings of the 14th International Conference, CAIP 2011, Seville, Spain, 29–31 August 2011; Proceedings, Part II 14; Springer: Cham, Switzerland, 2011; pp. 332–339. [Google Scholar]
  54. Hassner, T.; Itcher, Y.; Kliper-Gross, O. Violent flows: Real-time detection of violent crowd behavior. In Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012; pp. 1–6. [Google Scholar]
  55. Song, W.; Zhang, D.; Zhao, X.; Yu, J.; Zheng, R.; Wang, A. A novel violent video detection scheme based on modified 3D convolutional neural networks. IEEE Access 2019, 7, 39172–39179. [Google Scholar] [CrossRef]
  56. Tan, Z.; Xia, Z.; Wang, P.; Wu, D.; Li, L. SCTF: An efficient neural network based on local spatial compression and full temporal fusion for video violence detection. Multimed. Tools Appl. 2024, 83, 36899–36919. [Google Scholar]
  57. Arnab, A.; Dehghani, M.; Heigold, G.; Sun, C.; Lučić, M.; Schmid, C. Vivit: A video vision transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 6836–6846. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.