Next Article in Journal
Current Trends and Future Opportunities of AI-Based Analysis in Mesenchymal Stem Cell Imaging: A Scoping Review
Previous Article in Journal
Preclinical Application of Computer-Aided High-Frequency Ultrasound (HFUS) Imaging: A Preliminary Report on the In Vivo Characterization of Hepatic Steatosis Progression in Mouse Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Federated Self-Supervised Few-Shot Face Recognition

by
Nursultan Makhanov
,
Beibut Amirgaliyev
*,
Talgat Islamgozhayev
and
Didar Yedilkhan
Smart City Research Center, Astana IT University, Astana 010000, Kazakhstan
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(10), 370; https://doi.org/10.3390/jimaging11100370
Submission received: 11 September 2025 / Revised: 8 October 2025 / Accepted: 16 October 2025 / Published: 18 October 2025
(This article belongs to the Section Computer Vision and Pattern Recognition)

Abstract

This paper presents a systematic framework that combines federated learning, self-supervised learning, and few-shot learning paradigms for privacy-preserving face recognition. We use the large-scale CASIA-WebFace dataset for self-supervised pre-training using SimCLR in a federated setting, followed by federated few-shot fine-tuning on the LFW dataset using prototypical networks. Through comprehensive evaluation across six state-of-the-art architectures (ResNet, DenseNet, MobileViT, ViT-Small, CvT, and CoAtNet), we demonstrate that while our federated approach successfully preserves data privacy, it comes with significant performance trade-offs. Our results show 12–30% accuracy degradation compared to centralized methods, representing the substantial cost of privacy preservation. We find that traditional CNNs show superior robustness to federated constraints compared to transformer-based architectures, and that five-shot configurations provide an optimal balance between data efficiency and performance. This work provides important empirical insights and establishes benchmarks for federated few-shot face recognition, quantifying the privacy–utility trade-offs that practitioners must consider when deploying such systems in real-world applications.

1. Introduction

Face recognition has emerged as one of the most prominent applications of computer vision and biometric identification systems, experiencing a remarkable evolution from traditional handcrafted feature-based methods to sophisticated deep learning architectures [1,2]. The field has witnessed unprecedented progress with the advent of deep convolutional neural networks (CNNs), particularly ResNet [3]; DenseNet [4]; and, more recently, Vision Transformers [5,6], which have significantly improved recognition accuracy across diverse datasets and challenging real-world scenarios [7,8].
Despite these remarkable achievements, contemporary face recognition systems face several critical challenges that limit their practical deployment and widespread adoption. First, the centralized nature of traditional deep learning approaches requires the collection and aggregation of massive amounts of sensitive biometric data, raising substantial privacy concerns and regulatory compliance issues, particularly in light of stringent data protection regulations such as GDPR and CCPA [9,10]. Second, the computational and communication overhead associated with the training of large-scale models on centralized servers creates bottlenecks in distributed environments where data cannot be easily transferred due to bandwidth limitations or security constraints [11]. Third, the dependency on extensive manually annotated datasets for supervised learning poses significant challenges in terms of annotation costs, time requirements, and scalability to new domains or identities [12].
To address these fundamental limitations, three complementary learning paradigms have emerged as promising solutions. Federated Learning (FL) enables collaborative model training across distributed clients without requiring centralized data collection, thereby preserving data privacy while leveraging the collective knowledge of multiple participants [9,11]. Self-Supervised Learning (SSL) has demonstrated a remarkable capability in learning robust visual representations from unlabeled data by exploiting inherent data structures and relationships, reducing the dependency on manual annotations [13,14,15]. Few-Shot Learning (FSL) addresses the challenge of limited labeled data by allowing models to generalize to new classes or identities with minimal training examples, making it particularly suitable for scenarios where obtaining extensive labeled datasets is impractical [16,17].
The integration of these three paradigms presents a compelling opportunity to create a unified framework that simultaneously addresses privacy preservation, annotation efficiency, and communication optimization in face recognition systems. While federated learning ensures data privacy and enables distributed training, self-supervised learning reduces the reliance on labeled data during the pre-training phase, and few-shot learning enables efficient adaptation to new identities with minimal supervision. This synergistic combination is particularly relevant for real-world applications such as surveillance systems, access control, and mobile authentication, where data privacy, computational efficiency, and rapid deployment are paramount concerns.
Despite the individual success of federated learning [10,11], self-supervised learning [14,15,18], and few-shot learning [19,20,21] in face recognition, their integrated potential remains largely unexplored in the literature. Previous work has primarily focused on addressing these challenges in isolation, failing to take advantage of the complementary strengths of these paradigms. For example, while Kim et al. [11] proposed federated learning for face recognition via self-supervised intrasubject learning, their approach did not incorporate few-shot learning capabilities for efficient adaptation to new identities. Similarly, existing few-shot face recognition methods [20,22] typically assume centralized training scenarios and do not address concerns about privacy preservation or communication efficiency inherent in federated environments.
The integration of these three paradigms introduces several technical challenges that require careful consideration. First, the heterogeneity of data distributions across federated clients can significantly impact the effectiveness of self-supervised pre-training, as the learned representations may not generalize well across different client populations. Second, the communication overhead in federated settings becomes more pronounced when dealing with large-scale self-supervised models, necessitating efficient compression and aggregation strategies. Third, the few-shot learning component must be designed to work effectively with representations learned through federated self-supervised pre-training, which may exhibit different characteristics compared to centrally trained models.
To address these challenges, this paper proposes a novel unified framework that seamlessly integrates federated learning, self-supervised learning, and few-shot learning for privacy-preserving face recognition. Our approach consists of two main phases: (1) a self-supervised pre-training phase using the large-scale CASIA-WebFace dataset [23] to learn robust face representations without manual annotations and (2) a federated few-shot learning phase that efficiently adapts the pre-trained model to downstream face recognition tasks on the LFW dataset [24] with minimal labeled examples while maintaining privacy constraints.
We develop a systematic framework that combines three established paradigms as federated learning, self-supervised learning, and few-shot learning for face recognition. While the integration is primarily sequential rather than deeply synergistic, our comprehensive evaluation across multiple architectures provides valuable insights into the practical deployment of such systems. We acknowledge that this represents an engineering contribution rather than a fundamental algorithmic breakthrough, as we use standard SimCLR for self-supervised pre-training, standard FedAvg for federated learning, and standard prototypical networks for few-shot learning without deep algorithmic innovations.
Through extensive experimental evaluation across six state-of-the-art architectures, we demonstrate that while our federated approach preserves data privacy, it comes with significant performance trade-offs. Our results show 12–30% accuracy degradation compared to centralized methods, representing the substantial cost of privacy preservation. This significant performance degradation raises important questions about the practical viability of such approaches and represents a fundamental limitation that practitioners must carefully consider when deploying federated face recognition systems.
The main contributions of this work can be summarized as follows:
  • Comprehensive empirical analysis: We provide the first systematic evaluation of the combination of federated learning, self-supervised learning, and few-shot learning for face recognition across six state-of-the-art architectures, establishing benchmarks for future research.
  • Architectural suitability insights: We demonstrate that traditional CNNs (ResNet and DenseNet) show superior robustness to federated constraints compared to transformer-based architectures, with implications for practical deployment decisions.
  • Privacy–utility trade-off quantification: We establish benchmarks showing 12–30% accuracy degradation as the cost of privacy preservation, providing guidance for practical deployment decisions and highlighting the substantial limitations of current federated approaches.
  • Engineering framework: We develop a working system that successfully integrates three complementary paradigms, serving as a foundation for future algorithmic innovations in federated few-shot learning, while acknowledging the primarily sequential nature of this integration.
The remainder of this paper is organized as follows. Section 2 reviews related work on federated learning, self-supervised learning, and a few-shot learning method for face recognition. Section 3 formally defines the problem and establishes the mathematical framework. Section 4 presents our proposed unified approach with detailed algorithmic descriptions. Section 5 describes the experimental setup and datasets used for evaluation. Section 6 presents comprehensive experimental results and analysis. Finally, Section 7 concludes the paper and discusses future research directions.

2. Related Works

2.1. Face Recognition

Face recognition has evolved significantly from traditional methods to deep learning approaches. Early techniques, such as geometric feature analysis and texture-based methods (e.g., local binary patterns), relied on manually engineered features for tasks such as detection, extraction, and verification [1,2]. These methods struggled with environmental variations such as lighting changes and occlusions. The advent of deep learning, particularly convolutional neural networks (CNNs), revolutionized the field by automating feature extraction, achieving state-of-the-art accuracy through models such as VGG-Face, FaceNet, and ArcFace [7,8,12]. Hybrid approaches that combine classical and deep learning methods further improved robustness.
Recent advancements focus on addressing scalability and environmental challenges. Federated learning (FL) frameworks such as FedFS enable decentralized training of personalized face recognition models without centralized data aggregation, enhancing privacy through intra-subject self-supervised learning [9,11]. Techniques utilizing cosine similarity and adaptive soft labels have improved the robustness to pose variations and occlusions [12]. In addition, generative adversarial networks (GANs) are being used to synthesize training data in FL settings, mitigating data scarcity while preserving privacy [10].
Limitations persist in three key areas:
  • Federated Learning: Communication overhead and heterogeneity across edge devices hinder model convergence, while secure aggregation methods remain computationally intensive.
  • Few-Shot Learning: Deep learning models require large datasets, limiting their effectiveness in scenarios with minimal training samples.
  • Self-Supervised Learning: While frameworks like FedFS reduce dependency on labeled data, integrating self-supervised objectives with FL introduces complexity in maintaining consistent representation spaces across clients. Furthermore, ethical concerns around bias in training data and high computational demands underscore the need for lightweight and fairer architectures.

2.2. Self-Supervised Learning

Self-supervised learning (SSL) has emerged as a transformative paradigm for the learning of robust representations without manual annotations, leveraging intrinsic data structures to generate supervisory signals. Early SSL approaches focused on pretext tasks like image rotation prediction [15,25] and contrastive learning frameworks that maximize similarity between augmented views of data while repelling dissimilar pairs [26,27]. Cross-modal methods, such as cyclic translation and joint image–text pre-training, further expanded SSL’s scope by aligning representations across modalities. These techniques reduce reliance on labeled datasets, making SSL particularly valuable in domains with a scarcity of annotations.
Recent advances in SSL emphasize multimodal and domain-specific innovations. Contrastive frameworks like VG-SSL integrate geo-related image pairs (GeoPair) to enhance visual geo-localization accuracy, achieving performance comparable to that of supervised methods [27]. In medical imaging, SSL optimizes metrics such as the area under the curve (AUC) to balance sensitivity and specificity, which are critical for diagnostic reliability [13]. Multimodal SSL models such as those that combine TimeSformer visual backbones with language models demonstrate superior video captioning performance when trained jointly on images and videos [28]. These advancements highlight SSL’s adaptability to diverse data types and tasks.
Face recognition applications include domain adaptation and keypoint detection. SSL mitigates domain changes by aligning source- and target-domain embeddings through maximization of similarity, improving recognition accuracy under dataset biases [14]. For facial landmark detection in thermal imaging, SSL pretraining with rotation prediction and keypoint regression tasks achieves high precision despite limited labeled data [15]. Semi-supervised frameworks such as self-training LDA integrate affinity propagation to iteratively refine face subspaces using unlabeled samples, outperforming traditional supervised methods in low-data regimes. These applications underscore SSL’s ability to address data scarcity and generalization challenges in face recognition systems [18,29].

2.3. Few-Shot Learning

Few-shot learning (FSL) addresses data scarcity by leveraging prior knowledge to rapidly adapt models to new tasks with minimal labeled examples. Traditional approaches relied on hand-crafted features and shallow architectures, but deep learning frameworks now dominate through metric-based, optimization-based, and model-based meta-learning paradigms [16,17].
Meta-learning approaches focus on task generalization through episodic training.
Metric-based methods such as prototypical networks [19] classify samples by computing distances to class prototypes in learned embedding spaces, while matching networks [30] employ attention mechanisms for similarity scoring.
Optimization-based methods (e.g., MAML and Reptile [30]) learn model initialization parameters that enable fast adaptation to new tasks via gradient updates. The Reptile method shows superior accuracy (F1: 0.92) compared to MAML (F1: 0.87) in face recognition tasks [30].
Model-based methods use memory-augmented networks to store and retrieve task-specific information dynamically, enhancing adaptability [17,31].
Applications to face recognition demonstrate FSL’s viability in low-data regimes.
Siamese Networks with contrastive loss achieve 90% accuracy on five-way one-shot tasks by learning discriminative embeddings from single examples [20,21]. Hybrid frameworks combining deep features (e.g., Dlib) with transfer learning reduce dependency on labeled data, achieving 84% accuracy in 50-way face verification [22]. MetaFake [32] detects unseen face forgery techniques with 89.3% accuracy using task-specific prototypes and local feature aggregation, outperforming conventional supervised methods by 12%. Challenges persist in handling intra-class variation and scaling to large identity sets. Although methods like squeeze-and-excitation networks [21] enhance feature sensitivity, computational costs for secure aggregation in federated FSL remain prohibitive.

2.4. Combined Approaches

Combined approaches to face recognition seek to take advantage of the strengths of different paradigms to overcome individual limitations, enhancing robustness, accuracy, and applicability in diverse scenarios. These strategies often integrate traditional methods with deep learning; fuse multiple feature extraction techniques; or blend different learning paradigms like federated, few-shot, and self-supervised learning. For example, early methods combined geometric and texture-based features to improve recognition accuracy under varying lighting conditions. More recently, hybrid approaches utilizing deep neural networks for feature extraction combined with machine learning classifiers such as support vector machines (SVMs) have shown promise [33].
In the context of federated learning with self-supervised objectives, authors in [34] proposed Calibre, a personalized federated learning framework designed to achieve both fairness and accuracy across clients. Calibre method introduces a contrastive prototype adaptation mechanism that calibrates SSL representations by maintaining a balance between generic and client-specific information. The framework addresses the challenge of fuzzy class boundaries in SSL representations under non-IID data distributions through client-specific prototype regularizers and an aggregation algorithm guided by prototypes across clients. Theoretical analysis based on information theory formalizes the generality–personalization tradeoff, providing foundations for design decisions. Extensive experiments on the CIFAR-10, CIFAR-100, and STL-10 datasets demonstrate that Calibre achieves state-of-the-art performance in both mean accuracy and fairness across various non-IID settings.
Current integrated approaches focus on feature fusion, network optimization, and multi-modal data integration. Feature fusion combines hand-crafted features (e.g., LBP and HOG) with deep-learned features to capture complementary information, often improving performance compared to using either feature set alone [33,35]. However, these approaches typically assume centralized data availability and do not address the challenges of decentralized learning environments. For network optimization, Binary Particle Swarm Optimization (BPSO) can refine neural network hidden layers to extract relevant features, improving performance on datasets like FERET and LFW [36]. While Calibre demonstrates effective integration of SSL and FL, it does not explicitly address the few-shot learning scenario, which is critical for face recognition with limited labeled data per identity.
Limitations of current integrated approaches include the following:
  • Computational Complexity: The combination of multiple features and the training of deep networks increases computational demands, limiting real-time applications and deployment on resource-constrained devices.
  • Data Dependency: Although some methods reduce the dependency on labeled data, performance still depends on the quality and quantity of training data, especially in federated and few-shot learning scenarios.
  • Generalization Issues: Models trained on specific datasets may struggle to generalize to new environments or demographics, necessitating domain adaptation techniques or continuous learning strategies.
Our work addresses these limitations by introducing a framework that integrates federated learning, few-shot learning, and self-supervised learning to improve face recognition in privacy-sensitive and data-scarce environments. Building upon approaches like Calibre, we extend the paradigm of prototype-based learning to the federated few-shot setting, specifically addressing the unique challenges of face recognition, including identity preservation across clients, the handling of facial variations in unconstrained environments, and constructing meaningful few-shot episodes from disjoint identity distributions. By leveraging intra-subject self-supervision within a federated setting, we aim to reduce the reliance on labeled data while preserving user privacy. Furthermore, we address the challenge of heterogeneity across edge devices by incorporating lightweight architectures and efficient communication protocols. This approach positions our work at the intersection of multiple cutting-edge paradigms, offering a practical and scalable solution for real-world face recognition applications.

3. Problem Formulation

In this section, we formally define the federated face recognition problem and establish the mathematical framework for our approach. We consider a scenario where multiple clients possess private face datasets and aim to collaboratively train a global face recognition model without sharing their raw data.

3.1. Notations and Definitions

Let C = { C 1 , C 2 , , C N } denote a set of N clients participating in the federated learning system and S represent the central server. Each client ( C i ) has a local dataset ( D i = { ( x i , j , y i , j ) } j = 1 | D i | , where x i , j R H × W × 3 represents a face image and y i , j Y i is the corresponding identity label). Let D global = i = 1 N D i denote the global dataset composed of all client datasets, with a total size of | D global | = i = 1 N | D i | . Each client ( C i ) maintains its local dataset ( D i ) with | D i | samples.
We define the global model parameters as θ R d , where d is the parameter dimension. The face recognition model consists of a feature extractor ( f θ : R H × W × 3 R d f ) that maps input images to d f -dimensional feature representations and a classifier head ( g ϕ : R d f R | Y | ) for identity classification, where | Y | is the total number of identities across all clients.
For the few-shot learning setting, we define a support set ( S = { ( x s ( n ) , n ) } n = 1 N ) containing K labeled examples per class for N classes and a query set ( Q = { ( x q ( m ) , y q ( m ) ) } m = 1 M ) containing M test examples. The few-shot learning task is formulated as an N-way K-shot classification problem.
Federated Learning Problem: Given N clients with private datasets ( { D i } i = 1 N ), the objective is to learn a global model ( θ * ) that minimizes federated loss:
θ * = arg min θ L f e d ( θ ) = i = 1 N | D i | | D | L i ( θ )
where L i ( θ ) = 1 | D i | ( x , y ) D i ( θ ; x , y ) is the local loss function in the client (i) and ( θ ; x , y ) is the loss for a single data point.
Self-Supervised Learning Objective: For the pre-training phase, we employ contrastive learning to learn meaningful face representations without labels. Given an input image (x), we generate two augmented views ( x 1 = t 1 ( x ) and x 2 = t 2 ( x ) ) using transformations ( t 1 , t 2 T ). The contrastive loss for a positive pair ( i , j ) is expressed as follows:
L s s l ( i , j ) = log exp ( sim ( z i , z j ) / τ ) k = 1 2 B 1 ( k i ) exp ( sim ( z i , z k ) / τ )
where z i = g ( f θ ( x i ) ) is the projected representation, sim ( u , v ) = u T v / ( u v ) is the cosine similarity, τ is the temperature parameter, and B is the batch size.
Few-Shot Learning Task: Given a support set ( S ) and a query set ( Q ), the objective of learning of a few shots is to minimize the classification error on the query set. Using prototypical networks, the class prototype for class n is computed as follows: c n = 1 | S n | ( x i , y i ) S n f θ ( x i ) , where S n = { ( x i , y i ) S y i = n } is the support set for class n. The classification probability for a query sample ( x q ) is expressed as follows:
p ( y = n | x q ) = exp ( d ( f θ ( x q ) , c n ) ) n = 1 N exp ( d ( f θ ( x q ) , c n ) )
where d ( · , · ) is the Euclidean distance function.

3.2. System Model

Our federated face recognition system operates under the following architectural framework and constraints:
The system follows a star topology, where the central server (S) coordinates the training process among N distributed clients. Each client ( C i ) maintains its local dataset ( D i ) and computational resources for local model training. The server aggregates model updates from selected clients and broadcasts the updated global model.
Training Protocol: The training process consists of three phases:
  • Self-Supervised Pre-training: We pre-train a feature extractor using unlabeled face images through contrastive learning.
  • Federated Few-Shot Fine-tuning: The pre-trained model is fine-tuned using limited labeled samples in a federated few-shot learning framework.
  • Evaluation: The final model is evaluated on the following face recognition tasks.
Communication Model: In each communication round (t), the server selects a subset of clients ( S t C ) based on availability and resource constraints. Selected clients download the current global model parameters ( θ t ), perform local training for E epochs, and upload their model updates ( Δ θ i t = θ i t + 1 θ t ) to the server. The server then aggregates these updates using weighted averaging:
θ t + 1 = θ t + i S t | D i | j S t | D j | Δ θ i t .
The system enforces strict privacy preservation requirements.
  • Data Locality: Raw face images never leave the client devices, ensuring that D i remains private to the client ( C i ).
  • Model Privacy: Only model parameters or gradients are shared, not intermediate activations or embeddings that could leak sensitive information.
The system addresses multiple forms of heterogeneity:
  • Statistical Heterogeneity: Clients may have non-IID data distributions with varying identity compositions and demographic characteristics.
  • Temporal Heterogeneity: Clients may join or leave the system dynamically, and their availability patterns can vary.
Under standard assumptions of bounded gradients and Lipschitz continuity, the federated learning algorithm converges to a stationary point of the global objective function. The convergence rate depends on the degree of data heterogeneity, communication frequency, and local update steps.
This problem formulation establishes the theoretical foundation for our federated self-supervised few-shot face recognition approach, addressing the key challenges of privacy preservation, data heterogeneity, and limited labeled data in distributed face recognition systems.

4. Methodology

In this section, we present the detailed methodology of our federated face recognition system in Figure 1, which integrates self-supervised learning and few-shot learning techniques. The approach is designed to address the challenges of privacy preservation, data heterogeneity, and limited labeled data in face recognition tasks.

4.1. Federated Learning

We implement Federated Averaging (FedAvg) as our core training algorithm, which enables collaborative model training while keeping face images private on local devices. In FedAvg, a central server coordinates the training process across K clients without accessing their local data. Formally, in each round (t), the server selects a subset of clients ( S t { 1 , 2 , , K } ) and distributes the global model parameters ( w t ) to them. Each selected client (k) then performs local training using their private face dataset ( D k ) to minimize the loss function:  
min w L k ( w ) = 1 | D k | ( x , y ) D k ( w ; x , y )
where ( w ; x , y ) is the loss for a single data point. After completing E local epochs, each client sends their updated model parameters ( w k t + 1 ) back to the server, which aggregates them using a weighted average:
w t + 1 = k S t | D k | j S t | D j | w k t + 1 .
This approach enables our face recognition system to learn from diverse, distributed facial data while preserving privacy and addressing the heterogeneity challenges inherent in face datasets across different clients.

4.2. Self-Supervised Pre-Training

In our federated face recognition system, we adopt SimCLR (Simple Framework for Contrastive Learning of Visual Representations) as our self-supervised pre-training algorithm. SimCLR learns representations by maximizing agreement between differently augmented views of the same data example via a contrastive loss in the latent space. For a batch of N face images, we generate 2 N augmented examples by applying random transformations, i.e., t T (such as cropping, flipping, and color distortion), to each image. Each augmented example is then encoded through a base encoder ( f ( · ) ) and a projection head ( g ( · ) ) to obtain the representation ( z i = g ( f ( x i ) ) ). The contrastive loss for a positive pair ( i , j ) is defined as follows:
L i , j = log exp ( sim ( z i , z j ) / τ ) k = 1 2 N 1 ( k i ) exp ( sim ( z i , z k ) / τ )
where sim ( u , v ) = u T v / u v denotes the cosine similarity between two vectors and τ is a temperature parameter. In the federated setting, each client computes this loss locally based on their private data, and only model updates are shared during aggregation, preserving data privacy while enabling collaborative learning of facial representations.

4.3. Few-Shot Learning

Our few-shot learning approach utilizes prototypical networks, which learn a metric space where classification can be performed by computing distances to prototype representations of each class. In the N-way, K-shot setting, each class prototype ( c n ) is computed as the mean embedding of its support examples:
c n = 1 K ( x i , y i ) S n f θ ( x i )
where S n is the support set for class n and f θ is the embedding function. Given a query image ( x q ), classification is performed by calculating the probability distribution over classes based on softmax-normalized distances:
p θ ( y = n | x q ) = exp ( d ( f θ ( x q ) , c n ) ) n = 1 N exp ( d ( f θ ( x q ) , c n ) )
where d ( · , · ) is a distance function—typically, a Euclidean distance. This approach is particularly effective for face recognition with limited samples per identity, as it leverages the distance-based decision boundaries that align well with the embedding-space structure of facial features.

4.4. Federated Few-Shot Fine-Tuning

In our federated few-shot learning framework, we extend the prototypical network approach to the federated setting. After pre-training the model using self-supervised learning, we implement a fine-tuning protocol where clients collaborate to improve few-shot recognition performance. Formally, in each round (t) of federated fine-tuning, the server selects a subset of clients ( S t { 1 , 2 , , K } ) according to the selection probability ( p k ) for each client (k). Each selected client performs few-shot learning on their local data, computing class prototypes for their N k local classes:
c k , n t = 1 | S k , n | ( x i , y i ) S k , n f θ t ( x i )
where S k , n is the support set for class n in client k and f θ t is the embedding function with parameters ( θ t ) from the global model. Each client then computes the loss based on their query set ( Q k ) using these prototypes:
L k ( θ t ) = 1 | Q k | ( x q , y q ) Q k log exp ( d ( f θ t ( x q ) , c k , y q t ) ) n = 1 N k exp ( d ( f θ t ( x q ) , c k , n t ) ) .
The client updates its local model parameters via gradient descent:
θ k t + 1 = θ t η L k ( θ t )
where η is the learning rate. The server then aggregates these updates using a weighted average:
θ t + 1 = k S t | Q k | j S t | Q j | θ k t + 1 .
This iterative process continues until convergence, which is defined as occurring when the change in global model parameters falls below a threshold ( ϵ : θ t + 1 θ t < ϵ ). This process is detailed in Algorithm 1.
Algorithm 1 Federated Few-Shot Fine-tuning
  1:
Input: Pre-trained model parameters θ 0 , number of clients K, number of rounds T, local learning rate η , client selection probability { p k } k = 1 K
  2:
Output: Fine-tuned global model parameters θ T
  3:
for each round t = 0 , 1 , , T 1  do
  4:
    Server selects subset of clients S t according to probabilities { p k }
  5:
    Server sends global model parameters θ t to selected clients
  6:
    for each client k S t in parallel do
  7:
        for each local class n = 1 , 2 , , N k  do
  8:
             Compute prototype c k , n t = 1 | S k , n | ( x i , y i ) S k , n f θ t ( x i )
  9:
        end for
10:
        Compute loss L k ( θ t ) on query set Q k using prototypes
11:
        Update local model: θ k t + 1 = θ t η L k ( θ t )
12:
        Send updated model parameters θ k t + 1 to server
13:
    end for
14:
    Server aggregates updates: θ t + 1 = k S t | Q k | j S t | Q j | θ k t + 1
15:
    if  θ t + 1 θ t < ϵ  then
16:
        break                    ▹ Convergence achieved
17:
    end if
18:
end for
19:
return θ T

5. Experimental Setup

This section describes the experimental configuration used to evaluate our federated self-supervised few-shot face recognition approach. We detail the datasets, implementation specifics, baseline comparisons, and evaluation protocols.

5.1. Datasets

We utilize the CASIA-WebFace dataset [23] for self-supervised pre-training, which contains 494,414 face images of 10,575 identities collected from the Web. The dataset provides diverse facial variations in terms of pose, illumination, expression, and background, making it suitable for the learning of robust face representations. For federated learning simulation, we partition the dataset across 20 clients according to a Dirichlet distribution with a concentration parameter of α = 0.1 to simulate non-IID data heterogeneity. Each client is assigned a disjoint subset of identities, ensuring no identity overlap across clients.
After self-supervised pre-training, the global encoder obtained from the federated learning stage is used for federated few-shot fine-tuning and evaluation on the Labeled Faces in the Wild (LFW) dataset [24]. LFW contains 13,233 face images of 5749 identities collected from news articles on the Internet. The dataset is particularly challenging due to unconstrained conditions, including variations in lighting, pose, expression, and image quality. For the federated few-shot evaluation, we similarly employ 20 clients with an IID data distribution to ensure each client has sufficient examples for the construction of meaningful few-shot episodes while maintaining disjoint identities across clients.
Within each client, few-shot episodes are constructed locally following the standard N-way K-shot protocol. We evaluate 3-way and 5-way classification tasks with 3-shot, 5-shot, and 10-shot settings. For each episode, N identities are randomly sampled from the client’s local test set, with K images per identity selected for the support set and the remaining images used as queries. We ensure that the support and query sets contain disjoint images of the same identities. Each experimental configuration is evaluated with 100 episodes to ensure statistical significance.

5.2. Implementation Details

We evaluate our approach using multiple state-of-the-art architectures—namely, ResNet [3], DenseNet [4], MobileViT [37], ViT-Small [5], CvT [6], and CoAtNet [38]—to demonstrate generalizability.
For each architecture, we conduct experiments with two initialization strategies: (1) using models pre-trained on ImageNet and (2) using models trained from scratch. The feature extractor outputs 512-dimensional embeddings, which are used for both self-supervised pre-training and few-shot fine-tuning.
Our experimental configuration follows these specifications:
  • We select 5 of 20 clients randomly for each communication round, with each selected client performing 20 local training epochs.
  • We use the Adam optimizer with an initial learning rate of 1 × 10 3 for self-supervised pre-training and 1 × 10 4 for few-shot fine-tuning. The learning rate decays by a factor of 0.1 every 50 communication rounds.
  • The local batch size is set to 32 for pre-training and 16 for few-shot learning to accommodate memory constraints.
  • We conduct 200 rounds for self-supervised pre-training and 20 rounds with 3 local epochs for federated few-shot fine-tuning.
  • For contrastive learning, we set τ = 0.07 based on empirical validation.
During self-supervised pre-training, we apply the following augmentation techniques to generate positive pairs:
  • Random horizontal flipping with a probability of 0.5;
  • Random rotation within ± 15 degrees;
  • Color jittering (brightness: 0.4; contrast: 0.4; saturation: 0.4; hue: 0.1)
  • Random grayscale conversion with a probability of 0.2;
  • Gaussian blur with a kernel size of 3 and a probability of 0.5;
  • Random resized cropping to 224 × 224 pixels.
For few-shot fine-tuning, we apply minimal augmentation, consisting only of random horizontal flipping and normalization to preserve the integrity of facial features.
Our training follows a three-stage protocol:
  • Self-Supervised Pre-training Stage: We train the feature extractor using SimCLR on unlabeled CASIA-WebFace data for 200 communication rounds.
  • Federated Few-Shot Fine-tuning Stage: The pre-trained model is fine-tuned using prototypical networks on labeled LFW data for 20 communication rounds.
  • Evaluation Stage: The final model is evaluated on a few-shot recognition tasks using the constructed episodes.

5.3. Baselines

We compare our federated self-supervised few-shot learning approach with several baseline methods to demonstrate its effectiveness.
Centralized Few-Shot Learning: This baseline represents the upper-bound performance where all data is centrally available. We train prototypical networks on the entire CASIA-WebFace dataset for pre-training, followed by few-shot fine-tuning on LFW. This approach uses the same network architectures and hyperparameters but without federated constraints.
Federated Self-Supervised Few-Shot Learning: This hybrid approach performs centralized self-supervised pre-training on CASIA-WebFace, followed by federated few-shot fine-tuning on LFW. This baseline helps isolate the contribution of federated pre-training. Each client trains independently on their local data without any collaboration. This baseline demonstrates the importance of federated collaboration, especially when local data is limited.

5.4. Evaluation Metrics

We evaluated our approach using the following metrics. We report the mean accuracy and the 95% confidence intervals in 100 episodes for each few-shot configuration (3-way/5-way with 3-shot/5-shot/10-shot configurations). This provides a comprehensive evaluation of recognition performance under different difficulty levels. Furthermore, we measure the total communication cost in terms of the transmitted model parameters and the number of communication rounds required to achieve the target performance levels. All experiments are conducted on eight NVIDIA L4 GPUs with 23GB memory, and we obtain the maximum results with fixed seeds to ensure reproducibility.

6. Results and Analysis

This section presents a systematic evaluation of our federated self-supervised few-shot face recognition framework. We begin by presenting the overall experimental setup and performance metrics, followed by detailed analyses of architectural comparisons, the impact of federated learning constraints, pre-training strategies, and few-shot learning dynamics. Finally, we discuss practical implications for real-world deployment.

6.1. Experimental Setup and Overall Performance

We evaluate six state-of-the-art architectures—ResNet, DenseNet, MobileViT, ViT-Small, CvT, and CoAtNet—under both centralized and federated learning paradigms. Each architecture is tested with and without ImageNet pre-training across 3-way and 5-way classification tasks, using 3-shot, 5-shot, and 10-shot configurations. Table 1 and Table 2 present the comprehensive results for centralized and federated settings, respectively.
The centralized setting (Table 1) consistently outperforms the federated counterpart (Table 2) across all architectures and configurations, as expected, given the availability of complete data and the absence of communication constraints. However, the federated approach demonstrates competitive performance while preserving privacy, with performance gaps varying significantly across architectures and initialization strategies. Among all evaluated architectures, MobileViT with ImageNet pre-training achieves the highest performance in the centralized setting, reaching 99.92% accuracy in the five-way, five-shot configuration. This exceptional performance is attributed to MobileViT’s hybrid architecture, which effectively combines the inductive biases of convolutional layers with the global modeling capabilities of transformers for face recognition tasks. Conversely, ViT-Small exhibits the poorest performance in centralized settings, particularly in three-way tasks (54.20% for 3-shot), indicating that pure transformer architectures require more extensive pre-training or larger model sizes to achieve competitive performance in few-shot scenarios.

6.2. Architectural Comparison and Robustness

The performance characteristics of different architectural families reveal important insights into their suitability for federated few-shot face recognition. ResNet and DenseNet demonstrate relatively modest performance degradation when transitioning from centralized to federated settings (approximately 12–15% average drop), indicating their robustness to federated constraints. This resilience is attributed to their well-established architectural designs and proven effectiveness in distributed learning scenarios. Both architectures maintain consistent performance across different shot configurations, with ResNet achieving 85.20% and DenseNet reaching 85.30% on federated three-way, three-shot tasks without pre-training. Their inductive biases, particularly translation equivariance and local connectivity, appear well-suited for face recognition under distributed constraints.
Pure transformer architectures, as exemplified by ViT-Small, struggle significantly in few-shot scenarios. In centralized three-way, three-shot settings, ViT-Small achieves only 54.20% accuracy with pre-training, suggesting that the global attention mechanism may not be optimal when training data are limited. However, ViT-Small shows interesting behavior in federated settings, achieving 91.30% accuracy in 3-way, 10-shot configurations, indicating potential for performance improvement with sufficient support examples. This variability suggests that transformers require careful configuration in few-shot federated contexts.
Hybrid architectures that combine convolutional and attention mechanisms (MobileViT, CvT, and CoAtNet) exhibit diverse trade-offs. MobileViT achieves exceptional performance in centralized settings but experiences substantial degradation in federated scenarios. In particular, MobileViT drops from 85.30% to 78.02% in a five-way, three-shot configuration when trained from scratch, suggesting that lightweight architectures may be more sensitive to data heterogeneity and limited local training data. Despite these challenges, MobileViT with ImageNet pre-training maintains the highest federated performance (91.36% in a 5-way, 10-shot configuration), demonstrating the value of strong initialization. CvT shows strong dependence on pre-training, with dramatic performance differences between initialized and randomly initialized models. The performance drop from 78.64% to 40.26% in the federated five-way, three-shot configuration without pre-training highlights the critical importance of initialization strategies for this architecture in distributed learning. CoAtNet demonstrates more balanced performance across different settings, maintaining relatively stable accuracy without exhibiting extreme sensitivity to either pre-training or federated constraints. This stability suggests that CoAtNet’s architectural design principles are well-suited for diverse learning scenarios.
Contrary to expectations, model complexity does not directly correlate with federated performance degradation. Lightweight models like MobileViT show substantial performance drops, while more complex models like CoAtNet maintain relatively stable performance. This suggests that architectural design principles, rather than parameter count, are more critical to federated learning success.

6.3. Impact of Federated Learning Constraints

The transition from centralized to federated learning reveals several fundamental challenges that affect model performance differently across architectures. The federated results reveal several key challenges inherent to distributed learning. First, the non-IID data distribution across clients creates statistical heterogeneity that impedes convergence to optimal solutions. Second, limited communication rounds and local training epochs constrain the model’s ability to fully leverage collective knowledge across clients. Third, the privacy-preserving nature of federated learning prevents direct data sharing, limiting the model’s exposure to diverse facial variations that are crucial for robust face recognition.
Different architectures demonstrate varying degrees of adaptation to federated constraints. MobileViT with ImageNet pre-training maintains relatively high performance in federated settings (91.36% in the 5-way 10-shot, configuration), suggesting that pre-trained representations provide a strong foundation that can be effectively fine-tuned in distributed scenarios. This finding underscores the importance of transfer learning in mitigating federated learning challenges. Conversely, architectures without appropriate initialization struggle significantly. CvT without pre-training shows dramatic performance degradation (from 78.64% to 40.26% in the five-way, three-shot configuration), highlighting the critical importance of initialization strategies in federated learning. This sensitivity suggests that certain architectural designs require stronger inductive biases or better initialization to overcome data heterogeneity in federated settings.

6.4. Effect of Pre-Training Strategies

The impact of ImageNet pre-training varies substantially across centralized and federated paradigms, revealing architecture-dependent patterns. In centralized settings, the impact of ImageNet pre-training is architecture-dependent. MobileViT shows the most dramatic improvement with pre-training (from 85.30% to 99.80% in the five-way, three-shot configuration), indicating that lightweight architectures particularly benefit from pre-trained representations. This substantial improvement (14.50 percentage points) demonstrates that pre-trained features effectively compensate for the limited capacity of lightweight models.
Surprisingly, ResNet and DenseNet show a minimal or even negative impact from ImageNet pre-training in some configurations, suggesting that self-supervised pre-training on domain-specific face data (CASIA-WebFace) may be more beneficial than generic ImageNet features for face recognition tasks. For instance, ResNet without pre-training achieves 97.23% accuracy compared to 96.63% with pre-training in the three-way, three-shot configuration, indicating that domain-specific self-supervised learning provides more relevant features for face recognition.
The federated learning paradigm reveals different pre-training dynamics. For most architectures, ImageNet pre-training provides modest improvements or maintains comparable performance. However, CvT shows a stark contrast, where pre-training becomes crucial for reasonable performance (83.76% vs. 40.26% in the five-way, three-shot configuration). This 43.50-percentage-point difference represents the largest pre-training effect observed in our experiments, emphasizing that certain hybrid architectures critically depend on strong initialization to overcome federated learning challenges, particularly under limited data availability for each client.
The mixed results regarding the effectiveness of ImageNet pre-training suggest that domain-specific self-supervised pre-training on face data may be more valuable than generic pre-training for face recognition tasks. This finding supports our approach of using SimCLR on the CASIA-WebFace dataset for pre-training, as it provides face-specific representations that are more aligned with the downstream task requirements.

6.5. Few-Shot Learning Dynamics

The analysis of different shot configurations reveals important insights about learning behavior and optimal data efficiency. Generally, increasing the number of shots improves performance across most configurations, but the improvement patterns vary significantly. In centralized settings, most architectures show substantial improvements from 3-shot to 5-shot configurations, with diminishing returns from 5-shot to 10-shot configurations. For example, the accuracy of ResNet with ImageNet pre-training improves from 97.90% to 99.64% (from the three-shot to five-shot configuration) but only marginally to 99.80% (from the 5-shot to 10-shot configuration) in 5-way classification. This saturation pattern suggests that the five-shot configuration may represent an optimal balance between data efficiency and performance for most architectures, with additional support examples providing limited marginal benefit.
Interestingly, some architectures perform better in five-way than three-way tasks, which contradicts the typical expectation that fewer classes should be easier to distinguish. This phenomenon is particularly evident in CoAtNet without pre-training (97.80% in the five-way, three-shot configuration vs. 74.93% in the three-way, three-shot configuration). This 22.87-percentage-point difference may be attributed to the specific characteristics of the LFW dataset and episode construction, where five-way tasks potentially provide more diverse and representative samples that facilitate better prototype formation in the embedding space.
In federated settings, the few-shot learning dynamics become more complex. Performance improvements with an increased number of shots are less consistent, and some configurations show performance degradation with more shots. For instance, ResNet with ImageNet pre-training shows decreasing performance from 5-shot (89.66%) to 10-shot (86.26%) configurations in 3-way tasks. This unexpected degradation suggests that federated learning may suffer from overfitting to local data distributions when provided with more shots, highlighting the need for better regularization strategies in federated few-shot learning. The non-IID nature of client data may cause models to overfit to local biases rather than learning generalizable representations.

6.6. Practical Implications and Deployment Guidelines

Based on our comprehensive experimental analysis, we provide actionable recommendations for the deployment of federated face recognition systems in diverse real-world scenarios. For scenarios where communication efficiency is paramount, the ResNet and DenseNet architectures provide the best balance between performance and robustness. These architectures achieve approximately 85% accuracy in federated three-way, three-shot settings while maintaining reasonable parameter counts and stable convergence. When computational resources are limited, MobileViT with careful pre-training strategies may be suitable, despite its federated performance challenges. With proper ImageNet initialization, MobileViT can achieve over 91% accuracy on federated 5-way, 10-shot tasks while offering substantial reductions in communication overhead. For applications requiring the highest possible accuracy and where communication costs are less critical, hybrid architectures like CoAtNet offer promising alternatives. CoAtNet demonstrates consistent performance across diverse settings, making it suitable for scenarios where prediction reliability is more important than communication efficiency.
Our results suggest that domain-specific self-supervised pre-training is more valuable than generic ImageNet pre-training for face recognition tasks. Organizations deploying federated face recognition systems should prioritize collecting diverse unlabeled face data for self-supervised pre-training rather than relying solely on ImageNet initialization. This recommendation is particularly important for the ResNet and DenseNet architectures, which show minimal benefit or negative transfer from ImageNet pre-training while demonstrating strong performance with domain-specific self-supervised learning on the CASIA-WebFace dataset.
The five-shot configuration appears to provide an optimal balance between data efficiency and performance across most architectures and settings. This finding is practically significant, as it suggests that federated face recognition systems can achieve reasonable performance with relatively few labeled examples per identity, reducing the annotation burden on participating clients. Specifically, our results indicate that moving from a 5-shot to 10-shot configuration provides minimal accuracy gains (typically less than 2%) while doubling the annotation requirements, making the five-shot configuration the preferred configuration for cost-effective deployment.
To mitigate the performance degradation observed in federated settings, we recommend (1) implementing robust aggregation schemes that account for data heterogeneity across clients; (2) incorporating regularization techniques to prevent overfitting to local distributions, particularly in higher-shot configurations; and (3) employing domain-specific pre-training to provide strong initialization that reduces sensitivity to federated constraints. These strategies are especially critical for lightweight and hybrid architectures that show higher sensitivity to distributed learning challenges. The different architectures have varying parameter counts, directly affecting communication costs in federated learning, and our observations suggest that convolutional architectures tend to show more stable convergence, whereas transformer-based models may require more communication rounds to achieve comparable performance, with important implications for the total communication cost and the feasibility of deployment.

6.7. Comparison with Existing Literature

To provide context for our results, we compare our findings with existing work in federated face recognition and few-shot learning, acknowledging the challenges in direct comparison due to differences in experimental setups, datasets, and evaluation protocols.
Kim et al. [11] reported federated face recognition results using self-supervised intra-subject learning, achieving approximately 85–90% accuracy on LFW in centralized few-shot settings. Our centralized results with traditional CNNs (ResNet and DenseNet) indicate comparable or superior performance (95–99% in few-shot settings), validating our experimental setup. However, their federated results showed less performance degradation (5–10%) compared to our 12–30% drop, suggesting that their intra-subject approach may be more robust to federated constraints than our inter-subject contrastive learning approach.
In the broader few-shot face recognition literature, existing methods report varying performance depending on the dataset and protocol. Chanda et al. [20] achieved 90% accuracy on five-way, one-shot tasks using Siamese networks—comparable to our centralized results but significantly higher than our federated performance (60–85% depending on the architecture). Similarly, Ranadive et al. [22] reported 84% accuracy in 50-way face verification, which presents a more challenging scenario than our 3-way and 5-way tasks, making direct comparison difficult but highlighting the complexity of few-shot face recognition across different experimental settings.
Our approach to self-supervised learning aligns with recent trends in the field. Lin et al. [14] demonstrated the effectiveness of domain-specific pre-training for face recognition, and our finding that domain-specific self-supervised pre-training often outperforms ImageNet pre-training supports their observations. This consistency across different studies validates the importance of task-specific representation learning in face recognition applications.
The substantial performance degradation we observe (12–30%) in federated settings is notably higher than existing federated learning works in other domains, which typically report 5–15% degradation. This disparity suggests that face recognition may be particularly sensitive to the data heterogeneity and communication constraints inherent in federated learning, representing a significant challenge for practical deployment that extends beyond general federated learning limitations.
Our architectural analysis reveals interesting contrasts with current trends in centralized face recognition. While transformer-based models often achieve state-of-the-art results in centralized settings, our findings show that traditional CNNs (ResNet and DenseNet) outperform transformer-based architectures in federated few-shot scenarios. This discrepancy indicates that architectural preferences may differ significantly between centralized and federated learning environments, with implications for system design in distributed settings.
Several factors limit direct comparison with existing literature. Differences in datasets and evaluation protocols make quantitative comparison challenging, while many existing federated face recognition works do not report few-shot performance metrics. Additionally, the specific combination of federated learning, self-supervised learning, and few-shot learning has limited prior work for direct comparison. These limitations underscore the need for standardized benchmarks in federated few-shot face recognition to enable more meaningful comparisons across different approaches and facilitate progress in this emerging research area.

6.8. Limitations and Future Directions

Our analysis reveals several critical limitations that fundamentally challenge the practical viability of our approach and point to important future research directions.
The most significant limitation is the substantial 12–30% performance degradation in federated settings, which raises serious questions about practical deployment. This performance gap is not merely a technical challenge to be optimized away but represents a fundamental trade-off inherent in privacy-preserving distributed learning. For applications requiring high accuracy (e.g., security systems and access control), this degradation may render the approach unsuitable for real-world deployment. Future work must honestly address whether federated face recognition can ever achieve the accuracy levels required for critical applications.
Our approach suffers from several methodological shortcomings that limit its contribution. The integration of the three paradigms is primarily sequential rather than synergistic, missing opportunities for deeper algorithmic innovation. We use standard aggregation methods (weighted averaging) without addressing the fundamental challenges of statistical heterogeneity across clients. The lack of novel algorithmic contributions means that the work primarily demonstrates the limitations of existing techniques rather than advancing the state of the art.
The communication overhead and convergence challenges observed in our experiments suggest that the approach may not scale effectively to larger federated networks. The requirement for multiple communication rounds and the sensitivity to client heterogeneity create practical barriers to deployment. Future work should address whether federated few-shot learning can achieve acceptable performance with realistic communication constraints and client availability patterns.
Our evaluation is primarily limited to the LFW dataset following self-supervised pre-training on CASIA-WebFace. While this provides a standardized benchmark for comparison, it constrains our ability to assess model performance across diverse demographic distributions, image quality variations, and environmental conditions. Future work should validate these findings across additional face recognition benchmarks to establish broader generalizability. An important direction for future research involves extending the evaluation to diverse face recognition benchmarks, including AgeDB-30 for age-invariant recognition and CFP-FP for pose robustness. Such multi-dataset validation would provide a more comprehensive understanding of how federated constraints affect model performance across different recognition challenges.
Given these limitations, future research should focus on (1) developing fundamentally new approaches that achieve better privacy–utility trade-offs rather than incremental improvements to existing methods, (2) establishing theoretical frameworks that explain the performance degradation and guide algorithm design, (3) creating more realistic evaluation protocols that capture the complexity of real-world federated environments, and (4) honestly assessing whether federated face recognition can meet the accuracy requirements of practical applications or if alternative privacy-preserving approaches should be pursued.
While this work provides valuable empirical insights, the substantial performance limitations significantly reduce its practical impact. The 12–30% accuracy degradation represents a fundamental barrier to adoption that cannot be easily overcome through incremental improvements. Future work should prioritize developing approaches that can achieve acceptable accuracy levels while preserving privacy or clearly define the limited application domains where such performance trade-offs are acceptable.

7. Conclusions

This paper presents a systematic empirical study that combines federated learning, self-supervised learning, and few-shot learning paradigms for privacy-preserving face recognition. While our approach successfully combines three established paradigms, the integration is primarily sequential rather than deeply synergistic.
The most significant finding is the substantial privacy–utility trade-off inherent in federated learning: our results show consistent performance degradation of 12–30% compared to centralized methods, representing the substantial cost of privacy preservation. This significant performance degradation raises important questions about the practical viability of such approaches and represents a fundamental limitation that practitioners must carefully consider. Traditional convolutional networks (ResNet and DenseNet) demonstrate superior robustness to federated constraints, with modest 12–15% performance drops, while transformer-based architectures suffer more severe degradation, with some configurations losing over 30% accuracy.
Our empirical analysis reveals several important findings for practitioners. Domain-specific self-supervised pre-training on facial data often proves more beneficial than ImageNet pre-training, particularly for established architectures, validating the use of SimCLR on face-specific datasets. Additionally, five-shot configurations consistently provide optimal balance between data efficiency and recognition performance across most architectures and settings, suggesting that federated face recognition systems can achieve reasonable performance with relatively few labeled examples per identity.
This research makes valuable empirical contributions by establishing benchmarks for federated few-shot face recognition, quantifying privacy–utility trade-offs, and providing practical insights about architectural suitability for distributed learning scenarios. This work serves as a foundation for future algorithmic innovations in federated few-shot learning while honestly acknowledging the substantial challenges and limitations that remain in making such systems practically viable for real-world deployment.

Author Contributions

Conceptualization, N.M. and B.A.; methodology, N.M.; software, N.M. and T.I.; validation, B.A., T.I. and D.Y.; formal analysis, N.M.; investigation, N.M. and T.I.; resources, D.Y.; data curation, N.M. and T.I.; writing—original draft preparation, N.M.; writing—review and editing, N.M.; visualization, N.M.; supervision, B.A.; project administration, B.A. and D.Y.; funding acquisition, D.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Committee of Science of the Ministry of Science and Higher Education of the Republic of Kazakhstan (Grant No. BR24992852, “Intelligent models and methods of Smart City digital ecosystem for sustainable development and the citizens’ quality of life improvement”).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in Kaggle at https://www.kaggle.com/datasets/debarghamitraroy/casia-webface (accessed on 15 October 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zeng, H. A Review of Traditional Methods and Deep Learning for Face Recognition. Appl. Comput. Eng. 2024, 80, 9–17. [Google Scholar] [CrossRef]
  2. Salagean, G.L.; Leba, M. Face Recognition: A Literature Review. In Seventh International Scientific Conference ITEMA—Recent Advances in Information Technology, Tourism, Economics, Management and Agriculture; Udruženje ekonomista i menadžera Balkana: Belgrade, Serbia, 2023. [Google Scholar]
  3. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  4. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  5. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  6. Wu, H.; Xiao, B.; Codella, N.; Liu, M.; Dai, X.; Yuan, L.; Zhang, L. Cvt: Introducing convolutions to vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 22–31. [Google Scholar]
  7. Mane, S.B.; Shah, N.; Garje, V.; Tejwani, A. A Comprehensive Survey of Face Recognition Advancements. In Proceedings of the 2024 8th International Conference on Computing, Communication, Control and Automation (ICCUBEA), Pune, India, 23–24 August 2024; pp. 1–5. [Google Scholar]
  8. Chen, B.; Liao, X.; Zhu, H.; Gong, Z.; Li, Y. A Survey of Face Recognition Methods based on Deep Learning. Highlights Sci. Eng. Technol. 2022, 24, 191–197. [Google Scholar] [CrossRef]
  9. Woubie, A.; Solomon, E.; Attieh, J. Maintaining Privacy in Face Recognition Using Federated Learning Method. IEEE Access 2024, 12, 39603–39613. [Google Scholar] [CrossRef]
  10. Solomon, E.; Woubie, A. Federated Learning Method for Preserving Privacy in Face Recognition System. arXiv 2024, arXiv:2403.05344. [Google Scholar] [CrossRef]
  11. Kim, H.; Choi, H.; Kwak, Y. Federated Learning for Face Recognition via Intra-subject Self-supervised Learning. arXiv 2024, arXiv:2407.16289. [Google Scholar]
  12. Xie, Z. A Comprehensive Analysis of Face Recognition Based on Advanced Deep Learning Technology. Appl. Comput. Eng. 2024, 81, 104–111. [Google Scholar] [CrossRef]
  13. Taherdoost, H. Beyond Supervised: The Rise of Self-Supervised Learning in Autonomous Systems. Information 2024, 15, 491. [Google Scholar] [CrossRef]
  14. Lin, C.H.; Wu, B.F. Domain Adapting Ability Of Self-Supervised Learning For Face Recognition. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 479–483. [Google Scholar] [CrossRef]
  15. Aghaomidi, P.; Aram, S.; Bahmani, Z. Leveraging Self-Supervised Learning for Accurate Facial Keypoint Detection in Thermal Images. In Proceedings of the 2023 30th National and 8th International Iranian Conference on Biomedical Engineering (ICBME), Tehran, Iran, 30 November–1 December 2023; pp. 452–457. [Google Scholar]
  16. Tsoumplekas, G.; Li, V.; Sarigiannidis, P.; Argyriou, V. A Complete Survey on Contemporary Methods, Emerging Paradigms and Hybrid Approaches for Few-Shot Learning. arXiv 2025, arXiv:2402.03017. [Google Scholar]
  17. He, K.; Pu, N.; Lao, M.; Lew, M.S. Few-shot and meta-learning methods for image understanding: A survey. Int. J. Multimed. Inf. Retr. 2023, 12, 14. [Google Scholar] [CrossRef]
  18. Gan, H.; Sang, N.; Huang, R. Self-training-based face recognition using semi-supervised linear discriminant analysis and affinity propagation. J. Opt. Soc. Am. A 2013, 31, 1–6. [Google Scholar] [CrossRef]
  19. Snell, J.; Swersky, K.; Zemel, R.S. Prototypical Networks for Few-shot Learning. arXiv 2017, arXiv:1703.05175. [Google Scholar]
  20. Chanda, S.; AsishChakrapani, G.; Brun, A.; Hast, A.; Pal, U.; Doermann, D.S. Face Recognition—A One-Shot Learning Perspective. In Proceedings of the 2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), Sorrento, Italy, 26–29 November 2019; pp. 113–119. [Google Scholar]
  21. Albayati, A.M.; Chtourou, W.; Zarai, F. Attending to Channels in One-Shot Learning Face Recognition using Squeeze-and-Excitation Networks. In Proceedings of the 2022 International Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT), Ankara, Turkey, 20–22 October 2022; pp. 325–330. [Google Scholar]
  22. Ranadive, O.; Thakkar, D. k-Shot Learning for Face Recognition. Int. J. Comput. Appl. 2018, 181, 43–48. [Google Scholar] [CrossRef]
  23. Yi, D.; Lei, Z.; Liao, S.; Li, S.Z. Learning face representation from scratch. arXiv 2014, arXiv:1411.7923. [Google Scholar] [CrossRef]
  24. Huang, G.B.; Mattar, M.; Berg, T.; Learned-Miller, E. Labeled faces in the wild: A database forstudying face recognition in unconstrained environments. In Proceedings of the Workshop on Faces in ‘Real-Life’ Images: Detection, Alignment, and Recognition, Marseille, France, 12–18 October 2008. [Google Scholar]
  25. Ewerth, R.; Mühling, M.; Freisleben, B. Self-Supervised Learning of Face Appearances in TV Casts and Movies. In Proceedings of the Eighth IEEE International Symposium on Multimedia (ISM’06), San Diego, CA, USA, 11–13 December 2006; pp. 78–85. [Google Scholar]
  26. Khan, A.; Asmatullah, L.; Malik, A.; Khan, S.; Asif, H. A Survey on Self-supervised Contrastive Learning for Multimodal Text-Image Analysis. arXiv 2025, arXiv:2503.11101. [Google Scholar]
  27. Xiao, J.; Zhu, G.; Loianno, G. VG-SSL: Benchmarking Self-supervised Representation Learning Approaches for Visual Geo-localization. arXiv 2024, arXiv:2308.00090. [Google Scholar]
  28. Lialin, V.; Rawls, S.; Chan, D.; Ghosh, S.; Rumshisky, A.; Hamza, W. Scalable and Accurate Self-supervised Multimodal Representation Learning without Aligned Video and Text Data. In Proceedings of the 2023 IEEE/CVF Winter Conference on Applications of Computer Vision Workshops (WACVW), Waikoloa, HI, USA, 3–7 January 2023; pp. 390–400. [Google Scholar] [CrossRef]
  29. Zhao, X.; Evans, N.W.D.; Dugelay, J.L. Semi-supervised face recognition with LDA self-training. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 3041–3044. [Google Scholar]
  30. Priyanka; Kumar, S. Advancements in Meta-Learning Paradigms: A Comprehensive Exploration of Techniques for Few-Shot Learning in Computer Vision. In Proceedings of the 2024 International Conference on Intelligent Systems for Cybersecurity (ISCS), Gurugram, India, 3–4 May 2024; pp. 1–8. [Google Scholar]
  31. Gharoun, H.; Momenifar, F.; Chen, F.; Gandomi, A.H. Meta-learning approaches for few-shot learning: A survey of recent advances. arXiv 2023, arXiv:2303.07502. [Google Scholar] [CrossRef]
  32. Xu, N.; Feng, W. MetaFake: Few-shot Face Forgery Detection with Meta Learning. In Proceedings of the 2023 ACM Workshop on Information Hiding and Multimedia Security, Chicago, IL, USA, 28–30 June 2023. [Google Scholar]
  33. Althnian, A.; Aloboud, N.; Alkharashi, N.; Alduwaish, F.; Alrshoud, M.; Kurdi, H.A. Face Gender Recognition in the Wild: An Extensive Performance Comparison of Deep-Learned, Hand-Crafted, and Fused Features with Deep and Traditional Models. Appl. Sci. 2020, 11, 89. [Google Scholar] [CrossRef]
  34. Chen, S.; Su, N.; Li, B. Calibre: Towards fair and accurate personalized federated learning with self-supervised learning. In Proceedings of the 2024 IEEE 44th International Conference on Distributed Computing Systems (ICDCS), Jersey City, NJ, USA, 23–26 July 2024; pp. 891–901. [Google Scholar]
  35. Zrira, N.; Abouzahir, M.; Bouyakhf, E.H.; Benmiloud, I.; Himmi, M.M. Learning combined features for automatic facial expression recognition. Int. J. Knowl. Eng. Soft Data Paradig. 2019, 6, 153–169. [Google Scholar] [CrossRef]
  36. Charan, S.G. Face recognition using combined binary particle swarm optimisation and hidden layer of artificial neural network. Int. J. Adv. Intell. Paradig. 2019, 14, 70–79. [Google Scholar] [CrossRef]
  37. Mehta, S.; Rastegari, M. Mobilevit: Light-weight, general-purpose, and mobile-friendly vision transformer. arXiv 2021, arXiv:2110.02178. [Google Scholar]
  38. Dai, Z.; Liu, H.; Le, Q.V.; Tan, M. Coatnet: Marrying convolution and attention for all data sizes. Adv. Neural Inf. Process. Syst. 2021, 34, 3965–3977. [Google Scholar]
Figure 1. General architecture of federated self-supervised few-shot face recognition.
Figure 1. General architecture of federated self-supervised few-shot face recognition.
Jimaging 11 00370 g001
Table 1. Centralized ProtoNet results.
Table 1. Centralized ProtoNet results.
ModelImgNet
Pre-Trained
3-Way5-Way
3-Shot 5-Shot 10-Shot 3-Shot 5-Shot 10-Shot
ResNetyes96.6399.1099.6097.9099.6499.80
no97.2399.3399.8098.1499.7899.80
DenseNetyes96.3399.3799.6796.7499.6099.64
no95.7799.6799.8097.4299.6099.88
MobileViTyes99.4099.8099.7399.8099.9299.84
no86.4392.6396.2085.3094.1697.12
ViT smallyes54.2071.6078.2758.8283.5490.64
no58.1768.8776.7359.4483.4089.36
CvTyes62.0073.2773.2787.4897.7897.92
no38.5767.7774.8078.6493.8896.28
CoAtNetyes86.9096.1096.8097.6899.3699.16
no74.9378.5384.0797.8098.8894.00
Bold values indicate the highest accuracy achieved across all models for each shot configuration, while underlined with italic values denote the second-highest performance.
Table 2. Federated ProtoNet results.
Table 2. Federated ProtoNet results.
ModelImgNet
Pre-Trained
3-Way5-Way
3-Shot 5-Shot 10-Shot 3-Shot 5-Shot 10-Shot
ResNetyes84.5089.6686.2681.6484.8881.78
no85.2087.5684.3081.6885.2082.76
DenseNetyes83.4089.5082.4082.1086.0284.72
no85.3088.3682.2082.3086.2485.12
MobileViTyes80.3092.1390.7689.2091.1691.36
no79.8782.7382.7078.0280.0079.50
ViT smallyes62.6665.1091.3082.8060.7260.96
no79.8665.7388.5084.5859.1462.68
CvTyes78.2380.3679.8383.7690.2888.90
no52.8054.9058.6040.2646.1052.40
CoAtNetyes75.5384.2378.4084.1081.9681.64
no78.7081.0676.9083.9284.3877.82
Bold values indicate the highest accuracy achieved across all models for each shot configuration, while underlined with italic values denote the second-highest performance.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Makhanov, N.; Amirgaliyev, B.; Islamgozhayev, T.; Yedilkhan, D. Federated Self-Supervised Few-Shot Face Recognition. J. Imaging 2025, 11, 370. https://doi.org/10.3390/jimaging11100370

AMA Style

Makhanov N, Amirgaliyev B, Islamgozhayev T, Yedilkhan D. Federated Self-Supervised Few-Shot Face Recognition. Journal of Imaging. 2025; 11(10):370. https://doi.org/10.3390/jimaging11100370

Chicago/Turabian Style

Makhanov, Nursultan, Beibut Amirgaliyev, Talgat Islamgozhayev, and Didar Yedilkhan. 2025. "Federated Self-Supervised Few-Shot Face Recognition" Journal of Imaging 11, no. 10: 370. https://doi.org/10.3390/jimaging11100370

APA Style

Makhanov, N., Amirgaliyev, B., Islamgozhayev, T., & Yedilkhan, D. (2025). Federated Self-Supervised Few-Shot Face Recognition. Journal of Imaging, 11(10), 370. https://doi.org/10.3390/jimaging11100370

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop