Next Article in Journal
Beyond the Benchmark: A Customizable Platform for Real-Time, Preference-Driven LLM Evaluation
Previous Article in Journal
YOLO-LSDI: An Enhanced Algorithm for Steel Surface Defect Detection Using a YOLOv11 Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Federated Learning-Based Location Similarity Model for Location Privacy Preserving Recommendation

1
College of Computer Science and Technology, Zhengzhou University of Light Industry, Zhengzhou 450001, China
2
Shangu Cyber Security Technology Co., Ltd., Zhengzhou 450003, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(13), 2578; https://doi.org/10.3390/electronics14132578
Submission received: 24 March 2025 / Revised: 19 June 2025 / Accepted: 23 June 2025 / Published: 26 June 2025
(This article belongs to the Special Issue Big Data Security and Privacy)

Abstract

With the proliferation of mobile devices and wireless communications, Location-Based Social Networks (LBSNs) have seen tremendous growth. Location recommendation, as an important service in LBSNs, can provide users with locations of interest by analyzing their complex check-in information. Currently, most location recommendations use centralized learning strategies, which carry the risk of user privacy breaches. As an emerging learning strategy, federated learning is widely applied in the field of location recommendation to address privacy concerns. We propose a Federated Learning-Based Location Similarity Model for Location Privacy Preserving Recommendation (FedLSM-LPR) scheme. First, the location-based similarity model is used to capture the differences between locations and make location recommendations. Second, the penalty term is added to the loss function to constrain the distance between the local model parameters and the global model parameters. Finally, we use the REPAgg method, which is based on clustering for client selection, to perform global model aggregation to address data heterogeneity issues. Extensive experiments demonstrate that the proposed FedLSM-LPR scheme not only delivers superior performance but also effectively protects the privacy of users.

1. Introduction

With the exponential growth of global positioning system (GPS) trajectory data as a result of technological advances and the popularity of mobile devices, researchers have begun to focus on how to efficiently utilize this data to gain a deeper understanding of users’ complex mobile behaviors. As a result, location-based social networks (LBSNs) have been vigorously developed [1]. Through the location check-in function of mobile users, LBSNs closely integrate online virtual socialization with the offline real world [2], thus giving rise to a variety of location-based services. Among them, location recommendation plays a key role in providing superior user experience.
In the field of location-based recommendation, there are various alternative recommendation techniques such as content-based recommendation [3], collaborative filtering (CF), association rules [4], and utility-based recommendation [5]. CF is a widely used method in the recommendation field, known for its simplicity and efficiency. The method provides personalized recommendations to users by analyzing the similarities between their behaviors and preferences [6]. Specifically, CF-based location recommendation works by mapping user, location, and context information into a shared implicit space and using an inner product operation to compute the predictive score. Ultimately, the computed predictive scores are used to recommend tasks. Liu et al. [7] used a CF approach and considered groups as users to construct a group preference model by aggregating the preferences of each group member in order to improve the performance of the recommendation system. Zhang et al. [8] proposed a location-aware deep CF scheme. By using high-dimensional embedding vectors, a multi-layer perceptron, and a similarity adaptive corrector, this approach addresses the limitations of traditional CF methods, which are often constrained by low-dimensional and linear interactions.
However, for highly private and sensitive user information, the centralized information storage method used in the above study may carry the risk of leaking private user information. Especially after the enactment of the general data protection regulation (GDPR) [9] the requirements for personal data protection have become more stringent. Location service providers face the challenge of being unable to centrally store user data, making personalized recommendations difficult [10,11]. This situation is driving researchers and companies to seek new solutions that comply with the regulations. In recent years, federated learning within the field of distributed machine learning has garnered significant attention and become a central focus of extensive research [12,13,14,15]. This approach offers innovative solutions for location-based recommendations by enabling multiple participants to collaboratively train a shared machine learning model while maintaining data privacy. Federated learning effectively addresses the challenges of privacy, security, and distributed data, which are inherent in traditional machine learning methods. The core idea of federated learning is to enable participants to update model parameters using their local data and then send the updated model parameters to a central server for aggregation, thereby achieving continuous optimization and improvement of the model. In a federated learning framework, model parameters are transmitted between clients and the central server, effectively safeguarding user privacy. In order to protect users’ location data, this article uses a federated learning framework on the server side. However, in location-based recommendations using federated learning, each client has its own local dataset. Due to differences among clients, such as variations in user behavior and geographic locations, these local datasets exhibit different data distributions. In the case of non-independent and identically distributed (non-IID) data, the accuracy of the final aggregated global model is usually less than ideal. Due to the differences in data distribution among clients, each node tends to converge to a local optimum during local training, which may deviate from the direction of the global optimum. This can lead to biases in the models learned by individual nodes, thus severely affecting the accuracy of the aggregated global model [16,17]. Therefore, addressing the issue of data heterogeneity in location-based recommendations using federated learning becomes crucial.
Based on the above analysis, we propose a Federated Learning-Based Location Similarity Model for Location Privacy Preserving Recommendation (FedLSM-LPR) scheme. The scheme involves the collaborative work of a central server and multiple edge clients working together to carry out the training process on local data. First, each client uses a location-based similarity model for location recommendation. At the same time, a two-stage perturbation approach is used to protect the user’s data privacy. Second, the L2 norm regularization term between the local and global models is added to the loss function as a penalty to constrain the difference between the two models. Finally, we propose a clustering-based client selection method for the impact of non-IID data on federated learning location recommendation. By combining this method with the REPAgg aggregation method, the performance of federated learning location recommendation is effectively improved.
The main contributions of this paper are as follows:
  • The location similarity model based on federated learning is constructed to provide personalized location-related services for target users. The model will be employed on the client side in order to capture the subtle differences between locations and enhance the performance of location recommendations. Federated learning is utilized on the server side to ensure data privacy while aggregating the model.
  • The clustering-based client selected algorithm is proposed to further mitigate the impact of non-IID data on the framework of location recommendation. The algorithm performs client clustering and combines client selection based on local loss and model similarity. Additionally, the penalty term is utilized in the client loss function to constrain the difference between the local model and the global model.
  • Extensive experiments and analysis are conducted on two real datasets to validate the effectiveness of the proposed FedLSM-LPR scheme. The experimental results demonstrate that the performance is better than that of the other existing schemes.

2. Related Work

Location-based recommendations are a key application of location-based services (LBS), aiming to suggest places that align with users’ interests and preferences [18,19,20]. This helps users discover and better understand their surroundings. Location-based recommendations have garnered significant attention from researchers, with CF-based approaches being particularly noteworthy [21]. CF methods can be categorized into two types: user-based and location-based. User-based CF methods make recommendations by finding other users with similar interests and behavior patterns. This approach leverages the historical behaviors and preferences of users to identify those most similar to the target user and recommends locations that these similar users have liked to the target user. User-based CF methods are simple and intuitive, leveraging the similarities between users to make recommendations. This approach is well-suited for scenarios where user behaviors and preferences are relatively clear. Ye et al. [22] proposed a user-based CF recommendation scheme that utilizes location evaluation data and combines it with machine learning algorithms to compute friends with similarity to the target user to achieve location recommendation. However, user-based CF recommendation schemes have certain limitations. For instance, when dealing with new users, the lack of sufficient personal behavioral data makes it difficult to find similar users, thereby reducing the accuracy and reliability of the recommendations. In contrast, location-based CF methods make recommendations by analyzing the similarities between locations. This approach focuses on the relationships between locations rather than the similarities between users. By calculating the similarity between locations, it recommends locations similar to those the user has liked in the past. Zheng et al. [23] proposed a personalized location recommendation system utilizing GPS trajectory data. This system reveals the correlations between locations by learning users’ travel experiences and the sequence of location visits. As a location-based CF method, it overcomes the issues encountered in user-based CF methods. However, in today’s data-driven society, data security and privacy protection have become crucial [24]. The aforementioned research utilizes traditional centralized data storage methods, which may pose potential issues regarding data privacy and security. Traditional encryption methods, such as homomorphic encryption [25] and secret sharing [26], are challenging to implement widely in systems due to their computational costs.
To address the aforementioned limitations, McMahan et al. [27] first proposed the federated learning framework, introducing the Federated Averaging (FedAvg) algorithm to enable collaborative training of data distributed across different devices. As an emerging machine learning paradigm, federated learning aims to achieve collaborative learning on distributed data while protecting data privacy and security. In federated learning, data holders (e.g., personal devices, edge servers, institutional servers) retain their original data locally and perform model training, thereby avoiding the risk of sensitive data leakage caused by centralized storage. The basic process of federated learning is illustrated in Figure 1.
Federated learning faces several challenges in location-based recommendation systems. First, since user data is dispersed across different devices, it is a key challenge to efficiently integrate these data to improve the accuracy of the model. Second, federated learning models may also face the problem of model bias, i.e., the difference in distribution between local and global models, which can lead to bias in the model training process and reduce the generalization ability and recommendation accuracy of the models. In addition, although federated learning aims to enhance privacy protection, there may still be a risk of information leakage. How to ensure model performance and recommendation quality while guaranteeing user privacy still needs to be continuously explored. Together, these factors increase the complexity of realizing an efficient and accurate federated learning location recommendation system. In federated learning recommendation systems, Ammad-ud-din et al. [28] were the first to combine federated learning with CF to protect client data privacy, laying the foundation for further research. Chai et al. [29] proposed a scheme named FedMF, which combines federated learning and homomorphic encryption to protect user privacy. This scheme utilizes matrix factorization on the client side to make recommendations. Perifanis et al. [30] implemented federated learning recommendations using neural CF and proposed a privacy-preserving aggregation method to ensure the security of model training. Experimental results demonstrated the effectiveness of this system in both recommendation quality and privacy protection. Huang et al. [31] proposed a position recommendation scheme that combines matrix factorization and federated learning. This scheme constructs a geographic information matrix to quantify location information and combines it with a scoring matrix for matrix decomposition using the singular value decomposition technique. Lin et al. [32] proposed a federated recommendation scheme called FedRec for solving the rating prediction problem with explicit feedback. The scheme uses explicit feedback to extend the probability matrix decomposition to both batch and random styles. However, due to the non-IID nature of location data, the use of simple weighted averaging to aggregate client model parameters in the aforementioned studies may result in decreased model performance when handling such data. Dong et al. [33] proposed a personalized federated learning method using locality-sensitive hashing to address the impact of non-IID data in federated learning-based location recommendations. Li et al. [34] proposed a federated aggregation algorithm with dynamic loss to enhance the performance of weaker clients. Ding et al. [35] proposed a meta-learning inspired federated aggregation method for fast convergence, aiming to address data heterogeneity issues in federated recommendation systems and accelerate the convergence of global training. This method draws inspiration from the REPTILE algorithm, approximating global updates through Taylor expansion while maintaining gradient direction consistency to accelerate convergence. However, despite improving the efficiency of server-side aggregation, the federated aggregation method REPAgg still has limitations. First, it does not explicitly constrain the divergence between local client models and the global model, which may lead to bias accumulation in highly heterogeneous data environments. Second, the uniform random sampling strategy for client selection overlooks differences in data quality, potentially affecting model performance under extreme non-IID conditions.

3. Preliminaries

In this section, we provide a detailed explanation of the location-based similarity model. Following this, we introduce the core mechanism of the REPAgg aggregation method. Finally, we define the problem addressed in this study. Table 1 summarizes the key symbols used in this paper and their respective meanings.

3.1. Location-Based Similarity Model

The core idea of standard location-based CF [36] is that the similarity between location i and all the locations the user has interacted with in the past determines the user u’s prediction of the target location i. Such models infer users’ preferences and interests in unknown locations by mining the interaction patterns between users and locations. The formula for the prediction model of CF based on standard locations is shown as follows:
y ^ u i = j L u + l u j S i j ,
where L u + is the set of historical locations that the user interacted with, S i j is the similarity between location i and location j, and  l u j is user u’s preference for location j. User preferences for locations are typically represented as implicit and explicit feedback. Implicit feedback consists of indirect signals generated during the user’s interaction with the system, such as clicks, browsing history, and purchase records, which infer the user’s preferences. Explicit feedback is directly provided by the user, including ratings, reviews, likes, and dislikes.
Ning et al. [37] proposed the SLIM model, which learns location similarity by optimizing a recommendation-aware objective function. SLIM optimizes by minimizing the loss between two matrices: the original user-location interaction matrix and the reconstructed user-location interaction matrix generated by the location-based similarity model. However, this model is limited to learning similarity only between location pairs that share ratings and cannot identify transitive relationships between locations.
To address the limitations of the SLIM model, Kabbur et al. [38] proposed the FISM model. This model represents locations as low-dimensional embedding vectors and uses the inner product of the embedding vectors i and j to represent the similarity score. However, the FISM model overlooks the heterogeneity between different locations by assigning the same weight to all interacting locations. Treating all of a user’s historical locations equally can limit the model’s performance.
To address the aforementioned issues, He et al. [39] proposed the NAIS model, which possesses stronger representational capabilities. By incorporating an attention mechanism network, the NAIS model introduces additional parameters that better capture the differences between various locations. The NAIS prediction model can be represented by the following formula:
y ^ u i = m i T j L u + { i } α i j n j ,
α i j = exp f m i , n j k L u + { i } exp f m i , n k τ ,
where α i j is the similarity weights of location i and location j when predicting user u’s preference for target location i, τ is the smoothing index in the range of [0, 1], and  f m i , n j is the attention function of location i and location j. Because NAIS effectively identifies the heterogeneity between locations, we adopt it as the client model in federated learning for location-based recommendations.

3.2. Server Aggregation with REPAgg

REPTILE [40] is a meta-learning algorithm that iteratively optimizes model parameters through a combination of inner and outer updates. Its core objective is to achieve rapid adaptation to new tasks. Specifically, in the inner update phase, the algorithm performs multiple rounds of gradient descent on each specific task’s dataset to reduce the loss function value. This process mimics single-task training to facilitate rapid adaptation to new tasks. Following this, in the outer update phase, the model adjusts its parameters based on the differences between the parameters after the inner update and the initial parameters. By adding a proportion of these differences to the initial parameters, the model’s generalization ability to unknown tasks is enhanced. The internal and external updates can be represented by the following formula:
w z = w 0 α w 0 L w 0 , D z ,
w 0 = w 0 + β 1 | Z | z Z w z w 0 ,
where w 0 is the initialized model parameters of task z Z , w z is the updated model parameters of task z, α is the internally updated learning rate, β is the learning rate for external updates. w 0 L w 0 , D z is the gradient of the loss function L w 0 , D z with respect to the model parameter w 0 , and  D z is the training dataset of task z.
Ding et al. [35] applied the idea of REPTILE to federated learning and changed the traditional federated aggregation method, which is formulated as follows:
w G t + 1 = w k t + λ 1 | S | k S Δ k t ,
where w k t is the model parameter of client k, which is the global model obtained by the server via the federated aggregation algorithm at t 1 epoch, Δ k t is the model update obtained by client k via local training, λ is the learning rate, S is the set of selected clients, and  w G t + 1 is the global model obtained by the server via the federated aggregation algorithm at t epoch. Due to the server’s inability to grasp the overall distribution of user characteristics, this method may reduce the impact of model parameter updates for locations that are infrequently interacted with. Therefore, this federated aggregation approach is not suitable for recommendation tasks based on a location similarity model. Ding et al. [35] modified the above federated aggregation method and proposed the REPAgg aggregation method to better fit the needs of recommendation tasks based on a location similarity model. The formula for the REPAgg aggregation method is shown as follows:
M t + 1 = M t + λ k S Δ k t m 1 , k S Δ k t m 2 , , k S Δ k t m q ,
N t + 1 = N t + λ k S Δ k t n 1 , k S Δ k t n 2 , , k S Δ k t n q ,
where M and N are the location feature matrices, and the update of the location feature vectors for client k is denoted by Δ k t m i and Δ k t n i .

3.3. Problem Definition

To protect the privacy of each client’s data, this paper employs a federated learning framework. In federated learning-based location recommendation, each client independently stores and processes its own interaction data, which is never uploaded to the server. The server’s role is to organize and coordinate the work of the clients without directly accessing their data. This allows clients to collaboratively train a global model while ensuring that personal data remains private, thus achieving the goal of providing personalized location recommendations without compromising privacy. To further enhance privacy, clients apply a two-stage perturbation process, which disturbs both location data and model updates, preventing attackers from inferring sensitive information or reconstructing the original data from the model updates. We assume that clients trust each other in this federated learning setup. In federated learning-based location recommendation, we let u U denote the user, i L denote the location, with the number of users as p and the number of locations as q. The interactions between users and locations are represented in a simple binary manner: 1 indicates that the user has interacted with the location in the past, and 0 indicates otherwise. The formula for user interaction is shown as follows:
y u i = 1 , i L u + 0 , i L u + ,
where y u i is the interaction between user u and location i, and  L u + is the historical interaction position of user u.

4. Proposed Framework

In this section, we will provide a detailed overview of the overall architecture and methods used in the FedLSM-LPR scheme.

4.1. Model Overview

The overall architecture of the FedLSM-LPR scheme is illustrated in Figure 2. This architecture combines a location similarity model with federated learning to train on decentralized data, thereby constructing a global model while ensuring the privacy and security of user data.
The overall implementation steps of FedLSM-LPR are as follows: Step 1: Before federated learning begins, all clients calculate the rating frequency distribution based on their local data and send this information to the server. Upon receiving the rating frequency distributions from the clients, the server uses the K-means algorithm to cluster these distributions, resulting in the clients’ cluster labels C. Step 2: The server initializes the global model using a Gaussian distribution and randomly samples clients in each epoch to participate in the model training. The server then distributes the initialized global model w G 0 to the selected clients. Step 3: After receiving the initialized model from the server, each client employs a loss function with a penalty term and trains the model on its local dataset using a location similarity model. During the training process, clients protect their local data with label perturbation. After E local optimization steps, clients apply model update perturbation to the generated model updates and upload the locally computed loss and the perturbed model updates to the server. Step 4: Upon receiving the local losses and perturbed model updates from each client, the server employs the REPAgg aggregation method, which is based on cluster-based client selection, to aggregate the global model. Step 5: The server distributes the aggregated global model to each client. Repeat steps 3–5 until the model converges. The pseudo-code of the FedLSM-LPR scheme is shown in Algorithm 1.
Algorithm 1 FedLSM-LPR
Input: 
Client set U, Number of Iterations for Local Training E, Number of Iterations for Global Training T, Local and Global learning rates η and λ , Clusters c, Number of selected clients K, Selection of client ratio based on model similarity f 1 , Selection of client ratio based on loss f 2
Output: 
Global model w G
  1:
for each client k U  do
  2:
   Receive rating frequency distribution Q k from client k
  3:
end for
  4:
Cluster clients into C = C 1 , C 2 , , C c via K-means( Q k )
  5:
Server initializes global model w G 0 = { M , N , f ( . , . ) } and other hyper-parameters
  6:
for each global epoch t = 0 , 1 , , T  do
  7:
    S Randomly sample K clients from U
  8:
   Broadcast w G t to all k S
  9:
   for each client k S  do
10:
      Δ ¯ k t , L k ClientLocalUpdate ( k , w G t )
11:
   end for
12:
   for cluster C i C  do
13:
     //Obtain selected client set Sel C i using Client Selection algorithm
14:
      S e l C i C l i e n t S e l e c t i o n ( C i , f 1 , f 2 )
15:
     Aggregate model updates for selected clients
16:
   end for
17:
   //Aggregation of REPAgg model using Formulas (14) and (15)
18:
    w G t + 1 M o d e l A g g r e g a t i o n ( w G t , Δ ¯ C i t , λ )
19:
end for
20:
Function ClientLocalUpdate( k , w G t ):
21:
   //Accept server model
22:
    w k t w G t
23:
   for each local epoch e = 0 , 1 , , E  do
24:
      Use label perturbation to perturb training data
25:
       w G t L o c a l O p t i m i z a t i o n ( w G t , L u + , η )
26:
   end for
27:
    Δ k t = w k t w G t
28:
   Obtain Δ ¯ k t through using model update perturbation to perturb Δ k t
29:
    L k = C o m p u t e L o s s ( D k , w k t )
30:
   return  Δ ¯ k t , L k

4.2. Client-Side Regularization

FedIS does not take into account the potential differences between the local model and the global model when dealing with client-side local optimization. In FedIS, for the local data of client k S , the following formula is used for optimization:
L = 1 H i L u + log σ y ^ u i + i L u log σ 1 y ^ u i + ψ w k t 2 ,
where w k t is the local model of client k which is initialized by the global model w G t obtained from the previous epoch through model aggregation, H is the number of training samples, σ ( · ) is a sigmoid function, and  ψ is a parameter controlling the strength of regularization.
Since each client trains its model based solely on its own data without considering consistency with the global model, this can result in significant deviations between the locally optimized models and the global model. In practical applications, such deviations can cause instability in the global model’s performance, especially when dealing with highly non-IID data.
To solve the above problem, we use the L2 norm regularization term between the local and global models as a penalty term to prevent the local models from drifting too far from the global model. We incorporate this penalty term into Formula (10) to address the issue of local model drift. Consequently, client k uses its local data and optimizes the local model using the loss function specified in the following formula:
L = 1 H i L u + log σ y ^ u i + i L u log σ 1 y ^ u i + ψ w k t 2 + τ 2 w G t w k t 2 ,
where w k t is the local model of the client k, w G t is the global model, and τ is the regularization parameter used to control the degree of difference between the local model and the global model. It helps avoid overfitting on local data and improves the model’s generalization ability in federated learning environments. The value of parameter τ will affect the performance of the model: if the value of τ is too high, it will excessively restrict local updates and reduce personalization ability; if the τ value is too small, it may make it difficult for the global model to converge.
Overall, the above approach provides the following key benefits: (i) Constraining the extent of model updates. In a federated learning environment, each client trains its model locally based on its own dataset. Due to the differences in the characteristics and distributions of the datasets across clients, the extent of model updates can vary, leading to significant changes in model parameters at certain nodes. The penalty term can constrain the extent of these personalized model updates, preventing excessive parameter adjustments and more effectively managing the scope of model updates. (ii) Enhancing model generalization. By introducing regularization techniques into the loss function, the growth of model parameters can be effectively constrained, reducing the model’s sensitivity to specific training samples. This approach not only helps to mitigate the issue of overfitting but also enhances the model’s predictive capabilities when dealing with unseen data.

4.3. Clustering-Based Client-Selected REPAgg Aggregation Approach

In practical applications, differences in client identity, behavior, and environment often result in data that exhibits non-IID characteristics. These characteristics can lead to significant variations in model performance across different devices, thereby affecting the effectiveness of model training. In this paper, we use a clustering-based client-selected REPAgg aggregation method to address the above challenges. The method consists of three parts: (i) Client Clustering; (ii) Client Selection; (iii) Model Aggregation. Next, we will introduce each of these three parts in detail.
(i)
Client Clustering
Each client k calculates the frequency distribution of different ratings based on its local data and sends this distribution to the server. Upon receiving the rating frequency distributions from each client, the server uses these rating frequency distributions Q k as the basis for clustering, dividing the clients into c different clusters, denoted as C = C 1 , C 2 , , C c . Specifically, the server computes the Euclidean distance between the rating frequency distributions Q k of each client and uses a K-means clustering algorithm to categorize the clients into c clusters. The rating frequency distributions collected by the server do not include the historical location ids of user interactions, which means that the server cannot determine at which locations the users have rated. The Euclidean distance with respect to different rating frequency distributions Q k is defined as follows:
dist Q i , Q j = d = 1 D Q i d Q j d 2 .
(ii)
Client Selection
In the client selection phase, we primarily base the selection on model similarity and local loss. Specifically, for client k within a cluster, the model similarity is measured by calculating the cosine similarity between the local model and the global model. The cosine similarity ranges from −1 to 1. A value close to 1 indicates that the angle between the two vectors is close to 0 degrees, suggesting that the two models are very similar. Unlike the Euclidean distance metric, cosine similarity emphasizes consistency in the direction of the parameters rather than differences in numerical magnitude or location. The cosine similarity between the local model and the global model is calculated using the following formula:
cos w k t , w G t = w k t , w G t w k t w G t ,
where w k t is the local model of client k, w G t is the global model, and  cos ( · , · ) is the cosine similarity function. The larger the cosine similarity, the more the local model w k t tends to favor the global model w G t in direction. Local loss is the loss value computed by the client based on its own data. For cluster C i C , the server selects clients using a proportional method based on two metrics: the local loss uploaded by the clients and the model similarity. Algorithm 2 summarizes the process of client selection by the server.
Algorithm 2 Client Selection
Input: 
Client model similarity set sim k k C i , Client loss set loss k k C i , Selection of client ratio based on model similarity f 1 , Selection of client ratio based on loss f 2
Output: 
Selected client set S e l C i = { S e l e c t e d }
1:
sim k k C i Sort sim k k C i // Sort the client model similarity set
2:
loss k k C i Sort loss k k C i // Sort the client loss set
3:
sim * f 1 · sim k k C i // Select clients with high similarity in the top f 1
4:
if loss k sim *  then
5:
    loss loss k
6:
end if
7:
loss * f 2 · loss // Select clients with high loss in the top f 2
8:
{ S e l e c t e d } { s i m * } + { l o s s * }
When selecting clients, a strategy of choosing clients with larger losses can be used to quickly optimize the weakest parts of the model, promoting overall convergence. Clients with larger losses typically indicate that the model performs poorly on their data. Therefore, by training more on these clients, the model can learn features that are difficult to capture. However, only selecting clients with larger losses for aggregation may lead to a decline in model performance. To address this issue, we can use model similarity to guide the client selection process. High model similarity means that the direction of the local model is closer to the global model. This indicates that these clients typically have better local data representation and training effectiveness. Therefore, selecting clients with high model similarity to participate in global model updates can improve the overall model performance.
In addition, client selection on the basis of clustering has the following advantage: by clustering the frequency distribution of clients’ ratings, clients with similar rating habits are grouped into the same cluster. This operation allows for the clustering of clients that may have similar data distributions into a single cluster. Subsequently, representative clients are selected from the cluster based on the metrics. This approach ensures that a variety of data types are included in each iteration and mitigates the effects of non-IID data.
(iii)
Model Aggregation
In the model aggregation phase, we use the REPAgg aggregation method for global model aggregation. Unlike FedIS’s REPAgg-only approach, while REPAgg can mitigate the impact of non-IID data on model performance to some extent, it is not as effective when dealing with data with high heterogeneity. To address this issue, we combine client clustering, a client selection method based on loss function and model similarity, with REPAgg. Specifically, by clustering clients, FedLSM-LPR groups clients with similar data distributions and selects representative clients within each cluster to participate in model aggregation. This approach can better handle the differences in individual client data, thus effectively reducing the impact of non-IID data on model performance. We implement client clustering and selection methods, which necessitates modifying Formulas (7) and (8). Below are the revised formulas:
M t + 1 = M t + λ · C i C k Sel C i Δ k t m 1 , k Sel C i Δ k t m 2 , , k Sel C i Δ k t m q ,
N t + 1 = N t + λ · C i C k Sel C i Δ k t n 1 , k Sel C i Δ k t n 2 , , k Sel C i Δ k t n q ,
where Δ k t m i and Δ k t n i are updates to the location feature vectors, C = C 1 , C 2 , , C c is the set with c client clusters, and λ denotes the learning rate.

4.4. Two-Stage Perturbation

For federated learning location recommendation, if the model updates received by the server are not properly processed or encrypted, then malicious server administrators or third parties may infer user interaction information from these model updates through reverse engineering techniques. Therefore, we use a two-stage perturbation approach to protect the model updates uploaded from the client to the server to ensure the security and privacy of user data. Two-stage perturbation consists of label perturbation and model update perturbation.
In the client training phase, label perturbation is used to flip some of the basic fact labels by using randomized answer methods. Randomized answer methods can be used to protect input data in sensitive domains. We perform label perturbation by using the following formula:
ψ y i * = 1 = φ 0 , y i = 0 φ 1 , y i = 1 ,
where y i is the true label, φ 0 and φ 1 are the perturbation rates of y i = 0 and y i = 1 , φ 1 = 1 φ 0 , y i * is the perturbation labels; and ψ y i * = 1 is the probability of y i * = 1 .
Label perturbation perturbs the client’s raw data, which in turn perturbs the client’s model updates. Despite the above measures, the direct uploading of model updates from the client to the server may still lead to the leakage of user privacy. In order to further improve the level of client privacy protection, we use the model update perturbation technique to perturb the model updates, as shown in the following formula:
Δ ¯ k t = Δ k t · Bernoulli φ 2 ,
where φ 2 is the model update perturbation rate, B e r n o u l l i ( · ) is the Bernoulli function that generates the Bernoulli matrix with a certain stochastic probability and has the same shape as Δ k t , and Δ k t is the client’s model update.

4.5. Privacy Analysis

We will use differential privacy [41] to provide formal privacy protection for label perturbation and model update perturbation. The definition of local differential privacy is as follows:
Definition 1.
There are two adjacent datasets W and W , as well as a mechanism M that outputs certain results. For any output set Y, ( ϵ , δ ) -differential privacy requires the following inequality to hold:
Pr [ M ( W ) = Y ] e ϵ Pr [ M ( W ) = Y ] + δ ,
where Pr [ M ( W ) = Y ] and Pr [ M ( W ) = Y ] denote the probability that mechanism M outputs a certain outcome Y on dataset W and W , ϵ is the privacy budget, δ is an additional parameter that controls the probability of allowed privacy leakage.
By controlling the size of ϵ , it is possible to adjust the balance between privacy protection and accuracy of data analysis results. A smaller ϵ provides stronger privacy protection but may reduce query accuracy; a larger ϵ provides weaker privacy protection but better preserves the precision of the data. Label perturbation uses instantaneous random responses to perturb the training labels and satisfies differential privacy [42]. The theorem is as follows:
Theorem 1.
The label perturbation method satisfies ( ϵ , 0 ) -local difference privacy and its privacy budget ϵ can be expressed as follows:
ϵ = ln 1 φ 0 φ 0 ,
where φ 0 is the perturbation rate, which controls the intensity of the perturbation.
The above formula shows that there is a relationship between privacy budget ϵ and perturbation rate φ 0 . The higher the perturbation rate, the lower the privacy budget. In addition, the model update perturbation also ensures differential privacy [43], and the specific theorem is as follows:
Theorem 2.
For dataset V = { v 1 , v 2 , , v n } , we define s ( j ) = min k [ n ] 1 n i = 1 , i k n v i ( j ) and Γ = min j [ φ 2 ] s ( j ) . Then, ( ϵ m , δ m ) -local differential privacy is satisfied on V, and the model update perturbation is conditioned on:
δ m = φ 2 · e Ω ( ϵ m 2 Γ n ) ,
where φ 2 is the model update perturbation rate.
According to the extended optimal combination theorem [44], the combination of label perturbation and model update perturbation still satisfies the differential privacy requirement. The theorem is shown below:
Theorem 3.
Suppose there is a sequence of randomization algorithms ( M 1 , M 2 , , M n ) , each of which M j is capable of guaranteeing ( ϵ j , δ j ) -differential privacy. Under these conditions, the entire sequence of algorithms will satisfy j = 1 n ϵ j , δ w -differential privacy, where the total privacy loss δ w can be calculated by the following formula:
δ w = n e j = 1 n ϵ j max { δ j } j = 1 n .
We control the privacy intensity by adjusting the label perturbation rate and the model update perturbation rate. In this process, it is crucial to introduce perturbations rationally, both to ensure the effectiveness of the model and to ensure that data privacy is effectively protected.

5. Experiment

5.1. Experimental Settings

5.1.1. Dataset

In this study, the Foursquare dataset is utilized as the foundation for our experiments. This dataset encompasses 2,153,471 users, 1,143,092 locations, 1,021,970 check-ins, 27,098,490 social connections, and 2,809,581 user ratings. We selected the segment of the dataset that pertains to New York, USA, covering user check-ins from 12 April 2012 to 16 February 2013, approximately 10 months. This portion provides detailed records of user check-ins in New York City, including related business information and social activities across various locations such as restaurants, cafes, museums, and more. To ensure the accuracy of our analysis, we excluded users with fewer than five check-ins. After filtering, the dataset contains 1083 users and 38,333 unique locations, totaling 91,024 interactions between users and locations, with a data sparsity of 0.9978.

5.1.2. Evaluation Metrics

In terms of performance evaluation, this study employs two classical evaluation metrics for location recommendation systems: Hit Rate (HR) and Normalized Discounted Cumulative Gain (NDCG). These metrics measure whether the locations of interest to users appear at the top positions of a given recommendation list. NDCG is an evaluation metric used for ranking and recommendation systems, taking into account the position and relevance of the locations. We report the HR and NDCG at X = 10 . For testing, the data from the client’s last interaction is used, along with all previous interaction data for training. Additionally, we randomly select 99 locations from those that have not interacted with the client and add them to the test set for evaluation.

5.1.3. Parameter Settings

In our experiments, we set the number of sampled clients in federated learning K to 100, the local learning rate η to 0.01, the global model learning rate λ in model aggregation to 0.9, the number of local training iterations E to 5, the number of global training iterations T to 1000, the clustering size c to 4, client selection parameter f 1 based on model similarity is set to 50%, client selection parameter f 2 based on loss function is set to 50%, label perturbation rate φ 0 is set to 0.1, and model update perturbation rate φ 2 is set to 0.6. We searched for values of τ from {0.001, 0.01, 0.1, 0.5} and ultimately determined that the model performs best when τ is 0.01.

5.1.4. Baseline Schemes

In order to demonstrate the effectiveness of the FedLSM-LPR scheme, the following eight schemes are selected for comparison in this paper:
  • MF-ALS [45]: This scheme builds on the traditional matrix factorization algorithm by considering the problem of modeling the user’s implicit feedback and using singular value decomposition inside the implicit feedback dataset.
  • NAIS [39]: This scheme introduces an attention mechanism to compute the attention weights between locations, which improves the recommendation performance based on the location similarity model.
  • FCF [28]: This scheme integrates CF and federated learning for the first time, and at the same time confirms the applicability of federated learning in the field of personalized recommendation, which lays the foundation for subsequent research.
  • FedMF [29]: This scheme is a privacy-preserving recommendation system based on security matrix factorization, which protects the user’s privacy and security by using federated learning and homomorphic encryption.
  • FedNCF [30]: This scheme uses neural CF to generate high-quality recommendations and employs SecAvg, a secure aggregation protocol, to protect the security of user privacy.
  • FedBPR [46]: This scheme introduces a factorization model based on matrix factorization within the federated learning framework, allowing users to retain control over their data, thereby effectively enhancing data privacy protection.
  • FedVAE [47]: This scheme combines a variational auto-encoder(VAE) with federated learning techniques to build a distributed CF recommendation system by learning deep feature representations of users and locations on individual clients.
  • FedIS [35]: This scheme proposes a novel federated learning aggregation method, REPAgg, to address the heterogeneity of data characteristics across different clients in federated learning.

5.2. Experimental Results and Comparative Analysis

This article provides a comprehensive comparison of various solutions, with their performance detailed in Table 2.
The analysis of the experimental data leads us to the following conclusions:
By comparing two solutions using centralized data storage: MF-ALS and NAIS, the experimental results indicate that the NAIS solution, which employs attention mechanisms and neural networks, outperforms the traditional matrix factorization MF-ALS solution. This result suggests that NAIS, by leveraging neural networks to learn positional similarity, can uncover more complex nonlinear relationships and latent patterns, making it more effective than traditional models when handling complex data. However, it is important to note that both solutions utilize centralized data storage, which may pose a risk to user privacy.
In federated learning schemes, the FedVAE scheme demonstrates superior performance compared to traditional matrix factorization-based recommendation schemes by utilizing VAE. As a powerful generative model, VAE can learn the latent distribution of input data and generate new data instances. This characteristic enables FeVAE to more effectively handle various data types and complex data distributions.
The FedIS scheme was compared with two schemes for centralized storage, MF-ALS and NAIS. The results show that FedIS does not differ much from MF-ALS in two performance metrics, HR@10 and NDCG@10, but FedIS performs significantly worse than NAIS in the comparison with NAIS. This is due to the fact that NAIS employs a centralized data storage, which enables direct access to the complete global data, and utilizes the attention mechanism and neural network to fully explore the user-item complex interactions between users and items, thus obtaining more accurate recommendation results. In contrast, FedIS, as a federated learning scheme, relies on decentralized local data for its training process and cannot directly share the original data, so it has a natural disadvantage in data utilization efficiency. A deeper look reveals that our proposed FedLSM-LPR outperforms the NAIS scheme in two performance metrics, HR@10 and NDCG@10. The main reasons are as follows: (1) Through clustering-based client selection, clients with similar data distributions are grouped together, making the data within each client group in each training round more homogeneous and significantly reducing gradient conflict during aggregation. In contrast, although NAIS can utilize centralized data for global optimization, it cannot adapt to the heterogeneous nature of distributed data that naturally exists in real-world scenarios. (2) By analyzing the training loss and model similarity of clients, it prioritizes clients with consistent and reliable contributions to participate in aggregation, effectively filtering out low-quality or malicious local updates, whereas centralized NAIS lacks this ability to dynamically assess data quality. (3) Introducing a regularization term in the client loss function to force local training to balance personalization and global consistency. This design retains the privacy advantage of federated learning and circumvents the problem of global performance degradation due to local overfitting in traditional federated learning. The centralized training of NAIS is unable to achieve this distributed regularization effect, although it can directly optimize the global objective.
The FedIS scheme demonstrates superior performance compared to other federated learning baseline schemes. This is primarily because, unlike FedIS, the other schemes only use a simple weighted average aggregation method on the server side, failing to fully consider the impact of data heterogeneity on model performance in federated learning for location recommendation tasks. Additionally, the FedIS scheme employs a more advanced NAIS model on the client side, effectively enhancing the model’s recommendation performance.
We compared the scheme proposed in this paper with the FedIS scheme and found that our proposed scheme exhibits significant advantages based on the experimental results. FedIS only adopts the REPAgg aggregation method, which alleviates the impact of data heterogeneity on federated learning location recommendation systems to a certain extent, but its recommendation accuracy may decrease when the data from different clients varies greatly, and it cannot adequately cope with complex real-world data scenarios. Our proposed FedLSM-LPR scheme improves the deficiencies in FedIS by introducing a client regularization method. This regularization technique helps to control the discrepancy between the local model and the global model, especially when facing highly heterogeneous datasets, and can effectively reduce the excessive deviation of the local model from the global model.FedLSM-LPR also further optimizes the client-side clustering and selection strategies. Based on clustering, the scheme selects clients to participate in aggregation by evaluating the loss of clients and the similarity between models, thus ensuring that the data features within each subgroup are relatively consistent. This approach allows models to be aggregated in such a way that updates do not cancel each other out due to excessive data differences, thus improving the efficiency of aggregation. In addition, selecting high-loss clients is equivalent to automatically identifying the “shortcomings” of the current model during the training process and focusing resources on optimizing the poorly performing local models. At the same time, choosing clients with high model similarity ensures the stability of the update and prevents individual abnormal clients from deviating from the global model.
Figure 3 and Figure 4 compare the performance of FedLSM-LPR and FedIS on the evaluation dataset as the number of federated learning epochs increases. The figure shows that as the number of epochs increases, the global model better captures the location features distributed across various edge clients, leading to improved performance. During the initial training phase, both FedLSM-LPR and FedIS exhibit rapid performance improvements. However, FedLSM-LPR converges significantly faster, achieving high performance with fewer epochs. This indicates that FedLSM-LPR can more efficiently learn and integrate local data from different clients.

5.3. Model Efficiency

FedLSM-LPR and FedIS are trained from the same initial model, and both eventually converge. Under identical training settings, as illustrated in Figure 3 and Figure 4, FedLSM-LPR achieves convergence with fewer epochs and outperforms FedIS in terms of recommendation performance. Thanks to the clustering-based client selection method and the L2 norm regularization term, the model achieves stable performance in fewer training epochs.
In addition, we calculated the communication cost of FedLSM-LPR. Compared to FedIS’s 952 KB, our scheme achieves a lower communication cost of 933 KB. This advantage is due to the introduction of a clustering-based client selection mechanism, which selects only representative clients for training in each round, effectively reducing communication burden while ensuring model performance.

5.4. Analysis of Parameters

In this paper, we analyze the impact of the main hyperparameters on the performance of FedLSM-LPR, which include, the number of local training iterations E, the percentage of clients selected based on model similarity f 1 , the percentage of clients selected based on loss f 2 , the label perturbation rate φ 0 , and the model updating perturbation rate φ 2 . By tuning these hyperparameters, we can investigate their impact on the performance of the scheme.
In FedLSM-LPR, the hyperparameter E is used to control the number of iterations during local training for each client. The client receives the global model from the server in each round of federated learning and then performs E iterations locally to update the local model parameters. After completing E iterations, the client uploads the model update parameters to the server for the next round of federated learning. This hyperparameter can affect how much each client contributes to the model in each round of federated learning, which in turn affects the overall model performance and speed of convergence. A larger value of E results in each client training more times, resulting in a more adequate update of the model in each round of federated learning, which helps to increase the convergence speed of the model. However, a larger E value means that more model parameters need to be transmitted per training round, increasing the communication overhead. In addition, too many iterations per client in local training may lead to overfitting on local training data, reducing the generalization ability of the model. Conversely, a smaller value of E may result in insufficient updates during each training round, delaying the model’s convergence and affecting its performance. In addition, insufficient contribution from each client may lead to underutilization of some client data, affecting the overall model performance improvement. Therefore, in this paper, we find a balance point by adjusting the size of E to fit the FedLSM-LPR scheme. As shown in Figure 5, the FedLSM-LPR scheme obtains the highest performance when E is 5.
In distributed learning systems, the values of hyperparameters f 1 and f 2 play a key role. f 1 denotes the client with high similarity between the local and global models in the pre-selection f 1 . f 2 denotes the client with high local loss in the pre-selection f 2 . If these parameters are set too small, it may make it difficult for the model to perform well in environments containing heterogeneous data. This is because not enough samples can be captured to reflect the global data characteristics. Conversely, if these parameters are set too large, this will incur more communication costs and potentially increase the proportion of noisy and unreliable data, thus reducing model performance. Therefore, choosing the right f 1 and f 2 is crucial for obtaining good model performance. In this paper, we keep f 1 at 50% to analyze the effect of different f 2 on the model performance. In addition, when the parameter f 1 is set to 100%, all clients are selected to participate in the global model aggregation. This is a special case, i.e., there is no additional processing for client selection. Figure 6 shows that FedLSM-LPR obtains the highest performance when parameter f 1 is set to 50% and f 2 is set to 50%.
To more effectively protect user privacy and prevent information leakage, this paper adopts a two-stage perturbation method. Parameter φ 0 represents the client label perturbation rate, while parameter φ 2 represents the model update perturbation rate. These two parameters respectively control the degree of perturbation applied to client data and model updates. This study analyzes the impact of parameters φ 0 and φ 2 on the performance of FedLSM-LPR by varying their values. From Figure 7, it can be observed that when the model update perturbation rate remains constant, the performance of FedLSM-LPR significantly decreases as the label perturbation rate increases. This decline occurs because label perturbation alters user information, negatively impacting FedLSM-LPR’s performance. On the other hand, when the label perturbation rate is kept constant, the performance of FedLSM-LPR first decreases and then increases as the model update perturbation rate rises. This is because a higher model update perturbation rate results in more parameters being discarded, which weakens the impact of label perturbation. Excessive perturbation can effectively reduce the risk of user privacy breaches but inevitably harm FedLSM-LPR’s performance. Conversely, if the perturbation rate is set too low, the model’s performance level can be maintained, but user privacy is exposed to higher risks and becomes more vulnerable to malicious attackers. Therefore, balancing privacy protection and model performance is crucial.

6. Conclusions

LBSNs greatly enhance users’ lives. However, due to privacy concerns and stringent legal regulations, location recommendation faces substantial challenges. We propose a Federated Learning-Based Location Similarity Model for Location Privacy Preserving Recommendation (FedLSM-LPR) scheme. Unlike traditional centralized data storage, the FedLSM-LPR scheme utilizes users’ local data for collaborative learning, addressing privacy leakage issues. We employ a location similarity model on the client side and incorporate an L2 norm regularization term in the loss function. To protect client data and model updates, we use a two-stage perturbation method, ensuring user privacy. To mitigate the impact of non-IID data, we propose a clustering-based client selection method that selects representative clients within clusters for global model aggregation. Experimental results show that the FedLSM-LPR scheme outperforms other federated learning-based location recommendation schemes in terms of recommendation performance. Future research will focus on incorporating the temporal features of user-location interactions as these interactions change over time. This approach can enhance the model’s ability to predict user behavior and provide personalized recommendations. In addition, future work will focus on further enhancing privacy protection. We will focus on potentially malicious server attacks, such as those masquerading as a non-existent federation, data poisoning attacks, and member extrapolation, and design appropriate protection mechanisms against these attacks.

Author Contributions

Conceptualization, L.Z.; methodology, L.Z.; software, J.M.; validation, J.M.; formal analysis, Y.L.; investigation, L.Y.; resources, F.Z.; data curation, J.G.; writing—original draft preparation, L.Z. and J.M.; writing—review and editing, J.M.; visualization, L.Z.; supervision, F.Z.; project administration, L.Z.; funding acquisition, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key Technologies R&D Program of Henan Province (252102211088, 242102210135), in part by the Henan Postgraduate Joint Training Base Project (YJS2022JD08).

Data Availability Statement

The data used to support the findings of the study are available within the article.

Conflicts of Interest

Author Jingzhong Gu was employed by the company Shangu Cyber Security Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Bahgat, E.M.; Abo-alian, A.; Rady, S.; Gharib, T.F. Efficient Trajectory Prediction Using Check-In Patterns in Location-Based Social Network. Big Data Cogn. Comput. 2025, 9, 102. [Google Scholar] [CrossRef]
  2. Wang, H.; Xin, M. A Graph-Enhanced Dual-Granularity Self-Attention Model for Next POI Recommendation. Electronics 2025, 14, 1387. [Google Scholar] [CrossRef]
  3. Van Dat, N.; Van Toan, P.; Thanh, T.M. Solving distribution problems in content-based recommendation system with gaussian mixture model. Appl. Intell. 2022, 52, 1602–1614. [Google Scholar] [CrossRef]
  4. Fauzan, F.; Nurjanah, D.; Rismala, R. Apriori association rule for course recommender system. Indones. J. Comput. (Indo-JC) 2020, 5, 1–16. [Google Scholar]
  5. Lee, D.; Nam, K.; Han, I.; Cho, K. From free to fee: Monetizing digital content through expected utility-based recommender systems. Inf. Manag. 2022, 59, 103681. [Google Scholar] [CrossRef]
  6. Darraz, N.; Karabila, I.; El-Ansari, A.; Alami, N.; El Mallahi, M. Integrated sentiment analysis with BERT for enhanced hybrid recommendation systems. Expert Syst. Appl. 2025, 261, 125533. [Google Scholar] [CrossRef]
  7. Liu, Y.; Yin, M.; Zhou, X. A collaborative filtering algorithm with intragroup divergence for POI group recommendation. Appl. Sci. 2021, 11, 5416. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Yin, C.; Wu, Q.; He, Q.; Zhu, H. Location-aware deep collaborative filtering for service recommendation. IEEE Trans. Syst. Man Cybern. Syst. 2019, 51, 3796–3807. [Google Scholar] [CrossRef]
  9. Albrecht, J.P. How the GDPR will change the world. Eur. Data Prot. Law Rev. 2016, 2, 287. [Google Scholar] [CrossRef]
  10. Gui, R.; Gui, X.; Zhang, X. A trajectory privacy protection method based on the replacement of points of interest in hotspot regions. Comput. Secur. 2025, 150, 104279. [Google Scholar] [CrossRef]
  11. Li, L.; Lin, F.; Xiahou, J.; Lin, Y.; Wu, P.; Liu, Y. Federated low-rank tensor projections for sequential recommendation. Knowl.-Based Syst. 2022, 255, 109483. [Google Scholar] [CrossRef]
  12. Adnan, M.; Syed, M.H.; Anjum, A.; Rehman, S. A Framework for privacy-preserving in IoV using Federated Learning with Differential Privacy. IEEE Access 2025, 13, 13507–13521. [Google Scholar] [CrossRef]
  13. Banabilah, S.; Aloqaily, M.; Alsayed, E.; Malik, N.; Jararweh, Y. Federated learning review: Fundamentals, enabling technologies, and future applications. Inf. Process. Manag. 2022, 59, 103061. [Google Scholar] [CrossRef]
  14. Zhong, L.; Zeng, J.; Wang, Z.; Zhou, W.; Wen, J. SCFL: Spatio-temporal consistency federated learning for next POI recommendation. Inf. Process. Manag. 2024, 61, 103852. [Google Scholar] [CrossRef]
  15. Cong, Y.; Zeng, Y.; Qiu, J.; Fang, Z.; Zhang, L.; Cheng, D.; Liu, J.; Tian, Z. FedGA: A greedy approach to enhance federated learning with Non-IID data. Knowl.-Based Syst. 2024, 301, 112201. [Google Scholar] [CrossRef]
  16. Hu, H.; Kothari, A.N.; Banerjee, A. A Novel Algorithm for Personalized Federated Learning: Knowledge Distillation with Weighted Combination Loss. Algorithms 2025, 18, 274. [Google Scholar] [CrossRef]
  17. Pei, J.; Liu, W.; Li, J.; Wang, L.; Liu, C. A review of federated learning methods in heterogeneous scenarios. IEEE Trans. Consum. Electron. 2024, 70, 5983–5999. [Google Scholar] [CrossRef]
  18. Sun, L.; Zheng, Y.; Lu, R.; Zhu, H.; Zhang, Y. Towards privacy-preserving category-aware POI recommendation over encrypted LBSN data. Inf. Sci. 2024, 662, 120253. [Google Scholar] [CrossRef]
  19. Li, M.; Zheng, W.; Xiao, Y.; Zhu, K.; Huang, W. Exploring temporal and spatial features for next POI recommendation in LBSNs. IEEE Access 2021, 9, 35997–36007. [Google Scholar] [CrossRef]
  20. Zeng, J.; Tao, H.; Tang, H.; Wen, J.; Gao, M. Global and local hypergraph learning method with semantic enhancement for POI recommendation. Inf. Process. Manag. 2025, 62, 103868. [Google Scholar] [CrossRef]
  21. Do, P.M.T.; Nguyen, T.T.S. Semantic-enhanced neural collaborative filtering models in recommender systems. Knowl.-Based Syst. 2022, 257, 109934. [Google Scholar] [CrossRef]
  22. Ye, M.; Yin, P.; Lee, W.C.; Lee, D.L. Exploiting geographical influence for collaborative point-of-interest recommendation. In Proceedings of the 34th international ACM SIGIR Conference on Research and Development in Information Retrieval, Beijing, China, 24–28 July 2011; pp. 325–334. [Google Scholar]
  23. Zheng, Y.; Xie, X. Learning location correlation from gps trajectories. In Proceedings of the 2010 Eleventh International Conference on Mobile Data Management, Kansas City, MO, USA, 23–26 May 2010; pp. 27–32. [Google Scholar]
  24. Liu, B.; Lv, N.; Guo, Y.; Li, Y. Recent advances on federated learning: A systematic survey. Neurocomputing 2024, 597, 128019. [Google Scholar] [CrossRef]
  25. Xie, Q.; Jiang, S.; Jiang, L.; Huang, Y.; Zhao, Z.; Khan, S.; Dai, W.; Liu, Z.; Wu, K. Efficiency optimization techniques in privacy-preserving federated learning with homomorphic encryption: A brief survey. IEEE Internet Things J. 2024, 11, 24569–24580. [Google Scholar] [CrossRef]
  26. Gunasekaran, E.; Muthuraman, V. A new double layer multi-secret sharing scheme. China Commun. 2024, 21, 287–309. [Google Scholar] [CrossRef]
  27. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; Arcas, B.A.y. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  28. Ammad-Ud-Din, M.; Ivannikova, E.; Khan, S.A.; Oyomno, W.; Fu, Q.; Tan, K.E.; Flanagan, A. Federated collaborative filtering for privacy-preserving personalized recommendation system. arXiv 2019, arXiv:1901.09888. [Google Scholar]
  29. Chai, D.; Wang, L.; Chen, K.; Yang, Q. Secure federated matrix factorization. IEEE Intell. Syst. 2020, 36, 11–20. [Google Scholar] [CrossRef]
  30. Perifanis, V.; Efraimidis, P.S. Federated neural collaborative filtering. Knowl.-Based Syst. 2022, 242, 108441. [Google Scholar] [CrossRef]
  31. Huang, J.; Tong, Z.; Feng, Z. Geographical POI recommendation for Internet of Things: A federated learning approach using matrix factorization. Int. J. Commun. Syst. 2022, 35, e5161. [Google Scholar] [CrossRef]
  32. Lin, G.; Liang, F.; Pan, W.; Ming, Z. FedRec: Federated Recommendation With Explicit Feedback. IEEE Intell. Syst. 2021, 36, 21–30. [Google Scholar] [CrossRef]
  33. Dong, Q.; Liu, B.; Zhang, X.; Qin, J.; Wang, B. Sequential POI Recommend Based on Personalized Federated Learning. Neural Process. Lett. 2023, 55, 7351–7368. [Google Scholar] [CrossRef]
  34. Li, Z.; Bilal, M.; Xu, X.; Jiang, J.; Cui, Y. Federated learning-based cross-enterprise recommendation with graph neural networks. IEEE Trans. Ind. Inform. 2022, 19, 673–682. [Google Scholar] [CrossRef]
  35. Ding, X.; Li, G.; Yuan, L.; Zhang, L.; Rong, Q. Efficient federated item similarity model for privacy-preserving recommendation. Inf. Process. Manag. 2023, 60, 103470. [Google Scholar] [CrossRef]
  36. Sarwar, B.; Karypis, G.; Konstan, J.; Riedl, J. Item-based collaborative filtering recommendation algorithms. In Proceedings of the 10th International Conference on World Wide Web, Hong Kong, China, 1–5 May 2001; pp. 285–295. [Google Scholar]
  37. Ning, X.; Karypis, G. Slim: Sparse linear methods for top-n recommender systems. In Proceedings of the 2011 IEEE 11th International Conference on Data Mining, Vancouver, BC, Canada, 11–14 December 2011; pp. 497–506. [Google Scholar]
  38. Kabbur, S.; Ning, X.; Karypis, G. Fism: Factored item similarity models for top-n recommender systems. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, London, UK, 19–23 August 2013; pp. 659–667. [Google Scholar]
  39. He, X.; He, Z.; Song, J.; Liu, Z.; Jiang, Y.G.; Chua, T.S. NAIS: Neural attentive item similarity model for recommendation. IEEE Trans. Knowl. Data Eng. 2018, 30, 2354–2366. [Google Scholar] [CrossRef]
  40. Nichol, A.; Achiam, J.; Schulman, J. On first-order meta-learning algorithms. arXiv 2018, arXiv:1803.02999. [Google Scholar]
  41. Dwork, C.; Roth, A. The algorithmic foundations of differential privacy. Found. Trends® Theor. Comput. Sci. 2014, 9, 211–407. [Google Scholar] [CrossRef]
  42. Jiang, J.Y.; Li, C.T.; Lin, S.D. Towards a more reliable privacy-preserving recommender system. Inf. Sci. 2019, 482, 248–265. [Google Scholar] [CrossRef]
  43. Jain, P.; Kulkarni, V.; Thakurta, A.; Williams, O. To drop or not to drop: Robustness, consistency and differential privacy properties of dropout. arXiv 2015, arXiv:1503.02031. [Google Scholar]
  44. Vadhan, S.; Wang, T. Concurrent composition of differential privacy. In Proceedings of the Theory of Cryptography: 19th International Conference, TCC 2021, Raleigh, NC, USA, 8–11 November 2021; Proceedings, Part II. Springer: Cham, Switzerland, 2021; pp. 582–604. [Google Scholar]
  45. Hu, Y.; Koren, Y.; Volinsky, C. Collaborative filtering for implicit feedback datasets. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 263–272. [Google Scholar]
  46. Anelli, V.W.; Deldjoo, Y.; Di Noia, T.; Ferrara, A.; Narducci, F. User-controlled federated matrix factorization for recommender systems. J. Intell. Inf. Syst. 2022, 58, 287–309. [Google Scholar] [CrossRef]
  47. Polato, M. Federated variational autoencoder for collaborative filtering. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 20 September 2021; pp. 1–8. [Google Scholar]
Figure 1. Basic procedure of federated learning.
Figure 1. Basic procedure of federated learning.
Electronics 14 02578 g001
Figure 2. Overall overview of the FedLSM-LPR scheme.
Figure 2. Overall overview of the FedLSM-LPR scheme.
Electronics 14 02578 g002
Figure 3. HR@10 for each Epoch on the evaluation dataset for the FedLSM-LPR and FedIS schemes.
Figure 3. HR@10 for each Epoch on the evaluation dataset for the FedLSM-LPR and FedIS schemes.
Electronics 14 02578 g003
Figure 4. NDCG@10 for each Epoch on the evaluation dataset for the FedLSM-LPR and FedIS schemes.
Figure 4. NDCG@10 for each Epoch on the evaluation dataset for the FedLSM-LPR and FedIS schemes.
Electronics 14 02578 g004
Figure 5. Performance of FedLSM-LPR scheme under different number of iterations of local training.
Figure 5. Performance of FedLSM-LPR scheme under different number of iterations of local training.
Electronics 14 02578 g005
Figure 6. Performance of FedLSM-LPR for different loss based client selection rates.
Figure 6. Performance of FedLSM-LPR for different loss based client selection rates.
Electronics 14 02578 g006
Figure 7. Effect of different perturbation rates on model performance.
Figure 7. Effect of different perturbation rates on model performance.
Electronics 14 02578 g007
Table 1. Glossary of notations.
Table 1. Glossary of notations.
SymbolDefinition
p , q Number of users and locations
u , U User and the user set
i , L Location and the location set
L u + Historical interacted locations of user u
M , N Location feature matrix
m i , n i Location feature vector of location i
ENumber of iterations for local training
TNumber of global training iterations
y ^ u i Prediction of user u for location i
Δ k t Updates of client k at tth epoch
w G t Global model parameters at tth epoch
w k t Local model parameters at tth epoch
Table 2. Performance of different schemes on Foursquare.
Table 2. Performance of different schemes on Foursquare.
SchemesHR@10NDCG@10
MF-ALS0.42940.2751
NAIS0.43030.2773
FCF0.42850.2748
FedMF0.42750.2749
FedNCF0.41090.2560
FedBPR0.42570.2754
FedVAE0.42840.2756
FedIS0.42940.2758
FedLSM-LPR0.43420.2776
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, L.; Mu, J.; Yu, L.; Liu, Y.; Zhu, F.; Gu, J. Federated Learning-Based Location Similarity Model for Location Privacy Preserving Recommendation. Electronics 2025, 14, 2578. https://doi.org/10.3390/electronics14132578

AMA Style

Zhu L, Mu J, Yu L, Liu Y, Zhu F, Gu J. Federated Learning-Based Location Similarity Model for Location Privacy Preserving Recommendation. Electronics. 2025; 14(13):2578. https://doi.org/10.3390/electronics14132578

Chicago/Turabian Style

Zhu, Liang, Jingzhe Mu, Liping Yu, Yanpei Liu, Fubao Zhu, and Jingzhong Gu. 2025. "Federated Learning-Based Location Similarity Model for Location Privacy Preserving Recommendation" Electronics 14, no. 13: 2578. https://doi.org/10.3390/electronics14132578

APA Style

Zhu, L., Mu, J., Yu, L., Liu, Y., Zhu, F., & Gu, J. (2025). Federated Learning-Based Location Similarity Model for Location Privacy Preserving Recommendation. Electronics, 14(13), 2578. https://doi.org/10.3390/electronics14132578

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop