Next Article in Journal
Binary Secretary Bird Optimization Algorithm for the Set Covering Problem
Previous Article in Journal
Recovery of Implied Volatility in a Spatial-Fractional Black–Scholes Equation Under a Finite Moment Log Stable Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Heuristic Approach to Competitive Facility Location via Multi-View K-Means Clustering with Co-Regularization and Customer Behavior

by
Thanathorn Phoka
1,2,†,
Praeploy Poonprapan
3 and
Pornpimon Boriwan
3,*,†
1
Department of Computer Science and Information Technology, Faculty of Science, Naresuan University, 99 Village No. 9, Tha Pho, Muang District, Phitsanulok 65000, Thailand
2
Center of Excellence in Nonlinear Analysis and Optimization, Naresuan University, 99 Village No. 9, Tha Pho, Muang District, Phitsanulok 65000, Thailand
3
Department of Mathematics, Faculty of Science, Khon Kaen University, 123 Village No. 16 Mittraphap Rd., Nai-Muang, Muang District, Khon Kaen 40002, Thailand
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2025, 13(15), 2481; https://doi.org/10.3390/math13152481 (registering DOI)
Submission received: 18 June 2025 / Revised: 27 July 2025 / Accepted: 30 July 2025 / Published: 1 August 2025
(This article belongs to the Section E: Applied Mathematics)

Abstract

Solving competitive facility location problems can optimize market share or operational efficiency in environments where multiple firms compete for customer attention. In such contexts, facility attractiveness is shaped not only by geographic proximity but also by customer preference characteristics. This study presents a novel heuristic framework that integrates multi-view K-means clustering with customer behavior modeling reinforced by a co-regularization mechanism to align clustering results across heterogeneous data views. By jointly exploiting spatial and behavioral information, the framework clusters customers and facilities into meaningful market segments. Within each segment, a bilevel optimization model is applied to represent the sequential decision-making of competing entities—where a leader first selects facility locations, followed by a reactive follower. An empirical evaluation on a real-world dataset from San Francisco demonstrates that the proposed approach, using optimal co-regularization parameters, achieves a total runtime of approximately 4.00 s—representing a 99.34% reduction compared to the full CFLBP-CB model (608.58 s) and a 99.32% reduction compared to a genetic algorithm (585.20 s). Concurrently, it yields an overall profit of 16,104.17, which is an approximate 0.72% increase over the Direct CFLBP-CB profit of 15,988.27 and is only 0.21% lower than the genetic algorithm’s highest profit of 16,137.75. Moreover, comparative analysis reveals that the proposed multi-view clustering with co-regularization outperforms all single-view baselines, including K-means, spectral, and hierarchical methods. This superiority is evidenced by an approximate 5.21% increase in overall profit and a simultaneous reduction in optimization time, thereby demonstrating its effectiveness in capturing complementary spatial and behavioral structures for competitive facility location. Notably, the proposed two-stage approach achieves high-quality solutions with significantly shorter computation times, making it suitable for large-scale or time-sensitive competitive facility planning tasks.

1. Introduction

The competitive facility location problem (CFLP) focuses on determining optimal sites for new facilities with the objective of maximizing the market share acquired by these facilities (refer to survey papers [1,2,3]). Various models addressing this problem can be categorized based on the nature of competition, the characteristics of the location space, and the behavior of customers.
The location space in the CFLP can take several forms: it may be discrete, with new facilities selected from a finite set of candidate sites [4]; structured as a network, allowing facility placement at network nodes or even along edges [5]; or continuous, permitting facilities to be established at any point within a defined region [6].
Customer behavior is another crucial factor in these models. Various customer response patterns are typically captured using attraction functions, which quantify how customers are influenced by competing facilities. The most widely used customer choice mechanisms are known as the proportional and binary rules [7]. Under the proportional rule, customers distribute their patronage among all facilities based on the relative attractiveness of each (see, for example, [6,8,9]). In contrast, the binary rule assumes that each customer chooses to patronize only the most attractive facility [4,10,11]. While these rules account for much of the observed customer behavior in competitive location models, in some cases, variations or combinations may be needed to better capture real-world complexities.
Furthermore, competition in facility location models can be classified as static, with foresight, or dynamic. In static competition, existing facilities owned by competitors are already present, and they do not adjust their positions when a new firm enters the market [12]. In the foresight scenario, potential competitors are expected to enter the market soon after the new facilities are established—the leader first selects locations to maximize market share, anticipating subsequent facility placements by the follower [13]. In dynamic competition, firms continually reassess and update their facility locations over time [14].
Among these types of competition, dynamic competition is particularly notable for the strategic interplay between rival firms, which is effectively modeled using bilevel optimization frameworks. In such models, a leader first makes facility location decisions to maximize market share or profitability, and a follower then responds by optimizing its own placements [15]. Bilevel modeling enables a more realistic representation of competitive dynamics compared to single-level formulations, making it particularly suitable for problems involving interdependent decision structures. To support complex decision-making in large-scale competitive facility location problems—particularly those involving extensive customer or facility datasets—clustering techniques such as K-means clustering, known for their simplicity and widespread use, are frequently employed. These methods partition the data into k clusters characterized by high intra-cluster similarity and low inter-cluster similarity, thereby simplifying the problem structure and enhancing computational efficiency.
Conventional K-means clustering is inherently constrained by its single-view assumption, treating all features with equal importance through uniform weighting. This simplification often fails to reflect the realities of complex datasets, where feature dimensions typically contribute unequally to the clustering structure. As a result, applying equal weights can obscure meaningful patterns and compromise clustering quality—particularly when some attributes are more informative than others. In the context of the CFLP, standard K-means clustering is commonly applied using either spatial coordinates or customer preferences, or by concatenating them into a single feature vector. Such an approach disregards the distinct nature and relative importance of each data view; yet both facilities and customers are defined by multiple complementary attributes. Overlooking this multi-view structure can result in suboptimal cluster formations and, consequently, inferior location decisions.
To overcome these limitations, multi-view clustering has emerged as a powerful paradigm in machine learning, enabling the integration of multiple heterogeneous data views into a unified clustering framework [16]. Multi-view clustering techniques enable a more comprehensive understanding of the fundamental framework of massive datasets, resulting in it being useful in image analysis, bioinformatics, and social network research. The initial research on multi-view clustering was originally performed by Bickel and Scheffer [17]. Throughout the span of two decades, various types of multi-view clustering techniques were provided considering the findings of that remarkable research [18].
Given its widespread adoption and computational efficiency, K-means clustering serves as a practical foundation for implementing multi-view clustering with co-regularization. This approach jointly leverages multiple feature spaces and encourages consistency across them through co-regularization mechanisms. While multi-view clustering has demonstrated strong empirical performance in various domains, its application to competitive facility location problems remains relatively unexplored, presenting a novel and promising direction for research.
This study addresses an important research gap by introducing a novel approach that integrates multi-view K-means clustering into a bilevel facility location framework. In particular, we incorporate geographic data, the preference views of customers, and facility data to create high-quality, behavior-aware clusters that serve as the basis for hierarchical optimization. This integration improves clustering quality and interpretability in several key ways:
  • It captures richer customer and facility representations by jointly leveraging geographic and preference data.
  • It reveals latent patterns that are not discernible within any single view—for example, clusters with distinct preference profiles despite spatial proximity.
  • It enforces alignment between views through co-regularization, thereby yielding clusters that are meaningful across multiple data dimensions.
  • It provides robustness against noise or missing values in individual views by using complementary information from others.
  • It supports decision interpretability by producing clusters grounded in both geographic and behavioral factors.
This research introduces a novel heuristic framework for the competitive facility location problem with customer behavior (CFLP-CB) by integrating multi-view K-means clustering with co-regularization to ensure consistency across multiple data views. By clustering facilities and customers based on heterogeneous feature spaces and applying bilevel optimization within each cluster, the framework effectively captures leader–follower dynamics while maintaining strategic realism and computational tractability. The proposed methodology is validated on a real-world dataset from San Francisco, incorporating synthetic behavioral data to reflect diverse customer preferences. Empirical results show that this approach not only significantly reduces computation time but also improves market share performance when compared to baseline K-means clustering combined with bilevel optimization. Overall, the framework provides a scalable, interpretable, and data-driven solution for the CFLP in complex, multi-attribute decision environments.
This research work is systematically organized to reflect the development of this approach: Section 2 reviews the relevant literature, identifying a gap in current methods where multi-view behavioral integration remains underexplored. Section 3 presents the problem definition and the proposed bilevel optimization model, which includes a behavior-aware attraction function that jointly considers spatial proximity and preference alignment. Section 4 describes the algorithmic framework, detailing the multi-view K-means clustering method with inter-view co-regularization and the bilevel optimization process for facility assignment. Section 5 presents a detailed case study of a retail franchise expansion in San Francisco where clustering is based on geographic and preference features. Section 6 discusses the strategic implications of the results, and Section 7 concludes the work with insights and directions for future research, such as incorporating real customer preference data and exploring advanced multi-objective extensions.

2. Related Work

The facility location problem has long been studied in both cooperative and competitive contexts. In competitive environments, firms must strategically select facility locations while anticipating the responses of their competitors. These interdependent decisions are best modeled using bilevel optimization frameworks, where a leader makes the first move and a follower responds. This sequential structure captures the asymmetric information, anticipatory behavior, and competitive dynamics inherent in real-world market entry.
The roots of competitive location modeling can be traced to Hotelling’s seminal 1929 work [19], which applied game theory to the analysis of spatial competition. Over the decades, this concept was extended through the development of leader–follower models, most notably the Stackelberg framework [20], which formalized sequential decision-making in competitive environments. A recent comprehensive survey by Drezner and Eiselt [21] offers a detailed classification of competitive location models, distinguishing between game-theoretic approaches (which jointly consider location and pricing decisions) and operations research models (which focus on optimizing new facility placement among existing competitors). Their review highlights the integration of realistic features such as facility attractiveness, customer choice behavior, price sensitivity, and market uncertainty. They emphasize that future research should continue to address challenges posed by customer heterogeneity, demand elasticity, and behavioral attraction rules, particularly in data-driven environments.
Building on these foundations, several studies have proposed bilevel models that incorporate increasingly realistic representations of customer decision behavior. For example, Biesinger et al. [22] formulated six bilevel models that reflect combinations of binary, proportional, and mixed attraction mechanisms under both essential and nonessential demand conditions. They solved these models using a hybrid evolutionary algorithm integrated with mixed-integer linear programming (MILP) evaluation techniques. Similarly, Casas-Ramírez et al. [23] proposed a hybrid heuristic approach to efficiently solve the bilevel p-median problem with ordered customer preferences, demonstrating improved solution quality and scalability compared to traditional exact methods. The key characteristics of recent competitive facility location studies are summarized in Table 1 to highlight their objectives, solution methods, and application domains.
These studies highlight both the expressiveness of bilevel optimization for modeling strategic competition and the computational burden it imposes when applied to large-scale, real-world datasets. To reduce this complexity, researchers have increasingly employed clustering-based heuristics, particularly K-means clustering, to partition the problem into smaller subproblems. While effective at improving tractability, most K-means-based models rely on a single view of the data, typically spatial proximity, and fail to incorporate multidimensional characteristics such as customer preferences, behavior, or demographics.
To address this limitation, the machine learning literature offers multi-view clustering (MVC)—a family of techniques designed to integrate heterogeneous data representations into a cohesive clustering framework. As reviewed by Zhou et al. [16], MVC methods such as co-regularized spectral clustering, graph learning, and contrastive representation learning have demonstrated strong performance in domains like bioinformatics, image recognition, and document analysis. These methods improve robustness, accuracy, and interpretability by leveraging both shared and unique features across views. Despite their maturity in data science applications, MVC techniques remain largely unexplored in operations research, particularly for CFLPs involving strategic interactions and customer diversity.
This research addresses that gap by proposing a novel heuristic method that integrates multi-view K-means clustering with inter-view co-regularization and a bilevel optimization process for facility assignment in CFLPs. The proposed approach first clusters customers and facilities using both geographic and preference data, and then applies bilevel optimization within each cluster to model leader–follower dynamics. By unifying modern clustering techniques with strategic optimization, this method offers a scalable, data-aware, and behaviorally realistic framework for location planning in competitive, real-world environments.

3. Problem Statement and Mathematical Model

3.1. Motivating Example

To motivate the creation of the proposed mathematical formulation, we begin with a simplified example that illustrates the core structure of the CFLP as a bilevel optimization problem.
Example 1. 
Consider a competitive market in which two firms seek to find optimal locations to serve customer demand. Figure 1 illustrates the data for this scenario: blue circles represent customer demand points, while red circles denote the set of candidate facility sites from which both firms can choose to establish their facilities. In this situation, the first firm to select its locations is designated as the leader, and the subsequent firm is referred to as the follower.
In the CFLP, formulated as a bilevel optimization problem, optimal facility locations are determined by maximizing a profit function based on customer demands for both the leader and follower. As an example, if both firms establish two facilities, the classical CFLP computes placement by minimizing costs, depicted by the shortest dashed lines in Figure 1b. These lines represent customer–facility assignments driven by objectives like market share, accessibility, or total travel cost. In contrast, our proposed method, illustrated in Figure 1a, first applies a clustering technique to group customer demands. Optimal facility locations are then selected from candidate sites within each cluster, subsequently enabling profit function computation.

3.2. Proposed Bilevel Optimization Model with Preference-Aware Attraction

This subsection introduces the notational framework adopted in this study and formally defines the CFLBP-CB. A complete list of notations used throughout the manuscript is provided in the Notation and Definition section. The CFLBP-CB addresses a market environment in which a new entrant, referred to as the leader, seeks to establish p facilities within a region already occupied by an incumbent competitor, regarded as the follower, who is allowed to respond optimally by locating r additional facilities. Both the leader and the follower select facilities from a predetermined set of potential locations, with neither party having fixed facility placements at the outset. Customers in this study are assumed to exhibit a proportional behavior rule, which means their demand is allocated among all available facilities.
The mathematical formulation of the CFLP under proportional customer behavior, originally introduced by Suárez-Vega et al. [34], presumes that customers distribute their demand among all accessible facilities in accordance with the proportional attractiveness of each facility, as depicted in Figure 1b. Kochetov et al. [24] further examined a specific variant characterized by proportional essential customer behavior and developed a bilevel mixed-integer nonlinear programming (MINLP) framework to model the hierarchical, sequential decision-making process involving a market-entering leader and a competitive follower. Motivated by the studies [24,34], we propose an enhanced bilevel model that integrates a preference-based dimension into the customer–facility assignment mechanism. To do this, we define p i j as a similarity parameter representing the preference alignment between facility i and customer j. This parameter is incorporated into the attraction coefficient v i j , enabling more accurate capture of the behavioral complexity in customer–facility relationships. As a result, the proposed formulation simultaneously considers spatial proximity and individual customer preferences. The proposed bilevel optimization model for the CFLBP-CB is detailed as follows.
Leader’s Problem (Upper Level):
max x j J w j · i I v i j x i i I v i j x i + i I v i j y i *
s . t . i I x i = p
x i { 0 , 1 } , i I
Here, y * = ( y 1 * , y 2 * , , y m * ) represents the optimal solution to the follower’s problem, defined below.
Follower’s Problem (Lower Level):
max y j J w j · i I v i j y i i I v i j x i + i I v i j y i
s . t . i I y i = r
y i { 0 , 1 } , i I
The objective functions (1) and (4) maximize the market share (i.e., fulfilled demand) for the leader and follower, respectively. The customer attraction coefficient is allocated based on the preference similarity coefficient, defined as follows:
v i j = p i j d i j + ϵ = m M min 1 , Preference m , i Preference m , j d i j + ϵ
where M denotes the set of preference dimensions (e.g., product categories, service features), and Preference m , i and Preference m , j represent the value of the mth preference dimension for facility i and customer j, respectively. Specifically, Preference m , i denotes the facility’s capability to serve the mth preference dimension of customer j, while Preference m , j reflects the importance or demand level of the mth preference dimension from the customer’s perspective. This formulation ensures that p i j captures the behavioral alignment between customer needs and facility offerings across all dimensions, while normalizing to avoid overweighting any single preference component. Here, d i j is the distance between facility i and customer j, p i j is the preference similarity ratio, and ϵ is a small positive constant used to avoid division by zero.
Constraints (2) and (5) ensure that the leader and follower open exactly p and r facilities, respectively. In the lower-level problem, the leader’s facility decisions x i are treated as fixed parameters.

3.3. Linear Reformulation of the Follower Problem

To efficiently solve the lower-level problem, we adopt a linearization strategy originally proposed by Kochetov et al. [24], which reformulates the nonlinear objective using a set of auxiliary variables. Specifically, two new types of variables are introduced: z j , which represents the inverse of the total attraction experienced by customer j, and y i j , which denotes the portion of demand from customer j served by the follower’s facility at location i. These are defined as follows:
z j = 1 i I v i j x i + i I v i j y i j J
y i j = w j z j v i j y i i I , j J
To avoid confusion, we clarify the roles of decision variables y i and y i j . The binary variable y i indicates whether the follower opens a facility at location i, while the continuous variable y i j represents the portion of customer j’s demand allocated to facility i, conditional on y i = 1 . The auxiliary variable z j serves as the reciprocal of the denominator in the original fractional objective, capturing the total attraction from both leader and follower facilities. These transformations enable a linear representation of the original nonlinear model, facilitating its reformulation as a mixed-integer linear program (MILP) for the follower’s subproblem given fixed decisions x i from the leader.
The resulting MILP is expressed as follows:
max j J i I y i j
s . t . i I y i j + w j z j i I v i j x i w j j J
y i j w j y i i I , j J
y i j w j v i j z j y i j + W ( 1 y i ) i I , j J
y i j 0 , z j 0 i I , j J
Constraints (11) ensure that the total fulfilled demand from customer j (by both the leader and the follower) does not exceed w j . Constraints (12) restrict y i j to zero when the follower does not open a facility at location i, while (13) enforces consistency with the nonlinear relation in (9). A sufficiently large constant W ensures that these constraints are valid for all feasible solutions.
To ensure the validity of the linearization, a sufficiently large constant W must be selected such that constraints (13) correctly enforce the conditional structure of Equation (9). Specifically, W must satisfy the following:
W max j J w j · max i I , j J v i j · max j J z j
where the term max j ( z j ) is bounded above by max j 1 / i I v i j x i , as implied by constraint (11). This ensures that the upper bound on y i j is valid when y i = 0 and sufficiently tight when y i = 1 . Choosing W conservatively guarantees that the MILP formulation remains equivalent to the original nonlinear model defined by Equations (1)–(4).

3.4. Theoretical Equivalence of the Reformulated Model

To ensure that the MILP, as (10)–(14), accurately represents and finds the optimal solution for the follower’s problem in (4)–(6), we provide the following theoretical results.
Lemma 1. 
Let ( y i * , y i j * , z j * ) be an optimal solution to the MILP defined by constraints (10)–(14) for a fixed leader decision variable x = ( x i ) i I { 0 , 1 } | I | . Then, for every customer j J , the auxiliary variable z j * satisfies
z j * = 1 i I v i j x i + i I v i j y i * .
Proof. 
From constraint (11), we have, for each j J ,
i I y i j * + w j z j * i I v i j x i w j .
From constraints (12) and (13), we also have
y i j * = w j v i j z j * y i * , i I , j J .
Substituting this into the inequality (16) gives
i I w j v i j z j * y i * + w j z j * i I v i j x i w j w j z j * i I v i j y i * + i I v i j x i w j .
Dividing both sides by w j > 0 , we obtain
z j * i I v i j x i + i I v i j y i * 1 ,
which implies
z j * 1 i I v i j x i + i I v i j y i * .
To complete the proof, we show that this inequality holds with equality at optimality. As the MILP objective is
j J i I y i j * = j J w j z j * i I v i j y i * ,
which is increasing in z j * , if z j * were strictly less than the upper bound above, then increasing z j * slightly would improve the objective without violating constraint (11), contradicting the optimality. Therefore, the inequality must be tight:
z j * = 1 i I v i j x i + i I v i j y i * , j J .
   □
Theorem 1. 
Let x = ( x i ) i I { 0 , 1 } | I | be a fixed leader decision satisfying constraint (2). Suppose that ( y i * , y i j * , z j * ) is an optimal solution to the MILP formulation defined by (10)–(14), along with the facility count constraint i I y i = r and integrality condition y i { 0 , 1 } . Then, the binary vector y * = ( y i * ) is an optimal solution to the original nonlinear follower problem:
max y { 0 , 1 } | I | j J w j · i I v i j y i i I v i j x i + i I v i j y i
subject to the facility constraint i I y i = r .
Proof. 
Let x = ( x i ) i I { 0 , 1 } | I | be the fixed leader decision, and let ( y i * , y i j * , z j * ) be an optimal solution to the MILP formulation given by (10)–(14). Our goal is to show that the binary vector y * = ( y i * ) is an optimal solution to the original nonlinear follower problem.
  • Step 1: Recovering the structure of the nonlinear objective from MILP constraints.
From constraint (13), for all i I , j J , we know that y i j w j v i j z j y i . At optimality, this inequality is tight when y i * = 1 , giving the following:
y i j * = w j v i j z j * .
When y i * = 0 , the constraint implies y i j * 0 , and, since y i j * 0 , we have the following:
y i j * = 0 .
Thus, in both cases, we can write the following:
y i j * = w j v i j z j * y i * , i I , j J .
Now substitute (18) into constraint (11):
i I y i j * + w j z j * i I v i j x i = i I w j v i j z j * y i * + w j z j * i I v i j x i
= w j z j * i I v i j x i + i I v i j y i * .
Substituting Equation (20) into constraint (11) yields the following:
z j * 1 i I v i j x i + i I v i j y i * .
As z j * appears positively in the objective function (and the problem is a maximization), and according to Lemma 1, any solution where inequality (21) is not tight would contradict the optimality. Thus, at optimality, the following applies:
z j * = 1 i I v i j x i + i I v i j y i * , j J .
  • Step 2: Establishing optimality via contradiction.
Assume, for the sake of contradiction, that y * is not an optimal solution to the nonlinear follower problem. Then, there exists another feasible vector y ^ { 0 , 1 } | I | such that i I y ^ i = r and
j J w j · i I v i j y ^ i i I v i j x i + i I v i j y ^ i > j J w j · i I v i j y i * i I v i j x i + i I v i j y i * .
Following the same reasoning as in Step 1, one could construct variables y ^ i j = w j v i j z ^ j y ^ i , where
z ^ j = 1 i I v i j x i + i I v i j y ^ i .
This implies that ( y ^ , y ^ i j , z ^ j ) is feasible for the MILP, and it yields an objective value:
j J i I y ^ i j = j J w j · i I v i j y ^ i i I v i j x i + i I v i j y ^ i ,
which is strictly greater than the MILP objective at ( y i * , y i j * , z j * ) , contradicting the optimality of the original MILP solution.
Therefore, our assumption must be false. Hence, y * is an optimal solution to the nonlinear follower problem.    □

3.5. Numerical Illustration of Preference Similarity p i j

As the proposed CFLBP-CB model introduces a new parameter, p i j , the following example illustrates its conceptual role and practical interpretation.
Example 2. 
Consider a simplified scenario with three potential facilities and four customers. Each customer has a unit demand and expresses preferences across two dimensions, representing demand for product categories 1 and 2, respectively. Each facility, in turn, is characterized by capability scores in the same two dimensions, indicating how well it can fulfill the corresponding customer demand.
Applying the definition of the preference similarity ratio p i j given in Equation (7) to the data in Table 2, we compute p i j for each customer–facility pair. For example, the value for Customer 1 and Facility 1 is calculated as follows:
p 11 = min 1 , 0.9 1.0 · min 1 , 0.7 1.0 = 0.9 · 0.7 = 0.63 .
The complete matrix of preference similarity ratios is shown in Table 3.
The results presented in Table 3 illustrate the degree of alignment between customer preferences and facility capabilities across all customer–facility pairs. A higher value of p i j indicates a stronger match between a customer’s demand and a facility’s supply in terms of the two product categories. Notably, Facility 3 achieves a perfect similarity score of p i j = 1.0 for all customers, reflecting its ability to fully satisfy customer preferences in both dimensions. In contrast, Facility 1 and Facility 2 exhibit varying degrees of mismatch depending on the customer. For example, Facility 1 yields the lowest similarity score with Customer 1 ( p 11 = 0.63 ), due to relatively weaker supply in both product categories compared to the customer’s demand. Meanwhile, Facility 2 achieves a perfect match with Customer 2 ( p 22 = 1.0 ) but provides only partial alignment with other customers.
These results highlight the importance of incorporating the preference similarity ratio p i j in the CFLBP-CB model. By quantifying the compatibility between customer demand and facility supply, p i j directly influences customer attraction and allocation decisions in the bilevel framework. Facilities with higher p i j values are more likely to attract demand, reinforcing the model’s ability to account for heterogeneous preferences and supply heterogeneity in competitive location decisions.

4. Algorithmic Framework

This section outlines the detailed procedures for the clustering process and the optimization process. Algorithm 1 is used to partition a large set of candidate facility locations and customers into coherent clusters. Once this clustering is complete, Algorithm 2 is applied to determine the optimal facility locations within each cluster.

4.1. Multi-View K-Means Clustering with Co-Regularization

To address the heterogeneity in customer and facility features, we propose a generalized multi-view K-means clustering framework that integrates multiple data views—such as spatial coordinates and preference characteristics—into a unified clustering model. Motivated by Kumar et al. [35], our method applies co-regularization directly to the cluster assignment level, a departure from the original approach, which regularizes the agreement between eigenvector embeddings. By doing so, we make the approach more scalable and interpretable for structured data settings, such as those involving customer and facility attributes in the CFLBP-CB.
Algorithm 1: Generalized multi-view K-means clustering with co-regularization
Input: X ( 1 ) , X ( 2 ) , , X ( V ) : V views of the data, each X ( v ) R n × d v ;
K: Number of clusters;
λ v : Co-regularization penalty for each view v;
 max_iter: Maximum number of iterations;
Output: Final unified cluster assignments L ^ and centroids μ k ( v ) for each view v.
Mathematics 13 02481 i001
Algorithm 2: Generalized optimization assignment with minimum facility constraint
Input  : Final cluster labels L ^ from Algorithm 1;
      Dataset D with facility and customer data;
      View matrices X ( 1 ) , X ( 2 ) , , X ( V ) ;
      Optimization model CFLBP - CB ;
      Required minimum number of facilities per cluster m.
Output:  For each cluster k, the optimal facility locations assigned to the leader and follower, denoted as f k ( leader ) and f k ( follower ) , respectively.
Mathematics 13 02481 i002
The silhouette method [36], introduced by Rousseeuw (1987), is a widely used technique for evaluating the effectiveness of clustering methods by calculating the silhouette coefficient. This coefficient measures the appropriateness of assigning a data point to a cluster considering both its cohesion within the cluster and its separation from other clusters. The silhouette coefficient s i for a data point i is defined as follows:
s i = b i a i max ( a i , b i )
where
  • a i is the average distance from point i to all other points in the same cluster;
  • b i is the average distance from point i to all points in the nearest neighboring cluster.
The silhouette coefficient ranges from −1 to 1. A value close to 1 indicates that the point is well-matched to its own cluster and poorly matched to neighboring clusters. A value close to 0 suggests that the point is on or near the boundary between two clusters. A value close to −1 implies that the point is likely misclassified, as it is closer to a neighboring cluster than to its own.
The clustering process of Algorithm 1, which is introduced in this work, follows the steps below:
  • Step 1: Initialization
In the initialization step, for each view v, cluster labels are randomly assigned to each sample. Specifically, for each view v, the cluster labels L ( v ) are initialized randomly from the set { 1 , , K } n , where each sample i in view v is assigned a random cluster label between 1 and K.
After initializing the cluster labels, the next task is to compute the initial centroids. For each view v and each cluster k, the centroid μ k ( v ) is computed by averaging the data points that have been assigned to that cluster. The centroid is given by the following formula:
μ k ( v ) = 1 | L k ( v ) | x i L k ( v ) X i ( v )
where L k ( v ) represents the set of data points assigned to cluster k in view v, and X i ( v ) is the feature vector of sample i in view v.
  • Step 2: Iterative clustering process
In this step, the algorithm iteratively updates cluster assignments and centroids. It consists of two main subprocesses: the assignment step and the centroid update step. These subprocesses are repeated until convergence is achieved.
  • Step 2.1: Assignment step  
In the assignment step, for each view v, each data point i is assigned to the cluster that minimizes a penalized distance. The distance between the data point x i ( v ) and the centroid μ k ( v ) of cluster k in view v is first calculated as follows:
D k = x i ( v ) μ k ( v ) 2
Next, for each other view u v , a co-regularization penalty is applied. If the sample is assigned to a different cluster in view u, the distance is increased by λ v , where λ v is the co-regularization penalty for view v:
if L i ( u ) k : D k D k + λ v
This penalty encourages consistency across views by discouraging assignments that conflict between views. After calculating the distances for all clusters, the data point i is assigned to the cluster that minimizes D k across all views:
L i ( v ) arg min k D k
This ensures that each sample is assigned to the closest cluster in each view while respecting the co-regularization penalty.
  • Step 2.2: Centroid update step  
After the assignment step, the centroids are recomputed for each view and cluster. For each view v and each cluster k, the new centroid μ k ( v ) is calculated as the mean of the data points assigned to the cluster:
μ k ( v ) = 1 | L k ( v ) | x i L k ( v ) X i ( v )
This ensures that the centroids reflect the updated assignments. After updating the centroids, if the cluster assignments do not change compared to the previous iteration, the algorithm terminates, indicating convergence. If the assignments have changed, the algorithm proceeds to the next iteration.
  • Step 3: Post-processing (majority voting and silhouette tie-breaking)  
After the algorithm converges, post-processing is performed to finalize the cluster assignments. Specifically, the silhouette score is computed for each sample i in each view v to assess the clustering quality. The silhouette score s i ( v ) for a sample i in view v is computed based on the cluster assignments L ( v ) and is defined as follows:
s i ( v ) = silhouette ( x i , L ( v ) ) = b i ( v ) a i ( v ) max a i ( v ) , b i ( v )
where silhouette ( · ) denotes the silhouette coefficient of the data point x i with respect to the cluster labels L ( v ) . Here, a i ( v ) is the average distance from sample i to all other samples in the same cluster in view v, and b i ( v ) is the average distance from sample i to all samples in the nearest neighboring cluster in view v.
Next, majority voting is applied across all views. For each sample i, the number of votes for each cluster label k is counted across all views. The number of votes for cluster k is given by the following:
votes k = v = 1 V I ( L i ( v ) = k )
where I is the indicator function, which counts the votes for each label. If there is a tie in the majority vote, meaning multiple clusters have the same number of votes, the tie is resolved by selecting the cluster with the highest silhouette score among the tied labels. The final cluster assignment for sample i is then given by the following:
L ^ i arg max k { s i ( v ) | v = 1 , , V }
This ensures that, in the case of ties, the cluster with the best local clustering quality across views is selected.
The proposed clustering algorithm is presented in detail in Algorithm 1.
Example 3. 
Consider a dataset consisting of three data points, where each sample x i R 4 is composed of two distinct views: spatial coordinates and preference attributes. Specifically, let each point x i = ( x i ( 1 ) , x i ( 2 ) ) , where x i ( 1 ) R 2 denotes the spatial view and x i ( 2 ) R 2 denotes the preference view. The dataset is defined as follows:
x 1 = ( 0 , 0 , 1.00 , 0.00 ) , x 2 = ( 0 , 1 , 0.15 , 1.00 ) , x 3 = ( 10 , 10 , 0.10 , 1.00 ) .
The objective is to partition the data into K = 2 clusters. We compare the outcomes of (i) traditional K-means clustering applied to the concatenated features in R 4 , and (ii) the proposed multi-view K-means clustering algorithm with co-regularization. For the multi-view approach, the following co-regularization penalty parameters are used:
λ ( 1 ) = 5.0 ( spatial view ) , λ ( 2 ) = 0.1 ( preference view ) .
As a baseline, we first present the clustering result obtained by applying the traditional K-means algorithm.
Step 1: Initialization. Let the initial cluster centroids be selected as follows:
μ 1 ( 0 ) = x 1 = ( 0 , 0 , 1.00 , 0.00 ) , μ 2 ( 0 ) = x 3 = ( 10 , 10 , 0.10 , 1.00 ) .
Step 2: Assignment. To determine the cluster membership for each data point, we compute the squared Euclidean distance from each point to the initial centroids:
x 1 μ 1 ( 0 ) 2 = 0 , x 1 μ 2 ( 0 ) 2 = 201.81 , x 2 μ 1 ( 0 ) 2 = 2.7225 , x 2 μ 2 ( 0 ) 2 = 181.0025 ,
x 3 μ 1 ( 0 ) 2 = 201.81 , x 3 μ 2 ( 0 ) 2 = 0 .
Each data point is then assigned to the cluster whose centroid it is closest to. Consequently, the initial cluster assignments are
C 1 ( 1 ) = { x 1 , x 2 } , C 2 ( 1 ) = { x 3 } ,
where C 1 ( 1 ) and C 2 ( 1 ) denote Clusters 1 and 2 determined by the initial centroids, respectively.
Step 3: Centroid update. The cluster centroids are then updated by computing the mean of the data points assigned to each cluster. Specifically, the updated centroids are given by the following:
μ 1 ( 1 ) = 1 2 ( x 1 + x 2 ) = ( 0 , 0.5 , 0.575 , 0.5 ) , μ 2 ( 1 ) = x 3 = ( 10 , 10 , 0.10 , 1.00 ) .
Step 4: Reassignment check. To determine whether the cluster assignments have changed, we recompute the squared Euclidean distances from each data point to the updated centroids:
x 1 μ 1 ( 1 ) 2 = 0.6806 , x 1 μ 2 ( 1 ) 2 = 201.81 , x 2 μ 1 ( 1 ) 2 = 0.6806
x 2 μ 2 ( 1 ) 2 = 181.05 x 3 μ 1 ( 1 ) 2 = 190.73 , x 3 μ 2 ( 1 ) 2 = 0 .
Since each point remains closest to its previously assigned centroid, no reassignment occurs. Thus, the algorithm has converged. Therefore, the final cluster assignment obtained by the traditional K-means algorithm is
C 1 ( 2 ) = { x 1 , x 2 } , C 2 ( 2 ) = { x 3 } .
We next present the clustering result obtained by applying the proposed multi-view K-means clustering algorithm.
Step 1: Initial assignments. We begin by specifying the initial cluster assignments for each view. Let
L ( 1 ) = [ 1 , 1 , 2 ] , L ( 2 ) = [ 1 , 2 , 2 ] ,
where L ( 1 ) and L ( 2 ) denote the cluster labels in Views 1 and 2, respectively. This indicates that, in the spatial view ( v = 1 ), data points x 1 and x 2 are initially assigned to Cluster 1, while x 3 is assigned to Cluster 2. Simultaneously, in the preference view ( v = 2 ), x 1 is assigned to Cluster 1, and both x 2 and x 3 are assigned to Cluster 2.
Step 2: Compute view-specific centroids. Based on the initial cluster assignments, we compute the centroids separately for each view by averaging the data points assigned to each cluster:
μ 1 ( 1 ) = 1 2 ( x 1 ( 1 ) + x 2 ( 1 ) ) = [ 0 , 0.5 ] , μ 2 ( 1 ) = x 3 ( 1 ) = [ 10 , 10 ] , μ 1 ( 2 ) = x 1 ( 2 ) = [ 1.00 , 0.00 ] , μ 2 ( 2 ) = 1 2 ( x 2 ( 2 ) + x 3 ( 2 ) ) = [ 0.125 , 1.00 ] .
Step 3: Assignment computation for x 2 .
View 1 (Spatial):
x 2 ( 1 ) μ 1 ( 1 ) 2 = 0.25 + λ ( 1 ) = 5.25 , x 2 ( 1 ) μ 2 ( 1 ) 2 = 181 + 0 = 181 .
View 2 (Preference):
x 2 ( 2 ) μ 1 ( 2 ) 2 = 1.7225 + 0 = 1.7225 , x 2 ( 2 ) μ 2 ( 2 ) 2 = 0.000625 + λ ( 2 ) = 0.100625 .
Step 4: View-wise reassignment.
View 1: D 1 ( 1 ) = 5.25 < D 2 ( 1 ) = 181 L 2 ( 1 ) = 1 ;
View 2: D 2 ( 2 ) = 0.100625 < D 1 ( 2 ) = 1.7225 L 2 ( 2 ) = 2 .
Step 5: Unified clustering decision. Since the views disagree on the assignment of x 2 , we apply majority voting. As labels L 2 ( 1 ) = 1 , L 2 ( 2 ) = 2 have the same number of votes, we need to apply silhouette scores of x 2 separately in each view to assign the final cluster. By using the definition of the silhouette scores, we obtain s 2 1 = 0.926 and s 2 2 = 0.962 . This implies that x 2 is assigned to Cluster 2. In terms of x 1 , x 3 , these two points have the same label across the two views, and thereby the majority vote x 1 is in Cluster 1 and x 1 is in Cluster 2. Then, the final cluster label is
L ^ = [ 1 , 2 , 2 ] .

4.2. Optimization Assignment

The generalized optimization assignment with a minimum facility constraint (see Algorithm 2) is designed to optimize the CFLBP-CB across multiple clusters while ensuring that each cluster contains at least the specified minimum number of facilities. The procedure begins by initializing the unified set of K clusters obtained from Algorithm 1. In this phase, the algorithm operates in two main subprocesses: cluster-wise optimization, where facility assignments are optimized within each cluster, and updating unassigned facilities, where any remaining facilities are allocated to satisfy the required constraints.  
Step 1: Cluster-wise optimization  
The first step involves optimizing the assignment of facilities and customers within each cluster. For each cluster c, the customers C c and facilities F c associated with that cluster are extracted. If the number of facilities in the cluster is less than the required minimum m, the algorithm skips the optimization for that cluster. Otherwise, the optimization model is executed using the bilevel optimization problem, the CFLBP-CB. Once the CFLBP-CB model is solved, the facilities within the cluster are marked as used by setting the Used flag in the dataset. Finally, the cluster c is appended to the list of assigned clusters.  
Step 2: Update unassigned facilities  
After the cluster-wise optimization, the algorithm updates the list of unassigned clusters by removing the clusters that have already been assigned. It then identifies the unused facilities that have not been assigned to any cluster. For each unused facility f i , the algorithm temporarily assigns it to each unassigned cluster and computes the silhouette scores for all views. The silhouette score is used to assess the quality of the facility’s assignment to each cluster. The facility is then assigned to the cluster that yields the highest silhouette score, ensuring that the facility is placed in the most appropriate cluster based on both cohesion and separation. The updated cluster assignments for the facilities are then reflected in the dataset. Finally, Step 1 (cluster-wise optimization) is repeated.
The details of the generalized optimization assignment algorithm with the minimum facility constraint can be found in Algorithm 2.

5. Case Study

The experiments are conducted on a CPU with a maximum clock speed of 5.00 GHz. The K-means algorithm is custom implemented, while spectral and hierarchical clustering methods are sourced from Scikit-learn 1.6.1, benefiting from optimized code. The optimization model is formulated in Pyomo 6.9.2 and solved using MindtPy bundled with Pyomo 6.9.2 with the Outer Approximation (OA) strategy, employing GLPK 5.0 for mixed-integer subproblems and IPOPT 3.14.17 for nonlinear ones. These open-source solvers are selected for their accessibility and reproducibility. However, GLPK is less efficient than commercial solvers for large MILPs, and IPOPT may face challenges with nonconvex or nonsmooth problems. Additionally, the OA strategy can be computationally intensive due to repeated MILP-NLP solving cycles.
To evaluate the effectiveness of the proposed heuristic framework, we conduct a case study using a real-world dataset provided by the PySAL spatial analysis library (available at https://pysal.org/spopt/notebooks/facloc-real-world.html, accessed on 1 August 2024) The dataset includes geographic information for potential facility sites and customer demand sites distributed throughout the city of San Francisco. This urban-scale scenario provides a realistic testbed for assessing competitive facility location strategies under spatial and behavioral considerations.
The details of candidate facility locations and customer sites are summarized in Table 4 and Table 5, respectively.

5.1. Franchise Expansion into New Territories

This case study simulates a retail expansion scenario in which a parent company (the leader) opens five company-owned stores, and then franchisees (the followers) open five additional stores in response. The objective is to optimize joint market coverage and profitability while avoiding cannibalization between company-owned and franchised outlets. Based on this setup and using the real-world dataset summarized in Table 4 and Table 5, the case study involves | I | = 16 candidate facility locations and | J | = 205 customer sites distributed across the city of San Francisco.
In this setting, we assume two customer preference dimensions. The first preference ( P r e f e r e n c e 1 ) is Price Sensitivity, which reflects the extent to which a customer values affordability. The second preference ( P r e f e r e n c e 2 ) is Service Expectation, which captures the importance a customer places on high service quality. The customer attraction coefficient is then defined as
v i j = p i j d i j + ϵ = min 1 , P r e f e r e n c e 1 , i P r e f e r e n c e 1 , j × min 1 , P r e f e r e n c e 2 , i P r e f e r e n c e 2 , j d i j + ϵ ,
where i { 1 , 2 , , 16 } represents candidate facility locations and j { 1 , 2 , , 205 } represents customer sites.

5.2. Results

To address the experimental setup described in Section 5.1, we utilize the proposed two-stage solution framework. In the first stage, clustering is performed using Algorithm 1, which employs a behavior-aware multi-view K-means approach with co-regularization to ensure consistency across different data views. In the second stage, Algorithm 2 is applied to solve the CFLBP-CB within each cluster, effectively capturing the leader–follower dynamics characteristic of competitive decision-making.

5.2.1. Clustering Results

To implement Algorithm 1, which performs behavior-aware multi-view K-means clustering with co-regularization, we set the number of clusters to K = 5 and define two data views ( V = 2 ), corresponding to geographic and preference information. Since explicit customer preference data are not available in the original dataset, we synthetically generate behavioral attributes to construct P r e f e r e n c e 1 , i , P r e f e r e n c e 2 , i , P r e f e r e n c e 1 , j , and P r e f e r e n c e 2 , j . The co-regularization penalties are configured as λ 1 = 0.025 for the spatial view and λ 2 = 0.075 for the behavioral view, ensuring that both data sources contribute appropriately to the clustering process.
The initial outcomes of the multi-view K-means clustering are depicted in Figure 2. The left subplot displays clustering based on geographic proximity (X–Y), while the right subplot shows clustering based on preference features (Preferences 1–2). In both subplots, points are colored according to their assigned clusters: blue for Cluster 0, green for Cluster 1, brown for Cluster 2, gray for Cluster 3, and cyan for Cluster 4. The centroids of each cluster are indicated by markers in the same color as their respective clusters. The geographic clusters in the left subplot demonstrate strong spatial compactness, particularly for Cluster 2. In contrast, clusters in the preference space (right subplot) are more dispersed and overlapping, highlighting greater variability in these attributes. While noticeable differences exist between cluster assignments in the spatial and behavioral views, the multi-view K-means clustering approach introduces a degree of alignment, providing a solid foundation for further cluster integration.
To further explain the discrepancies between views, Figure 3 illustrates inconsistencies in cluster assignments, where black-circled points denote customers whose groupings differ between geographic and preference perspectives. Such mismatches, arising when spatially close customers exhibit distinct behavioral profiles or vice versa, underscore the challenge of simultaneously achieving spatial and behavioral consistency within a single clustering solution. This observation motivates a refinement stage, which integrates the two views using majority voting and silhouette scores to produce a more unified clustering outcome, as presented in Figure 4. Here, cluster membership is indicated by color, with consensus achieved through the aforementioned refinement process. The unified results highlight that even customers within the same geographic cluster can exhibit distinct behavioral preferences, reflecting real-world complexities in competitive facility location problems. To quantify these unified clusters, Table 6 reports the balanced distribution of customers and facilities after refinement, confirming the effectiveness of the proposed clustering strategy (Algorithm 1). Collectively, Figure 2, Figure 3 and Figure 4 and Table 6 demonstrate that our multi-view K-means clustering framework, incorporating co-regularization and refinement, achieves superior inter-view consistency and improved intra-cluster compactness compared to single-view approaches. This unified clustering not only resolves ambiguous assignments effectively but also provides a stable and interpretable input structure for subsequent facility location optimization, ensuring alignment between spatial accessibility and customer preferences.

5.2.2. Bilevel Optimization Results

To determine the optimal facility locations for the parent company (leader) and its franchisees (follower), we apply Algorithm 2 to the unified clusters generated by Algorithm 1. The CFLBP-CB model is solved within each cluster using Algorithm 2, with the parameters set to max_iter = 100, ϵ = 1 × 10 6 , and a required minimum of m = 2 candidate sites per cluster (since each cluster must have at least two candidate sites, one for the leader and one for the follower). Table 7 summarizes the resulting performance metrics, including overall profit, leader profit, follower profit, leader market share, and follower market share, for Clusters 0 through 4.
The results demonstrate that profit outcomes vary notably across clusters, influenced by the spatial and behavioral characteristics of customers. For example, Cluster 4 achieves the highest overall and leader profits (4307.10 and 2740.41, respectively), indicating a strong alignment with customer preferences or a larger customer base, followed closely by Cluster 0. In contrast, Clusters 2 and 3 exhibit comparatively lower overall and follower profits, suggesting either smaller populations or less favorable positioning for the follower. These outcomes highlight that leader profit tends to remain relatively stable and substantial across clusters, whereas follower profit is more sensitive to competitive facility placement, as reflected by the lower values in Clusters 1, 2, and 3. This pattern is characteristic of competitive facility location problems solved via bilevel optimization. Note that the follower market share, shown in the last column of Table 7, is effectively the complement of the leader’s market share.
Figure 5 illustrates the distribution of customer clusters and the corresponding optimal facility locations under two distinct views: geographic features (X–Y) and preference features (Preferences 1–2). Each color signifies a customer cluster, and the triangles mark the selected facility locations that are optimized to capture the most market share within each cluster. The figure demonstrates that facilities are generally placed in areas with higher customer densities, particularly in clusters where customers are more tightly grouped. There is clear alignment between customer distributions and optimal facility sites, indicating the effectiveness of the multi-view K-means clustering approach in supporting profitable facility placement. This geographic and preference alignment underpins the profit patterns observed in Table 7, emphasizing that strategic clustering and facility placement are key drivers of market success in competitive location problems.

5.2.3. Sensitivity of Multi-View K-Means Clustering to Co-Regularization Parameters

To evaluate the robustness of the proposed framework, we conduct a sensitivity analysis by varying the co-regularization parameters λ 1 and λ 2 , which control the influence of spatial and behavioral views, respectively, during the multi-view K-means clustering stage. Table 8 reports the overall profit values obtained from the subsequent bilevel optimization for each tested parameter configuration.
The results demonstrate that, when λ 1 = 0.025 and λ 2 = 0 , the clustering process is biased toward spatial alignment and disregards behavioral consistency, yielding the lowest overall profit of 14,577.37. In contrast, when the behavioral view is given moderate influence ( λ 1 = 0.005 , λ 2 = 0.045 ), the profit increases substantially to 15,493.84, indicating the value of incorporating customer preferences in the clustering stage. However, pushing the behavioral influence further ( λ 2 = 0.105 ) while keeping spatial influence minimal ( λ 1 = 0.0025 ) results in a slightly lower profit (15,279.92), suggesting that excessive bias toward one view can reduce overall performance.
Notably, the highest overall profit of 15,812.44 is achieved when both co-regularization terms are set to zero ( λ 1 = 0 , λ 2 = 0 ), implying that independent clustering of views followed by decision-level integration (through majority voting and silhouette-based refinement) may be sufficient to produce a well-aligned and effective cluster structure. These findings emphasize that, while co-regularization can help enforce alignment across data views, it must be applied judiciously. Over-regularization may constrain the natural clustering tendencies of individual views and diminish optimization outcomes. A balanced or even unpenalized approach, in combination with robust post-clustering refinement, may offer the best trade-off in practice.

5.2.4. Parameter Tuning

It is evident from Table 8 that the choice of co-regularization parameters significantly influences the achievable overall profit. To further illustrate and generalize the effect of varying λ 1 and λ 2 across a broader parameter space, we provide a surface plot in Figure 6.
Figure 6 presents a 3D surface plot that depicts the relationship between the co-regularization parameters λ 1 and λ 2 and the overall profit attained by the proposed framework. The plot reveals that the profit varies nonlinearly across the parameter space, with several local maxima and minima. Notably, the highest overall profit is observed for specific combinations of moderate values of both λ 1 and λ 2 , rather than at their extreme ends. This underscores the importance of carefully tuning the balance between spatial and behavioral regularization during the clustering process. In addition, regions with steep gradients indicate that the model’s performance can be sensitive to small changes in these parameters, underscoring the importance of careful parameter selection. Employing strategies such as grid search can help identify the values of λ that yield the highest overall profit, thereby ensuring stable model performance.

5.2.5. Results from Single-View Clustering

Table 9 reports the bilevel optimization outcomes across three single-view clustering algorithms—K-means, spectral, and hierarchical—each applied to three types of input feature spaces: geographic only (2D), preference only (2D), and combined geographic–preference (4D). For each configuration, five clusters are formed, with corresponding leader and follower profits, total profit, and computation times for the clustering and optimization steps. This comparative analysis reveals how clustering strategy and feature representation influence profitability and leader–follower dynamics.
K-means clustering applied to geographic features yields the highest total profit (15,562.87), with Cluster 3 producing the greatest cluster-level return (4717.91), including the top follower profit (1772.84). Preference-based clustering results in a comparable total profit (15,561.16), though it displays stronger leader dominance in several clusters, suggesting reduced market competition due to finer preference-based segmentation. In contrast, the combined 4D feature clustering underperforms (14,405.02), indicating that richer input dimensions may introduce redundancy or fragmentation, thereby reducing clustering coherence and overall efficiency.
Spectral clustering on preference features achieves the highest total profit among all configurations (15,603.99), with substantial leader gains (12,439.67), particularly in Cluster 0 (5313.10). Geographic-based spectral clustering performs slightly worse (14,996.05), though it still offers effective spatial partitioning. Interestingly, spectral clustering with combined features also reaches 15,603.99, suggesting that this method better handles high-dimensional data compared to K-means clustering by exploiting latent structures through spectral embedding.
Hierarchical clustering demonstrates superior computational efficiency, with clustering times consistently under 0.004 s. The preference-only variant yields the highest profit (15,376.47), followed by the geographic-only variant (15,292.71) and, lastly, the 4D view (14,765.25). Notably, hierarchical clustering leads to more balanced follower profits across clusters. In the preference-only model, follower profits are relatively high and evenly distributed (e.g., 1401.62 in Cluster 0 and 902.14 in Cluster 4), reflecting improved market competition.
Spectral clustering applied to preference features yields the highest total profit (15,603.99), demonstrating its superior capability in segmenting customers according to preference-related characteristics. K-means clustering with geographic features yields a comparable profit (15,562.87), while offering strong spatial cohesion and accessibility. Hierarchical clustering stands out for its computational efficiency—with clustering times consistently below 0.004 s—and for producing a more balanced profit distribution between leaders and followers, as indicated by the narrower profit gaps across clusters. In contrast, clustering with combined geographic and preference features (4D) consistently underperforms across all methods, emphasizing the limitations of traditional algorithms in handling high-dimensional feature spaces. These findings underscore the importance of aligning clustering techniques with the underlying data structure to optimize performance in competitive facility location settings.
To support these findings, Figure 7, Figure 8 and Figure 9 illustrate the optimal facility placements derived from each single-view clustering method—including K-means, spectral, and hierarchical clustering—applied to three different input feature sets: geographic (2D), preference (2D), and combined (4D). Each subfigure presents two projections, the spatial view (left) and the preference view (right), enabling visualization of how clusters formed in one feature space manifest in the other.
Figure 7 depicts optimal facility placements derived from single-view K-means clustering using three input feature sets: geographic (2D), preference (2D), and a combined 4D representation.
  • Subplot (a): Geographic features (2D). Clustering based on geographic attributes yields compact and well-separated spatial clusters. Facility placements (triangles) are centrally located within their respective clusters, supporting strong geographic accessibility. However, the corresponding preference view exhibits substantial overlap among clusters, indicating poor alignment with customer preferences.
  • Subplot (b): Preference features (2D).Clustering in the preference domain produces distinct and well-separated clusters, with facility locations closely aligned with preference centers. Nevertheless, the geographic projection reveals spatial dispersion and overlap, which may hinder efficient service delivery and cost-effective deployment.
  • Subplot (c): Combined geographic and preference features (4D). The integration of both feature types leads to incoherent clustering in the spatial view, with facilities misaligned and positioned far from cluster centers. While the preference view maintains some alignment, overall cluster quality deteriorates.
Figure 8 presents the optimal facility placements obtained from single-view spectral clustering applied to the geographic (2D), preference (2D), and combined (4D) feature sets.
  • Subplot (a): Geographic features (2D). Clustering based on geographic attributes results in well-separated spatial clusters, with facilities (triangles) embedded within each group, supporting accessibility. However, the corresponding preference view reveals significant cluster mixing, indicating poor alignment with customer behavioral attributes.
  • Subplot (b): Preference features (2D). Clustering in the preference domain produces clearly defined clusters and well-aligned facilities, demonstrating effective segmentation. Yet the spatial projection shows considerable dispersion and overlap, reducing efficiency in physical service delivery.
  • Subplot (c): Combined geographic and preference features (4D).The combined feature input yields scattered and incoherent clusters in both spatial and preference views, with facilities often located far from cluster centers, suggesting poor integration of the two feature spaces.
Figure 9 illustrates the optimal facility placements derived from single-view hierarchical clustering applied to three distinct input feature sets:
  • Subplot (a): Geographic features (2D).The clustering results exhibit compact and spatially coherent groups. Facilities (triangles) are centrally located within clusters, enhancing spatial accessibility. However, in the corresponding preference space, clusters display substantial overlap, reflecting poor alignment with customer preferences.
  • Subplot (b): Preference features (2D). The preference view reveals reasonably well-separated clusters, and most facility placements align with the cluster centers. An exception is Cluster 4, where the facility is not centrally located, indicating inconsistency. Meanwhile, the spatial view shows scattered and overlapping clusters, undermining geographic efficiency.
  • Subplot (c): Combined geographic and preference features (4D). Clusters appear poorly defined in both spatial and preference views. Facility placements are inconsistently positioned and often located far from the respective cluster centers, reducing both behavioral targeting and spatial coverage.
These findings highlight a fundamental limitation shared across single-view clustering methods (K-means, spectral, and hierarchical): optimizing in one feature space often compromises performance in the other. While each method offers distinct advantages—graph-based similarity in spectral clustering, interpretability in hierarchical clustering, and spatial cohesion in K-means clustering—their effectiveness in competitive facility location depends critically on feature-space alignment. The performance degradation observed in high-dimensional or misaligned settings underscores the importance of consistent and carefully selected input features to ensure both spatial accessibility and behavioral segmentation.

5.2.6. Results from Multi-View K-Means Clustering

Table 10 reports the bilevel optimization results under various co-regularization settings ( λ 1 , λ 2 ) in the multi-view K-means clustering framework. Each setting reflects a different weighting between geographic ( λ 1 ) and preference ( λ 2 ) views, influencing the resulting cluster structures and facility placement performance.
The configuration ( λ 1 = 0.015 , λ 2 = 0.120 ) yields the highest total profit of 16,318.94, indicating that strong emphasis on preference alignment, when moderately balanced with spatial structure, can enhance market segmentation and profitability. Similarly, the setting ( λ 1 = 0.0225 , λ 2 = 0.075 ) achieves a comparably high profit of 16,310.62 with shorter clustering time (0.1255 s), and demonstrates well-distributed leader and follower profits across clusters—particularly a strong performance in Cluster 4.
By contrast, overly imbalanced or weak regularization settings lead to lower performance. For instance, ( λ 1 = 0.025 , λ 2 = 0.015 ) , which places less emphasis on preference consistency, yields a reduced total profit of 15,791.85. Likewise, the absence of regularization ( λ 1 = 0 , λ 2 = 0 ) results in 15,812.44, suggesting that uncoordinated clustering across views weakens the model’s ability to align customer segments effectively.
Overall, the results emphasize the importance of balanced co-regularization in multi-view clustering. Moderate integration of geographic and preference features supports more coherent cluster structures, which translate into improved leader–follower profitability and efficient computation times. These findings highlight the need for careful tuning of regularization parameters in competitive facility location models.

5.2.7. Comparison Between Single-View and Multi-View K-Means Clustering Results

To evaluate the effectiveness of the proposed multi-view clustering framework with co-regularization, we compare its performance against single-view clustering methods—K-means, spectral, and hierarchical—applied to geographic, preference, and combined (4D) feature sets. The single-view results indicate that geographic-only and preference-only clustering yield the highest total profits (15,562.87 and 15,561.15, respectively), with relatively short clustering times (0.03–0.06 s) but longer optimization times (3.10–6.50 s). In contrast, using a combined 4D input leads to lower total profit (14,405.02) and the longest optimization time (up to 6.50 s), suggesting that naïvely merging features can dilute distinct structural information and degrade performance.
In comparison, the multi-view K-means clustering approach preserves the distinctiveness of geographic and preference views while promoting alignment between them. As shown in Table 11, several co-regularization settings outperform all single-view models in terms of total profit. For example, the configuration ( λ 1 = 0.025 , λ 2 = 0.075 ) achieves the highest profit (16,417.64), with a clustering time of 0.1708 s and an optimization time of 2.9801 s. Similarly, other configurations, such as ( λ 1 = 0.0225 , λ 2 = 0.075 ) and ( λ 1 = 0.015 , λ 2 = 0.120 ) , yield profits exceeding 16,300 with comparably short runtimes. Even simple settings such as ( λ 1 = 0.000 , λ 2 = 0.000 ) outperform the best single-view baseline, demonstrating robustness across parameter choices. It is worth noting that these co-regularization parameters ( λ 1 , λ 2 ) can be selected through strategies such as grid search, as discussed in Section 5.2.4.
Notably, multi-view methods generally require longer clustering times (0.09–0.17 s) due to co-regularization but achieve shorter optimization times (2.98–3.52 s), resulting in competitive or superior overall runtimes compared to single-view methods.
These findings emphasize some key advantages of the proposed framework:
  • Multi-view clustering effectively integrates complementary spatial and behavioral patterns, yielding better facility–customer groupings for optimization.
  • Co-regularization provides a tunable mechanism to balance intra-view fidelity and inter-view consistency, improving clustering quality and optimization outcomes.
  • Despite slightly increased clustering overhead, the total computation time is reduced due to more efficient optimization on coherent clusters.
In summary, multi-view clustering with co-regularization not only improves profit and solution quality in competitive facility location problems but also enhances computational efficiency by reducing downstream optimization time.

6. Discussion

6.1. Effectiveness of Multi-View Clustering

As shown in Table 9, the multi-view K-means clustering approach with co-regularization consistently outperforms single-view methods in terms of overall profit. While single-view clustering using either geographic or preference features can yield competitive results, the multi-view method achieves superior outcomes by preserving the structural integrity of each view and promoting inter-view consistency. The subsequent refinement step—incorporating majority voting and silhouette scores—further enhances cluster coherence and interpretability.

6.2. Insights from Co-Regularization Sensitivity

Sensitivity analysis of the co-regularization parameters λ 1 (geographic view) and λ 2 (preference view) indicates that balanced inter-view alignment is crucial for solution quality. Configurations favoring preference alignment while preserving spatial coherence (e.g., λ 1 = 0.025 , λ 2 = 0.075 ) yield the highest profit. In contrast, settings with no regularization or strong bias toward one view result in suboptimal or unbalanced clustering. These findings underscore the importance of tuning co-regularization to effectively integrate spatial and preference features in practice.

6.3. Comparison of Computational Efficiency with Bilevel Optimization

From a computational complexity standpoint, the efficiency of the proposed framework can be approximated by
O T ( n + m ) V K ( d + V ) + ( n + m ) 2 d V + n / k 1 · n / k 1 1 · k · n k · m k + T solve ( 10 ) ,
where the first component, T ( n + m ) V K ( d + V ) + ( n + m ) 2 d V , captures the computational cost of the multi-view clustering process. This term grows polynomially with respect to the number of customers n, candidate facilities m, feature views V, and clusters K. The second component,
n / k 1 · n / k 1 1 · k · n k · m k + T solve ( 10 ) ,
represents the computational effort required for solving the CFLBP-CB bilevel optimization problem within each cluster. This can be algebraically reformulated as follows:
n k · n k 1 · k · n k · m k + T solve ( 10 ) n 3 m k 3 + n 2 k T solve ( 10 ) ,
which reveals its polynomial nature in terms of n, m, and k. As n , the asymptotic behavior of the overall runtime in Equation (25) becomes
O n 3 m k 3 ,
demonstrating that the optimization step increasingly dominates the total computation time in large-scale instances, while the clustering components remain comparatively lightweight.
In contrast, the computational complexity of directly solving the full CFLBP-CB bilevel optimization model is substantially higher and can be estimated as
O ( n m + T solve ( 10 ) ) · n k · n k k ,
which grows exponentially with the number of selected facilities k. This expression arises from evaluating all possible combinations of selecting k leader facilities from n facility location candidates and, subsequently, k follower facilities from the remaining n k locations.
To illustrate the exponential nature of this term, we apply a standard upper bound on the binomial coefficient:
n k n e k k ,
where e is the base of the natural logarithm. Substituting this into the expression for the full bilevel optimization complexity, we obtain the following approximation:
T solve ( 10 ) · n e k k · e ( n k ) k k .
When k n , the term ( n k ) can be approximated by n, yielding the following:
T solve ( 10 ) · n e k 2 k .
This form clearly illustrates the exponential growth with respect to the number of selected facilities k. Consequently, as n , the asymptotic behavior of the overall runtime in Equation (29) becomes
O n e k 2 k .
This exponential behavior arises even under fixed binary leader and follower decisions ( x i , y i ), indicating the inherent intractability of directly solving the full CFLBP-CB model for large-scale instances.
Therefore, by comparing Equations (28) and (30), it is evident that the proposed framework incurs a significantly lower computational cost than directly solving the full CFLBP-CB model. This comparison reinforces the practical efficiency and scalability of the clustering-based decomposition strategy for addressing large-scale competitive facility location problems.

6.4. Evaluating Clustering-Induced Profit Bias

To assess the potential bias introduced by restricting demand allocation within clusters, we compare the profit outcomes generated by the proposed framework with those obtained by evaluating its facility placement decisions under the full CFLBP-CB model. As shown in Table 12, the highest test-set profit under the full model (16,104.18) is achieved using the co-regularization parameters ( λ 1 , λ 2 ) = ( 0.0050 , 0.1500 ) . This setting differs from the framework-optimal parameters ( 0.0250 , 0.0750 ) , which produce the highest profit within the clustering-based model but yield a lower profit of 15,770.30 when assessed on the full model.
While the configuration ( λ 1 , λ 2 ) = ( 0.0050 , 0.1500 ) achieves superior performance when evaluated under the full CFLBP-CB model and exhibits strong consistency along with a high silhouette score in the spatial (X–Y) view, it simultaneously yields a negative silhouette score ( 0.0773 ) in the preference view. As shown in the preference-space projection in Figure 10a, the resulting clusters appear to be behaviorally fragmented, with significant overlap among different cluster labels. This indicates that, although the spatial structure is well-formed, the behavioral segmentation is weak, suggesting that customers with divergent preferences are grouped together due to spatial proximity.
On the other hand, considering the quality of clustering across the parameter configurations presented in Table 12, the best overall clustering structure is obtained using the co-regularization parameters ( λ 1 , λ 2 ) = ( 0.0050 , 0.0900 ) , as illustrated in Figure 10b. This setting yields more coherent clusters in both the spatial and preference views, reflecting a better balance between geographic proximity and behavioral similarity. These results highlight the importance of carefully selecting co-regularization parameters in accordance with the decision-maker’s priorities—whether optimizing for global performance or ensuring high-quality, behaviorally consistent segmentation.
The correlation analysis between objective values obtained from the proposed clustering-based model and those from the full CFLBP-CB model reveals varying degrees of consistency across evaluation criteria. As shown in Table 13, the overall objective exhibits a weak correlation ( r = 0.2271 ), suggesting potential shifts in global profit rankings under the full model. The follower’s profit shows a moderate correlation ( r = 0.2506 ), reflecting sensitivity to inter-cluster competition. In contrast, stronger correlations are observed for the leader’s profit and the follower’s market share ( r = 0.5868 ), indicating that strategic positioning and market influence are relatively well preserved.
These findings imply that, while the clustering-based framework may introduce bias at the aggregate level, it remains effective for subproblems involving individual stakeholders. In particular, when the decision-maker’s focus is on leader profitability or follower market share, the proposed method offers outcomes that are reasonably aligned with those of the full model.

6.5. Comparative Metaheuristic: Genetic Algorithm

A genetic algorithm (GA) is implemented using the DEAP 1.4.3 library with the following parameter configuration: The evolutionary process spans 10 generations, including the initial generation, maintaining a constant population size of 20 individuals. Genetic operators are subsequently applied, with the crossover probability set to 0.8 and the mutation probability set to 0.2. Additionally, an individual bit-flip probability of 0.05 is used for mutation, and a tournament size of 3 is employed for parent selection. Crucially, a dedicated repair function is integrated into the evaluation process to ensure that all generated individuals strictly adhere to problem-specific constraints. These constraints mandate that the leader opens exactly p facilities ( x i = p ), the follower opens exactly r facilities ( y i = r ), and no single facility location is simultaneously opened by both players ( x i + y i 1 for all locations i). This comprehensive parameter configuration collectively controls the balance between the exploration of the search space and its exploitation by the GA, thereby aiming to identify optimal facility locations.
As presented in Table 14, the GA demonstrates a clear convergence behavior, with both the average and maximum overall profit steadily increasing across the nine generations. The algorithm commences with an average overall profit of 14,941.60 and converges to an average of 16,039.75, while the maximum overall profit reaches 16,137.75 by the final generation. The computation time per generation varies, generally ranging from approximately 66 to 94 s.

6.6. Comparative Evaluation of Optimization Methods

The results presented in Table 15 compare the performance of three distinct solution approaches to the CFLBP-CB: the proposed multi-view clustering-based framework, direct optimization of the full bilevel CFLBP-CB model, and a metaheuristic method based on a GA. The proposed framework, which integrates Algorithms 1 and 2, emphasizes computational efficiency and interpretability by conducting localized optimization within spatially and behaviorally coherent clusters. In contrast, the full bilevel model provides a benchmark for exact global optimization, while the GA serves as a heuristic baseline, exploring the solution space more broadly through a global search.
Table 15 clearly demonstrates the substantial computational efficiency of the proposed multi-view framework. Achieving solutions within a total runtime of approximately 4 s—comprising less than 0.2 s for the clustering stage and around 3 s for the optimization phase—this approach significantly outperforms existing methods. Specifically, it yields an approximate 99.34% reduction in runtime compared to the Direct CFLBP-CB model, which requires 608.5756 s for optimization. Furthermore, its efficiency is also pronounced compared to the GA, demonstrating an approximate 99.32% reduction when benchmarked against the 585.2 s required by the GA to first surpass the proposed method’s overall profit (at generation 6).
In terms of solution quality, the multi-view framework demonstrates varying overall profit levels when evaluated under the full model, dependent on the co-regularization parameters. Specifically, an overall profit of 15,770.30 is achieved with ( λ 1 , λ 2 ) = ( 0.0250 , 0.0750 ) , while a higher profit of 16,104.17 is obtained with ( λ 1 , λ 2 ) = ( 0.0050 , 0.1500 ) . This variability highlights the influence of parameter selection on the solution’s overall performance. At its optimal configuration, employing co-regularization parameters of ( λ 1 , λ 2 ) = ( 0.0050 , 0.1500 ) , the proposed framework yields an overall profit of 16,104.17. This result proves to be remarkably competitive with the GA, exhibiting a marginal difference of approximately 0.21% when compared to that of the GA’s highest achieved profit of 16,137.75. Furthermore, this proposed framework’s overall profit of 16,104.17 represents an approximate 0.72% increase compared to the Direct CFLBP-CB model, which yields a profit of 15,988.27.
Although the GA achieves a slightly higher profit, this advantage comes at the cost of significantly longer runtime. The proposed framework, by contrast, combines computational efficiency with interpretable, behaviorally consistent clustering, making it a scalable and practical solution for real-world decision environments that require both performance and speed.

6.7. Implications for Competitive Facility Planning

The proposed framework offers a scalable and interpretable solution for data-driven facility location in competitive environments. By incorporating a bilevel optimization model within each cluster, it captures realistic leader–follower dynamics, allowing dominant providers to anticipate and respond strategically to competitors. The attraction coefficients v i j integrate spatial proximity and customer preferences, aligning facility placements with diverse customer segments.
While the preference attributes used in this study are synthetic, they serve as a proof of concept for behavior-aware optimization. In practical applications, these features can be replaced with empirical preference or behavioral data (e.g., purchasing history, stated preferences), enabling the framework to support real-world competitive planning where customer behavior significantly influences service demand.

6.8. Limitations and Future Work

While the synthetic generation of preference attributes allows for simulating realistic customer heterogeneity, future work should incorporate actual demographic, socioeconomic, or transaction-level data to enhance empirical realism. Additionally, extending the current bilevel framework to include capacity constraints, pricing decisions, or stochastic demand, alongside exploring advanced multi-view clustering techniques, would significantly improve its real-world applicability.

7. Conclusions and Future Work

This paper introduces a novel heuristic framework for solving the competitive facility location problem with customer behavior (CFLBP-CB), integrating behavior-aware multi-view K-means clustering with co-regularization and bilevel optimization. Addressing key limitations of traditional location models, the proposed method accounts for both spatial proximity and customer behavioral preferences, thereby enabling more nuanced and effective facility placement decisions in competitive settings. The proposed framework adopts a two-stage structure. In the first stage, multi-view clustering with co-regularization is employed to partition the market based on both geographic and behavioral features. In the second stage, a bilevel optimization model is applied within each cluster to capture the hierarchical decision-making dynamics between the leader and the follower.
Our empirical evaluation, using a real-world San Francisco dataset, demonstrates the substantial efficacy of the proposed two-stage approach. It significantly reduces computation time by over 99.34% (from 608.58 to 4.00 s) and improves overall profit by approximately 0.72% (from 15,988.27 to 16,104.17) compared to directly solving the full bilevel optimization model. Furthermore, its integrated multi-view clustering with co-regularization outperforms all single-view baselines (K-means, spectral, and hierarchical), yielding an approximate 5.21% increase in overall profit and reduced optimization time by effectively capturing complementary spatial and behavioral structures. Notably, when benchmarked against the GA metaheuristic, our proposed method demonstrates a highly competitive overall profit. At its optimal co-regularization setting, an overall profit of 16,104.17 is achieved, which is only 0.21% lower than the highest recorded profit of the GA (16,137.75). This compelling combination that can deliver high-quality solutions while demanding significantly fewer computational resources renders the proposed approach exceptionally well-suited for large-scale or time-sensitive competitive facility planning tasks.
A sensitivity analysis provides further insights, underscoring the critical importance of carefully tuning the co-regularization parameters to achieve an optimal balance between spatial coherence and behavioral alignment. The analysis also indicates that, while refinement steps (e.g., majority voting and silhouette evaluation) can correct certain misalignments, strategic parameter selection fundamentally enhances both clustering quality and subsequent optimization outcomes.
Several avenues for future research are identified to further advance this framework. These include incorporating authentic real-world customer behavior data and extending the bilevel CFLBP-CB model to encompass additional decision variables like service capacities or pricing strategies. Furthermore, more advanced multi-view clustering approaches and alternative clustering techniques—such as spectral and hierarchical clustering—could be explored for their compatibility with co-regularization frameworks. Beyond these extensions, the developed framework holds potential for broader application in diverse domains, such as healthcare access, retail expansion, and public service planning, thereby broadening its relevance in evolving market conditions.

Author Contributions

Conceptualization, P.B. and T.P.; methodology, P.B. and T.P.; software, T.P.; validation, P.B., P.P. and T.P.; formal analysis, P.B. and T.P.; investigation, P.B. and T.P.; resources, P.B. and T.P.; data curation, P.B. and T.P.; writing—original draft preparation, P.B. and T.P.; writing—review and editing, P.B. and T.P.; visualization, T.P.; supervision, P.B. and T.P.; funding acquisition, P.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Fundamental Fund of Khon Kaen University. This research received funding support from the National Science, Research and Innovation Fund (NSRF).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset employed in this case study is publicly available at https://pysal.org/spopt/notebooks/facloc-real-world.html (accessed on 1 August 2024).

Acknowledgments

This research was funded by the Fundamental Fund of Khon Kaen University, Thailand.

Conflicts of Interest

The authors declare no conflicts of interest.

Notation and Definition

The following notations are adopted throughout this manuscript.
ISet of potential facility locations, indexed by i { 1 , 2 , , n } .
JSet of customers, indexed by j { 1 , 2 , , m } .
pNumber of facilities to be opened by the leader.
rNumber of facilities to be opened by the follower.
w j Demand of customer j, representing the maximum potential turnover that can be captured by a serving facility.
d i j Distance between facility location i and customer j.
p i j Preference similarity ratio between facility i and customer j.
v i j Attraction coefficient combining spatial proximity (via inverse distance) and
behavioral similarity (via p i j ), defined as v i j = p i j d i j + ϵ .
ϵ A small positive constant used to avoid division by zero.
x i Binary variable equal to 1 if the leader opens a facility at location i;
0 otherwise.
y i Binary variable equal to 1 if the follower opens a facility at location i;
0 otherwise.

References

  1. Drezner, T. A Review of Competitive Facility Location in the Plane. Logist. Res. 2014, 7, 114. [Google Scholar] [CrossRef]
  2. Eiselt, H.A.; Marianov, V.; Drezner, T. Competitive Location Models. In Location Science; Laporte, G., Nickel, S., Saldanha da Gama, F., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 391–429. [Google Scholar]
  3. ReVelle, C.; Eiselt, H.; Daskin, M. A Bibliography for Some Fundamental Problem Categories in Discrete Location Science. Eur. J. Oper. Res. 2008, 184, 817–848. [Google Scholar] [CrossRef]
  4. Fernández, P.; Pelegrín, B.; Lančinskas, A.; Žilinskas, J. New Heuristic Algorithms for Discrete Competitive Location Problems with Binary and Partially Binary Customer Behavior. Comput. Oper. Res. 2017, 79, 12–18. [Google Scholar] [CrossRef]
  5. Grohmann, S.; Urošević, D.; Carrizosa, E.; Mladenović, N. Solving Multifacility Huff Location Models on Networks Using Metaheuristic and Exact Approaches. Comput. Oper. Res. 2017, 78, 537–546. [Google Scholar] [CrossRef]
  6. Fernández, J.; Boglárka, G.; Redondo, J.L.; Ortigosa, P.M. The Probabilistic Customer’s Choice Rule with a Threshold Attraction Value: Effect on the Location of Competitive Facilities in the Plane. Comput. Oper. Res. 2019, 101, 234–249. [Google Scholar] [CrossRef]
  7. Serra, D.; Eiselt, H.A.; Laporte, G.; ReVelle, C.S. Market Capture Models under Various Customer-Choice Rules. Environ. Plan. B Plan. Des. 1999, 26, 741–750. [Google Scholar] [CrossRef]
  8. Drezner, T.; Drezner, Z. Finding the Optimal Solution to the Huff-Based Competitive Location Model. Comput. Manag. Sci. 2004, 1, 193–208. [Google Scholar] [CrossRef]
  9. Fernández, P.; Pelegrín, B.; Lančinskas, A.; Žilinskas, J. Exact and Heuristic Solutions of a Discrete Competitive Location Model with Pareto-Huff Customer Choice Rule. J. Comput. Appl. Math. 2021, 385, 113200. [Google Scholar] [CrossRef]
  10. Campos, C.M.; Santos-Peñate, D.R.; Moreno, J.A. An Exact Procedure and LP Formulations for the Leader–Follower Location Problem. TOP 2010, 18, 97–121. [Google Scholar] [CrossRef]
  11. Pelegrín, B.; Fernández, P.; García, M.D. On Tie Breaking in Competitive Location under Binary Customer Behavior. Omega 2015, 52, 156–167. [Google Scholar] [CrossRef]
  12. Plastria, F. Static Competitive Facility Location: An Overview of Optimisation Approaches. Eur. J. Oper. Res. 2001, 129, 461–470. [Google Scholar] [CrossRef]
  13. Plastria, F.; Vanhaverbeke, L. Discrete Models for Competitive Location with Foresight. Comput. Oper. Res. 2008, 35, 683–700. [Google Scholar] [CrossRef]
  14. Farahani, R.Z.; Abedian, M.; Sharahi, S. Competitive Facility Location. In Facility Location: Concepts, Models, Algorithms and Case Studies; Farahani, R.Z., Hekmatfar, M., Eds.; Springer: Berlin, Germany, 2009; pp. 347–372. [Google Scholar]
  15. Shan, W.; Yan, Q.; Chen, C.; Zhang, M.; Yao, B.; Fu, X. Optimization of Competitive Facility Location for Chain Stores. Ann. Oper. Res. 2019, 273, 187–205. [Google Scholar] [CrossRef]
  16. Zhou, L.; Du, G.; Lü, K.; Wang, L.; Du, J. A Survey and an Empirical Evaluation of Multi-View Clustering Approaches. ACM Comput. Surv. 2024, 56, 187. [Google Scholar] [CrossRef]
  17. Yang, Y. Protecting Attributes and Contents in Online Social Networks. Ph.D. Thesis, University of Kansas, Lawrence, KS, USA, 2014. [Google Scholar]
  18. Ye, F.; Chen, Z.; Qian, H.; Li, R.; Chen, C.; Zheng, Z. New Approaches in Multi-View Clustering. In Recent Applications in Data Clustering; IntechOpen: London, UK, 2018. [Google Scholar] [CrossRef]
  19. Hotelling, H. Stability in Competition. Econ. J. 1929, 39, 41–57. [Google Scholar] [CrossRef]
  20. Von Stackelberg, H. Marktform und Gleichgewicht; Springer: Berlin, Germany, 1934. [Google Scholar]
  21. Drezner, Z.; Eiselt, H.A. Competitive Location Models: A Review. Eur. J. Oper. Res. 2024, 316, 5–18. [Google Scholar] [CrossRef]
  22. Biesinger, B.; Hu, B.; Raidl, G. Models and Algorithms for Competitive Facility Location Problems with Different Customer Behavior. Ann. Math. Artif. Intell. 2016, 76, 93–119. [Google Scholar] [CrossRef]
  23. Casas-Ramírez, M.-S.; Camacho-Vallejo, J.-F. Solving the p-Median Bilevel Problem with Order through a Hybrid Heuristic. Appl. Soft Comput. 2017, 60, 73–86. [Google Scholar] [CrossRef]
  24. Kochetov, Y.; Kochetova, N.; Plyasunov, A. A Matheuristic for the Leader–Follower Facility Location and Design Problem. In Proceedings of the 10th Metaheuristics International Conference (MIC 2013), Singapore, 4–8 August 2013; Lau, H., Van Hentenryck, P., Raidl, G., Eds.; Citeseer: Princeton, NJ, USA, 2013; Volume 32, pp. 32/1–32/3. [Google Scholar]
  25. Rahmani, A.; Hosseini, M. A Competitive Stochastic Bi-Level Inventory Location Problem. Int. J. Manag. Sci. Eng. Manag. 2021, 16, 209–220. [Google Scholar] [CrossRef]
  26. Beresnev, V.L.; Melnikov, A.A. Computation of an Upper Bound in the Two-Stage Bilevel Competitive Location Model. J. Appl. Ind. Math. 2022, 16, 377–386. [Google Scholar] [CrossRef]
  27. Latifi, S.E.; Tavakkoli-Moghaddam, R.; Fazeli, E.; Arefkhani, H. Competitive Facility Location Problem with Foresight Considering Discrete-Nature Attractiveness for Facilities: Model and Solution. Comput. Oper. Res. 2022, 146, 105900. [Google Scholar] [CrossRef]
  28. Yu, W. Robust Competitive Facility Location Model with Uncertain Demand Types. PLoS ONE 2022, 17, e0273123. [Google Scholar] [CrossRef]
  29. Parvasi, S.P.; Taleizadeh, A.A.; Cárdenas-Barrón, L.E. Retail Price Competition of Domestic and International Companies: A Bi-Level Game Theoretical Optimization Approach. RAIRO-Oper. Res. 2023, 57, 291–323. [Google Scholar] [CrossRef]
  30. Zhou, Y.; Kou, Y.; Zhou, M. Bilevel Memetic Search Approach to the Soft-Clustered Vehicle Routing Problem. Transp. Sci. 2023, 57, 701–716. [Google Scholar] [CrossRef]
  31. Calvete, H.I.; Galé, C.; Iranzo, J.A.; Hernández, A. A Bilevel Approach to the Facility Location Problem with Customer Preferences under a Mill Pricing Policy. Mathematics 2024, 12, 3459. [Google Scholar] [CrossRef]
  32. Lin, Y.H.; Tian, Q.; Yu, Y. Bilevel Competitive Facility Location and Design under a Nested Logit Model. Comput. Oper. Res. 2024, 183, 107146. [Google Scholar]
  33. Legault, R.; Frejinger, E. A Model-Free Approach for Solving Choice-Based Competitive Facility Location Problems Using Simulation and Submodularity. INFORMS J. Comput. 2025, 37, 603–622. [Google Scholar] [CrossRef]
  34. Suárez-Vega, R.; Santos-Peñate, D.; García, D.P. Competitive Multifacility Location on Networks: The (r|Xp)-Medianoid Problem. J. Reg. Sci. 2004, 44, 569–588. [Google Scholar] [CrossRef]
  35. Kumar, A.; Rai, P.; Daumé, H. Co-Regularized Multi-View Spectral Clustering. Adv. Neural Inf. Process. Syst. 2011, 24, 1413–1421. [Google Scholar]
  36. Rousseeuw, P.J. Silhouettes: A Graphical Aid to the Interpretation and Validation of Cluster Analysis. J. Comput. Appl. Math. 1987, 20, 53–65. [Google Scholar] [CrossRef]
Figure 1. Visualization of a CFLP and CFLBP-CB. (a) Illustrates the spatial distribution of customer locations ( C j ) and candidate facility sites ( F i ). locations C j and candidate facility sites F i . (b) Optimal facility placement for the leader ( p = 2 ) and the follower ( r = 2 ) in the CFLP. The dashed lines indicate customer–facility assignments: purple and green lines represent assignments to leader facilities, while pink and blue lines represent assignments to follower facilities. (c) Optimal facility placement for the leader ( p = 2 ) and the follower ( r = 2 ) in the CFLBP-CB. This subfigure illustrates the solution under the CFLBP-CB model, which incorporates customer preference similarity into the customer–facility assignment process.
Figure 1. Visualization of a CFLP and CFLBP-CB. (a) Illustrates the spatial distribution of customer locations ( C j ) and candidate facility sites ( F i ). locations C j and candidate facility sites F i . (b) Optimal facility placement for the leader ( p = 2 ) and the follower ( r = 2 ) in the CFLP. The dashed lines indicate customer–facility assignments: purple and green lines represent assignments to leader facilities, while pink and blue lines represent assignments to follower facilities. (c) Optimal facility placement for the leader ( p = 2 ) and the follower ( r = 2 ) in the CFLBP-CB. This subfigure illustrates the solution under the CFLBP-CB model, which incorporates customer preference similarity into the customer–facility assignment process.
Mathematics 13 02481 g001
Figure 2. Intermediate clustering results from the multi-view K-means algorithm with co-regularization. (Left) Clustering based on geographic information (X–Y). (Right) Clustering based on preference information (Preferences 1–2). Colored points represent data instances (customers and facilities), while colored crosses denote the cluster centroids in each view.
Figure 2. Intermediate clustering results from the multi-view K-means algorithm with co-regularization. (Left) Clustering based on geographic information (X–Y). (Right) Clustering based on preference information (Preferences 1–2). Colored points represent data instances (customers and facilities), while colored crosses denote the cluster centroids in each view.
Mathematics 13 02481 g002
Figure 3. Visualization of clustering differences between the two data views. Colored dots represent data points (customers and facilities) assigned to different clusters, while colored crosses indicate the corresponding cluster centroids. (Left) Clustering results based on geographic information (X–Y). (Right) Clustering results based on preference information (Preferences 1–2). Black circles indicate data points that are assigned to different clusters across the two views, highlighting inconsistencies caused by divergent feature characteristics.
Figure 3. Visualization of clustering differences between the two data views. Colored dots represent data points (customers and facilities) assigned to different clusters, while colored crosses indicate the corresponding cluster centroids. (Left) Clustering results based on geographic information (X–Y). (Right) Clustering results based on preference information (Preferences 1–2). Black circles indicate data points that are assigned to different clusters across the two views, highlighting inconsistencies caused by divergent feature characteristics.
Mathematics 13 02481 g003
Figure 4. Final clustering results after applying multi-view K-means clustering with co-regularization, followed by majority voting and silhouette-based refinement. Colored dots represent data points (customers and facilities) assigned to different clusters, while colored crosses indicate the corresponding cluster centroids. (Left) Clustering result based on geographic information (X–Y). (Right) Clustering result based on preference information (Preferences 1–2).
Figure 4. Final clustering results after applying multi-view K-means clustering with co-regularization, followed by majority voting and silhouette-based refinement. Colored dots represent data points (customers and facilities) assigned to different clusters, while colored crosses indicate the corresponding cluster centroids. (Left) Clustering result based on geographic information (X–Y). (Right) Clustering result based on preference information (Preferences 1–2).
Mathematics 13 02481 g004
Figure 5. Optimal facility locations within the multi-view K-means clustering framework, identified using spatial (X–Y) and preference (Preferences 1–2) information. Colored points indicate customer locations by cluster, while colored triangles represent the corresponding optimal facility placements in each view. This visualization highlights the alignment between customer distributions and optimal facility siting, which contributes to the profit outcomes reported in Table 7.
Figure 5. Optimal facility locations within the multi-view K-means clustering framework, identified using spatial (X–Y) and preference (Preferences 1–2) information. Colored points indicate customer locations by cluster, while colored triangles represent the corresponding optimal facility placements in each view. This visualization highlights the alignment between customer distributions and optimal facility siting, which contributes to the profit outcomes reported in Table 7.
Mathematics 13 02481 g005
Figure 6. Illustrations of the impact of co-regularization parameters on the proposed framework: (a) Side-view surface plot illustrating the influence of co-regularization parameters λ 1 and λ 2 on overall profit. (b) Top-view surface plot illustrating the influence of co-regularization parameters λ 1 and λ 2 on overall profit.
Figure 6. Illustrations of the impact of co-regularization parameters on the proposed framework: (a) Side-view surface plot illustrating the influence of co-regularization parameters λ 1 and λ 2 on overall profit. (b) Top-view surface plot illustrating the influence of co-regularization parameters λ 1 and λ 2 on overall profit.
Mathematics 13 02481 g006
Figure 7. Optimal facility placements obtained using single-view K-means clustering across different input features. Colored points indicate customer locations by cluster, while colored triangles represent the corresponding optimal facility placements in each view. Each subfigure presents two projections: (Left) representation of optimal facility locations in spatial features (X–Y) and (Right) representation of optimal facility locations in preference features (Preferences 1–2). (a) K-means: geographic (2D); (b) K-means: preference (2D); (c) K-means: geographic + preference (4D).
Figure 7. Optimal facility placements obtained using single-view K-means clustering across different input features. Colored points indicate customer locations by cluster, while colored triangles represent the corresponding optimal facility placements in each view. Each subfigure presents two projections: (Left) representation of optimal facility locations in spatial features (X–Y) and (Right) representation of optimal facility locations in preference features (Preferences 1–2). (a) K-means: geographic (2D); (b) K-means: preference (2D); (c) K-means: geographic + preference (4D).
Mathematics 13 02481 g007
Figure 8. Optimal facility placements obtained using single-view spectral clustering across different input features. Colored points indicate customer locations by cluster, while colored triangles represent the corresponding optimal facility placements in each view. Each subfigure presents two projections: (Left) representation of optimal facility locations in spatial features (X–Y) and (Right) representation of optimal facility locations in preference features (Preferences 1–2). (a) Spectral: geographic (2D); (b) Spectral: preference (2D); (c) Spectral: geographic + preference (4D).
Figure 8. Optimal facility placements obtained using single-view spectral clustering across different input features. Colored points indicate customer locations by cluster, while colored triangles represent the corresponding optimal facility placements in each view. Each subfigure presents two projections: (Left) representation of optimal facility locations in spatial features (X–Y) and (Right) representation of optimal facility locations in preference features (Preferences 1–2). (a) Spectral: geographic (2D); (b) Spectral: preference (2D); (c) Spectral: geographic + preference (4D).
Mathematics 13 02481 g008
Figure 9. Optimal facility placements obtained using single-view hierarchical clustering across different input features. Colored points indicate customer locations by cluster, while colored triangles represent the corresponding optimal facility placements in each view. Each subfigure presents two projections: (Left) representation of optimal facility locations in spatial features (X–Y) and (Right) representation of optimal facility locations in preference features (Preferences 1–2). (a) Hierarchical: geographic (2D); (b) Hierarchical: preference (2D); (c) Hierarchical: geographic + preference (4D).
Figure 9. Optimal facility placements obtained using single-view hierarchical clustering across different input features. Colored points indicate customer locations by cluster, while colored triangles represent the corresponding optimal facility placements in each view. Each subfigure presents two projections: (Left) representation of optimal facility locations in spatial features (X–Y) and (Right) representation of optimal facility locations in preference features (Preferences 1–2). (a) Hierarchical: geographic (2D); (b) Hierarchical: preference (2D); (c) Hierarchical: geographic + preference (4D).
Mathematics 13 02481 g009
Figure 10. Optimal facility placements obtained using single-view K-means clustering across different input features. Colored points indicate customer locations by cluster, while colored triangles represent the corresponding optimal facility placements in each view. Each subfigure presents two projections: (Left) representation of optimal facility locations in spatial features (X–Y) and (Right) representation of optimal facility locations in preference features (Preferences 1–2). (a) Optimal facility placements obtained using the proposed framework with multi-view K-means clustering under co-regularization parameters ( λ 1 , λ 2 ) = ( 0.0050 , 0.1500 ) . (b) Optimal facility placements obtained using the proposed framework with multi-view K-means clustering under co-regularization parameters ( λ 1 , λ 2 ) = ( 0.0050 , 0.0900 ) .
Figure 10. Optimal facility placements obtained using single-view K-means clustering across different input features. Colored points indicate customer locations by cluster, while colored triangles represent the corresponding optimal facility placements in each view. Each subfigure presents two projections: (Left) representation of optimal facility locations in spatial features (X–Y) and (Right) representation of optimal facility locations in preference features (Preferences 1–2). (a) Optimal facility placements obtained using the proposed framework with multi-view K-means clustering under co-regularization parameters ( λ 1 , λ 2 ) = ( 0.0050 , 0.1500 ) . (b) Optimal facility placements obtained using the proposed framework with multi-view K-means clustering under co-regularization parameters ( λ 1 , λ 2 ) = ( 0.0050 , 0.0900 ) .
Mathematics 13 02481 g010
Table 1. Summary of recent studies on the competitive facility location problem (CFLP).
Table 1. Summary of recent studies on the competitive facility location problem (CFLP).
StudyObjective FunctionSolution MethodApplication Domain
Drezner (2014)  [1]Maximize market share with location and quality differentiationBilevel model with best-response dynamicsRetail competition with quality-sensitive customers
Biesinger et al. (2017) [22]Maximize market share with different types of customer behavior rulesMILPRetail competition
Casas-Ramírez et al. (2017) [23]Minimize the total cost between facility and customerReformulation of bilevel optimization to single-level optimization and heuristicsRetail competition, service networks
Kochetov et al. (2018) [24]Maximize market share under proportional customer behaviorBilevel MINLP, heuristicsService networks
Rahmani et al. (2021) [25]Maximize expected profit under demand uncertainty and competitionBranch-and-cut algorithmInventory and distribution systems
Beresnev et al. (2022) [26]              Maximize leader’s profit in a two-stage bilevel modelMILP reformulation to compute upper boundCompetitive retail location
Latifi et al. (2022) [27]Maximize follower’s market share with discrete facility attractiveness and foresightBilevel integer programming with exact algorithm and dominance rulesCompetitive retail location planning
Yu et al. (2022) [28]Maximize leader’s market share under demand uncertaintyTwo-stage robust optimizationRetail facility planning
Parvasi et al. (2023) [29]Maximize domestic firm’s profit via price-setting competitionBilevel game-theoretic optimizationInternational retail price competition
Zhou et al. (2023) [30]Minimize total routing cost under soft customer clusteringBilevel memetic algorithm with savings heuristic and local searchLogistics and vehicle routing
Calvete et al. (2024) [31]Maximize net profit and total customer preferenceMulti-objective optimizationSupply chain planning
Lin et al. (2024) [32]Maximize own revenue considering nested customer preferencesBilevel optimization, nested logit modelRetail stores, parcel lockers, park-and-ride stations
Legault et al. (2025) [33]Maximize expected market share under random utility-based customer choiceSubmodular optimization, simulation-based reformulationRetail location with choice uncertainty
Table 2. Customer preferences and facility capabilities. Dashes (—) indicate values that are not applicable.
Table 2. Customer preferences and facility capabilities. Dashes (—) indicate values that are not applicable.
DemandSupply
ID Product Category 1 Product Category 2 Product Category 1 Product Category 2
( Preference 1 , j ) ( Preference 2 , j ) ( Preference 1 , i ) ( Preference 2 , i )
Customer 11.01.0
Customer 20.80.9
Customer 31.00.7
Customer 40.90.9
Facility 10.90.7
Facility 20.80.9
Facility 31.01.0
Table 3. Preference similarity ratio matrix p i j . .
Table 3. Preference similarity ratio matrix p i j . .
Facility∖CustomerCustomer 1Customer 2Customer 3Customer 4
Facility 10.63000.77780.90000.7778
Facility 20.72001.00000.80000.8889
Facility 31.00001.00001.00001.0000
Table 4. Candidate facility locations in the city of San Francisco 1.
Table 4. Candidate facility locations in the city of San Francisco 1.
Candidate LocationLatitudeLongitude
Location 137.7724–122.5100
Location 237.7538–122.4889
Location 1637.7971–122.3989
1 Source: PySAL facility location dataset.
Table 5. Customer demand sites in the city of San Francisco 1.
Table 5. Customer demand sites in the city of San Francisco 1.
Customer IDLatitudeLongitudePopulation
137.6508−122.48874135
237.6600−122.48354831
20537.7823−122.41658188
1 Source: PySAL facility location dataset.
Table 6. Number of customers and facilities assigned to each cluster.
Table 6. Number of customers and facilities assigned to each cluster.
ClusterNumber of CustomersNumber of Facilities
0432
1352
2325
3445
4512
Table 7. Overall, leader, and follower profits for each cluster.
Table 7. Overall, leader, and follower profits for each cluster.
ClusteringOverall ProfitLeader Profit/Leader Market ShareFollower ProfitFollower Market Share
Equation (1) + Equation (10) Equation (1) Equation (10) Equation (4)
 Cluster 03673.642394.471279.171905.52
 Cluster 13208.622421.07787.551078.93
 Cluster 22152.801650.62603.421650.62
 Cluster 33075.482430.24645.241969.76
 Cluster 44307.102740.411566.692359.59
Total profit16,417.64
Time (s)
 Cluster0.1708
 Optimize2.9801
Table 8. Sensitivity of overall profit with respect to co-regularization parameters λ 1 and λ 2 .
Table 8. Sensitivity of overall profit with respect to co-regularization parameters λ 1 and λ 2 .
Co-Regularization ParametersOverall Profit
λ 1 λ 2
0.025014,577.37
0.0050.04515,493.84
0.00250.10515,279.92
0015,812.44
Table 9. Profit outcomes of the bilevel optimization model under three single-view clustering algorithms (K-means, spectral, hierarchical) applied to different input feature views: geographic only (2D), preference only (2D), and combined (4D). Each entry includes clustering and optimization time, as well as per-cluster and total profits for leader and follower facilities.
Table 9. Profit outcomes of the bilevel optimization model under three single-view clustering algorithms (K-means, spectral, hierarchical) applied to different input feature views: geographic only (2D), preference only (2D), and combined (4D). Each entry includes clustering and optimization time, as well as per-cluster and total profits for leader and follower facilities.
Clustering Feature ViewClustering AlgorithmTime (s)OverallLeaderFollower
Cluster Optimize Profit Profit Profit
Geographic Only (2D)K-means
     Cluster 0 2487.711646.93840.78
     Cluster 1 4492.303514.33977.96
     Cluster 2 1748.421109.41639.01
     Cluster 3 4717.912945.071772.84
     Cluster 4 2116.531503.62612.91
Total profit 0.03493.101915,562.8710,719.374843.50
Preference Only (2D)K-means
     Cluster 0 4115.543336.64778.90
     Cluster 1 2587.652342.99244.66
     Cluster 2 2695.602229.22466.38
     Cluster 3 3894.702351.021543.69
     Cluster 4 2267.661607.27660.39
Total profit 0.06044.619115,561.1611,867.133694.03
Geographic + Preference (4D)K-means
     Cluster 0 3706.873031.13675.74
     Cluster 1 3746.442993.62752.82
     Cluster 2 2554.851895.87658.98
     Cluster 3 1944.561016.50928.06
     Cluster 4 2452.301783.51668.79
Total profit 0.06654.916214,405.0210,720.633684.39
Geographic Only (2D)Spectral
     Cluster 0 4418.973640.63778.34
     Cluster 1 3622.132663.59958.54
     Cluster 2 2617.911620.06997.84
     Cluster 3 1988.181241.54746.64
     Cluster 4 2348.861888.13460.73
Total profit 0.03634.664814,996.0511,053.953942.10
Preference Only (2D)Spectral
     Cluster 0 5996.145313.10683.04
     Cluster 1 2422.522061.96360.57
     Cluster 2 2437.391972.63464.76
     Cluster 3 2137.911384.10753.81
     Cluster 4 2610.021707.89902.14
Total profit 0.03306.309615,603.9912,439.673164.32
Geographic + Preference (4D)Spectral
     Cluster 0 2259.981645.98614.00
     Cluster 1 2555.682189.37366.32
     Cluster 2 4686.463531.551154.91
     Cluster 3 3180.111974.971205.14
     Cluster 4 2046.261486.62559.64
Total profit 0.03306.309615,603.9912,439.673164.32
Geographic Only (2D)Hierarchical
     Cluster 0 4336.973398.04938.92
     Cluster 1 3559.182648.60910.57
     Cluster 2 2545.231745.23800.00
     Cluster 3 1663.021055.61607.41
     Cluster 4 3402.492345.991056.50
Total profit 0.00342.786415,292.7110,979.304313.41
Preference Only (2D)Hierarchical
     Cluster 0 3576.642175.021401.62
     Cluster 1 2784.321921.26863.06
     Cluster 2 3359.662748.92610.74
     Cluster 3 4460.883824.70636.19
     Cluster 4 1311.061102.98208.09
Total profit 0.00354.612115,376.4711,534.483841.98
Geographic + Preference (4D)Hierarchical
     Cluster 0 3789.642875.07914.56
     Cluster 1 2553.211930.72622.49
     Cluster 2 4141.972942.291199.68
     Cluster 3 2587.482101.52485.96
     Cluster 4 1692.951267.86425.09
Total profit 0.00354.502814,765.2511,117.483647.77
Table 10. Bilevel optimization results under different co-regularization settings ( λ 1 , λ 2 ) using multi-view clustering.
Table 10. Bilevel optimization results under different co-regularization settings ( λ 1 , λ 2 ) using multi-view clustering.
Co-Regularization ParametersClusterLeader ProfitFollower Profit
λ 1 λ 2
0.02250.075Cluster 02632.181389.18
Cluster 12565.93831.55
Cluster 21557.11600.06
Cluster 31879.81547.71
Cluster 42740.411566.69
Total profit16,310.62
Time (s)
Cluster0.1255
Optimize3.0761
0.0150.120Cluster 01993.431020.81
Cluster 12899.40873.65
Cluster 22571.18395.45
Cluster 31776.11700.12
Cluster 42644.411444.39
Total profit16,318.94
Time (s)
Cluster0.0915
Optimize3.4215
0.0250.015Cluster 02444.47574.85
Cluster 12037.711016.45
Cluster 22725.751107.30
Cluster 31318.40726.50
Cluster 42501.471338.96
Total profit15,791.85
Time (s)
Cluster0.1001
Optimize3.4982
0.0000.000Cluster 02505.52909.47
Cluster 13522.011365.49
Cluster 21711.16869.78
Cluster 32151.39949.43
Cluster 41192.56635.64
Total profit15,812.44
Time (s)
Cluster0.1750
Optimize3.5193
Table 11. Profit and runtime comparison between single-view and multi-view clustering methods.
Table 11. Profit and runtime comparison between single-view and multi-view clustering methods.
Clustering Feature View/
Co-Regularization Parameters
Clustering AlgorithmTime (s)OverallLeaderFollower
Cluster Optimize Profit Profit Profit
Geographic Only (2D)Single-view0.03493.101915,562.8710,719.374843.50
Preference Only (2D)K-means clustering0.06044.619115,561.1611,867.133694.03
Geographic + Preference (4D) 0.06654.916214,405.0210,720.633684.39
Geographic Only (2D)Single-view0.03634.664814,996.0511,053.953942.10
Preference Only (2D)Spectral clustering0.03306.309615,603.9912,439.673164.32
Geographic + Preference (4D) 0.03306.309615,603.9912,439.673164.32
Geographic Only (2D)Single-view0.00342.786415,292.7110,979.304313.41
Preference Only (2D)Hierarchical clustering0.00354.612115,376.4711,534.483841.98
Geographic + Preference (4D) 0.00354.502814,765.2511,117.483647.77
( λ 1 , λ 2 ) = ( 0.025 , 0.075 ) Multi-view0.17082.980116,417.6411,535.574882.07
( λ 1 , λ 2 ) = ( 0.0225 , 0.075 ) K-means clustering0.12553.076116,310.6211,375.444935.18
( λ 1 , λ 2 ) = ( 0.015 , 0.120 ) 0.09153.421516,318.9411,884.534434.41
( λ 1 , λ 2 ) = ( 0.025 , 0.015 ) 0.10013.498215,791.8511,027.804764.05
( λ 1 , λ 2 ) = ( 0.000 , 0.000 ) 0.17503.519315,812.4411,082.644729.80
Table 12. Evaluation of the top-10 profit rankings when applying the optimal facility placements from the proposed framework to the full CFLBP-CB model.
Table 12. Evaluation of the top-10 profit rankings when applying the optimal facility placements from the proposed framework to the full CFLBP-CB model.
Co-Regularization ParametersConsistencySilhouetteCohesionProfit
X–Y Pref 1–2 X–Y Pref 1–2 Framework Full Model
( 0.0050 , 0.1500 ) 2210.3505–0.07730.73931.501415,504.5416,104.18
( 0.0175 , 0.0000 ) 1650.02080.40590.66742.001414,969.9316,071.82
( 0.0100 , 0.1350 ) 1970.20740.04080.70081.657314,694.0215,907.97
( 0.0050 , 0.0900 ) 1200.39510.15090.78251.857915,638.2715,882.79
( 0.0075 , 0.0750 ) 1170.33740.19090.77441.895215,559.7215,882.79
( 0.0100 , 0.0600 ) 1260.18210.28540.72031.934715,742.9715,835.59
( 0.0250 , 0.0750 ) 2050.00730.18720.65651.846916,417.6415,770.30
( 0.0125 , 0.0150 ) 1030.17650.39140.7151.99715,548.1715,758.24
( 0.0125 , 0.1500 ) 2200.0274–0.07850.65631.526914,760.9615,751.88
Table 13. Correlation between objective values obtained from the proposed framework and the full CFLBP-CB model.
Table 13. Correlation between objective values obtained from the proposed framework and the full CFLBP-CB model.
Evaluation CriteriaCorrelation Coefficient
Overall objective0.2271
Leader profit0.5868
Follower profit0.2506
Follower market share0.5868
Table 14. Genetic algorithm optimization progress.
Table 14. Genetic algorithm optimization progress.
GenerationEvaluationsAverage Overall ProfitMaximum Overall ProfitTime (Seconds)
02014,941.6015,595.0491.77
11815,264.7715,661.8289.97
21615,424.8715,991.5480.94
31915,565.0215,991.5493.35
41315,756.8116,028.8466.15
51815,930.9816,099.7689.54
61515,963.9316,104.4073.48
71415,971.3216,104.4069.96
81216,030.2316,104.4058.89
91516,039.7516,137.7574.06
Table 15. Profit and runtime comparison between the proposed framework with co-regularization ( λ 1 , λ 2 ) and the baseline CFLBP-CB model. Underlined values indicate results that are directly comparable across methods.
Table 15. Profit and runtime comparison between the proposed framework with co-regularization ( λ 1 , λ 2 ) and the baseline CFLBP-CB model. Underlined values indicate results that are directly comparable across methods.
Solution FrameworkCo-RegularizationTime (s)OverallLeaderFollowerFollower
Cluster Optimize Profit Profit Profit Market Share
Multi-view framework(0.0250, 0.0750)0.17082.980116,417.6411,535.574882.078964.43
Multi-view solution on full model 15,770.3011,390.204380.109109.80
Multi-view framework(0.0050, 0.1500)0.10423.896315,504.5411,055.314449.229444.69
Multi-view solution on full model 16,104.1711,798.484305.698701.51
Direct CFLBP-CB 608.575615,988.2711,254.594733.689245.41
GA
gen 6 585.200016,104.4011,594.584509.828905.42
gen 9 788.110016,137.7511,872.014265.738627.98
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Phoka, T.; Poonprapan, P.; Boriwan, P. A Heuristic Approach to Competitive Facility Location via Multi-View K-Means Clustering with Co-Regularization and Customer Behavior. Mathematics 2025, 13, 2481. https://doi.org/10.3390/math13152481

AMA Style

Phoka T, Poonprapan P, Boriwan P. A Heuristic Approach to Competitive Facility Location via Multi-View K-Means Clustering with Co-Regularization and Customer Behavior. Mathematics. 2025; 13(15):2481. https://doi.org/10.3390/math13152481

Chicago/Turabian Style

Phoka, Thanathorn, Praeploy Poonprapan, and Pornpimon Boriwan. 2025. "A Heuristic Approach to Competitive Facility Location via Multi-View K-Means Clustering with Co-Regularization and Customer Behavior" Mathematics 13, no. 15: 2481. https://doi.org/10.3390/math13152481

APA Style

Phoka, T., Poonprapan, P., & Boriwan, P. (2025). A Heuristic Approach to Competitive Facility Location via Multi-View K-Means Clustering with Co-Regularization and Customer Behavior. Mathematics, 13(15), 2481. https://doi.org/10.3390/math13152481

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop