This study proposes a weight optimization method that integrates expert-defined constraints with data-driven methodologies. Dual-source training labels integrate field reference data and expert knowledge: field reference data reflect the combined influence of multiple criteria, while AHP–WLC generates expert evaluation results with WLC thresholds determined via I-KMEANS clustering. A multilayer perceptron (MLP) is employed to model the relationship between input criteria combinations and the hybrid training labels, with SHAP interpretability analysis applied to quantify the contribution of each criterion. This framework balances expert knowledge with data-driven patterns.
2.4.1. AHP–WLC Label Generation
- (1)
Expert-Driven Weight Determination Via AHP
The standard workflow for determining weights using the AHP consists of defining the set of evaluation criteria, constructing pairwise comparison matrices based on expert judgment, and verifying matrix consistency. Valid weight results are outputted only when passing the consistency check (CR < 0.1) [
60]. The judgment matrices are constructed based on
Table 3.
The obtained judgment matrix is contained in
Table 4.
Calculate weights using the geometric mean method.
where
aij is an element in the pairwise comparison matrix.
Using the derived weights, compute the maximum eigenvalue
λmax of the pairwise comparison matrix.
Compute the Consistency Index (
CI) using the obtained maximum eigenvalue.
Compute the Consistency Ratio (
CR) using Consistency Index (
CI) and Random Index (
RI) values.
Verify if CR exceeds 0.1. If CR < 0.1, the consistency verification is passed. Adopt the weights obtained via the geometric mean method as final criteria weights.
- (2)
Determination of WLC Thresholds via I-KMEANS
Pearson correlation coefficients, which measure linear relationships between continuous variables based on covariance and standard deviations, were calculated among criteria. Criterion threshold determination methods were assigned according to correlation levels: criteria with strong correlations underwent joint data-driven partitioning, while weakly correlated criteria were partitioned using expert knowledge and data distribution patterns.
Data-driven unsupervised learning algorithms face two critical limitations: Lack of prior constraints. Vulnerable to sensitivity in discrete variable distributions, leading to degraded clustering performance and interpretability. Unregulated thresholds. Evaluation dimensions often include non-restrictive thresholds that deviate significantly from rational ranges (e.g., PV construction in slopes > 7° remains feasible but substantially increases costs).
This study introduces domain knowledge-driven threshold filtering rules to construct expert informed constraints. By presetting valid domains in feature space, the method applies empirical constraints to input data, achieving two key benefits: Mitigates cognitive bias risks from fully expert-dependent weighting (e.g., AHP) and enhances unsupervised algorithm robustness and semantic traceability.
The traditional K-means algorithm partitions the feature space by minimizing within-cluster variance and represents cluster centroids as a two-dimensional matrix, where each column corresponds to an evaluation criterion. However, in renewable energy suitability evaluation, criteria are subject to strict monotonicity constraints (e.g., higher GHI values should correspond to higher suitability scores). Traditional cluster centroids often produce disordered centroid arrangements that violate these monotonicity constraints.
This study proposes an Isotonic K-means (I-KMEANS)algorithm, which enforces monotonicity by alternating between unconstrained optimization steps and constrained projection operations.
The I-KMEANS algorithm is inspired by the projected gradient method in constrained nonlinear optimization, establishing a dual-loop iterative structure (see
Figure 5). The outer loop performs standard K-means updates of cluster centroids, analogous to gradient descent, while the inner loop performs constraint projection, where isotonic regression is employed to enforce monotonicity constraints on the unconstrained centers from the outer loop. To prevent identical column values, small perturbation is applied to duplicate elements, followed by SLSQP-based column scaling to preserve similarity to the original distribution while ensuring sufficient variance. This sequence of operations constitutes a single iteration of the inner loop.
The projection steps can be decomposed into the following operations:
- (i)
Isotonic regression: Isotonic regression modifies each column of the unconstrained cluster centroids individually to satisfy the specified monotonicity constraints. For individual columns, the adjustment objective can be expressed as:
In this study, the PAV (Pool Adjacent Violators) algorithm was utilized to perform isotonic regression. Initially, each observation is treated as an independent block, and a monotonic sequence is constructed by iteratively merging adjacent blocks that violate the monotonicity constraint using weighted averaging. The process can be formally expressed as
In the equation,
Bj denotes the
jth block,
wi represents the weight of the
ith sample,
yi is the value of the
ith sample,
µj refers to the mean of block
Bj, and
µnew indicates the mean of the newly merged block. One iteration of the PAV algorithm is completed by assigning
µnew to all values in the merged violating blocks. The process ends when all blocks meet the required monotonicity.
- (ii)
Result perturbation: Equal group sizes across successive iterations frequently result in identical values along the same dimension. A perturbation strategy is applied to identical entries in the isotonic regression output, following this rule:
- (iii)
Result scaling: Final results are derived through column-wise scaling of the perturbed centroids.
Parameters
a and
b in the scaling formula are optimized on a per-column basis using the Sequential Least-Squares Quadratic Programming (SLSQP) algorithm, with the objective of minimizing the deviation between the scaled and original values while simultaneously maximizing the variance within each column. Predefined weighting coefficients are introduced to balance the dual objectives, thereby ensuring that the scaled data retains key structural characteristics of the original data while promoting more uniform distributions across columns.
where
λ1,
λ2 are hyperparameters controlling the variance and distribution constraint weights, respectively.
Vmin is the minimum variance threshold, and ∆
min is the minimum spacing requirement (both are hyperparameters).
The algorithm optimizes scaling parameters a and b per column by approximating the objective function with quadratic models and constraints with linear ones. It iteratively updates the solution along the optimal search direction, ensuring objective reduction and constraint satisfaction. This results in optimal scaling that preserves data fidelity while improving feature discriminability. Once the parameters a and b are determined for all columns of the centroid matrix, scaling is applied to complete the projection step, thereby concluding a single inner-loop iteration. A full iteration consists of an outer loop and a corresponding inner loop. The iteration terminates when the displacement of all cluster centroids between two successive iterations falls below a predefined threshold ε. Three thresholds are calculated using the four cluster centroids derived from I-KMEANS, with the specific formula expressed as
- (3)
Result Computation Via WLC
Criteria classification thresholds were determined based on a combination of I-KMEANS clustering and expert knowledge, to facilitate multi-criteria weighting and aggregation using the AHPWLC framework. For supply–demand suitability evaluation, conditional normalization eliminated land price criterion impacts prior to result computation.
2.4.2. Training Sample Generation
The field reference data and the AHP–WLC-derived supply–demand suitability scores were independently normalized within each grid cell.
where
Ei is the field reference data of the
ith grid cell;
Si is the AHP–WLC suitability score of the
ith grid cell.
The two normalized datasets are subsequently aggregated using a weighted summation.
Then normalize the weighted summation results to obtain the final output. In this paper,
w1 is set to be 0.4, and
w2 is set to be 0.6.
Training samples are constructed by combining fused labels with normalized values of the criteria involved in MLP training within their corresponding grid cells, formatted as [label, x1, x2, …, xn]. Twenty percent of the total samples are randomly selected as validation data to evaluate model performance after each training epoch. These validation samples are held throughout the training process.
2.4.3. Refined Weight Determination Via MLP
The multilayer perceptron (MLP), a classic feedforward neural network, achieves high-dimensional feature representation through hierarchical nonlinear transformations. Its modular architecture includes an input layer, multiple hidden layers using ReLU or Sigmoid activation functions, and a task-oriented output layer. The model extracts abstract features through successive affine transformations and nonlinear activations in hidden layers, trained using backpropagation to minimize a loss function via gradient descent [
61].
This study constructs a three-layer MLP architecture using ReLU activations to model nonlinear relationships. The model quantifies prediction deviations via Mean Squared Error (MSE) and optimizes parameters through a backpropagation algorithm, learning a nonlinear mapping from inputs to targets. The architecture effectively captures complex geographical correlations in PV suitability through synergistic mechanisms combining hidden-layer feature abstraction (dimensional expansion) with output-layer dimensionality reduction mapping. The specific structure can be expressed mathematically.
Dimension specifications:
Input vector: x ∈ Rd (d-dimensional features)
Hidden layer: W(1) ∈ Rh×d (weight matrix), b(1) ∈ Rh (bias vector)
Output layer: W(2) ∈ Ro×h (weight matrix), b(2) ∈ Ro (bias vector)
g(·): Hidden layer Activation (E.G., Relu: g(z) = max(0, z))
In MLP training, parameter updates are governed by a training strategy, comprising optimizers and schedulers. This study employs the Adam optimizer [
62] and cosine annealing scheduler [
63]. The Adam optimizer [
62] dynamically adjusts parameter update step sizes using first and second moment estimates of gradients, while the cosine annealing scheduler [
63] periodically modulates learning rates by simulating a cosine function curve, enabling the model to escape local optima and enhance generalization capability.
For the trained model, SHAP (Shapley Additive explanations) values [
64] are computed to quantify the contribution of each input feature to the model’s predictions. Rooted in Shapley values from cooperative game theory, SHAP quantifies feature impacts by computing weighted integrations of marginal contributions.
Specifically for deep neural networks, the Deep Explainer algorithm (an enhanced variant of Deep LIFT) is employed. This method calculates feature attribution values through backpropagated gradients.
The recalibrated weights are computed by summing the weight allocations of participating computation criteria, with the land price criterion explicitly excluded from this process. This summation constitutes the final recalibrated weight values.