1. Introduction
Global climate change, ecological degradation, and energy resource scarcity profoundly affect the survival and development of human society [
1]. The world has reached a consensus on reducing fossil energy consumption and vigorously developing clean energy; increasing the share of safe, green, and reliable low-carbon energy has become the trend and direction of global energy development [
2]. To achieve the dual-carbon goals (carbon peaking and carbon neutrality), China is accelerating adjustments to its energy structure and promoting a green, low-carbon energy transition [
3]. In this context, alongside the vigorous development of renewable energies, such as hydropower, wind, and solar, nuclear power has also been prioritized as a stable baseload source [
4]. As of the end of the first quarter of 2025, China’s nuclear installed capacity has exceeded 55 GW, playing a critical role in ensuring grid stability and meeting peak-regulation requirements.
Therefore, in the forthcoming development phase, it is imperative to vigorously develop renewable energies, such as hydropower, wind, and solar, to continuously promote a high proportion of clean energy, to drive sustained growth in renewable energy development, to focus on reducing fossil energy exploitation and utilization, and to gradually realize both the incremental and stock replacement of energy in China [
5,
6]. As of the end of 2024, China had constructed more than 94,000 dams of various types, including 29 mega hydropower stations each with installed capacity exceeding one million kilowatts, all ranking first in the world. Specifically, by the end of the first quarter of 2025, China’s grid-connected installed capacity of wind and solar power reached 1480 million kilowatts, historically surpassing the installed capacity of thermal power, with wind power at 536 million kilowatts and photovoltaic power at 946 million kilowatts.
In the context of the large-scale integration of wind and solar power, as hydropower becomes coupled with these renewable sources, it must frequently adjust its output in response to wind and solar generation fluctuations, rendering the traditional independent operation modes of “water-determined power” and “power-determined water” no longer applicable [
7,
8]. Therefore, how to account for the impact of large-scale renewable energy integration on hydropower scheduling and to effectively balance the power system’s renewable energy absorption requirements with multi-energy comprehensive benefit realization represents a research hotspot and challenge in the field of hydropower optimal scheduling [
9]. Wen et al. [
10] addressed the capacity and operation optimization problem for cascade hydropower stations with high levels of wind and solar integration by proposing a multi-scale nested joint operation model, which—under consideration of wind and solar forecast uncertainty—ensures long-term generation benefits while enhancing short-term supply reliability. Wang et al. [
11] proposed a long-term scheduling decision model driven by forecasts to tackle the challenges posed by wind and solar fluctuations and hydropower resource utilization decisions in long-term operation, thereby improving hydropower’s response capability to renewable uncertainty and optimizing the overall system economy. Jin et al. [
12], focusing on cascade hydropower operation strategies under varying wind and photovoltaic penetration levels, proposed typical-day aggregation and hybrid planning optimization schemes that effectively perform peak shaving and valley filling and reduce hydropower output fluctuations under different penetration scenarios. Angarita et al. [
13] proposed a stochastic optimization method to address wind-power output forecasting uncertainty, aiming to maximize the combined generation benefit of hydropower and wind power.
In order to maximize the comprehensive benefits of hydropower station operation management, it is necessary to comprehensively coordinate all parties’ requirements during scheduling. On the water-use side, hydropower scheduling should fully consider integrated basin water-resource regulation to meet differentiated demands across various temporal and spatial scales and types, such as flood control [
14], water supply [
15], power generation [
16], navigation [
17], and ecological protection [
18]; on the electricity-use side, hydropower scheduling should exploit its flexibility to smooth generation–load fluctuations and to satisfy various electricity and energy demands, such as peak shaving and frequency regulation [
19], and power supply with energy storage [
20]. Therefore, hydropower station scheduling is essentially a complex multi-objective optimization problem, and common solution methods mainly include the weighting method [
21] and multi-objective optimization algorithms [
22]. The weighting method assigns weights to objectives based on their relative importance, aggregates them to convert the multi-objective problem into a single-objective, and then applies single-objective solution techniques. Salazar et al. [
23] addressed multiple objectives in a hydropower network under different hydrological conditions—maximizing annual power generation while minimizing unit water consumption per generation—by proposing a nonlinear multi-objective optimization model and using the weighting method to combine generation and water-use efficiency objectives into a single objective. Jia et al. [
24] proposed a multi-objective optimization model for inter-regional reservoir scheduling to balance interests among water supply, power generation, and ecological releases, and applied the weighting method to aggregate multiple objectives. This approach of converting a multi-objective problem into a single-objective one via weighting has the advantages of computational simplicity and intuitive results [
25]; however, it requires setting objective weights in advance, which involves subjectivity, and yields only a single solution, failing to satisfy the decision-makers’ needs for different preferences [
26].
Compared with the weighting method, intelligent optimization algorithms offer stronger global search capabilities and can obtain a Pareto-optimal solution set that fully reflects multi-objective conflicts, thereby providing decision-makers with diverse alternatives aligned with different preference structures [
27]. Consequently, the use of multi-objective optimization algorithms to solve complex reservoir scheduling problems has attracted widespread attention in both domestic and international research. However, under the new context of multi-energy complementary scheduling—characterized by multidimensional temporal and spatial scales, multiple scheduling stakeholders, and heterogeneous regulation capabilities across power stations—traditional optimization algorithms struggle to effectively handle the complex coupled constraints inherent in multi-energy scheduling problems [
28]. Therefore, numerous scholars have continuously enhanced classical algorithms to meet practical problem-solving needs and to achieve higher solution efficiency and accuracy. For example, Hu et al. [
29] addressed the long-term cascade hydropower scheduling problem with multi-objective generation and ecological flow requirements by proposing an adaptive multi-objective particle swarm optimization algorithm based on decomposition and dominance mechanisms, which improves both the diversity and convergence of the solution set. Yazdi and Moridi [
30] tackled the multi-objective design-parameter optimization problem for cascade reservoir systems by introducing a non-dominated sorting differential evolution algorithm, achieving an effective Pareto front search under multiple objectives while reducing computational cost. Hooshyar et al. [
31] addressed the challenge of an excessively large decision space in multi-reservoir optimal dispatch by proposing a multi-agent reinforcement learning approach that integrates an aggregation/decomposition (AD-RL) strategy, thereby mitigating high computational complexity and slow convergence issues. Azizipour et al. [
32] targeted the shortcomings of conventional and evolutionary methods—namely their tendency to become trapped in local optima and to exhibit low efficiency—in multi-reservoir system optimization by introducing a hybrid cellular automata–simulated annealing (CA-SA) method, which enhances convergence speed and solution quality.
The existing multi-energy coordination models often focus exclusively on either the generation side’s output capability or the grid side’s load fluctuations, and such one-sided perspectives struggle to balance multiple objectives, like generation benefit and grid peak-regulation absorption [
33,
34]. Accordingly, this study proposes an optimal scheduling model for hydropower–wind–solar complementarity with dual objectives: maximizing the total generation of the hydropower–wind–solar system and minimizing the mean square error of the residual load on the grid. To solve this model efficiently, we introduce the LMSPEAII algorithm, which integrates multiple improvement strategies into SPEAII. Finally, K-medoids clustering is applied to the large Pareto-optimal solution set to extract representative solutions for systematic analysis, from which this study’s conclusions are drawn.
3. Efficient Model Solving Based on Improved Strength Pareto Algorithm
To efficiently solve the reservoir optimization scheduling problem and to deeply reveal the trade-off between generation-side output and grid-side mean square error of residual load, this paper proposes an improved algorithm LMSPEAII based on the Strength Pareto Evolutionary Algorithm II (SPEAII). When applied to reservoir scheduling, SPEAII and other mainstream multi-objective evolutionary algorithms often exhibit slow convergence, insufficient solution set diversity, and susceptibility to local optima. The improved algorithm introduces strategies, such as Latin hypercube initialization [
41], hybrid distance framework, and adaptive mutation mechanism, effectively enhancing both global exploration and local exploitation capabilities. Consequently, the obtained solutions display broader coverage and more fully characterize the trade-off between generation-side output and grid-side mean square error of the residual load. In the context of reservoir scheduling, this not only significantly improves the optimization efficiency of scheduling strategies but provides richer and more diverse candidate solutions for achieving the dual objectives of economic performance and system stability, thereby offering more robust theoretical support and technical assurance for practical engineering decision-making.
3.1. Principle of Standard SPEA II Algorithm
- (1)
Initializing
The initialization phase of SPEAII is similar to that of most evolutionary algorithms: the population
of size N is generated via random sampling, and simultaneously generate an empty archive set
. For each individual and for each decision-variable dimension, sampling is performed according to the following formula:
where
denotes the generated position,
and
are the lower and upper bounds of that decision variable, and
is a random number drawn from [0, 1].
- (2)
Fitness assignment
The current population
and the archive
are merged into the union set
. First, a fitness value is assigned to each individual in the union set
. The fitness
is obtained by adding the individual’s raw fitness value
and its density estimation value
. The raw fitness
itself is obtained by accumulating the individual’s strength values
. The calculation formulas for each part are as follows:
where
denotes the strength value of individual
, representing the number of other individuals in the union set
that it dominates;
denotes the total number of individuals in the union set.
where
is the density estimation value of individual
;
is the distance between individual
and its
nearest neighbor in the objective space;
is the number of objectives (i.e., the dimensionality of the objective space); and
denotes the value of individual
on the
objective.
The final fitness value
of individual
is calculated as follows:
- (3)
Environment selection and truncation operations
All individuals in the union set R with a fitness value less than 1 are selected for retention,. If the number of individuals with fitness less than 1 exceeds the archive size threshold
, a truncation operation is performed to reduce the number to the archive limit. The truncation process begins by applying min–max normalization to each objective dimension in the current archive to eliminate the influence of differing units across objectives. The normalization is performed as follows:
where
denotes the normalized value of individual
on the
objective;
and
represent the minimum and maximum values of the
objective among all individuals, respectively.
Then, in the normalized objective space, the Euclidean distance between individuals is calculated using Equation (21). For each individual iii, its distances to all other individuals are sorted in ascending order as
. When the number of nondominated individuals (i.e., individuals with fitness < 1) exceeds the archive capacity, the truncation process is repeatedly performed as follows: identify the individual with the smallest, first nearest-neighbor distance
, the most crowded individual—remove it from the archive, and then update the distance matrix and re-sort distances until the number of individuals equals the archive threshold. This is calculated as follows:
where
and
are the normalized values of individuals
and
on the
objective, respectively.
- (4)
Genetic operation stage
Individuals are selected from the current population and the external archive using tournament selection to serve as parents for Simulated Binary Crossover (SBX). The selected parent individuals are then split randomly and uniformly into two groups, and , to facilitate the crossover operation. The offspring generated through crossover will subsequently undergo mutation. The detailed steps for each phase are as follows:
- (1)
Crossover Phase: A random number between 0 and 1 is generated to determine whether SBX crossover occurs, based on the crossover probability . If crossover is to occur, the crossover distribution factor is calculated using the following formula:
where
is the crossover distribution index, which is a predefined constant controlling the extent of offspring variation around the parents;
is a random number between 0 and 1, used to determine the calculation method for
.
Furthermore, the computed
is used to generate offspring according to the following operations:
where
is a random integer taking the value of either 0 or 1; the sign of
determines the direction of crossover, which is randomized. This allows the offspring to be generated on either side of the parent solutions, thereby enhancing population diversity.
Furthermore, the crossover factor
is adjusted as follows: a random number between 0 and 1 is generated. If the number is greater than 0.5, then
; otherwise,
. This ensures that the computed crossover factor has a 50% chance of being set to 1. The two offspring are then generated as linear combinations of the parent genes, with the crossover factor
determining the contribution ratio from each parent. The offspring are computed as follows:
where
and
are the genes of the parent individuals,
and
are the offspring generated through the crossover operation.
- (2)
Mutation Phase: The offspring and generated in the crossover phase are combined into a matrix . Whether each offspring undergoes mutation is determined by comparing a randomly generated number between 0 and 1 to the mutation probability . Here, denotes the number of decision variables. The mutation degree for each individual is controlled by the mutation step size, which is calculated using the polynomial mutation formula as follows:
where
is a random number from 0 to 1,
is the upper bound of the decision variable,
is the lower bound of the decision variable,
is the mutation distribution index, controls the size of the mutation step, is the set value.
After calculating the mutation step size, the offspring
and
undergo mutation according to the following formula:
where
denotes the gene of a certain offspring before mutation, and
denotes the corresponding gene of that offspring after mutation.
3.2. Improvement Strategy
- (1)
Latin hypercube initialization
Random sampling often leads to sample clustering in certain regions, thereby weakening the algorithm’s global exploration capability. To address this, the improved algorithm employs Latin hypercube sampling (LHS) for initialization and incorporates a maximin criterion to adjust sample point positions. LHS works by dividing the range of each decision variable into
N equally spaced intervals, and then randomly selecting one point from each interval without repetition. This ensures a more uniform and comprehensive distribution of sample points across the entire search space. The initialization formula for Latin hypercube sampling is given by the following equations:
where
represents the value of the
individual at the d decision variable;
and
are the lower and upper bounds of the
decision variable, respectively;
is a random number drawn from the interval [0, 1], introducing randomness within each subinterval;
is the length of each subinterval. The
k subinterval has endpoints
, and
is an index generated by a random permutation of
, ensuring that each index appears only once in each dimension.
- (2)
Hybrid distance architecture
In calculating the crowding degree between individuals, a Euclidean–Cosine hybrid distance framework is adopted to replace the traditional single Euclidean distance, as shown in Equation (34). This hybrid framework not only maintains the dispersion of absolute differences between solutions but effectively captures their directional diversity in the objective space. As a result, it improves the convergence and distribution uniformity of the Pareto front simultaneously [
42]. Moreover, since both Euclidean distance and cosine distance have a computational complexity of
, the additional cost introduced by their linear combination is negligible. This is calculated as follows:
where
is the normalized value of individual
on the
target,
and
are the normalized Euclidean and Cosine distances,
is the new distance calculation formula between individuals, and
is the weight parameter.
The weight parameter
is determined based on the relative proportion of the variances of the normalized Euclidean distance and cosine distance. A larger variance indicates that the corresponding distance metric exhibits a more dispersed numerical distribution within the current population, and thus has a stronger ability to distinguish the crowding characteristics of individuals. Therefore, the weight assignment is variance-based, emphasizing the more discriminative metric to improve the selection effectiveness. The specific calculation is as follows:
where
and
represent the variances of the normalized Euclidean distance and cosine distance, respectively.
- (3)
Adaptive mutation mechanism
In the original SPEAII algorithm, both the crossover probability
and mutation probability
are set to a fixed value of 1. This results in all parent individuals undergoing crossover and all offspring undergoing mutation. In the early stages of the algorithm, when population diversity is high, such large probabilities help rapidly explore the solution space and facilitate the combination of superior genes. However, as the evolution progresses and the population begins to converge, overly high crossover and mutation rates may disrupt high-quality genetic material, thereby reducing the convergence speed of the algorithm. To address this issue, dynamically adjusting the crossover and mutation probabilities during the evolutionary process enables the algorithm to leverage their specific roles at different stages: higher rates promote exploration in the early phase, while lower rates help maintain elite solutions and fine-tune convergence in the later phase. Therefore, the improved algorithm adopts adaptive control strategies for both crossover probability and mutation probability, aiming to enhance performance by balancing exploration and exploitation throughout the optimization process. The specific improvement strategies are detailed as follows.
where
represents the
individual’s
decision variable,
represents the mean of the
decision variable,
is the number of individuals in the population,
D is the number of decision variables, and
represents population diversity.
3.3. Improved Algorithm Solving Process
Figure 1 illustrates the process of the original algorithm and the added improvement strategies, the specific process of using the improved LMSPEAII algorithm to solve the reservoir scheduling model is as follows:
Step 1: A total of N individuals representing the decision variable—the reservoir water level process Z—are generated using the Latin hypercube sampling method, ensuring that the water level at each time period satisfies the corresponding water level constraints. Meanwhile, an empty archive set is initialized to store the non-dominated solutions identified during the evolutionary process.
Step 2: The current population and the archive set are merged to form a combined set . For each individual in the combined set, the water head , generation discharge , and power output at each time period are calculated based on the decision variable Z. Subsequently, the fitness value of each individual is evaluated using these computed outputs.
Step 3: Determine whether the number of individuals in the set exceeds the archive threshold based on fitness values, calculate the blending distance of each individual in the target space, sort the calculated blending distance of individuals in descending order, and ultimately retain the top individuals equal to the archive threshold.
Step 4: Select individuals from the current population and external archives through tournaments as parents for simulated binary crossover (SBX), divide the generated individuals into two parts and randomly and evenly, perform crossover operations on the parents to obtain offspring, and then perform mutation operations on the offspring.
Step 5: During the crossover process of the parent generation and the mutation process of the offspring, there may be some individuals who exceed the set boundaries due to random disturbances. Therefore, boundary correction is required for individuals after completing the crossover and mutation operations.
Step 6: Repeat steps Step 2 to Step 5 until the maximum number of iterations is reached, then terminate the iteration and obtain the current objective function values and decision variable variation process.
3.4. Performance Testing of LMSPEAII Algorithm
3.4.1. Algorithm Performance Evaluation Indicators
Multi-objective optimization algorithms aim to solve problems with multiple conflicting objectives, seeking to find a set of solutions that achieve or approximate the best values across all objective functions [
43]. To scientifically evaluate the performance of these algorithms, appropriate evaluation metrics need to be selected. These metrics reflect algorithm performance from different perspectives, such as convergence, diversity, and uniformity. Through quantitative assessment, different algorithms can be objectively compared on the same problem, guiding algorithm improvement and optimization. Therefore, selecting evaluation metrics to assess the performance of multi-objective algorithms is an important means to ensure the scientific validity, comparability, and practicality of research results. To comprehensively evaluate the performance of multi-objective algorithms, this study selects the inverted generational distance (IGD) [
44] and the hypervolume (HV) [
45] as evaluation indicators, reflecting algorithm performance from different dimensions during the solving process.
- (1)
Inverted Generational Distance (IGD)
The IGD represents the average distance from the optimal solution set to the obtained non-dominated solutions, and is mainly used to evaluate the convergence of multi-objective algorithms. A smaller IGD value indicates that the obtained non-dominated solution set is closer to the optimal solution set, demonstrating better convergence performance of the algorithm. Its calculation formula is as follows:
where
represents the true Pareto front,
denotes the approximate front solution set,
typically uses the Euclidean distance, and
is the total number of sample points in the true Pareto front.
- (2)
Hypervolume (HV)
The HV measures the volume of the objective space and can simultaneously evaluate the convergence and diversity of the non-dominated solution set obtained by the algorithm. Specifically, it calculates the hypervolume formed between the non-dominated solution set and a reference point. A larger HV value indicates that the solution set is closer to the true Pareto front in terms of convergence and diversity, demonstrating better algorithm performance. Its calculation formula is as follows:
where
is the approximate Pareto solution set,
represents the value of solution
in the
objective,
is the reference point (usually chosen to be worse than all solutions),
is the number of objective functions, and
denotes the Lebesgue measure (which corresponds to area in two dimensions, volume in three dimensions, and so forth).
3.4.2. Classic Multi-Objective Benchmark Testing Function
To verify the adaptability and robustness of the improved algorithm across different problem types, this section employs the Zitzler–Deb–Thiele test suite (ZDT) and Deb–Thiele–Laumanns–Zitzler test suite (DTLZ) series as benchmark test functions. The ZDT series (such as ZDT1 to ZDT6) features different shapes and distribution characteristics of objective functions and is primarily used to evaluate the algorithm’s convergence and solution set uniformity when handling convex, non-convex, continuous, and discrete Pareto fronts. The DTLZ series focuses more on problem scalability and the distribution of solution sets in high-dimensional objective spaces, revealing the algorithm’s performance in solving large-scale multi-objective optimization problems. These two types of functions respectively represent low-dimensional and high-dimensional multi-objective problems with diverse characteristics, thereby reflecting algorithm performance from multiple perspectives and difficulty levels.
In this paper, the SPEAII algorithm uses the default parameter settings recommended in the original publication [
46], without any additional tuning. Zitzler, Thiele, and colleagues have demonstrated the robustness of these settings across multiple test suites, with a crossover probability
of 0.9, a mutation probability
of
, a crossover distribution index
of 20, and a mutation distribution index
of 20. Their experiments have shown that this configuration delivers strong convergence and diversity performance across different problem scales and characteristics, eliminating the need for scenario-specific retuning. By contrast, the improved LMSPEAII algorithm relies entirely on on-the-fly parameter control strategies. Its parameters adjust themselves dynamically according to the current population’s distribution and search progress, with no manual presetting required. Such adaptive control methods are well established in the evolutionary computation literature, offering the advantage of responsive adaptation to each search phase and thereby greatly reducing—or even obviating—the need for manual parameter tuning.
To further demonstrate the excellent performance of the improved LMSPEAII algorithm in multi-objective problems, the original SPEAII algorithm and commonly used classical efficient algorithms, including Multi-Objective Evolutionary Algorithm based on Decomposition (MOEA/D), Non-dominated Sorting Genetic Algorithm III (NSGAIII), and Pareto Envelope-based Selection Algorithm II (PESAII), were selected as comparison algorithms. The population size and maximum number of iterations for all algorithms were uniformly set to 50 and 500, respectively, with default initial parameter settings for the comparison algorithms. To reduce variability caused by randomness and to enhance the statistical significance of the results, each test function was independently run 20 times. The performance evaluation metrics, IGD and HV, were calculated for each run, and the average and median values of these 20 runs were taken as representative metrics for analysis.
Table 1,
Table 2,
Table 3 and
Table 4 detail the metric values of the improved LMSPEAII and the other four algorithms on each benchmark test function (including the ZDT and DTLZ series). For intuitive comparison, the best-performing algorithm metrics in each table are highlighted in dark gray, while the second-best metrics are marked in light gray.
Figure 2 and
Figure 3 use radar charts to visually display the distribution of different performance metrics of each algorithm across the test functions.
From the experimental data, the improved LMSPEAII algorithm demonstrates significantly better performance than the original SPEAII algorithm across both the ZDT and DTLZ test function sets. Moreover, when compared with mainstream multi-objective optimization algorithms, such as MOEA/D, NSGAIII, and PESAII, LMSPEAII achieves the best results on over 75% of the performance metrics. These results clearly indicate that the improved LMSPEAII algorithm effectively enhances both convergence and solution set diversity, leading to a notable overall performance improvement on classical multi-objective benchmark functions.
5. Conclusions and Summary
To address the grid instability caused by integrating wind and solar power, this paper develops a hydropower–wind–solar multi-objective scheduling model that balances generation-side benefits and grid-side peak-regulation requirements. This paper further introduces a series of enhancements to propose the LMSPEAII algorithm for efficient solution of the model and employs K-medoids clustering to extract representative Pareto solutions, analyzing hydropower output and peak-regulation characteristics to demonstrate the applicability and robustness of both the improved algorithm and the proposed model under varying resource conditions.
- (1)
Effectiveness of Algorithmic Improvements: Based on the classical multi-objective optimization algorithm SPEAII, this study introduces three strategies—Latin hypercube initialization, a hybrid distance framework, and an adaptive mutation mechanism—to develop LMSPEAII. Comparative experiments on the ZDT and DTLZ benchmark suites against the original SPEAII and mainstream algorithms NSGAIII, MOEA/D, and PESAII, using the IGD and HV as performance metrics, demonstrate that LMSPEAII outperforms the other algorithms in diversity, convergence speed, and solution accuracy, validating the effectiveness of the proposed enhancements in boosting exploration capability and avoiding local optima.
- (2)
Rationality of Model Application: A multi-objective scheduling model was formulated with the goals of maximizing generation-side output and minimizing grid-side mean square error of the residual load. Two extreme resource scenarios—March (dry season with strong wind and solar) and September (flood season with weak wind and solar)—were selected to reveal the complementary and synergistic interactions among hydropower, wind, and solar. Solving this model with LMSPEAII and other multi-objective algorithms shows that LMSPEAII maintains superior applicability and performance under complex multi-energy dispatch conditions. Further analysis of the Pareto fronts indicates that hydropower’s flexible regulation capability enables effective trade-offs between generation efficiency and grid load smoothing, and that its dominant role and adjustment strategies adapt to differing water- and renewable-resource conditions, confirming the robustness of both the model and algorithm across varied resource contexts.
- (3)
Clustering for Simplified Analysis: From the 200 Pareto solutions generated by LMSPEAII, K-medoids clustering (with the elbow method determining the optimal number of clusters) was applied to both dispatch periods, and the medoid of each cluster was taken as its representative solution. An in-depth examination of these representative solutions’ hydropower–wind–solar output profiles and residual-load curves reveals the continuum of trade-off strategies from maximum generation to optimal peak regulation, and highlights how differing resource conditions influence dispatch strategies. This clustering approach not only uncovers the main scheduling patterns under various objective trade-offs but greatly reduces the effort of analyzing every Pareto solution individually, thereby improving research efficiency and clarity of interpretation.
In summary, this paper has developed a multi-objective scheduling model that balances generation efficiency and peak-regulation requirements, and has combined the LMSPEAII algorithm with K-medoids clustering analysis both to achieve efficient optimization of the hydropower–wind–solar complementary system and to reveal the intrinsic scheduling patterns under varying resource conditions, thereby providing a systematic methodology and empirical basis for integrated dispatch strategy formulation. With the continuous increase in renewable energy penetration, the grid’s sensitivity to power-balance disturbances also rises. At low wind–solar penetration levels, residual-load fluctuations are primarily absorbed by conventional units; however, once penetration exceeds a certain threshold, the inherent variability of renewables becomes the dominant driver of residual-load swings, causing the variance of the residual load to multiply. Under such conditions, the deployment of flexible regulation measures—such as energy storage, demand-side response, and reserve capacity—becomes critically important. In the model, minimizing the mean square error of the residual load thus reflects the optimal allocation and utilization of system regulation resources under a high-penetration renewable context. Looking ahead, the model can be extended to incorporate forecast uncertainties of wind and solar power as well as hydrological inflow variability within a stochastic or robust optimization framework to further enhance the reliability of scheduling strategies; concurrently, integrating electricity market mechanisms and energy storage systems can enable the coordinated optimization of economic performance and flexibility, supporting intelligent and sustainable operation under high-penetration renewable energy scenarios.