A Surrogate-Model-Based Approach for the Optimization of the Thermal Design Parameters of Space Telescopes

: The thermal design parameters of space telescopes are mainly optimized through traversal and iterative attempts. These optimization techniques are time consuming, rely heavily on the experience of the engineer, bear a large computational workload, and have difﬁculty in achieving optimal outcomes. In this paper, we propose a design method (called SMPO) based on an improved back-propagation neural network (called GAALBP) that builds a surrogate model and uses a genetic algorithm to optimize the model parameters. The surrogate model of a space telescope that measures the atmospheric density is established using GAALBP and then compared with surrogate models established using a traditional BP neural network and radial-basis-function neural network. The results show that the regression rate of the surrogate model based on the GAALBP reaches 99.99%, a mean square error of less than 2 × 10 − 6 , and a maximum absolute error of less than 4 × 10 − 3 . The thermal design parameters of the surrogate model are optimized using a genetic algorithm, and the optimization results are veriﬁed in a ﬁnite element simulation. Compared with the design results of the manually determined thermal design parameters, the maximum temperature of the CMOS is reduced by 5.33 ◦ C, the minimum temperature is increased by 0.39 ◦ C, and the temperature ﬂuctuation is reduced by a factor of 4. Additionally, SMPO displays versatility and can be used in various complex engineering applications to provide guidance for the better selection of appropriate parameters and optimization. and L.G.; methodology, software, W.Z.; Y.X., D.T. and L.G.; formal analysis, W.Z.; investigation, W.Z.; resources, L.G.; data curation, Z.J.; writing—original draft preparation, W.Z.; writing—review and editing, L.G.; visualization, W.Z.;


Introduction
Space telescopes are developing toward deep space exploration and maneuver to change orbit, with increasing demands on imaging quality, however, they experience changing and complex thermal environments [1,2]. The temperature of a telescope directly affects its imaging quality, and a reliable thermal design remains the basis for ensuring the stable operation of the telescope [3]. The thermal design of telescopes involves the iterative optimization of a large number of parameter combinations, which currently relies on the design experience of engineers and involves a process of repeated attempts. The process is time consuming and it is difficult to find an optimal solution. The development of methods that allow the rapid optimization of the thermal design parameters of telescopes has become an important issue [4], and techniques of parameter optimization have thus received much attention in recent years.
Scholars have investigated the parameter optimization of space telescopes, but there have been few studies on the optimization of the thermal design parameters of space telescopes. As examples, del Rio et al. [5] optimized the design parameters of X-ray mirrors using a genetic algorithm (GA) and Zhang et al. [6] used the inverse of the effective temperature of a star for a given flux density obtained using a stochastic particle swarm space telescope and compares the results with those obtained using the traditional method of manual parameter optimization. Section 4 presents the conclusions of the study.

Methodology of SMPO
The methodology of SMPO involves building a surrogate model using GAALBP and using a GA to find the best parameters, as shown in Figure 1. Appl. Sci. 2022, 12, x FOR PEER REVIEW 3 of 15 space telescope and compares the results with those obtained using the traditional method of manual parameter optimization. Section 4 presents the conclusions of the study.

Methodology of SMPO
The methodology of SMPO involves building a surrogate model using GAALBP and using a GA to find the best parameters, as shown in Figure 1. Part I: GAALBP network training This paper proposes an improved BP neural network. The improved network uses a GA to optimize the initial weights and thresholds of the network, the best individuals in the population are selected in a winner-takes-all manner [22], and the coding information of the best individuals is used for the initial weights and thresholds of the network. Therefore, before the training of the network, the GA coding length and fitness value need to be calculated and data preprocessing conducted. The calculations are as follows: (1) Calculation of the encoding length. The length S of the GA encoding is derived from the network topology determined by the feature dimensions of the input and output and the numbers of layers and nodes of the hidden layer. The calculation is ( )  (1) where nin is the number of neurons in the input layer, N is the number of layers in the hidden layer, ni is the number of neurons in the hidden layer (i = 1, 2,…, N), and nout is the number of neurons in the output layer. (2) Calculation of the fitness value. In the training phase of the network, the mean square deviation of the predicted values fitted to the network from the true values is taken as the fitness value of the GA. The maximum fitness value is zero when Treal and Tpre are equal. The calculation is where Tprei is the predicted temperature, Treali is the true temperature, and N is the number of test samples.
(3) Data preprocessing. The input data are normalized to eliminate the effect of variables of different orders of magnitude on the network training. Here, the optimal normalization method [23] is used to map the data to the range [−1, 1]. The calculation is Part I: GAALBP network training This paper proposes an improved BP neural network. The improved network uses a GA to optimize the initial weights and thresholds of the network, the best individuals in the population are selected in a winner-takes-all manner [22], and the coding information of the best individuals is used for the initial weights and thresholds of the network. Therefore, before the training of the network, the GA coding length and fitness value need to be calculated and data preprocessing conducted. The calculations are as follows: (1) Calculation of the encoding length. The length S of the GA encoding is derived from the network topology determined by the feature dimensions of the input and output and the numbers of layers and nodes of the hidden layer. The calculation is where n in is the number of neurons in the input layer, N is the number of layers in the hidden layer, n i is the number of neurons in the hidden layer (i = 1, 2, . . . , N), and n out is the number of neurons in the output layer.
(2) Calculation of the fitness value. In the training phase of the network, the mean square deviation of the predicted values fitted to the network from the true values is taken as the fitness value of the GA. The maximum fitness value is zero when T real and T pre are equal. The calculation is where T prei is the predicted temperature, T reali is the true temperature, and N is the number of test samples.
(3) Data preprocessing. The input data are normalized to eliminate the effect of variables of different orders of magnitude on the network training. Here, the optimal normalization method [23] is used to map the data to the range [−1, 1]. The calculation is where x min is the minimum value for the same dimension, and x max is the maximum value for the same dimension. The optimized weights and thresholds are assigned to the BP network, and the input data are used to train the network. During the training process, the learning rate changes adaptively with the relative change in the error, bearing the aim of keeping the learning stable while maintaining the largest possible learning step. If the error increases, a smaller learning rate is used in continuously searching for the direction of the gradient descent, and if the error increases by more than a certain percentage, the weights and thresholds of the previous round are discarded and the learning rate is reduced. This process improves the learning rate, however, when the learning rate is too high and the error reduction is not guaranteed, the learning rate is reduced until stable learning is restored. The improved BP network compensates for drawbacks such as the fixed learning rate of the BP network and can be trained to obtain a better surrogate model. The rule for correcting the learning rate with error is determined by manual debugging, and the method of adaptively adjusting the learning rate is shown in Figure 2. If the error in the current round increases by more than a factor of 1.04 relative to that in the previous round, the weights and thresholds of the current round are discarded, and the weights and thresholds in the next round are calculated using the values of the previous round, and the learning rate is reduced by a factor of 0.7. If the error in the current round is higher than that in the previous round but by less than a factor of 1.04, then the current weights, thresholds, and learning rate are retained. If the error continues to decrease, the learning rate is increased by a factor of 1.05. Appl where xmin is the minimum value for the same dimension, and xmax is the maximum value for the same dimension.
The optimized weights and thresholds are assigned to the BP network, and the input data are used to train the network. During the training process, the learning rate changes adaptively with the relative change in the error, bearing the aim of keeping the learning stable while maintaining the largest possible learning step. If the error increases, a smaller learning rate is used in continuously searching for the direction of the gradient descent, and if the error increases by more than a certain percentage, the weights and thresholds of the previous round are discarded and the learning rate is reduced. This process improves the learning rate, however, when the learning rate is too high and the error reduction is not guaranteed, the learning rate is reduced until stable learning is restored. The improved BP network compensates for drawbacks such as the fixed learning rate of the BP network and can be trained to obtain a better surrogate model. The rule for correcting the learning rate with error is determined by manual debugging, and the method of adaptively adjusting the learning rate is shown in Figure 2. If the error in the current round increases by more than a factor of 1.04 relative to that in the previous round, the weights and thresholds of the current round are discarded, and the weights and thresholds in the next round are calculated using the values of the previous round, and the learning rate is reduced by a factor of 0.7. If the error in the current round is higher than that in the previous round but by less than a factor of 1.04, then the current weights, thresholds, and learning rate are retained. If the error continues to decrease, the learning rate is increased by a factor of 1.05.
( 1). . .  In Figure 2, n.error is the error in round n, (n − 1).error is the error in round (n − 1); n.w and n.b are, respectively, the weight and threshold in round n; (n + 1).w and (n + 1).b are, respectively, the weight and threshold in round (n + 1); and ∆w, ∆b is the variation calculated from the error using the gradient descent method.
Part II: GA parameter optimization The GA is used to find the output extrema of the established surrogate model and the optimal solution corresponding to the extrema.
The GA initializes the generated population in the given parameter ranges. By setting the target output value, in the stage of seeking the extreme value, the difference between the target value and the actual output value is found and the negative of its absolute value is used as the fitness value fit of the GA, and the optimal solution is obtained by selecting the optimal individuals through crossover and mutation operations. Finally, the optimization result is substituted into the high-precision finite element model to verify whether In Figure 2, n.error is the error in round n, (n − 1).error is the error in round (n − 1); n.w and n.b are, respectively, the weight and threshold in round n; (n + 1).w and (n + 1).b are, respectively, the weight and threshold in round (n + 1); and ∆w, ∆b is the variation calculated from the error using the gradient descent method.
Part II: GA parameter optimization The GA is used to find the output extrema of the established surrogate model and the optimal solution corresponding to the extrema.
The GA initializes the generated population in the given parameter ranges. By setting the target output value, in the stage of seeking the extreme value, the difference between the target value and the actual output value is found and the negative of its absolute value is used as the fitness value fit of the GA, and the optimal solution is obtained by selecting the optimal individuals through crossover and mutation operations. Finally, the optimization result is substituted into the high-precision finite element model to verify whether the output satisfies the demand, and if not, the number of population individuals are increased iteratively until the target is met. The fitness value fit is calculated as where T pre is the predicted temperature and T real the true temperature.

Example Applications and Results
To verify its performance, SMPO is applied in the optimization of the thermal design parameters of an atmospheric density measurement space telescope (called the ADST), which was designed and manufactured in China, and the optimized parameters are substituted into high-precision finite element software for verification. A comparison is made with the results of the current conventional method of relying on an engineer's experience to optimize the design using the solving of partial differential equations, and the superiority of the optimization framework is thus verified.

Background of the ADST
The atmospheric density measurement space telescope (ADST) is tasked with monitoring the density of the stratospheric atmosphere at an orbital height of 280 km and an orbital inclination of 90 • . The ADST primarily includes a main frame, filter wheel assembly, mirror barrel, and detector-focal-plane assembly (including a complementary metal-oxide-semiconductor (CMOS) detector and CMOS board and two circuit boards with field-programmable gate arrays). The thermophysical finite element model of the ADST was developed using a nodal network [24] as shown in Figure 3. The internal heat source mainly reflects the heat generation of the CMOS of the ADST. The CMOS works intermittently (twice per track) with a power consumption of 1 W and has a preparation time less than 10 min and working time of less than 5 min. The heat consumption of the internal heat source is the same for the preparation state and working state. High-and low-temperature conditions and thermal control indicators are defined in Tables 1 and 2. The ADST telescope is installed inside the module, and only the light inlet is in contact with the space environment. The heat generated by the internal consumption of the ADST is not directly exchanged with the external environment, and the CMOS is extremely sensitive to temperature fluctuations. In this paper, SMPO is applied to optimize the thermal design parameters of the ADST telescope and thus control the temperature of the ADST. Table 1. Definitions of high-and low-temperature cases for the ADST.

Cases Definition of Cases
High Case ADST initial temperature 28 • C; due to the variable interval, the high case is analyzed according to 2 times continuous working, i.e., 30 min continuous working.

Low Case
ADST initial temperature 10 • C; internal heat source does not work; platform temperature control circuit works for a long time, other thermal control works 2 times per track, single work for 15 min, interval 30 min. Table 2. Thermal control index of the ADST.

Index Low Case
High Case heat source is the same for the preparation state and working state. High-and low-temperature conditions and thermal control indicators are defined in Tables 1 and 2. The ADST telescope is installed inside the module, and only the light inlet is in contact with the space environment. The heat generated by the internal consumption of the ADST is not directly exchanged with the external environment, and the CMOS is extremely sensitive to temperature fluctuations. In this paper, SMPO is applied to optimize the thermal design parameters of the ADST telescope and thus control the temperature of the ADST.  The CMOS works intermittently inside the cabin, and the heat generated by internal consumption is not exchanged directly with the external environment. After the preliminary thermal design, it was proposed to install the ADST adiabatically with its surrounding components. The surface of the ADST is blackened to reduce the effect of the cabin environment and other components on the temperature of the ADST. The heat generated by the CMOS is then transferred to the main frame through heat conduction and other means to achieve its temperature control. The heat transfer path of the CMOS is shown in Figure 4. The present paper takes the temperature of the ADST as the optimization target and selects 11 main thermal design parameters of the heat dissipation path as the parameters to be optimized, as shown in Table 3. The proposed SMPO is then used to control the temperature of the ADST. environment and other components on the temperature of the ADST. The heat generated by the CMOS is then transferred to the main frame through heat conduction and other means to achieve its temperature control. The heat transfer path of the CMOS is shown in Figure 4. The present paper takes the temperature of the ADST as the optimization target and selects 11 main thermal design parameters of the heat dissipation path as the parameters to be optimized, as shown in Table 3. The proposed SMPO is then used to control the temperature of the ADST.

Application of SMPO
Before being imported into the surrogate model for training, the data need to be normalized to eliminate the effects of variables of different orders of magnitude on the training results. The present paper adopts most-valued normalization to map the original data to the range of [−1, 1] according to Equation (2). The data distributions before and after the normalization are shown in Figure 5. It is observed that the normalized data distribution is more concentrated and contains a smaller difference between the variables, which are conducive to accelerating the convergence of the update of the network training weights. Appl

Application of SMPO
Before being imported into the surrogate model for training, the data need to be normalized to eliminate the effects of variables of different orders of magnitude on the training results. The present paper adopts most-valued normalization to map the original data to the range of [−1, 1] according to Equation (2). The data distributions before and after the normalization are shown in Figure 5. It is observed that the normalized data distribution is more concentrated and contains a smaller difference between the variables, which are conducive to accelerating the convergence of the update of the network training weights.
(a) (b) Figure 5. Comparison of the distribution of data before and after normalization (a. Data before normalization; b. Data after normalization).
The normalized data are imported into the GAALBP network, with 90% of the data used for training and 10% for testing. Eleven parameters are optimized as inputs and the predicted temperature of the CMOS is taken as the output. Hyperparameters of the network are set as presented in Table 4. Hyperparameters are mainly divided into GAALBP network hyperparameters and hyperparameters of the GA that optimize the initialization weights and thresholds of the network. The training process and results are shown in Figures 6 and 7, where it is seen that, the network training regression rate reaches 0.9999 and the maximum prediction error (the difference between the predicted value and true value) of the network is less than 4 × 10 −3 . The adaptive learning rate during training is shown in Figure 8. At the beginning of training, the learning rate continually increases, indicating that the error is gradually decreasing. In the later stages of training, the learning rate fluctuates around a value of 0.2, indicating that the training error has increased. However, the learning rate displays an overall decreasing trend, indicating that the training error decreases. The training effects of the adaptive learning rate and the improved BP network are compared for constant learning rates of 0.01, 0.1, 0.2, and 0.3, as shown in Figure 9. The test error for the adaptive learning rate is two orders of magnitude smaller than the test errors for the constant learning rates when the hyperparameters are the same, demonstrating the superiority of the adaptive learning rate.  The normalized data are imported into the GAALBP network, with 90% of the data used for training and 10% for testing. Eleven parameters are optimized as inputs and the predicted temperature of the CMOS is taken as the output. Hyperparameters of the network are set as presented in Table 4. Hyperparameters are mainly divided into GAALBP network hyperparameters and hyperparameters of the GA that optimize the initialization weights and thresholds of the network. The training process and results are shown in Figures 6 and 7, where it is seen that, the network training regression rate reaches 0.9999 and the maximum prediction error (the difference between the predicted value and true value) of the network is less than 4 × 10 −3 . The adaptive learning rate during training is shown in Figure 8. At the beginning of training, the learning rate continually increases, indicating that the error is gradually decreasing. In the later stages of training, the learning rate fluctuates around a value of 0.2, indicating that the training error has increased. However, the learning rate displays an overall decreasing trend, indicating that the training error decreases. The training effects of the adaptive learning rate and the improved BP network are compared for constant learning rates of 0.01, 0.1, 0.2, and 0.3, as shown in Figure 9. The test error for the adaptive learning rate is two orders of magnitude smaller than the test errors for the constant learning rates when the hyperparameters are the same, demonstrating the superiority of the adaptive learning rate.         The quality of the surrogate model directly determines how well the parameters are optimized in the subsequent step. Comparisons are made with the traditional BP network, GA-optimized BP (GABP) network, and radial basis function neural network [25] to verify the superiority of GAALBP. The training results are presented in Table 5 and Figure 10. The GABP network is a BP network with a GA-optimized constant learning rate; i.e., the only difference between the GABP and GAALBP is the difference in the learning rate variation. It is seen that the training error of GAALBP is 1% of that of the traditional BP network and 40% that of the GABP network whereas the mean square error of GAALBP is smaller than the mean square errors of the other networks. The results thus demonstrate the superiority of GAALBP.   The quality of the surrogate model directly determines how well the parameters are optimized in the subsequent step. Comparisons are made with the traditional BP network, GA-optimized BP (GABP) network, and radial basis function neural network [25] to verify the superiority of GAALBP. The training results are presented in Table 5 and Figure 10. The GABP network is a BP network with a GA-optimized constant learning rate; i.e., the only difference between the GABP and GAALBP is the difference in the learning rate variation. It is seen that the training error of GAALBP is 1% of that of the traditional BP network and 40% that of the GABP network whereas the mean square error of GAALBP is smaller than the mean square errors of the other networks. The results thus demonstrate the superiority of GAALBP.  Figure 9. Training error for different learning rates.
The quality of the surrogate model directly determines how well the parameters are optimized in the subsequent step. Comparisons are made with the traditional BP network, GA-optimized BP (GABP) network, and radial basis function neural network [25] to verify the superiority of GAALBP. The training results are presented in Table 5 and Figure 10. The GABP network is a BP network with a GA-optimized constant learning rate; i.e., the only difference between the GABP and GAALBP is the difference in the learning rate variation. It is seen that the training error of GAALBP is 1% of that of the traditional BP network and 40% that of the GABP network whereas the mean square error of GAALBP is smaller than the mean square errors of the other networks. The results thus demonstrate the superiority of GAALBP.  After establishing the surrogate model, the GA is used to optimize the input of the surrogate model to meet the thermal control index of the CMOS. The negative of the absolute value of the difference between the target temperature and predicted temperature is taken as the fitness value of GA training: where Tgoal is the target temperature and Tpre is the predicted temperature. A larger fitness value means that the predicted temperature is closer to the target temperature, meaning that the parameter optimization is better.
The GA hyperparameters are set as given in Table 6. The coding length of genetic individuals is equal to the number of parameters to be optimized (i.e., 11), and each individual coding contains all the information of the parameters to be optimized. The size of the initialized population is 100. The individuals are input to the surrogate model to calculate the fitness value, and the poorly adapted individuals are eliminated in the manner of survival of the fittest. Meanwhile, the highly adapted individuals are crossed and mutated to produce new individuals. New rounds of iterations are performed until the best adapted individuals in the population are selected. Method of Crossover Crossover heuristicXover One-hundred iterations are performed. The iterative process is presented in Table 7. The algorithm selects the best individual as having a fitness value of −0.0002, indicating that the difference between the temperature of the optimized CMOS and the CMOS index is not greater than 0.0002. After establishing the surrogate model, the GA is used to optimize the input of the surrogate model to meet the thermal control index of the CMOS. The negative of the absolute value of the difference between the target temperature and predicted temperature is taken as the fitness value of GA training: where T goal is the target temperature and T pre is the predicted temperature. A larger fitness value means that the predicted temperature is closer to the target temperature, meaning that the parameter optimization is better.
The GA hyperparameters are set as given in Table 6. The coding length of genetic individuals is equal to the number of parameters to be optimized (i.e., 11), and each individual coding contains all the information of the parameters to be optimized. The size of the initialized population is 100. The individuals are input to the surrogate model to calculate the fitness value, and the poorly adapted individuals are eliminated in the manner of survival of the fittest. Meanwhile, the highly adapted individuals are crossed and mutated to produce new individuals. New rounds of iterations are performed until the best adapted individuals in the population are selected. One-hundred iterations are performed. The iterative process is presented in Table 7. The algorithm selects the best individual as having a fitness value of −0.0002, indicating that the difference between the temperature of the optimized CMOS and the CMOS index is not greater than 0.0002. To verify the optimized results, the parameters optimized by the GA are substituted into software for Monte Carlo simulation using the finite element method for thermal design, and the optimized results are verified for the high-and low-temperature cases of the telescope. The verification results are presented in Figures 11-13. Figure 11a shows the temperature cloud of the CMOS heat dissipation component in the low-temperature case whereas Figure 11b shows the temperature cloud of the CMOS itself in the low-temperature case. Figure 12a shows the temperature cloud of the CMOS heat dissipation component in the high-temperature case whereas Figure 12b shows the temperature cloud of the CMOS itself in the high-temperature case. Figure 13 shows the temperature fluctuation of the CMOS itself in the high-and low-temperature cases. The figures reveal that in the low-temperature case, as the internal heat source essentially fails to work and the CMOS package in the cabin is less affected by external heat flow, the overall minimum temperature of the CMOS components exceeds 10 • C, the maximum temperature is less than 13.9 • C, the temperature of the CMOS itself is stable at 12.82 • C, and the temperature uniformity is within 0.01 • C. In the high-temperature case, the internal heat source works intermittently, the overall temperature uniformity of the CMOS components is within 0.93 • C, the minimum temperature of the CMOS itself is 31.10 • C, the maximum temperature is 32.69 • C, and the maximum temperature fluctuation is less than 1.6 • C, which meets the demands of the CMOS thermal control index.  To verify the optimized results, the parameters optimized by the GA are substituted into software for Monte Carlo simulation using the finite element method for thermal design, and the optimized results are verified for the high-and low-temperature cases of the telescope. The verification results are presented in Figures 11-13. Figure 11a shows the temperature cloud of the CMOS heat dissipation component in the low-temperature case whereas Figure 11b shows the temperature cloud of the CMOS itself in the low-temperature case. Figure 12a shows the temperature cloud of the CMOS heat dissipation component in the high-temperature case whereas Figure 12b shows the temperature cloud of the CMOS itself in the high-temperature case. Figure 13 shows the temperature fluctuation of the CMOS itself in the high-and low-temperature cases. The figures reveal that in the lowtemperature case, as the internal heat source essentially fails to work and the CMOS package in the cabin is less affected by external heat flow, the overall minimum temperature of the CMOS components exceeds 10 °C, the maximum temperature is less than 13.9 °C, the temperature of the CMOS itself is stable at 12.82 °C, and the temperature uniformity is within 0.01 °C. In the high-temperature case, the internal heat source works intermittently, the overall temperature uniformity of the CMOS components is within 0.93 °C, the minimum temperature of the CMOS itself is 31.10 °C, the maximum temperature is 32.69 °C, and the maximum temperature fluctuation is less than 1.6 °C, which meets the demands of the CMOS thermal control index.

Results
The optimization results of SMPO are compared with the results of the manual optimization of CMOS thermal parameters by engineers as presently performed in industry to verify the performance of SMPO. Additionally, the optimized results are substituted into finite element software to verify the performance of the parameters in the high-and low-temperature cases. The results are shown in Figure 14. In the low-temperature case, CMOS components essentially fail to work, there is no internal heat source, the CMOS temperature basically does not fluctuate, and the results obtained through SMPO optimization and those obtained through manual optimization both meet the requirements of the CMOS thermal control index, however, the overall temperature in the results obtained through SMPO is 0.5 °C higher than that in the results obtained through manual optimization, leaving a greater temperature margin in the low-temperature case. In the hightemperature case, the internal heat source works intermittently. Compared with the results obtained through manual optimization, the temperature derived through SMPO is

Results
The optimization results of SMPO are compared with the results of the manual optimization of CMOS thermal parameters by engineers as presently performed in industry to verify the performance of SMPO. Additionally, the optimized results are substituted into finite element software to verify the performance of the parameters in the high-and low-temperature cases. The results are shown in Figure 14. In the low-temperature case, CMOS components essentially fail to work, there is no internal heat source, the CMOS temperature basically does not fluctuate, and the results obtained through SMPO optimization and those obtained through manual optimization both meet the requirements of the CMOS thermal control index, however, the overall temperature in the results obtained through SMPO is 0.5 °C higher than that in the results obtained through manual optimization, leaving a greater temperature margin in the low-temperature case. In the hightemperature case, the internal heat source works intermittently. Compared with the results obtained through manual optimization, the temperature derived through SMPO is Temperature/°C Figure 13. CMOS temperature fluctuations in the high-and low-temperature cases.

Results
The optimization results of SMPO are compared with the results of the manual optimization of CMOS thermal parameters by engineers as presently performed in industry to verify the performance of SMPO. Additionally, the optimized results are substituted into finite element software to verify the performance of the parameters in the high-and low-temperature cases. The results are shown in Figure 14. In the low-temperature case, CMOS components essentially fail to work, there is no internal heat source, the CMOS temperature basically does not fluctuate, and the results obtained through SMPO optimization and those obtained through manual optimization both meet the requirements of the CMOS thermal control index, however, the overall temperature in the results obtained through SMPO is 0.5 • C higher than that in the results obtained through manual optimization, leaving a greater temperature margin in the low-temperature case. In the high-temperature case, the internal heat source works intermittently. Compared with the results obtained through manual optimization, the temperature derived through SMPO is 5.33 • C lower at the highest temperature of the CMOS and 0.39 • C lower at the lowest temperature of the CMOS. Additionally, the temperature fluctuation is reduced by a factor of 4, and the calculation time is reduced from several days to a few hours. The effectiveness and superiority of SMPO are thus demonstrated.

Conclusions
This paper proposes a surrogate model-based method for optimizing the thermal parameters of space telescopes, called SMPO. The method employs a BP neural network surrogate model based on the adaptive learning rate (called GAALBP) so as to bear a lower computational cost than the traditional thermal design approach of solving partial differential equations. Additionally, the proposed method uses a genetic algorithm (GA) to optimize the weights and thresholds of the BP network and thus improves the accuracy of the surrogate model. After the surrogate model is established, the genetic algorithm is used again to optimize the input of the network so that the output of the network approximates the target value.
In this paper, we established a thermophysical model of a space telescope (called the ADST), selected 11 parameters of the heat dissipation path of the CMOS detector as indicators to be optimized, applied the GAALBP network surrogate model to approximate the thermophysical model of the ADST, established the mapping relationship between the 11 indicators to be optimized and the temperature of the CMOS, and used a GA to optimize indicators ensuring the output meets CMOS temperature requirements. The theoretical and simulation results reveal that SMPO proposed in the paper outperforms traditional engineer-dependent optimization, in terms of the model evaluation accuracy and higher computational efficiency.
Space telescopes are developing toward the direction of modularization, rapid launch, and short design cycles. Optimization design methods such as those similar to SMPO that can quickly realize multi-parameter intelligence and automation is of particular importance. Moreover, SMPO is an optimization framework and an optimization idea. The rapid optimization process can be transplanted into other models to achieve rapid thermal design and batch implementation. SMPO is applicable to not only the optimization of the thermal design parameters of space telescopes but also post-processing and design optimization in other fields. However, the convergence of SMPO is not particularly stable, and the SMPO optimization framework does not automate the post-processing of data, with manual data conversion still a requirement. Therefore, it remains necessary to Temperature/°C Figure 14. Comparison of the performance of parameter optimization in high-and lowtemperature cases.

Conclusions
This paper proposes a surrogate model-based method for optimizing the thermal parameters of space telescopes, called SMPO. The method employs a BP neural network surrogate model based on the adaptive learning rate (called GAALBP) so as to bear a lower computational cost than the traditional thermal design approach of solving partial differential equations. Additionally, the proposed method uses a genetic algorithm (GA) to optimize the weights and thresholds of the BP network and thus improves the accuracy of the surrogate model. After the surrogate model is established, the genetic algorithm is used again to optimize the input of the network so that the output of the network approximates the target value.
In this paper, we established a thermophysical model of a space telescope (called the ADST), selected 11 parameters of the heat dissipation path of the CMOS detector as indicators to be optimized, applied the GAALBP network surrogate model to approximate the thermophysical model of the ADST, established the mapping relationship between the 11 indicators to be optimized and the temperature of the CMOS, and used a GA to optimize indicators ensuring the output meets CMOS temperature requirements. The theoretical and simulation results reveal that SMPO proposed in the paper outperforms traditional engineer-dependent optimization, in terms of the model evaluation accuracy and higher computational efficiency.
Space telescopes are developing toward the direction of modularization, rapid launch, and short design cycles. Optimization design methods such as those similar to SMPO that can quickly realize multi-parameter intelligence and automation is of particular importance. Moreover, SMPO is an optimization framework and an optimization idea. The rapid optimization process can be transplanted into other models to achieve rapid thermal design and batch implementation. SMPO is applicable to not only the optimization of the thermal design parameters of space telescopes but also post-processing and design optimization in other fields. However, the convergence of SMPO is not particularly stable, and the SMPO optimization framework does not automate the post-processing of data, with manual data conversion still a requirement. Therefore, it remains necessary to further improve the convergence and stability of the SMPO, and to achieve full automation of SMPO.