Abstract
Obtaining an antenna’s back-up structure (BUS) temperature field is an essential prerequisite for analyzing its thermal deformation. Thermodynamic simulation can obtain the structure’s thermal distribution, but it has low computational accuracy. There is a problem with cumbersome wiring and difficult maintenance of the temperature measurement system. This study developed an improved RIME-XGBoost model to realize the temperature prediction of the BUS of the Nanshan 26-m Radio Telescope (NSRT). The proposed model successfully predicts the NSRT’s BUS temperature distribution based solely on environmental sensing (ambient temperature, angle of solar radiation, antenna’s orientation, etc.). The relative prediction accuracy between the predicted and actual BUS temperature is 97.15%, and the predictive error is less than 0.897 K (root mean square error, RMSE). This research result provides an alternative method for the real-time reconstruction of the structure’s thermal distribution in large-aperture radio telescopes.
1. Introduction
With the increase in the aperture of radio telescopes and the improvement in observation frequency bands, there are higher requirements for the surface accuracy and pointing error of radio telescopes. However, large-aperture radio telescopes are susceptible to solar radiation, wind gusts, ambient temperature, relative humidity, and other environmental factors, which can degrade antenna performance. Among these factors, the non-uniform temperature field acting on the antenna structure is a critical factor in the decrease in surface accuracy and pointing accuracy. Greve et al. [1] found through finite element analysis of the IRAM (Institute of Radioastronomie Millimétrique) 30-m telescope that a 0.1 K change in the antenna structure temperature difference would result in a 0.005 mm surface error. Li et al. [2] conducted a simulation analysis on the effect of sunshine temperature deformation on a 65-m radio telescope’s surface accuracy, and the result shows that a 1 K temperature difference changes the surface accuracy component by about 0.09 mm. As reported by Ukita et al. [3], measurements performed on the 10-m ASTE (Atacama Submillimeter Telescope Experiment) telescope revealed that the pointing error induced by temperature difference is approximately 1.5 arcseconds per degree. Therefore, achieving high-precision and real-time acquisition of temperature changes in antenna structures is a necessary prerequisite for improving the performance of radio telescopes.
A series of relevant studies have been carried out on the thermal analysis of the radio telescope structure. Attoli et al. [4] analyzed the Sardinia Radio Telescope using the finite element method to determine the pointing error generated by each simulated thermal scenario. Nikolic et al. [5] used phase-retrieval holography measurement on the GBT (Green Bank Telescope) to demonstrate that thermal deformation usually exceeds gravitational deformation during daytime. Wei et al. [6] analyzed the temperature distribution of the primary reflector of a 70-m radio telescope using the finite element method (FEM). Meanwhile, through on-site measurements with an optical camera and a thermal imaging camera, they found that the temperature distribution of the primary reflector under solar radiation is very uneven, and the maximum of the root mean square temperature is 12.3 K. In summary, the main methods for thermal acquisition of antenna structures include thermodynamic simulation and on-site measurement. The BUS temperature field is challenging to obtain accurately through thermodynamic simulation. The temperature sensors can obtain the structure temperature, but problems include electromagnetic interference, difficult maintenance, complicated sensor wiring, and easy wear and tear. Finally, it is challenging to realize the long-term, stable, real-time acquisition of the BUS temperature. So far, there is no efficient and feasible method to realize the real-time acquisition of the thermal distribution.
In order to achieve real-time and high-precision acquisition of the BUS temperature information, we attempt to use surrogate models or neural networks to solve this problem. Indeed, researchers have attempted to realize real-time structure temperature prediction without temperature sensors by leveraging machine learning or neural network models based on long-term historical temperature data. For instance, Yan et al. [7] proposed a new multi-data-driven model based on reinforcement learning and deep learning for high-precision prediction of locomotive axle temperature. Hu et al. [8] proposed a multi-scale convolutional neural network to predict the final cooling temperature, with a prediction error of approximately ± 10% of the actual temperature. Li et al. [9] proposed a hybrid model that combines a physical model with a long short-term memory (LSTM) network for predicting temperature during spacecraft thermal testing. The temperature prediction accuracy of this method is MAE (mean absolute error) = 17.41 K, (the coefficient of determination) = 0.9988. However, there are no experimental cases yet attempting to use neural network models to achieve thermal acquisition of radio telescope structures.
In this study, we propose a model for the real-time reconstruction of the structure’s temperature field. Building the prediction model using a data-driven approach can replace complex or unknown mathematical equations. The working environment of the NSRT is taken as the input, which contains the NSRT’s orientation, ambient temperature, and relative humidity, the angle of solar radiation, etc. The measured data for 8 months is used as the model training dataset, and then the model parameters are optimized to obtain the best prediction model. Finally, the model realizes the real-time prediction of the radio telescope’s BUS temperature based on environment sensing. This method analyzes the weights of the factors affecting the BUS temperature distribution, realizing the prediction of the BUS temperature distribution solely based on the environmental sensing, without relying on the real-time temperature sensors.
2. Data Collection and Feature Analysis
2.1. Data Collection
The data collection system consists of a structure temperature measurement system and an environment monitoring system. The data were collected from 15 November 2021 to 15 May 2022 [10]. The time interval of the data is 10 s, which includes ambient temperature, relative humidity, barometric pressure, wind direction, wind speed, the antenna’s azimuth and elevation, solar altitude and azimuth, and the temperature of 45 locations on the NSRT’s BUS.
In total, 45 FBG (Fiber Bragg Grating) temperature sensors were installed on the NSRT’s BUS [11,12]. Fiber-Optic Grating Sensors have been widely applied in structure health monitoring scenarios due to their advantages, such as immunity to electromagnetic interference, small size, light weight, durability, and high bandwidth, which allow a significant number of sensors to operate in the same system [13,14]. The specific installation positions of the temperature sensors are shown in Figure 1, in which Figure 1a is the top view of the BUS and A-P indicates the 16 measurement channels of the 16 main radial beams. Figure 1b is the front view of a single radial beam M, and the dots are the mounting positions for these sensors. It can be seen that all of the sensors are deployed at equidistant points on the lower chord beams.
Figure 1.
Diagram of temperature sensors’ deployment position. (a) Top view. (b) Cross-sectional view.
The field installation of temperature sensors is shown in Figure 2. Figure 2a shows the manned spider crane positioned at the lower chord beam of the BUS to install the temperature sensor. The global diagram of the sensor mounting positions on the BUS is shown in Figure 2b. The optical lines and temperature sensors are fixed with moisture-proof, waterproof, heat-insulating, and wear-resistant grey PTFE (polytetrafluoroethylene) film tape. The local detail sensor mounting diagram is shown in Figure 2c.
Figure 2.
Diagram of temperature sensor field installation. (a) Manned crane for laying sensors. (b) Sensor layout global diagram. (c) Sensor layout detail diagram.
The temperature acquisition equipment in the NSRT high-frequency cabin is shown in Figure 3. The fiber-optic lines transmit the Bragg wavelength information collected by each temperature sensor to the high-frequency cabin, as shown in Figure 3a. The temperature value interface of all temperature sensors was obtained from 16 optical fibers processed by the FBG demodulator, as shown in Figure 3b. The data processing equipment of the FBG temperature measurement system is shown in Figure 3c.
Figure 3.
Diagram of the BUS temperature data collecting equipment. (a) Temperature sensor collection device. (b) Software interface of the BUS temperature. (c) Global view of the temperature collection device.
The environmental data collection equipment is a multi-functional integrated weather station built near the NSRT, which can detect wind direction, wind speed, ambient temperature, relative humidity, barometric pressure, and other meteorological elements. The overall weather station consists of three parts, the meteorological sensors, the data logger, and the data acquisition system, of which the meteorological sensors on the pole are shown in Figure 4.
Figure 4.
Diagram of multi-element meteorological station.
The solar altitude and elevation are calculated from the NSRT geographic coordinates and the current time using Python 3.9 code. The NSRT’s azimuth and elevation are collected from the NSRT real-time operating log and will not be repeated here.
2.2. Data Analysis
After collecting the BUS temperature and the working condition data, the feature analysis of each data variable is carried out to explore their correlation with the BUS temperature. The heat map of the Pearson coefficient matrix is used to measure the correlation between the features, as shown in Figure 5 [15]. Only the mean, maximum, and minimum BUS temperature values are counted to simplify the expression of the BUS temperature data.
Figure 5.
Heat map of the correlation between the BUS temperature and the ambient factors.
As can be seen from Figure 5, the BUS temperature is strongly related to the ambient temperature. The solar altitude and the relative humidity also have a certain correlation with the BUS temperature. The following analyzes the ambient temperature, relative humidity, and solar altitude in turn. Figure 6a shows the variation in the average BUS temperature and the ambient temperature over time, and Figure 6b shows the relationship between the BUS temperature and the ambient temperature during the day and night. From Figure 6, it is concluded that the overall BUS average temperature is highly correlated with the ambient temperature in a linear way [16], especially at night, and the goodness of fit of the is 0.92481.
Figure 6.
Diagram of the BUS temperature and ambient temperature. (a) Time-varying diagram of the BUS temperature and ambient temperature. (b) Relational graph of the BUS temperature and ambient temperature between day and night.
Table 1 shows the statistical result of the BUS temperature. From the data in Table 1, the maximum of the 45 points on the BUS is 28.28 K higher than the ambient temperature, and the average temperature of the whole structure is 12.29 K higher than the ambient temperature, which shows that the temperature field of the BUS is greatly affected by the change in the ambient temperature. In addition, the BUS temperature strongly correlates with the ambient temperature, and its large temperature drift mostly occurs at noon in summer.
Table 1.
Statistical result of the BUS temperature.
As for the effect of the ambient humidity on the BUS temperature, there is no apparent correlation. The effect of the solar altitude on the BUS temperature is related to the combination of the NSRT’s orientation (heating angle) and the rotation speed, making it challenging to find the pattern using mathematical statistics. As for the effect of wind, barometric pressure, and other factors on the BUS, we will not carry out their feature analysis here because the correlation between them is weak and the features are not prominent.
3. Development of the Model
3.1. The XGBoost Model
eXtreme Gradient Boosting (XGBoost) is a C++ implementation of the Gradient Boosting Machine algorithm by Tianqi Chen at the University of Washington [17], and it is highly suitable for obtaining structure temperature fields [18,19,20]. XGBoost makes a second-order Taylor expansion of the loss function and adds a regularization term to the loss function to control the complexity of the model and avoid overfitting. Meanwhile, XGBoost utilizes the central processing unit’s multithreaded parallel operation to improve computational efficiency greatly. The objective function of the XGBoost model can be written as follows:
where denotes the objective function, is the loss function, is the regularization term, also known as the complexity of the regression tree, M is the number of leaf nodes in the decision tree, denotes the node value, and Y and are hyperparameters to control the penalty strength.
3.2. The RIME Algorithm
The rime optimization algorithm (RIME) is a new heuristic optimization algorithm based on the physical phenomenon of rime-ice in nature [21]. The RIME algorithm implements exploration and exploitation behaviors in optimization by simulating the soft-rime and hard-rime growth process of rime-ice and constructing a soft-rime search strategy and a hard-rime puncture mechanism. In the initialization phase, the RIME algorithm initializes the whole rime-population . The rime-population consists of p rime-agents , and each rime-agent consists of q rime-particles . Thus, the rime-population can be directly represented by the rime-particles , as shown in Equation (2).
From Formula (2), it can be seen that to initialize rime-population , one needs to initialize its smallest unit rime-particles . is a candidate solution, and in this experiment, rime-particles represent a set of hyperparameters of the XGboost model. If there are d hyperparameters, then can be expressed as , where represents the dth hyperparameter, and the process of the hyperparameter is shown in Formula (3).
where is the lower bound of the dth hyperparameter. is the upper bound of the dth hyperparameter. is a random number within the range of 0 to 1.
Furthermore, is used to denote the growth state of each rime-agent, i.e., the fitness value of the agent in the meta-heuristic algorithm; the expression is shown Equation (4).
where f is the adaptation degree of each rime-particle.
Each rime-particle will move according to a certain law before condensation, and the free particles will move to the soft-rime neighborhood to condense, but cannot condense when exceeding a certain range. Aiming at the motion properties of the rime-particles, the RIME algorithm proposes the soft-rime search strategy, as shown in Equation (5).
where , is the new position of the updated particle, and p and q denote the q-th particle of the p-th rime-agent. is the q-th particle of the best rime-agent in the rime-population .
The increase in the soft-rime area enables the algorithm to cover the full space search quickly. Using the hard-rime puncture phenomenon to achieve dimensional crossover interchange between ordinary and optimal agents improves the ability to jump out of the local optimum. The formula for replacement between particles is shown in Equation (6).
where denotes the normalized value of the current agent fitness value, indicating the chance of the p-th rime-agent being selected, and is a random number within .
After the Soft Frost Search (exploration) and Hard Frost Perforation (exploitation) are completed, a forward greedy selection mechanism should be used to filter the updated solutions, thereby indirectly ensuring the balance effect between the two. The specific approach is to compare the updated fitness value of a rime-agent with the fitness value of a rime-agent before the update. If the updated fitness value is better than the value before the update, the individual (rime-agent ) will be replaced. On one hand, this mechanism actively replaces individuals, enabling the rime-population to continuously have good individuals, thereby improving the quality of the global solution. On the other hand, as the position of the rime-agents of the rime-population changes significantly in each iteration, there will inevitably be rime-agents that perform worse than the rime-population before the update. Therefore, this operation can be used to ensure that the rime-population evolves in a more optimal direction in each iteration. The specific determination formula is shown in Equation (7):
where is the fitness value of the new individual, while is the fitness value before the update.
3.3. Algorithm Improvement
In order to overcome the shortcomings of the RIME algorithm, such as poor initial sensitivity and easy overfitting, the RIME cluster initialization is optimized using chaotic mapping. In addition, the search agent uses refracted opposition-based learning to obtain the opposite solution in the solution space so as to enhance the global search ability and increase the possibility of the algorithm finding a better solution.
3.3.1. Sine Chaotic Map
Sensitivity to initial conditions and control parameters are the main features of chaotic maps. Generally, simple linear or non-linear equations are used to create chaotic behaviors, and Table 2 shows several classical forms of chaotic mapping.
Table 2.
Common chaotic mappings.
Different chaotic maps have their own characteristics, which are not repeated here. It should be emphasized that a sine map is simple and efficient, but also has the disadvantage of an uneven probability density distribution. However, the soft-rime search strategy of the RIME algorithm can make up for the shortcomings of the sine map. Therefore, the sine map is used to optimize the RIME cluster initialization instead of the random initialization. This increases the diversity of the rime-population and avoids falling into local optimization during the search process of the RIME algorithm. The improved algorithm for rime-population initialization is actually an improvement on its basic unit rime-particles . The specific improved process is shown in Formula (8).
where is the lower bound of the dth hyperparameter. is the upper bound of the dth hyperparameter. is a random number within the range of 0 to 1. is the sine map function.
3.3.2. Refractive Opposition-Based Learning
The refractive opposition-based learning strategy (ROBL) is a modification of the opposition-based learning strategy [22], which enhances the optimization algorithm’s accuracy by combining the principle of light refraction with the OBL’s advantage of calculating the opposite solution to expand the population search space.
As shown in Figure 7, the algorithmic core of the refractive opposition-based learning strategy draws on the principle of light refraction. is the angle of incidence, is the angle of refraction, and z and are two points within the coordinate axes [a,b]. L is the length of the incident ray. is the length of the refracted ray. The sinusoidal expressions for the angles of incidence and refraction are shown in Equation (9):
where denotes the final solution.
Figure 7.
Schematic diagram of the light refraction.
In this algorithm, when is less than 1, the search agent maps to other regions to increase the possibility of finding the global optimal region. When is greater than 1, the search agent obtains the inverse solution in the local region. Eventually, this way iteratively converges to the optimal solution. The ROBL strategy is applied to the RIME algorithm to improve the balance between the Soft Frost Search (exploration) and Hard Frost Perforation (exploitation). Before selecting the optimal solution for this iteration through the forward greedy selection mechanism, we first perform ROBL optimization on the updated individual (rime-agents ) fitness, as shown in Equation (10):
where k is the current iteration number, while K is the maximum number of iterations. is the individual fitness value before this iteration update. is the individual fitness value of this update. is the fitness value after applying the RBOL strategy.
Furthermore, the forward greedy selection mechanism in the RIME algorithm should also be modified accordingly. The modified formula is shown in Equation (11):
From Formula (10), it can be concluded that at the beginning of the iteration, the fitness of is very close to that of . This is set to accelerate the iteration process and make the rime-population approach the optimal solution more quickly. As the iteration progresses to a later stage, the rime-population’s iteration speed is controlled by , and ultimately a better solution is obtained.
3.4. Modeling the Improved RIME-XGBoost
This paper proposes an improved RIME-XGBoost model that uses a multi-strategy optimized algorithm. It integrates the processes of dataset partitioning, improved RIME-XGBoost model training, hyper-parameter optimization, and evaluation index calculation, as shown in Figure 8. The core pseudo-code of the improved RIME algorithm is demonstrated in Algorithm 1.
| Algorithm 1 Pseudo-code of improved RIME algorithm. |
|
Figure 8.
Flowchart of the improved RIME-XGBoost model.
The major steps are listed below: (1) The model uses three major factors as inputs: ambient factors, NSRT operating state, and the BUS temperature. Then we divide the data into training (80%) and testing (20%) datasets. (2) For modeling XGBoost, parameters such as the number of trees, the learning rate, and the maximum depth of the tree are selected as the objective optimization parameters. (3) The parameters selected for the model are optimize. Firstly, the RIME population is initialized by the sine map. Then the RIME algorithm is used to update the population. Eventually, the ROBL strategy is used to iteratively converge to the optimal solution. (4) The model’s fitness is evaluated according to the error values, i.e., MSE. If the model errors do not satisfy the stop condition, we continue to execute step (3) iteratively. The specific expression of the mean squared error () as the stop condition is entered as follows:
In Equation (12), is the actual value, denotes the predicted value, and is the number of samples. Generally speaking, a smaller indicates better model performance. (5) The optimal parameters of RIME-XGBoost are output to obtain the final model. Meanwhile, testing datasets are used to verify the model’s prediction accuracy. Finally, the final model is saved for direct use later on.
3.5. Statistical Criteria for Model Assessment
In order to evaluate the dependability of the model effectively in this research, several indicators such as the root mean square error (), the mean absolute error (), and the coefficient of determination () are used to represent the relationship between the actual BUS temperature value and the predicted value. Below, each of these evaluation indices is explained individually.
The root mean square error (RMSE) has the same dimension and unit as the target variable (i.e., its unit is degrees Celsius), facilitating intuitive analysis of error conditions. However, due to the squaring operation involved, it tends to amplify the impact of outliers. The larger the RMSE is, the greater the deviation between the predicted value and the actual value will be. Its calculation formula is shown in Formula (13):
where is the actual value, is the predicted value, is the average value, and is the number of samples.
Compared to RMSE, the mean absolute error (MAE) is robust to outliers and not significantly affected by extreme values, allowing it to truly reflect the average deviation level of most samples. It also has the same dimension and unit as the target variable (i.e., its unit is degrees Celsius). Its calculation formula is shown in Formula (14):
The coefficient of determination () measures the extent to which the model explains the variance in . It is commonly used for comparisons among different models, and its value ranges from 0 to 1. The closer the value is to 1, the better the model’s fitting performance. Its calculation formula is shown in Formula (15):
4. Model Validation and Results Analysis
4.1. Experimental Environment Configuration
The experiments results of this paper were processed in Python 3.9. The software environment was a Windows 10 64-bit operating system. The CPU was an Intel Core i9-9900K processor (Intel, Santa Clara, CA, USA). The memory was a Samsung DDR4 3200 MHz 32 GB RAM (Samsung, Suwon, Republic of Korea). The GPU was an NVIDIA GeForce RTX 3080 10 GB (Nvidia, Santa Clara, CA, USA).
4.2. Comparative Analysis with Common Models
In the context of model prediction, the model frameworks adopted for similar application scenarios should generally be quite similar. Indeed, some researchers have employed neural networks, and scholars have attempted to realize real-time temperature prediction of structures without temperature sensors by leveraging machine learning or neural network models based on long-term historical temperature data [7,8,9,19]. Although these models are all used for structure thermal prediction scenarios, the specific prediction objects differ significantly, leading to distinct differences in the training datasets of the models. Consequently, the scene migration adaptability of different models varies. Therefore, to comprehensively analyze the performance of the proposed model in this paper, a comparative analysis will be conducted between GRU (Gated Recurrent Unit), CNN (Convolutional Neural Network), LSTM (Long Short-term Memory), Transformer-BiLSTM, and the proposed model. Table 3 shows the model parameters required for each model. The training and testing results are presented in Table 4.
Table 3.
Common models’ parameter settings.
Table 4.
Common models’ prediction performance.
Combined with the performance results of each model in Table 4, the prediction accuracy and fitting ability are analyzed. The improved RIME-XGBoost model demonstrates the highest prediction accuracy and fitting capability across all evaluated metrics. It achieves an extremely low MAE of 0.331 K (train) and 0.621 K (test), an RMSE of 0.369 K (train) and 0.897 K (test), and an of 0.999 K (train) and 0.996 K (test) in the training and test stages, respectively. This indicates that the improved ensemble learning framework (XGBoost) excels in capturing complex patterns within the data. Following this, the Transformer-BiLSTM model exhibits remarkable performance, with MAE values of 1.08 K (train) and 1.166 K (test), RMSE values of 1.458 K (train) and 1.543 K (test), and values of 0.988 (train) and 0.984 (test). The integration of Transformer and BiLSTM enables effective capture of long-range dependencies and sequential features, leading to superior accuracy compared to traditional deep learning models. In contrast, traditional deep learning models such as GRU, CNN, and LSTM exhibit moderate performance. Among them, CNN has a relatively better (0.962 for training and 0.959 for testing), while LSTM performs the least competitively in terms of both error metrics and fitting ability.
From the analysis of the generalization ability of each model, the Transformer-BiLSTM model displays excellent generalization ability, as evidenced by the minimal differences between its training and testing metrics. The consistency in MAE, RMSE, and across stages suggests that it can adapt well to unseen data. Models like CNN and GRU also maintain reasonable generalization, with small gaps between their training and testing performances. This implies that these models have relatively stable performance when applied to new data. For the improved RIME-XGBoost model, although there is a noticeable difference between the training and testing metrics, its overall performance remains at a very high level, indicating strong yet slightly less consistent generalization compared to Transformer-BiLSTM. LSTM, on the other hand, exhibits the relatively weakest generalization ability among the models, with the most significant increase in error metrics from the training to the testing stage.
Through the analysis of each model architecture, ensemble learning, as exemplified by the improved RIME-XGBoost model, proves to be highly effective in achieving ultra-high prediction accuracy, making it a preferred choice when the highest precision is the primary goal. The fusion architecture of Transformer-BiLSTM combines the strengths of Transformer’s self-attention mechanism and BiLSTM’s sequential processing capability, resulting in both high accuracy and strong generalization, which is particularly suitable for tasks involving sequential data with long-range dependencies. Traditional deep learning models (GRU, CNN, and LSTM) offer decent performance but lack the breakthroughs in accuracy and generalization seen in the more advanced architectures. Among them, CNN demonstrates better performance in capturing spatial features, while GRU and LSTM handle sequential data with varying degrees of effectiveness, with GRU being relatively more efficient and stable.
In conclusion, the improved RIME-XGBoost stands out for its unparalleled prediction accuracy, Transformer-BiLSTM excels in balancing accuracy and generalization, and traditional models like CNN and GRU provide viable options for scenarios with moderate performance requirements. The choice of model should be tailored to specific task priorities, such as precision, generalization, or computational efficiency.
4.3. Performance Comparisons of XGBoost Based on Other Heuristic Optimization
As discussed in Section 4.2, the comprehensive predictive performance of the proposed model surpasses that of several comparable models. Section 4.3 evaluates the performance of the proposed model against other XGBoost models based on different heuristic algorithms. The goal is to investigate whether the hyperparameters optimized by improved RIME significantly outperform heuristic settings and to determine if improved RIME effectively addresses the limitations of XGBoost in hyperparameter selection. The heuristic algorithms employed in these comparative models are either those proposed in recent journal papers or widely used optimization algorithms. They include Particle Swarm Optimization (PSO) [23], Sparrow Search Algorithm (SSA) [24], Pelican Optimization Algorithm (POA) [25], Ivy Algorithm (IVY) [26], Hiking Optimization Algorithm (HOA) [27], RIME, and Improved RIME. Detailed information about them is shown in Table 5. Furthermore, regarding the parameter settings of these heuristic optimization algorithms, after parameter debugging, it was found that the number of hyperparameters is relatively small in this scenario. When the population size exceeds 10, there is no significant change in the model’s prediction accuracy. Therefore, for all the heuristic algorithms involved in the comparison, the population size is set to 20 and the maximum number of iterations is set to 50.
Table 5.
Detailed information about comparative heuristic algorithms.
As shown in Table 5, PSO is renowned for its fast convergence, which enables rapid solution exploration. It is also robust, maintaining stable performance across various problem landscapes, and is easy to implement. SSA has strong global search ability and can explore a wide range of solution spaces to avoid falling into local optimization. POA-XGBoost has a simple structure, which simplifies its understanding and deployment. It can also achieve efficient optimization to ensure that the XGBoost model is effectively optimized with minimal computational overhead. IVY is flexible and can adapt well to different problem forms. It is especially good at dealing with multimodal problems, that is, problems with multiple optimal solutions, and can efficiently explore the complex solution space. HOA has fast convergence, similar to Particle Swarm Optimization (PSO), and can effectively balance exploration (searching for new areas) and development (optimizing known good areas). It can be seen that these heuristic algorithms have their own advantages, but the performance in the specific context of realizing the temperature prediction of telescopes’ back-up structure based on XGBoost still needs to be analyzed experimentally. Table 6 shows the performance results of the model optimized by these heuristic algorithms.
Table 6.
Heuristic models’ prediction performance.
Comparing the performance of the models in Table 6 and Table 4, it can be seen that the XGboost model has good prediction accuracy in the thermal field of the reflector antenna’s BUS. All the heuristic algorithms selected in the experiment combined with the XGBoost model show that the performance of these models’ MAE is within 1 K, RMSE is within 1.5 K, and is more than 98%, which are significantly better than other mainstream models (such as GRU, CNN, LSTM, etc.). It can be concluded that in the specific data prediction scenario, the selection of the baseline model largely determines the final model architecture.
From the perspective of the overall prediction accuracy of the model, the performance of these XGBoost models optimized directly by heuristic algorithms is basically similar. It can be seen that the performance results of the new optimization algorithms proposed in recent years are better than those proposed in earlier years. Most importantly, the performance of the improved RIME-XGBoost model proposed in this paper is significantly better than other XGBoost models optimized directly by heuristic algorithms. The improved RIME-XGBoost also has the lowest MAE (0.331 K for the training set and 0.621 K for the test set) and RMSE (0.369 K for training set and 0.897 K for test set) in all models. In contrast, the error index of POA-XGBoost is relatively high (MAE on the test set is 1.07 K; RMSE is 1.498 K). In summary, the integration of heuristic algorithms with XGBoost yields promising predictive performance, and improved RIME-XGBoost emerges as the most robust model in terms of accuracy, generalization, and error control. The variability in performance across models underscores the importance of selecting or adapting heuristic algorithms based on the specific requirements of the prediction task.
4.4. Performance Verification of Improved Model Under Extreme Conditions
Through the comparative analysis of various models and optimization algorithms, it can be concluded that the improved RIME-XGBoost model proposed in this paper has excellent performance in the scenario of obtaining the antenna’s BUS temperature. Figure 9 shows the RMSE distribution between the BUS temperature values predicted by the improved RIME-XGBoost model and the actual temperature data in the field. As shown in Figure 9, the temperature prediction accuracy in the area with more dense temperature sensors is almost within 0.7 K, and the overall prediction accuracy of the BUS temperature is within 1 K.
Figure 9.
Schematic diagram of the prediction accuracy of the BUS temperature field.
In order to verify the performance of this model under extreme environmental conditions, we selected the three features with the strongest correlation as the key variables influencing the environmental conditions from Figure 5. These three variables are ambient temperature, relative humidity, and solar altitude. We divided these three feature variables into intervals to test the performance of the model under different environmental conditions. Table 7 represents the parameter range division information for extreme working conditions, and Table 8 presents the specific performance results of the model under each environmental condition. Figure 10 shows the visualization results of prediction accuracy.
Table 7.
Information on the parameter range division for extreme working conditions.
Table 8.
The performance of the model under extreme conditions.
Figure 10.
Schematic diagram of the BUS temperature field prediction accuracy under various extreme conditions.
From the analysis of the overall performance and stability of the model, the model demonstrates high predictive accuracy across most environmental conditions, as indicated by the consistently high values (all above 0.973). However, performance metrics (MAE and RMSE) exhibit variations under different extreme scenarios, suggesting that extreme environmental factors impose challenges on the model’s predictive capability.
As for the influence of ambient temperature variables on the model, when the operating condition is extremely cold or extremely hot, both temperature extremes lead to increased prediction errors (higher MAE and RMSE) and a slight decrease in . Notably, extremely hot conditions affect the model more significantly than extremely cold conditions, implying that the model is more sensitive to high-temperature extremes.
For humidity conditions, compared with medium humidity, extremely dry and extremely wet environments will reduce the performance of the model. Among these, extremely dry conditions cause the most significant increase in MAE and RMSE, indicating that extreme dryness is a more critical factor for the model’s temperature prediction accuracy than extreme wetness.
As for the influence of the solar altitude angle variable on the model, the error is higher than medium solar altitude (MAE = 0.462 K, RMSE = 0.493 K, = 0.999) for both low and high solar altitudes. High solar altitude has a slightly more pronounced effect on MAE and RMSE than low solar altitude, while remains relatively high (above 0.989) under both extreme solar altitude scenarios, suggesting that the model still maintains reasonable explanatory power even under such conditions.
In summary, the model exhibits strong performance under moderate environmental conditions but experiences measurable degradation in prediction accuracy under various extreme conditions (extremely cold/hot, extremely dry/wet, and extreme solar altitudes). Among these extreme conditions, extremely dry, extremely hot, and high solar altitude conditions are the most impactful on the model’s MAE and RMSE. These findings highlight the need for further model optimization or adaptation strategies to enhance robustness in extreme environments.
4.5. In-Depth Discussion of Model Performance
Through a comprehensive comparative analysis of the performance data of various models across Section 4.2, Section 4.3 and Section 4.4, it can be concluded that the selection of the baseline model is crucial in the initial stage of designing the model architecture for specific prediction scenarios. An inappropriate choice is likely to result in the designed model failing to achieve the expected prediction accuracy. Additionally, if a model exhibits poor accuracy on the training set, it will inevitably perform poorly when validated on the test set. Furthermore, a model with considerable prediction accuracy on the training set does not necessarily demonstrate excellent performance on the test set. This perspective is exemplified by the model proposed in this paper: although its performance in the training set is MAE = 0.331 K, RMSE = 0.369 K, and = 0.999, its performance in the test set is MAE = 0.621 K, RMSE = 0.897 K, = 0.996 (nevertheless, its performance on the test set remains optimal compared to other models).
From the performance results, the proposed model in this paper shows an excessively rapid decline in prediction accuracy error from the training set to the test set, indicating that the generalization ability of the proposed model needs to be improved. An analysis of the model’s generalization ability under extreme operating conditions reveals that the model exhibits unsatisfactory prediction accuracy in extremely dry, extremely cold, and extremely hot environments. After analyzing the dataset, we found that this may be related to changes in the physical parameters of heat transfer involved in the structure thermal scenarios. Under extremely dry or extremely cold conditions, the convective heat transfer coefficient of the air medium in the external environment undergoes significant changes, leading to alterations in the duration of the structure temperature response. Consequently, the temperature predicted by the model at such times may suffer from temporal misalignment. Although such cases are not common, they should be the focus of future research. In addition, regarding the model’s poor performance under extremely hot conditions, an analysis of the original collected data shows that the training dataset consists of data from November to June of the following year, thus lacking summer data. This deficiency ultimately results in the model’s subpar prediction performance in extremely hot environments.
5. Conclusions
High precision and real-time acquisition of structure temperature data of large aperture radio telescopes is essential to ensure the long-term stable use of the antenna. This paper proposes a multi-strategy optimized RIME-XGBoost model to predict the real-time temperature of the NSRT’s BUS based on the real-time environmental conditions near the NSRT. We combine meta-heuristic optimization with ensemble learning to address the challenge of acquiring the BUS temperature data in complex outdoor environmental conditions. The key findings include the following:
(1) In the temperature prediction scenario of the reflector antenna back-up structure, the improved rime XGBoost model shows excellent prediction results. Compared with the actual temperature sensor, the prediction accuracy of the model is 97.15%, and the prediction error (RMSE) is less than 0.897 K.
(2) The XGBoost model optimized by the heuristic algorithms has good prediction accuracy in the reflector antenna back-up structure temperature prediction scenario. The performance of these models is significantly better than that of mainstream models (GRU, CNN, LSTM, and Transformer-BiLSTM), and the performance of these models includes an MAE within 1 K, RMSE within 1.5 K, and more than 98%.
(3) During the operation and service of the radio telescope, the developed model completely gets rid of the dependence on the installation and maintenance costs of physical sensors and can achieve high-precision acquisition of bus temperature by relying only on the surrounding meteorological data and the real-time mechanical parameters of the telescope, laying the foundation for the subsequent antenna thermal control and antenna health monitoring. Future work will prioritize three directions:
(1) Increase the model’s generalization ability to overcome the problem of model performance degradation in extreme environments.
(2) Explore the variation law of structure temperature and heat transfer duration and enhance the effectiveness of the dataset.
(3) Deploy a lightweight variant on the embedded system to realize the real-time thermal monitoring of the BUS.
Author Contributions
Conceptualization, Q.X. and S.-J.Z.; methodology, Q.X. and H.W.; software, S.-J.Z.; validation, S.-J.Z., H.W., and F.X.; formal analysis, F.-L.H.; investigation, X.-M.C.; resources, F.X.; data curation, F.X.; writing—original draft preparation, S.-J.Z.; writing—review and editing, Q.X.; visualization, S.-J.Z.; supervision, X.-M.C. and F.-L.H.; project administration, Q.X.; funding acquisition, Q.X. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Tianshan Talent Training Program of Xinjiang Uygur Autonomous Region (2023TSYCCX0105); the Xinjiang Uygur Autonomous Region “Tianshan innovation team” (No. 2024D14003); the Western-Light Key Laboratory Project of the Western Cross-disciplinary Team (xbzg-zdsys-202320); the Scientific Instrument Developing Project of the Chinese Academy of Sciences (PTYQ2022YZZD01); and Shaanxi Antenna and Control Technology Key Laboratory open fund.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The data presented in this study are available on request from the corresponding author.
Conflicts of Interest
The authors declare no conflicts of interest.
Abbreviations
The following abbreviations are used in this manuscript:
| BUS | back-up structure |
| XGBoost | eXtreme Gradient Boosting |
| NSRT | Nanshan 26-m Radio Telescope |
| RMSE | root mean square error |
| IRAM | Institute of Radioastronomie Millimétrique |
| SRT | Sardinia Radio Telescope |
| GBT | Green Bank Telescope |
| FBG | Fiber Bragg Grating |
| PTFE | polytetrafluoroethylene |
| OBL | opposition-based learning |
| ROBL | refractive opposition-based learning |
| MSE | mean squared error |
| MAE | mean absolute error |
| ASTE | Atacama Submillimeter Telescope Experiment |
| OOF | Out of Focus |
| FEM | finite element method |
| CNN | convolutional neural network |
| LSTM | long short-term memory |
| BARO | barometric pressure |
| GRU | Gated Recurrent Unit |
References
- Greve, A.; Bremer, M. Thermal Design and Thermal Behaviour of Radio Telescopes and Their Enclosures; Springer: Berlin, Germany, 2010. [Google Scholar]
- Li, G.; Li, Z.C.; Mou, A.G. An Experimental Research on Sunshine Temperature Effect upon the Back Structure of the 65-m Radio Telescope. Acta Astron. Sin. 2013, 54, 189–198. [Google Scholar] [CrossRef]
- Ukita, N.; Ezawa, H.; Ikenoue, B.; Saito, M. Thermal and Wind Effects on the Azimuth Axis Tilt of the ASTE 10-m Antenna. Publ. Natl. Astron. Obs. Jpn. 2007, 10, 25–33. [Google Scholar]
- Attoli, A.; Stochino, F.; Buffa, F.; Poppi, S.; Serra, G.; Sanna, G.; Cazzani, A. Sardinia Radio Telescope structural behavior under solar thermal load. Structures 2022, 39, 901–916. [Google Scholar] [CrossRef]
- Nikolic, B.; Prestage, R.; Balser, D.; Chandler, C.; Hills, R. Out-of-focus holography at the Green Bank Telescope. Astron. Astrophys. 2007, 465, 685–693. [Google Scholar] [CrossRef]
- Wei, S.; Kong, D.; Wang, B.; Fu, L.; Xiao, W.; Yin, Z.; Li, Y.; Xiao, Z. Estimation of Pointing Errors of Large Radio Telescopes under Solar Radiation Based on Digital Twin. Symmetry 2024, 16, 668. [Google Scholar] [CrossRef]
- Yan, G.X.; Bai, Y.; Yu, C.Q.; Yu, C.M. A Multi-Factor Driven Model for Locomotive Axle Temperature Prediction Based on Multi-Stage Feature Engineering and Deep Learning Framework. Machines 2022, 10, 759. [Google Scholar] [CrossRef]
- Hu, X.; Zhang, D.H.; Tan, R.J.; Xie, Q. Controlled Cooling Temperature Prediction of Hot-Rolled Steel Plate Based on Multi-Scale Convolutional Neural Network. Metals 2022, 12, 1455. [Google Scholar] [CrossRef]
- Li, X.; Chen, C. Temperature Prediction for Aerospace Thermal Tests Based on Physical and LSTM Hybrid Model. Aerosp. 2024, 11, 964. [Google Scholar] [CrossRef]
- Zhang, S.J.; Xu, Q.; Wang, H.; Xue, F.; Cao, X.M. Predicting the Temperature Field Distribution of Radio Telescope Back-Up Structure on EVSC Unsupervised Feature Selection and MIMO-BP Neural Network. Acta Astron. Sin. 2025, 66, 13–23. [Google Scholar] [CrossRef]
- Hirayama, N.; Sano, Y. Fiber Bragg grating temperature sensor for practical use. Isa Trans. 2000, 39, 169–173. [Google Scholar] [CrossRef]
- Di Sante, R. Fibre Optic Sensors for Structural Health Monitoring of Aircraft Composite Structures: Recent Advances and Applications. Sensors 2015, 15, 18666–18713. [Google Scholar] [CrossRef]
- Zhao, Y.; Du, J.; Xu, Q.; Bao, H. Real-Time Monitoring of the Position and Orientation of a Radio Telescope Sub-Reflector with Fiber Bragg Grating Sensors. Sensors 2019, 19, 619. [Google Scholar] [CrossRef] [PubMed]
- Xu, Q.; Bao, H. An Online Measurement and Calibration Method for a Radio Telescope Sub-Reflector Support Structure Using Fiber Bragg Grating. Micromachines 2023, 14, 1093. [Google Scholar] [CrossRef]
- Benesty, J.; Chen, J.; Huang, Y.; Cohen, I. Pearson Correlation Coefficient; Springer: Berlin, Germany, 2009; pp. 1–4. [Google Scholar] [CrossRef]
- Wei, S.X.; Kong, D.Q.; Wang, Q.M. Non-uniform temperature distribution of the main reflector of a large radio telescope under solar radiation. Res. Astron. Astrophys. 2021, 21, 293. [Google Scholar] [CrossRef]
- Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
- Zhang, Z.H.; Ye, Q.; Fu, L.; Wang, J.Q.; Guo-Xiang, M.; Shen, Z.Q. Rapid Deformation Calculation for Large Reflector Antennas: A Surrogate Model Method. Res. Astron. Astrophys. 2022, 23, 015001. [Google Scholar] [CrossRef]
- Shan, J.; Che, Z.; Liu, F. Accurate Rotor Temperature Prediction of Permanent Magnet Synchronous Motor in Electric Vehicles Using a Hybrid RIME-XGBoost Model. Appl. Sci. 2025, 15, 3688. [Google Scholar] [CrossRef]
- Shaeri, M.R.; Ellis, M.C.; Randriambololona, A.M. Xgboost-based model for prediction of heat transfer coefficients in liquid cold plates. In Proceedings of the 8th Thermal and Fluids Engineering Conference (TFEC 2023), University of Maryland, College Park, MD, USA, 26–29 March 2023. [Google Scholar]
- Su, H.; Zhao, D.; Heidari, A.A.; Liu, L.; Zhang, X.; Mafarja, M.; Chen, H. RIME: A physics-based optimization. Neurocomputing 2023, 532, 183–214. [Google Scholar] [CrossRef]
- Tizhoosh, H.R. Opposition-based learning: A new scheme for machine intelligence. In Proceedings of the International Conference on Computational Intelligence for Modelling, Control and Automation (CIMCA 2005) and the International Conference on Intelligent Agents, Web Technologies and Internet Commerce (IAWTIC 2005), Vienna, Austria, 28–30 November 2005; pp. 695–701. [Google Scholar] [CrossRef]
- Sengupta, S.; Basak, S.; Peters, R.A. Particle Swarm Optimization: A Survey of Historical and Recent Developments with Hybridization Perspectives. Mach. Learn. Knowl. Extr. 2018, 1, 157–191. [Google Scholar] [CrossRef]
- Xue, J.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
- Trojovský, P.; Dehghani, M. Pelican Optimization Algorithm: A Novel Nature-Inspired Algorithm for Engineering Applications. Sensors 2022, 22, 855. [Google Scholar] [CrossRef]
- Zhang, K.; Yuan, F.; Jiang, Y.; Mao, Z.; Zuo, Z.; Peng, Y. A Particle Swarm Optimization-Guided Ivy Algorithm for Global Optimization Problems. Biomimetics 2025, 10, 342. [Google Scholar] [CrossRef] [PubMed]
- Oladejo, S.O.; Ekwe, S.O.; Mirjalili, S. The Hiking Optimization Algorithm: A novel human-based metaheuristic approach. Knowl.-Based Syst. 2024, 296, 111880. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).