A Simpliﬁed Climate Change Model and Extreme Weather Model Based on a Machine Learning Method

: The emergence of climate change (CC) is a ﬀ ecting and changing the development of the natural environment, biological species, and human society. In order to better understand the inﬂuence of climate change and provide convincing evidence, the need to quantify the impact of climate change is urgent. In this paper, a climate change model is constructed by using a radial basis function (RBF) neural network. To verify the relevance between climate change and extreme weather (EW), the EW model was built using a support vector machine. In the case study of Canada, its level of climate change was calculated as being 0.2241 (“normal”), and it was found that the factors of CO 2 emission, average temperature, and sea surface temperature are signiﬁcant to Canada’s climate change. In 2025, the climate level of Canada will become “a little bad” based on the prediction results. Then, the Pearson correlation value is calculated as being 0.571, which conﬁrmed the moderate positive correlation between climate change and extreme weather. This paper provides a strong reference for comprehensively understanding the inﬂuences brought about by climate change. Stata determine 11 significant indicators. Then, the CC model was established on the basis of an RBF neural network. It was found that the current climate change level of Canada is 0.2241, which indicates that the level of Canada’s climate change is normal on the whole. Meanwhile, the amount of ice coating had a negative effect on climate change, and “the average temperature” and “global carbon dioxide concentration” were quite significant factors, and the contribution of SST was recognized as well. Furthermore, a prediction of the possible trend of climate change can provide more useful guidance information, and considering differences in the model performance, two prediction methods were conducted at the same time, which were from the CC model and RBF neural network. It was found that the global climate level will break through the normal level in 2025 and the future climate change will develop in the direction of “a little bad” in the long term. Additionally, the forecast result indicated that the CC model is reliable and has good applicability. However, the prediction ability of the CC model is up to 34 years. When constructing the extreme weather model, due to the local and targeted characteristics of the indicators in the EW model, reliable precipitation and temperature indicators were extracted from prior literature. Then, the EW model is constructed on the basis of a SVM and the JD algorithm. In the final correlation analysis, the Selkirk Fortress region of Canada was selected as the research area due to its obvious extreme characteristics . After extracting corresponding data of its indicators from 1990 to 2018, the Pearson correlation value was selected to measure the correlation between climate change and extreme weather. The results showed that there was a moderate positive correlation; therefore, even if local extreme weather is frequent, global warming will continue as well. The above result is also strong evidence that can explain the apparent contradiction between local extreme cold and global warming. As the above results are studied from the perspective of a macro average, if the model is applied to a narrower area, the Pearson value will be larger . In fact, some similar machine learning models have already been adopted


Introduction
Over the past hundred decades, global temperatures have been rapidly rising. According to the latest data from NASA, the long-term trend of global temperature rise has become more apparent, as is indicated in Figure 1. From 1880 to 2018, where the warmest ten years begin at 2015, and 2016 is the hottest year since 1880 [1,2]. However, the impacts of climate change (CC) are not just a rise in temperature. There is much scientific evidence showing that the effects brought by climate change are diverse, such as warming oceans, decreased snow cover, sea-level rise, ocean acidification, extreme events, etc. [3]. The negative influence of climate change seems to appear more and more prominent. It creates not only new risks and exacerbates existing vulnerabilities, but also offers various challenges to human health and safety, quality of life, and the rate of economic growth [4]. In order to prevent the further development of climate change, some scientific associations have studied the root causes of climate change and it is undeniable that human activities are the main factors [5]. However, there are still many factors that should be highlighted. Simply studying the relationship between certain natural phenomena and climate change cannot reflect the impact of climate change from a macro perspective. Therefore, a comprehensive study involving various influential factors is critically needed [6].
factors [5]. However, there are still many factors that should be highlighted. Simply studying the relationship between certain natural phenomena and climate change cannot reflect the impact of climate change from a macro perspective. Therefore, a comprehensive study involving various influential factors is critically needed [6]. There is a large volume of published studies that have researched the influence of climate change. Some scientific methods have already been adopted in climate change research. Research using a process-oriented approach reveals the ecological response to recent climate change [7]. It discussed the phenology and physiology of organisms, the range and distribution of species, the composition of the community, and the dynamic characteristics of ecosystems. This confirms that climate change does have an influence on various species and human society, and it exposes a consistent mode of ecological change across the whole system [8]. Then, a scientific study on the sensitivity, potential impact, adaptability, and vulnerability of forests was conducted, which indicated that moderate increases in carbon dioxide concentration and temperature will promote forest growth and wood productivity in the short term. However, the enormous risks brought by extreme weather (EW) phenomena, such as drought and floods, is likely to reduce the wood productivity of the forest [9]. Another important finding has found that the impact of climate change on biological species is becoming more apparent than ever before. Using a climate envelope species distribution model to disclose the effect of climate change on wild crop relatives, 16-22% of the estimated species will go extinct in a specific migration scenario, and the range size of most species will reduce by over 50% [10]. In fact, the occurrence of climate change not only brings a huge challenge to the ecological environment and species diversity, but may also increase the spread of infectious diseases. This is because an essential consequence of climate change is global warming, and the rise of temperatures can accelerate the spread of some infectious diseases, such as the Ross River virus disease, which will likely threaten human health [11][12][13]. In addition, climate change is also closely related to community well-being, which suggests that the impact of climate change on human society is increasing [14]. Meanwhile, climate change adaptation policies caused by climate change also play an essential role in the stability of the country and society [15,16]. Obviously, climate change is no longer a matter of simple ecological environment change, but is also a primary issue related to social stability and national sustainable development.
However, the current impacts of climate change have not been quantified clearly. Most research approaches verify the existence of climate change using existing data and phenomena. It should be mentioned that the data on climate change factors are often small in magnitude [17,18]. Hence, the public is not sensitive to the effects of climate change. For instance, the 2019 United Nations Climate Change Conference had announced that the global average annual temperature rise rate is only 0.03 °C, a phenomenon called hiatus [1]; the small change rate of 0.03 °C can cause the public to mistakenly The global temperature (℃) The global temperature from 1880 to 2018 There is a large volume of published studies that have researched the influence of climate change. Some scientific methods have already been adopted in climate change research. Research using a process-oriented approach reveals the ecological response to recent climate change [7]. It discussed the phenology and physiology of organisms, the range and distribution of species, the composition of the community, and the dynamic characteristics of ecosystems. This confirms that climate change does have an influence on various species and human society, and it exposes a consistent mode of ecological change across the whole system [8]. Then, a scientific study on the sensitivity, potential impact, adaptability, and vulnerability of forests was conducted, which indicated that moderate increases in carbon dioxide concentration and temperature will promote forest growth and wood productivity in the short term. However, the enormous risks brought by extreme weather (EW) phenomena, such as drought and floods, is likely to reduce the wood productivity of the forest [9]. Another important finding has found that the impact of climate change on biological species is becoming more apparent than ever before. Using a climate envelope species distribution model to disclose the effect of climate change on wild crop relatives, 16-22% of the estimated species will go extinct in a specific migration scenario, and the range size of most species will reduce by over 50% [10]. In fact, the occurrence of climate change not only brings a huge challenge to the ecological environment and species diversity, but may also increase the spread of infectious diseases. This is because an essential consequence of climate change is global warming, and the rise of temperatures can accelerate the spread of some infectious diseases, such as the Ross River virus disease, which will likely threaten human health [11][12][13]. In addition, climate change is also closely related to community well-being, which suggests that the impact of climate change on human society is increasing [14]. Meanwhile, climate change adaptation policies caused by climate change also play an essential role in the stability of the country and society [15,16]. Obviously, climate change is no longer a matter of simple ecological environment change, but is also a primary issue related to social stability and national sustainable development.
However, the current impacts of climate change have not been quantified clearly. Most research approaches verify the existence of climate change using existing data and phenomena. It should be mentioned that the data on climate change factors are often small in magnitude [17,18]. Hence, the public is not sensitive to the effects of climate change. For instance, the 2019 United Nations Climate Change Conference had announced that the global average annual temperature rise rate is only 0.03 • C, a phenomenon called hiatus [1]; the small change rate of 0.03 • C can cause the public to mistakenly believe that climate change has stopped, and as such, the public will think that the government does not need to continue to put too much energy and money on the issues surrounding climate change. The lack of support of government work will bring certain risks to social stability, and the public's Symmetry 2020, 12, 139 3 of 29 misunderstanding of climate change will continue to exacerbate the negative impact of climate change. In addition, in some regions, the potential impact of climate change has been overlooked due to economic backwardness, and in other regions, the impact of climate change has been exaggerated because of the hype from media and other platforms. Having an unscientific understanding of climate change is useless for solving climate problems, since even if the real data reflected by climate change do not fluctuate greatly, its impact will not be small. The public needs to correctly understand the current level of climate change and its future developmental trends, and the gap between the real situation and the public's imagination needs to be balanced. Therefore, research on quantifying the effects of climate change is essential for a sustainable future.
The purpose of this study is to achieve the coordination and symmetry between the micro research and the macro research, the complex climate and weather phenomena are integrated with scientific theories through analysis and modeling. First, providing a mathematical model of climate change that is able to be explained and comprehended. Second, the constructed extreme weather model is unique in its ability to quantify the effects of extreme weather conditions. Third, explaining the relevance between climate change and extreme weather. This research aims to provide a new way to understand the impact of climate change better.
The content structure is arranged as follows: Part 1 is a comprehensive introduction. Part 2 introduces the methods of constructing the climate change model and the extreme weather model. Part 3 provides the results and discussion, and the conclusion is given in Part 4.

Assumptions
Some climate-related weather models usually involve many changeable factors. Some assumptions are put forward to make the model more scientific. Hypothesis 1. The trend of temperature changes is continuous all the time, and there is no sudden increase or decrease in temperature.

Hypothesis 2.
The official data obtained from the observatory is enough to represent the changes within the study area and the officially provided data is accurate.

Hypothesis 3.
The statistical data is reliable in terms of reflecting the true meaning of the indicator.

Hypothesis 4.
When discussing the temporal and spatial trends of temperature, the temporal and spatial variations of temperature are considered to be two relatively independent variables. Hypothesis 5. The original climate change system is a grey system. The selected impact indicators are independent of each other and there is no cross-impact or complex multi-collinearity. Hypothesis 6. The EW model can be decomposed into the sum of the extreme temperature model (TM) and the extreme precipitation model (PM). Additionally, the complex nonlinear relationships or other relationships that may exist between TM and PM are not considered.

Construct the Climate Change Model
In a prevailing view, a complete CC model should be capable of guiding the government and relevant policymakers toward clarifying the current climate status and possible future climate change trends [19]. Thus, the work toward obtaining a CC model can be divided into two parts: (1) determining the significant indicators and (2) constructing the CC model using some reliable mathematical methods. The value of the influence of climate change (C) is defined as the level of climate change, which belongs in the range 0-1. When the value of C is closer to 0, it means the state of the climate is more stable and better. When the value of C is closer to 1, it means the state of the climate is worse. Therefore, C can be defined as the following: where i is the ith impact indicator, i = 1, 2, . . . , 11, q i is the weight coefficient of the ith indicator, C i is the climate change value of the ith indicator, and adding the climate change values of all indicators gives the overall climate change value. x i is the normalized processed value of the ith index, which is obtained by mapping the index to the interval [0,1] after dimensionless processing. ε is the error term of the CC model. The expression of the important factor x i is given as follows: In order to obtain the CC model, determining the indicators involved in the CC model should be completed at first.

Determination of the Indicators
It is well accepted that there are numerous factors affecting climate change; however, this does not mean that each factor contributes significantly to climate change. To determine the important factors contributing to climate change, Laplacian eigenmapping (LM) was used as a nonlinear dimensionality reduction technique is adopted to facilitate the establishment of the CC model.
LM is a graph-theory-based method. It maps the randomly distributed points in space through analyzing the similarity between neighboring structures to obtain the data shape in low-dimensional space [20]. According to the local neighbor relationship existing in the high-dimensional space, the data set of the local neighbor relationship in the low-dimensional area is obtained.
The graph is composed of nodes and edges. The edge represents the value information. The degree of the node is the number of edges passing through the point. The degree of node i is the sum of the weights of all edges passing through the node. The definition matrix D is a diagonal matrix whose primary oblique element value is the degree of the node.
where d i is the weight of the edge between nodes i and j.
In this study, the directed graph of the LM was applied, which is a semi-positive definite matrix. All the eigenvalues are non-negative real numbers. The transform of dimensionality reduction is obtained by decomposing the eigenvalues of the LE matrix.
The selected data set x 1 , x 2 , . . . , x n are the vectors of space vector R l , and the space vector of the expected dimensionality reduction is defined as R m , m ≤ l. M is a manifold embedded in R l . The steps involved in applying the LM are as follows.

Establish the adjacency relationship of the construction diagram
If the distance between the sample points of x i and x j is small, an edge can be made for x i and x j . When judging the distance between two sample points, the criterion is the neighbor rule. If node i is in the nearest n neighbor node sets of the node j, then the nodes of i and j are considered to be close. If there are five nodes, the possible relationship can be described as shown in Figure 2, thereby the relationship between nodes can be determined.

Calculate the weight of the edge
If node exists on the edge of node , it means that the two nodes are connected and the weights of the edges between the two nodes are given by: where is a constant. If two nodes are unconnected, then = 0.
After getting the weight of each edge, then the directed graph can be obtained.

Feature mapping
Based on the directed graph, calculate the Laplacian matrix and solve the generalized eigenvalues and eigenvector values: = , (5) where represents the Laplacian matrix, is the generalized matrix, and is the eigenvalue.
If the solution of the generalized matrix is , , … , , the feature vector values are sorted in ascending order to obtain: 0 = ≤ ≤ … ≤ .
(6) If the eigenvalue of the point is zero, remove the point. Subsequently, the eigenvectors of the remaining points are constructed into the final projection matrix. The data is mapped into the lowdimensional space, thereby the dimensionality reduction processing of the indicator is complete.

Climate Model Based on a Radial Basis Function Neural Network
Inspired by the work of predecessors, a radial basis function (RBF) neural network is adopted to get the weight coefficient of each indicator. It is worth noting that the RBF neural network is suitable for managing the issues of weight connections. The relationship between the input vector and the output vector can be confirmed without the requirement of weight connections [21]. There is a linear assumption between the hidden layer and the output layer, namely that the output of the network is linear for tunable parameters. Thus, the speeds of learning and training are promoted in a way, and the learning rate is rather low. When the amount of involved data is quite large, the dominating advantage of the RBF neural network becomes clear. The significant steps of the RBF neural network are shown below.

Calculate the weight of the edge
If node i exists on the edge of node j, it means that the two nodes are connected and the weights of the edges w ij between the two nodes are given by: where t is a constant. If two nodes are unconnected, then w ij = 0. After getting the weight of each edge, then the directed graph can be obtained.

Feature mapping
Based on the directed graph, calculate the Laplacian matrix and solve the generalized eigenvalues and eigenvector values: where L represents the Laplacian matrix, f is the generalized matrix, and λ is the eigenvalue. If the solution of the generalized matrix is f 0 , f 1 , . . . , f k−1 , the feature vector values are sorted in ascending order to obtain: If the eigenvalue of the point is zero, remove the point. Subsequently, the eigenvectors of the remaining points are constructed into the final projection matrix. The data is mapped into the low-dimensional space, thereby the dimensionality reduction processing of the indicator is complete.

Climate Model Based on a Radial Basis Function Neural Network
Inspired by the work of predecessors, a radial basis function (RBF) neural network is adopted to get the weight coefficient of each indicator. It is worth noting that the RBF neural network is suitable for managing the issues of weight connections. The relationship between the input vector and the output vector can be confirmed without the requirement of weight connections [21]. There is a linear assumption between the hidden layer and the output layer, namely that the output of the network is linear for tunable parameters. Thus, the speeds of learning and training are promoted in a way, and the learning rate is rather low. When the amount of involved data is quite large, the dominating advantage of the RBF neural network becomes clear. The significant steps of the RBF neural network are shown below.
(1) The determination of parameters Symmetry 2020, 12, 139 6 of 29 1. Identify the initial input vector, output vector, and expected output vector: here X is the input vector and n represents the nth input layer. Y is the output vector, and O is the expected output vector. q is the qth output layer.
2. Initialize weights between the hidden layer and the output layer: where p is the pth hidden layer.
Based on center initialization theory, the weight initialization method from the hidden layer to the output layer is expressed below: where mink is the minimum value of all expected outputs in the kth output neuron in the training set and maxk is the maximum value of all expected outputs of the kth output neuron in the training set.

Initialize the central parameters of each neuron in the hidden layer
The initial value of the central parameter of the RBF neural network is given below: where p is the total number of neurons in the hidden layer and j = 1, 2, . . . , p.

Initialize the width vector
The width vector affects the range of the action of neurons regarding input information; the smaller the width, the narrower the shape of the action function of the corresponding hidden layer neurons. Its function is as follows: where d f is the width adjustment coefficient; its value is less than 1. The function of d f is to realize the ability to perceive and respond to local information easily and accurately.
(2) Calculate the output value, z j , of the jth neuron in the hidden layer: Symmetry 2020, 12, 139 where C j is the central vector in the hidden layer of the jth neuron. It is composed of the jth neuron in the hidden layer corresponding to the central component of all neurons in the input layer. D j is the width vector of the jth neuron in the hidden layer corresponding to C j .
(3) Calculate the output of neurons in the output layer: (4) Calculate the weight coefficient The training method for the weight parameters of the RBF neural network is the gradient descent method. The central, width, and adjusting weight parameters are adjusted to the best value by learning through adaptation, and the iterative calculation is as follows: where W k j (t) is the adjustment weight between the kth output neuron and the jth hidden layer neuron in the tth iteration calculation. c ji (t) is the adjustment weight between the jth output neuron and the ith hidden layer neuron in the tth iterative calculation. d ji (t) is the width corresponding to the centre c ji (t). Then, the evaluation function of the RBF neural network is constructed as follows: where E represents the evaluated result of the RBF neural network, O lk is the expected output value of the kth output neuron in the input sample l, and y lk is the network output value of the kth output neuron in the input sample l.
(5) Calculate the RMS of the network output: If RMS ≤ ε, the network training is finished. Otherwise, retrain the network from step 3.

Construct the Extreme Weather Model
Recently, global warming has been intensifying, but local extreme weather frequently occurs, which makes many people think that global warming has already stopped. The paradox is whether there is a conflict between the climate change model and the extreme weather model. Therefore, it is necessary to carry out deeper research on extreme weather. Furthermore, another goal of establishing an EW model is to separate the factors that can explain extreme weather from a series of disordered extreme weather indicators.

Determination of the Indicators
In this study, the EW model only covers extreme temperature conditions and extreme precipitation conditions. According to Kaas and Frich [22], the requirements for extreme weather indicators are "weak extremes, low noise, and more reliable." In the previous study of extreme weather, the extreme value threshold method is usually used to define whether a weather phenomenon is extreme. If the threshold value is exceeded, the weather condition is considered to be an extreme weather event. The threshold value is set according to the existing literature, studies, and standards published on the government website. Hence, this study finally selected 13 indicators that can be used to describe extreme weather. The selected indicators and the corresponding threshold values can be seen in Table 1.

Extreme Weather Model Based on Support Vector Machine
The primary purpose of establishing the extreme weather model was to separate the factors that can explain extreme weather from a series of disordered extreme weather indicators and assign corresponding weights to these explanatory variables to define extreme weather. A support vector machine (SVM) method was adopted to establish the EW model. Different from prior perceptron models, a SVM can classify a distance hyperplane more accurately, especially the point closest to the hyperplane distance. The core element of the SVM method is to keep the spots near the hyperplane away from the hyperplane. Then, maximize the interval such that the classification points are on both sides of the respective support vectors to achieve an optimized classification effect [23].
Assume the hyperplane has the form w T x + b, where it is required that a point close to the hyperplane moves far from the hyperplane. This can be described as follows: where x is the input vector, w is the weight vector, b is the bias, and γ is the function interval, which is used to determine whether the index classification is correct. At the same time, in order to unify the conditions, a constraint condition is added to the normal vector w, and γ is defined to measure the precise distance between the hyperplane and the observed value. In this part, the constraint function y i w T x i + b is defined as being 1, and therefore Equation (22) can be simplified to: Since the maximization problem of 1 w 2 is equivalent to the minimization problem of 1 2 w 2 2 , the optimization function of SVM is equivalent to Equation (24): The objective function 1 2 w 2 2 is convex, so the Lagrangian function can be used to solve the optimal norm problem. The mathematical form of the Lagrangian function is shown as follows: where α i represents the optimal coefficient of the hyperplane. Subsequently, optimization goals can be simplified as min max L(ω, b, α) by deriving the partial derivatives of w and b, respectively.
Based on Equation (26), the relationship between w and α can be determined. Therefore, the optimization function ψ(α) can be defined: Owing to ψ(α) only having a relationship with α, the problem of solving the objective function ψ(α) can be transformed to obtain the solution α: The solution process of Equation (28) requires the use of the sequential minimal optimization algorithm to obtain the corresponding values of α* and ω* [24]. Through the above processes, all points can be classified on the side of their respective support vectors. A linear ε insensitive loss function is selected to allow for error minimization. Then, the RBF neural network is used to regress the SVM. Due to the different parameter values set in the software, the regression function shape obtained is quite different [25]. Therefore, it remains a need to learn and train the sample multiple times, the best combination of parameters can be obtained eventually, and the regression parameters of the SVM can be obtained as well.

Data
In this study, Canada was the targeted country. Historical data on weather across Canada can be found at http://climate.weather.gc.ca/historical_data/search_historic_data_e.html. In fact, there are many climatic stations in every administrative region of Canada. In this study, due to the influence of regional differences on the research results, stations in every administrative region were selected. Additionally, due to the impact of factors, such as the scale of different climate stations, equipment accuracy, management, etc., the quality of data between each site varied a lot. Therefore, the quality of the data provided by the station is taken into consideration during station selection. The selected climate stations are shown in Figure 3 and Table 2.
Initially, data preprocessing on the original data was performed. First, all data was imported into the MATLAB (R2019a) software and the properties of the data strip with incomplete data was set, allowing for the existence of null values. The data obtained contained noise and clutter, and therefore data mining technology was used to perform cleaning, integration, conversion, discretization, and reduction on the filtered data. As is indicated in Figure 4, the data fluctuated greatly, and there were abnormal points. The maximum variance of some data was 2.1 when the deviation test was performed.   Initially, data preprocessing on the original data was performed. First, all data was imported into the MATLAB (R2019a) software and the properties of the data strip with incomplete data was set, allowing for the existence of null values. The data obtained contained noise and clutter, and therefore data mining technology was used to perform cleaning, integration, conversion, discretization, and reduction on the filtered data. As is indicated in Figure 4, the data fluctuated greatly, and there were abnormal points. The maximum variance of some data was 2.1 when the deviation test was performed.   Then, the data attributes were verified. The types of the data were digital, the interval was set to [−30, 20], and the length of the value ranged was up to one decimal place. Outliers were identified as points that were more than two standard deviations away from the mean of a given attribute. The clustering adjusted the existing deviation correction. The processed result is shown in Figure 5, where it is apparent that the processed data is smooth and stable.

The Climate Change Model
Using the existing literature and official government documents as a reference, some original indicators affecting climate change can be selected [26,27]. However, the quality of the information varies greatly, therefore information from the academic literature is dominant, which is supplemented by information obtained from other channels. First, when acquiring indicators on the basis of a range of information sources, the indicators should be sufficiently independent, and its  Then, the data attributes were verified. The types of the data were digital, the interval was set to [−30, 20], and the length of the value ranged was up to one decimal place. Outliers were identified as points that were more than two standard deviations away from the mean of a given attribute. The clustering adjusted the existing deviation correction. The processed result is shown in Figure 5, where it is apparent that the processed data is smooth and stable.  Then, the data attributes were verified. The types of the data were digital, the interval was set to [−30, 20], and the length of the value ranged was up to one decimal place. Outliers were identified as points that were more than two standard deviations away from the mean of a given attribute. The clustering adjusted the existing deviation correction. The processed result is shown in Figure 5, where it is apparent that the processed data is smooth and stable.

The Climate Change Model
Using the existing literature and official government documents as a reference, some original indicators affecting climate change can be selected [26,27]. However, the quality of the information varies greatly, therefore information from the academic literature is dominant, which is

The Climate Change Model
Using the existing literature and official government documents as a reference, some original indicators affecting climate change can be selected [26,27]. However, the quality of the information varies greatly, therefore information from the academic literature is dominant, which is supplemented by information obtained from other channels. First, when acquiring indicators on the basis of a range of information sources, the indicators should be sufficiently independent, and its corresponding data needs to be available. Then, the frequency of each indicator should be counted. The initial index selection results are shown in Table 3. Owing to the large workload, only some of official and literature sources are shown below. Cooling degree-days https://climatedataguide.ucar.edu/climatedata/outgoing-longwave-radiation-olr-hirs [52,53] It should be mentioned that although 17 original indicators were determined, the significance of each indicator was ambiguous. Therefore, it was necessary to use scientific mathematical methods to identify the important indicators, and only indicators that had a significant impact on the overall result were retained in the climate change model. First, Stata 15.0 and MATLAB (R2019a) software were applied to extract the corresponding data. Then, LM was used to select the indicators with significant contributions. In addition, the climate was viewed as a grey system with multivariate interaction; by considering the uncertainty and greyness of the system, the RBF neural network was used to determine the influence weight coefficient of each index, and the CC model was finally obtained.

Determination of the Climate Change Model
LM was applied to recognize the significant indicators. As the 17 original indicators were confirmed, the adjacency relationship of the construction diagram was subsequently established, and the conceptual diagram is shown in Figure 6. Using MATLAB software to conduct the feature mapping, the points with an eigenvalue of zero were removed to construct the final projection matrix using the eigenvectors of the remaining points. The dimension reduction processing of the indicator was completed when the data was mapped in a low-dimensional space. The results are shown in Figure 7, the original data was scattered randomly in three-dimensional space, and the ordered data in two-dimensional space was obtained after the dimensionality reduction process. Finally, the identified important factors are represented in Table  4.  Using MATLAB software to conduct the feature mapping, the points with an eigenvalue of zero were removed to construct the final projection matrix using the eigenvectors of the remaining points. The dimension reduction processing of the indicator was completed when the data was mapped in a low-dimensional space. The results are shown in Figure 7, the original data was scattered randomly in three-dimensional space, and the ordered data in two-dimensional space was obtained after the dimensionality reduction process. Finally, the identified important factors are represented in Table 4.
The RBF neural network was applied after all the indicators were determined. The processes of adopting the RBF neural network are shown as follows. Using MATLAB software to conduct the feature mapping, the points with an eigenvalue of zero were removed to construct the final projection matrix using the eigenvectors of the remaining points. The dimension reduction processing of the indicator was completed when the data was mapped in a low-dimensional space. The results are shown in Figure 7, the original data was scattered randomly in three-dimensional space, and the ordered data in two-dimensional space was obtained after the dimensionality reduction process. Finally, the identified important factors are represented in Table  4.   Monthly average evaporation rate x 6 7 Earth surface wind speed x 7 8 Earth surface cloud amount x 8 9 Average temperature x 9 10 Relative humidity x 10 11 Total carbon dioxide emissions x 11

Preparation of data
The MATLAB software was adopted to conduct the RBF neural network. The input and output variables for the RBF neural network were the selected 11 indicators. The eligible data of 11 indicators were put into MATLAB. Due to a large amount of data, considering the limitation of computer processing efficiency, the data was first subjected to fast Fourier transform (FFT) processing. Then, the data was standardized [54]. Each indicator belonged to the interval [−1, 1]. The feature vector of the FFT processed sequence was later used as the training sequence. In accordance with successful experience [55] and multiple attempts, in this paper, 60% of the data was used for training, 20% of data was used for testing, and 20% of data was used for validation. It was confirmed that the above proportion has the best training performance for the subsequent processes. Furthermore, the maximum number of training times was set to 1000, the learning precision was 0.01, and the model automatically generated the number of hidden nodes. The maximum threshold of the indicator was 1, and the minimum threshold was −1. First, a pre-experiment was undertaken to determine the appropriate number of neurons. Through several experiments, the number of neurons in the hidden layer was set to 10, R 2 ≥ 0.9, and the performance of the model reached 0.96. When the number of neurons was greater than 10, although the training effect was better in a way, the testing effect was poor. Therefore, it was reasonable to set the number of neurons to 10. The residual and the average residual were required to be less than 0.05, and the number of iterations was 12. The specific structure of the RBF neural network is shown in Figure 8.

Number
Indicator Variable The RBF neural network was applied after all the indicators were determined. The processes of adopting the RBF neural network are shown as follows.

Preparation of data
The MATLAB software was adopted to conduct the RBF neural network. The input and output variables for the RBF neural network were the selected 11 indicators. The eligible data of 11 indicators were put into MATLAB. Due to a large amount of data, considering the limitation of computer processing efficiency, the data was first subjected to fast Fourier transform (FFT) processing. Then, the data was standardized [54]. Each indicator belonged to the interval [−1, 1]. The feature vector of the FFT processed sequence was later used as the training sequence. In accordance with successful experience [55] and multiple attempts, in this paper, 60% of the data was used for training, 20% of data was used for testing, and 20% of data was used for validation. It was confirmed that the above proportion has the best training performance for the subsequent processes. Furthermore, the maximum number of training times was set to 1000, the learning precision was 0.01, and the model automatically generated the number of hidden nodes. The maximum threshold of the indicator was 1, and the minimum threshold was −1. First, a pre-experiment was undertaken to determine the appropriate number of neurons. Through several experiments, the number of neurons in the hidden layer was set to 10, ≥ 0.9, and the performance of the model reached 0.96. When the number of neurons was greater than 10, although the training effect was better in a way, the testing effect was poor. Therefore, it was reasonable to set the number of neurons to 10. The residual and the average residual were required to be less than 0.05, and the number of iterations was 12. The specific structure of the RBF neural network is shown in Figure 8.

Analysis of the error
The root mean square error (RMSE) is a primary index to evaluate the level of error [56]. The RMSE output using MATLAB software, as is indicated in Figure 9, quickly decreases and then tended infinitely toward zero after the second iteration, which suggests that the model worked well, and the model performance was optimal at the sixth iteration (1.0406). Then, the reliability R was output as R ≥ 0.85912, which indicates that the model was very reliable. In the end, after comparing the training data with the pre-retained verification set data, it can be seen that the two data sets were quite close. Hence, it was confidently confirmed that the training results were very stable.

Analysis of the error
The root mean square error (RMSE) is a primary index to evaluate the level of error [56]. The RMSE output using MATLAB software, as is indicated in Figure 9, quickly decreases and then tended infinitely toward zero after the second iteration, which suggests that the model worked well, and the model performance was optimal at the sixth iteration (1.0406). Then, the reliability R was output as R ≥ 0.85912, which indicates that the model was very reliable. In the end, after comparing the training data with the pre-retained verification set data, it can be seen that the two data sets were quite close. Hence, it was confidently confirmed that the training results were very stable.

End training
After the RBF neural network with preset parameters passed the error test, the model operation result could be output. Furthermore, there was a "weight connection one" between the input layer and the hidden layer, and a "weight connection two" between the hidden layer and the model calculation output layer; therefore, these two connections should be handled further. To verify the performance of the RBF neural network, confirmation of the response of the outputted 11 variables was required. As shown in Figure 10, the training, test, and validation sets were close to each other,

End training
After the RBF neural network with preset parameters passed the error test, the model operation result could be output. Furthermore, there was a "weight connection one" between the input layer and the hidden layer, and a "weight connection two" between the hidden layer and the model calculation output layer; therefore, these two connections should be handled further. To verify the performance of the RBF neural network, confirmation of the response of the outputted 11 variables was required. As shown in Figure 10, the training, test, and validation sets were close to each other, and the consistency shown by the three data sets was indicated by a low level of error. Therefore, the results of the operation of the RBF neural network were acceptable, and as such, the training could be terminated.
Symmetry 2020, 12, x FOR PEER REVIEW 16 of 30 and the consistency shown by the three data sets was indicated by a low level of error. Therefore, the results of the operation of the RBF neural network were acceptable, and as such, the training could be terminated.
where is the vth error term: 1 represents the error caused by the Pacific decadal oscillation, 2 represents the instability error caused by the El Niño phenomenon, 3 represents the instability error caused by the La Niña phenomenon, and 4 represents the sum of other possible instability errors.
Since there is no general standard for measuring the magnitude of climate change, the level of climate change needed to be defined after obtaining the CC model. This paper selected 189,742 sample data from different regions of the world from 1800 to 2019 to calculate the climate change value. Then, using statistical techniques for statistical analysis, the climate change values were given corresponding climate levels, as defined in Table 5.  The weight connection one and the weight connection two were collated and analyzed. After calculating some relevant parameters according to Equation (10), the CC model could be obtained as follows: c = 0.1132x 1 − 0.0965x 2 + 0.1031x 3 + 0.0905x 4 + 0.1163x 5 + 0.0814x 6 + 0.0681x 7 +0.0502x 8 + 0.1311x 9 + 0.0171x 10 where ε v is the vth error term: ε 1 represents the error caused by the Pacific decadal oscillation, ε 2 represents the instability error caused by the El Niño phenomenon, ε 3 represents the instability error caused by the La Niña phenomenon, and ε 4 represents the sum of other possible instability errors. Since there is no general standard for measuring the magnitude of climate change, the level of climate change needed to be defined after obtaining the CC model. This paper selected 189,742 sample data from different regions of the world from 1800 to 2019 to calculate the climate change value. Then, using statistical techniques for statistical analysis, the climate change values were given corresponding climate levels, as defined in Table 5. Averaging the known values of all the indicators of 13 stations, they were put into the established climate change model. Consequently, the current climate change level of Canada was calculated to be 0.2241. This indicates the level of Canada's climate change is normal, which is a very intuitive indication given that climate change in Canada is still acceptable.
Through analyzing the CC model, it can be found that the amount of ice coating had a negative effect on climate change, and "the average temperature" and "global carbon dioxide concentration" were significant factors. Their relationship was interpreted under the circumstance of the global climate condition, as indicated in Figure 11, where global temperature and carbon dioxide concentration both showed a significant upward trend from 1880 to 2018. Then, the average temperature of Canada from 1981 to 2019 was considered to obtain the temperature change rate (temperature load) and the average temperature of each region. When the temperature load was more significant than zero, the temperature had an upward trend [57,58]. All temperature loads and average temperatures are visualized on the map of Canada shown in Figure 12. Through analyzing the CC model, it can be found that the amount of ice coating had a negative effect on climate change, and "the average temperature" and "global carbon dioxide concentration" were significant factors. Their relationship was interpreted under the circumstance of the global climate condition, as indicated in Figure 11, where global temperature and carbon dioxide concentration both showed a significant upward trend from 1880 to 2018. Then, the average temperature of Canada from 1981 to 2019 was considered to obtain the temperature change rate (temperature load) and the average temperature of each region. When the temperature load was more significant than zero, the temperature had an upward trend [57,58]. All temperature loads and average temperatures are visualized on the map of Canada shown in Figure 12.   1888  1892  1896  1900  1904  1908  1912  1916  1920  1924  1928  1932  1936  1940  1944  1948  1952  1956  1960  1964  1968  1972  1976  1980  1984  1988  1992  1996  2000  2004  2008  2012  2016 The global temperature The global temperature from 1880 to 2018 In Figure 12, the temperature loads of the research areas were greater than zero. The normal state of Canada's climate change is a microcosm of global climate change which indicates the positive correlation between temperature and carbon dioxide. Furthermore, the area with the most significant temperature load was NU (0.0037), and the temperature loads of the southeast coastal areas were all great than zero, indicating that the region will continue to experience warming in the future. In the area of the western coast, the temperature of BC was abnormally high, which may be related to the occurrence of the El Niño and La Niña phenomena in recent years. Furthermore, the fluctuation of the temperature load was significant, which also showed a clear trend of upward fluctuation in threedimensional space.
Also, the contribution of sea surface temperature (SST) to climate change was quite large [59]. After obtaining the data of SST (data provided by the NOAA/OAR/ESRL PSD, Boulder, Colorado, USA, from their website at https://www.esrl.noaa.gov/psd/), the temperature profile of the SST worldwide from 1800 to 2019 was produced. As indicated in Figure 13, the larger the yellow area in the figure, the wider the coverage of the SST high-temperature range. Before 1880, the hightemperature center of the SST was concentrated around the area of the eastern Pacific Ocean, and then the central area of the high temperature continuously expanded [34]. The area with the highest SST was concentrated in the central area of the Pacific Ocean and the Indian Ocean. In the eastern part of Africa and the southern part of Asia and the Indian Ocean, the average SST has increased from [22,27] to [27,30]. The increase in SST near the west coast of Canada was consistent with the rise In Figure 12, the temperature loads of the research areas were greater than zero. The normal state of Canada's climate change is a microcosm of global climate change which indicates the positive correlation between temperature and carbon dioxide. Furthermore, the area with the most significant temperature load was NU (0.0037), and the temperature loads of the southeast coastal areas were all great than zero, indicating that the region will continue to experience warming in the future. In the area of the western coast, the temperature of BC was abnormally high, which may be related to the occurrence of the El Niño and La Niña phenomena in recent years. Furthermore, the fluctuation of the temperature load was significant, which also showed a clear trend of upward fluctuation in three-dimensional space.
Also, the contribution of sea surface temperature (SST) to climate change was quite large [59]. After obtaining the data of SST (data provided by the NOAA/OAR/ESRL PSD, Boulder, Colorado, USA, from their website at https://www.esrl.noaa.gov/psd/), the temperature profile of the SST worldwide from 1800 to 2019 was produced. As indicated in Figure 13, the larger the yellow area in the figure, the wider the coverage of the SST high-temperature range. Before 1880, the high-temperature center of the SST was concentrated around the area of the eastern Pacific Ocean, and then the central area of the high temperature continuously expanded [34]. The area with the highest SST was concentrated in the central area of the Pacific Ocean and the Indian Ocean. In the eastern part of Africa and the southern part of Asia and the Indian Ocean, the average SST has increased from [22,27] to [27,30]. The increase in SST near the west coast of Canada was consistent with the rise in the average temperature in the coastal area, thereby the contribution of SST to climate change cannot be neglected.
Symmetry 2020, 12, x FOR PEER REVIEW 19 of 30 in the average temperature in the coastal area, thereby the contribution of SST to climate change cannot be neglected.

Predictions of the CC Model
A prediction of the possible trend of climate change can provide more useful guiding information, and it can help the public, government, and policymakers have a holistic concept of climate change. Meanwhile, to verify the predicted performance of the established model, two different predictions were conducted, and the prediction differences were compared in detail.
First, the CC model was used to obtain the climate change values from 1800 to 2019. was used as the input variable of the RBF neural network. The maximum number of training times was set to 500 times, the learning precision was 0.01, and the model automatically generated the number of hidden nodes. The prediction time was from 2019 to 2044 (the next 25 years). The maximum and minimum thresholds of the indicator were 1 and −1, respectively. After several pre-experiments, the number of neurons was set to 7 and the number of iterations was 9. When the residual and the average residual was less than 0.05, the training was terminated; the resulting prediction result is called "Prediction 1." Second, the original data set was divided as a test set, a training set, and a verification set. The test set was processed by the RBF neural network directly. The maximum number of training times was set to 1000, the learning precision was 0.01, the model automatically generated the number of hidden nodes, and the forecast time was the next 25 years. The maximum and minimum thresholds of the indicator were 1 and −1, respectively. The number of hidden layer neurons and iterations were both set to 10, and the performance of the model was the best (1.002) under the above settings. When the average residual was less than 0.05, the training was terminated. The RBF neural network was used to output the initial predicted values, then the CC model was applied to process the above values to output the final predicted results ("Prediction 2"). The results of the two predictions are shown in Figure 14.

Predictions of the CC Model
A prediction of the possible trend of climate change can provide more useful guiding information, and it can help the public, government, and policymakers have a holistic concept of climate change. Meanwhile, to verify the predicted performance of the established model, two different predictions were conducted, and the prediction differences were compared in detail.
First, the CC model was used to obtain the climate change values C from 1800 to 2019. C was used as the input variable of the RBF neural network. The maximum number of training times was set to 500 times, the learning precision was 0.01, and the model automatically generated the number of hidden nodes. The prediction time was from 2019 to 2044 (the next 25 years). The maximum and minimum thresholds of the indicator were 1 and −1, respectively. After several pre-experiments, the number of neurons was set to 7 and the number of iterations was 9. When the residual and the average residual was less than 0.05, the training was terminated; the resulting prediction result is called "Prediction 1." Second, the original data set was divided as a test set, a training set, and a verification set. The test set was processed by the RBF neural network directly. The maximum number of training times was set to 1000, the learning precision was 0.01, the model automatically generated the number of hidden nodes, and the forecast time was the next 25 years. The maximum and minimum thresholds of the indicator were 1 and −1, respectively. The number of hidden layer neurons and iterations were both set to 10, and the performance of the model was the best (1.002) under the above settings. When the average residual was less than 0.05, the training was terminated. The RBF neural network was used to output the initial predicted values, then the CC model was applied to process the above values to output the final predicted results ("Prediction 2"). The results of the two predictions are shown in Figure 14. It can be seen that the climate level will gradually transfer from "normal" to "a little bad." In 2025, the global climate level will break through the normal level, and the future climate change will develop in the direction of "a little bad" in the long term. Additionally, the forecast result of "Prediction 1" was close to that of "Prediction 2", indicating that the CC model was reliable and had a good applicability. In 2053, "Prediction 1" shows a downward trend, while "Prediction 2" shows a sudden upward trend. The possible reason for this phenomenon is the scale effect caused by the data. Thereby the CC model's prediction stability remained for 34 years. Since climate change is a quite complex phenomenon, through the limited amount of data, there will be numerous uncertainties in predicting changes over decades. The subsequent optimization of the model can be conducted by improving the stable prediction performance of the model in the long term.

The Extreme Weather Model
In this study, we used a SVM to construct the EW model. The modelling process was divided into the following parts.   It can be seen that the climate level will gradually transfer from "normal" to "a little bad." In 2025, the global climate level will break through the normal level, and the future climate change will develop in the direction of "a little bad" in the long term. Additionally, the forecast result of "Prediction 1" was close to that of "Prediction 2", indicating that the CC model was reliable and had a good applicability. In 2053, "Prediction 1" shows a downward trend, while "Prediction 2" shows a sudden upward trend. The possible reason for this phenomenon is the scale effect caused by the data. Thereby the CC model's prediction stability remained for 34 years. Since climate change is a quite complex phenomenon, through the limited amount of data, there will be numerous uncertainties in predicting changes over decades. The subsequent optimization of the model can be conducted by improving the stable prediction performance of the model in the long term.

The Extreme Weather Model
In this study, we used a SVM to construct the EW model. The modelling process was divided into the following parts. It can be seen that the climate level will gradually transfer from "normal" to "a little bad." In 2025, the global climate level will break through the normal level, and the future climate change will develop in the direction of "a little bad" in the long term. Additionally, the forecast result of "Prediction 1" was close to that of "Prediction 2", indicating that the CC model was reliable and had a good applicability. In 2053, "Prediction 1" shows a downward trend, while "Prediction 2" shows a sudden upward trend. The possible reason for this phenomenon is the scale effect caused by the data. Thereby the CC model's prediction stability remained for 34 years. Since climate change is a quite complex phenomenon, through the limited amount of data, there will be numerous uncertainties in predicting changes over decades. The subsequent optimization of the model can be conducted by improving the stable prediction performance of the model in the long term.

The Extreme Weather Model
In this study, we used a SVM to construct the EW model. The modelling process was divided into the following parts.   (2) Generate a scatter plot Since the values of a single category were difficult to classify and did not have a general law, the index was projected to [−2, 6] under each category. Figure 16 shows that the red reference points were very close to the green reference points. In order to facilitate the division and avoid the problem that the classification boundary was not clear due to a large amount of data, this study only generated ten regular distribution data points based on each category. The length of each category indicator was 244 × 12 months = 2928 months. (2) Generate a scatter plot Since the values of a single category were difficult to classify and did not have a general law, the index was projected to [−2, 6] under each category. Figure 16 shows that the red reference points were very close to the green reference points. In order to facilitate the division and avoid the problem that the classification boundary was not clear due to a large amount of data, this study only generated ten regular distribution data points based on each category. The length of each category indicator was 244 × 12 months = 2928 months. (3) Mark the category A vector file was created to mark the class of each point as a label for each vector. The vector took a value of −1 or 1, indicating that the class belonged to precipitation or temperature.

(4) Cross-validation
Cross-validation means that most of the samples were modelled, and a small number of samples were used to predict the established model and to solve the forecast error of a small number of samples [60]. To reduce the error of cross-validation, Bayesian optimization was used to reduce the loss due to cross-validation.  (3) Mark the category A vector file was created to mark the class of each point as a label for each vector. The vector took a value of −1 or 1, indicating that the class belonged to precipitation or temperature.

(4) Cross-validation
Cross-validation means that most of the samples were modelled, and a small number of samples were used to predict the established model and to solve the forecast error of a small number of samples [60]. To reduce the error of cross-validation, Bayesian optimization was used to reduce the loss due to cross-validation. After the processing of SVM, all indicators could be divided into two categories: precipitation indexes and temperature indexes, which are the red and green points shown in Figure 18, respectively. Through the classification, a clear division of the indicator attributes was prepared for After the processing of SVM, all indicators could be divided into two categories: precipitation indexes and temperature indexes, which are the red and green points shown in Figure 18, respectively. Through the classification, a clear division of the indicator attributes was prepared for SVM regression processing.

(6) Classify indicators
After the processing of SVM, all indicators could be divided into two categories: precipitation indexes and temperature indexes, which are the red and green points shown in Figure 18, respectively. Through the classification, a clear division of the indicator attributes was prepared for SVM regression processing. 2. The regression of the SVM This paper proposes a calculation method based on the extreme weather fluctuation value (JD) in the unit grid to reflect extreme weather fluctuations. The value of JD is between 0 and 1. The closer the JD value is to 1, the greater the weather fluctuations and the more extreme the weather conditions. The definition of JD is as follows: × e 0.0001i + δ 1 , where h i is the original value; y i is the processed value; max {h i } and min{h i } are the maximum and minimum values of the indicator data set under the category, respectively; y i indicates the normalized result of all indicators; JD 1 indicates the temperature indexes; JD 2 indicates the precipitation indexes; n indicates the total number; σ indicates the standard deviation of the indicator; and e 0.0001i represents a time-varying exponent term. δ 1 represents the disturbance term of the temperature and δ 2 represents the disturbance term of the precipitation, both of which are regarded as normally distributed random noise. Using the processed indicators as inputs, with an output value of JD, the indicator data set from 1800 to 2009 was selected as the training sample set and the indicator data set from 2010 to 2019 was used as the verification set. The linear ε insensitive loss function was selected for controlling the error minimization, and the radial basis kernel function was applied in the SVM regression. Finally, the regression parameters of SVM could be obtained. The extreme temperature model and the extreme precipitation model are shown below: y 1 = 0.0176h 1 − 0.0361h 2 + 0.1073h 3 + 0.0144h 4 + 0.1334h 5 + 0.1621h 6 + 0.105h 7 + δ 1 , y 2 = 0.0126h 8 + 0.1763h 9 + 0.1019h 10 + 0.0104h 11 + 0.0133h 12 + 0.1327h 13 where δ 1 is the sum of possible noise associated with temperature and δ 2 is the sum of possible noise related to precipitation. Furthermore, ample data was calculated to obtain the distribution of the extreme temperature values and extreme precipitation values. Using statistical analysis, the interval classifications of the extreme temperature indicators and the extreme precipitation indicators were found and are shown in Table 6.

Correlation Analysis
The occurrence of a climate change hiatus is confusing and creates the illusion that climate change has stopped [61]. In January 2019, two-thirds of the United States and Canada had large hail events, and there was unusual extreme cold weather. Hence, many people held a view that global warming has stopped and the government does not have to spend too much energy and expenses on global-warming-related work [62,63]. In order to help the public better understand the relationship between global warming and local extreme cold weather, this study selected the Selkirk Fortress region of Canada as an example to analyze its correlation. The reason for choosing the Selkirk Fortress region is that it has obvious characteristics of extreme weather, and the data for climate change and extreme weather of the Selkirk Fortress region is available and complete. Through the above discussion, the CC model and the EW model were related to different factors. The key interface was the time when analyzing its correlation. The study period was from 1990 to 2018, and the statistics of h 1 -h 7 from September to January in the Selkirk Fortress area were selected.
Through applying the CC model and the EW model, the climate change values and extreme cold weather values could be obtained. Then, the Pearson correlation value ρ CT . was chosen to judge the relevance relationship, where −1 ≤ ρ CT ≤ 1. When ρ CT is positive, it indicates that there is a positive correlation. Otherwise, it suggests that there is a negative correlation [64]. The magnitude of the correlation is presented in Table 7. Due to the large amount of data, this paper used the SPSS 17.0 to calculate the correlation value. After importing the obtained values into the SPSS software, the value of ρ CT was found to be 0.571, which indicates that there was a moderate positive correlation. This means that when local extreme cold weather occurs frequently, climate change will become active; as a result, the extreme cold weather will become more serious, and the climate change value will increase as well. It worth noting that the typical manifestation of progressive climate change is global warming. Therefore, while the extreme local cold occurs, global warming may occur; therefore, the local extreme cold and global warming are not contradictory. The public should have confidence in the government's policy decisions and make their contributions through energy conservation and emissions reductions throughout their daily lives. In the context of abnormal global climate change, focusing on long-term benefits is far more important than short-term benefits, which inevitably requires the joint efforts of people all over the world.

Discussion
The work surrounding climate change and extreme weather is quite complex and tough, which is because climate and weather are the result of a combination of spatial factors at a particular time. Climate impacts have been identified, but their magnitude is still vague. This research aimed to provide reliable evidence for the impact of current climate change and extreme weather, and provides the public with a dialectical perspective on the current climate and weather conditions. The work in this research was conducted by establishing climate change and extreme weather models. As is indicated in Figure 19, the research could not only evaluate current conditions, but also predicted the change of future climate change. Furthermore, the relationship between climate change and extreme weather were recognized through the scientific method. In the case study, Canada was selected as our research object. The work was divided into three parts: (1) constructing the climate change model, then making a prediction; (2) constructing the extreme weather model; and (3) conducting a correlation analysis. When constructing CC model, Stata and MATLAB software were used to obtain the corresponding data, and LM was used to determine 11 significant indicators. Then, the CC model was established on the basis of an RBF neural network. It was found that the current climate change level of Canada is 0.2241, which indicates that In the case study, Canada was selected as our research object. The work was divided into three parts: (1) constructing the climate change model, then making a prediction; (2) constructing the extreme weather model; and (3) conducting a correlation analysis. When constructing CC model, Stata and MATLAB software were used to obtain the corresponding data, and LM was used to determine 11 significant indicators. Then, the CC model was established on the basis of an RBF neural network. It was found that the current climate change level of Canada is 0.2241, which indicates that the level of Canada's climate change is normal on the whole. Meanwhile, the amount of ice coating had a negative effect on climate change, and "the average temperature" and "global carbon dioxide concentration" were quite significant factors, and the contribution of SST was recognized as well. Furthermore, a prediction of the possible trend of climate change can provide more useful guidance information, and considering differences in the model performance, two prediction methods were conducted at the same time, which were from the CC model and RBF neural network. It was found that the global climate level will break through the normal level in 2025 and the future climate change will develop in the direction of "a little bad" in the long term. Additionally, the forecast result indicated that the CC model is reliable and has good applicability. However, the prediction ability of the CC model is up to 34 years. When constructing the extreme weather model, due to the local and targeted characteristics of the indicators in the EW model, reliable precipitation and temperature indicators were extracted from prior literature. Then, the EW model is constructed on the basis of a SVM and the JD algorithm. In the final correlation analysis, the Selkirk Fortress region of Canada was selected as the research area due to its obvious extreme characteristics. After extracting corresponding data of its indicators from 1990 to 2018, the Pearson correlation value was selected to measure the correlation between climate change and extreme weather. The results showed that there was a moderate positive correlation; therefore, even if local extreme weather is frequent, global warming will continue as well. The above result is also strong evidence that can explain the apparent contradiction between local extreme cold and global warming. As the above results are studied from the perspective of a macro average, if the model is applied to a narrower area, the Pearson value will be larger.
In fact, some similar machine learning models have already been adopted in researching the issues regarding climate change. For example, support vector machines (SVMs), artificial neural networks (ANNs), and generalized regression neural network (GRNN) have been applied to assess the impact of climate change on the above-ground biomass. However, due to the existence of systematic errors and the lack of consideration of time series, the uncertainty of the model is large [65]. Another study adopted an ANN method to determine the influence of climate drivers on sand-deposition in semi-arid regions [66]. ANN was also adopted to provide globally consistent estimates of precipitation [67]. However, the shortcoming of the above research is that the output model may be less accurate due to insufficient training times. The biggest difference between this study and other studies lies in the following three points: (1) The model is more stable. The RBF neural network can rapidly improve the performance of the model according to the number of training times. The prediction ability of the CC model based on the RBF neural network and LM can perform better. Furthermore, the model established in this paper can directly extract the attributes of the original data and give feedback information, and the computer operation efficiency of the CC and EW models were significantly improved, and the inefficient and high-error processing status caused by the large amount of data was solved. (2) The model considers the time series of the data. The data used in this study was the complete data set with a long time series to make full use of the original attributes of the data and avoid the occurrence of data distortion caused by multi-model processing. (3) This research is more comprehensive. All possible factors affecting climate change and extreme weather were considered in the research, instead of focusing on at most a few factors. Therefore, the macro guidance of this research is more significant.
The CC and EW models were established on the basis of relevant machine learning theory, which was quite an innovative method. The forecasting time scale of the CC model was found to be 34 years, which is realizable and stable in a way. The main reason for the limitation was the variability of the climate change system and the scale effect of the data. Future scholars can optimize CC models through other neural network theories. Furthermore, it should be mentioned that the research in this article was based on the macro average state, so it may be inconsistent with local special cases, where the large research area and scattered data are the most likely reasons. To improve the local application performance of the model, future scholars can improve the model performance by continuously optimizing the parameters.

Conclusions
This research aimed to help the public better understand the current situation of climate change and provide useful information for the government. The research was conducted by constructing the CC and EW models. In the research area of Canada, the CC model was constructed using LM and a RBF neural network, and the current climate change level of Canada was calculated to be 0.2241, which indicates that the overall level of climate change of Canada is still normal. However, the climate change level of Canada will transfer from "normal" to "a little bad" in 2025 on the basis of the prediction result. Furthermore, it was found that the CO 2 level, average temperature, and SST were significant factors to climate change. In order to figure out the relationship between climate change and extreme weather, the EW model was established using an SVM. Additionally, the EW model could be divided into a TM model and PM model. Then, the Pearson correlation analysis was applied, and its value was calculated to be 0.571, which confirmed that there was a moderate positive correlation. This research provides guidance for government policy formulation and correct public awareness, which is not only conducive to the government's better allocation of fiscal expenditure through more reasonable policies, but also helps the public to make their own efforts through having a correct perception. Climate change can only be solved through the joint efforts of the government and the public.
However, there are some areas where the above established models can be improved. The prediction accuracy of the CC model lasts for 34 years. Beyond this predicted time threshold, the model exhibited instability. In addition, when establishing the extreme weather model, it was assumed that the extreme temperature model and the extreme precipitation model were independent of each other. Actually, there are complex nonlinear relationships, which may have a certain impact on the accurate application of the model. Future scholars can optimize the model by improving the prediction stability of the model in the long term and process the temperature and precipitation data directly to obtain a more accurate mathematical model. This paper provides a new perspective for the public to understand climate change and extreme weather while providing a useful reference for governments and relevant policymakers.