Uncertainty Quantification of Material Properties in Ballistic Impact of Magnesium Alloys

The design and development of cutting-edge light materials for extreme conditions including high-speed impact remains a continuing and significant challenge in spite of steady advances. Magnesium (Mg) and its alloys have gained much attention, due to their high strength-to-weight ratio and potential of further improvements in material properties such as strength and ductility. In this paper, a recently developed computational framework is adopted to quantify the effects of material uncertainties on the ballistic performance of Mg alloys. The framework is able to determine the largest deviation in the performance measure resulting from a finite variation in the corresponding material properties. It can also provide rigorous upper bounds on the probability of failure using known information about uncertainties and the system, and then conservative safety design and certification can be achieved. This work specifically focuses on AZ31B Mg alloys, and it is assumed that the material is well-characterized by the Johnson–Cook constitutive and failure models, but the model parameters are uncertain. The ordering of uncertainty contributions for model parameters and the corresponding behavior regimes where those parameters play a crucial role are determined. Finally, it is shown that how this ordering provides insight on the improvement of ballistic performance and the development of new material models for Mg alloys.


Introduction
The assessment and design of armor plates for protecting humans and vehicles against high-speed impact has long been of interest in military and aerospace applications. Among a large number of protection materials, magnesium (Mg) and its alloys have gained much attention, due to their high strength-to-weight ratio and potential of further improvements in material properties such as strength and ductility [1][2][3]. The density of Mg is approximately 35% lower than that of aluminum and approximately 77% lower than that of steel. Therefore, Mg alloys are the lightest metallic material that have high potential for weight reduction, thereby decreasing the amount of fuel used in military and aerospace applications [4,5]. However, compared with other conventional materials, such as aluminum and steel, fewer studies have been performed on the relationship between material properties of Mg alloys and its ballistic response, particularly under extreme conditions, i.e., high strain-rates and temperatures [6][7][8].
The ballistic impact of materials and structures is characterized by a very complex mechanical-thermal coupled process which mainly depends on the material properties and geometrical parameters of the target and the projectile, e.g., strength, toughness, shape and size [9][10][11]. These extreme complexities, which involve high non-linearity, singularity and dependence on a large number of parameters, render it impossible to derive a closed-form analytical solution. As a result, numerical methods, e.g., finite element methods, have attracted more attention and been employed extensively in the modeling of ballistic impact problems [12][13][14]. In the numerical modeling of ballistics, the responses of the materials to mechanical and thermal loading conditions are always described by constitutive and failure models, which supplies a stress-strain relation at multiple temperatures and strain rates in order to formulate governing equations along with the kinematic and conservation laws. These constitutive and failure models are usually empirical , representing the primary link between experimental inputs and predicted outputs and hence constituting the strongest source of physical fidelity in a given calculation. The data used to calibrate material models can be achieved either traditionally by laboratory experiments for simple specimens such as split Hopkinson bar [15,16], or by sub-grid scale simulations such as crystal plasticity [17,18]. Regardless of the source, analysts are most commonly restricted to a limited set of model forms, either by their simulation tool of choice or by the significant effort required to formulate, implement and characterize a new model. In practice, this restriction limits how well a fixed set of material parameters can fully represent a broad range of complex constitutive and failure behaviors. Therefore, the model parameters, characterizing multiple types of material properties, must be allowed to vary over a certain range, due to many factors including the potential complexity of response, any stochastic response characteristics, and the paucity of experimental data. These uncertainties render deterministic analysis of limited value. Instead, it becomes necessary to estimate the likely spread of performance metrics and relevant design requirement in order to provide an adequate design margin and meet specifications with sufficient confidence in the modeling of ballistic problems.
Within computational science, uncertainty quantification (UQ) is a family of powerful solution strategies that aim to characterize the variability of a given analysis and the spread in the predicted performance of a system [19][20][21][22][23]. The work presented in this paper focuses on systems in which the main source of uncertainty is an imperfect knowledge of material properties, as described by parameterized constitutive and failure models. The specific approach by which uncertainties are quantified is through the largest deviation in the performance measure resulting from a finite variation in the corresponding material properties, due to McDiarmid [24] belong to a general class known as concentration-ofmeasure (CoM) inequalities [25]. Upper bounds of the probability that the system fails to perform within the design margin are also determined. Such bounds are rigorous, i.e., they are sure to be conservative and result in safe designs. These bounds are also optimal if all the known information about uncertainties and system is leveraged using the optimal uncertainty quantification (OUQ) strategy [26]. This characterization of systems using CoM and OUQ requires only knowledge of limiting ranges of the input variables, and not their full probability distribution as is the case of Bayesian methods. The upper bound performance characteristics are computed by exercising an existing deterministic code in order to sample the mean response of the system and to calculate the largest deviation in output in order to identify worst-case combinations of parameters. The CoM and OUQ strategies have been used in various applications including design of a thermalhydraulic reactor [27], design of a fractal electrical circuit [28], ballistic impact of alumnium alloys [29][30][31] and sub-ballistic impact of Mg alloys [32,33].
The AZ31B type of Mg alloys is chosen in this study since AZ31B is widely used in aerospace and automotive applications [34][35][36][37]. The goal of this work is to investigate how the uncertainties in the constitutive and failure properties of AZ31B affect its ballistic performance when the material is subject to high-speed impact, and then to provide insight on how to improve such performance and develop new material models for AZ31B. For simplicity, it is specifically hypothesized that the material behavior of AZ31B is known to be well-described by the Johnson-Cook constitutive and failure models [38,39], but the corresponding material constants are only imperfectly characterized and within speficic ranges. Then, both the resulting uncertainty in the performance and the mean performance measure are computed using the DAKOTA Version 6.7 software package [40] of the Sandia National Laboratories. For all simulation parameters not considered as random variables, the values are considered as specified simulation conditions for evaluation. This strategy of fixing the boundary and initial conditions for assessment is entirely analogous to design testing for impact resistance [41], wherein performance is evaluated relative to a targeted set of pre-characterized impact conditions. Simulations of such conditions are carried out over a range of impact velocities using the commercial finite-element package LS-DYNA [42] on a single converged mesh. The ballistic tests conform to the form of the Recht-Ipson model, and the material failure mechanisms for AZ31B are noteworthy including spalling, plugging, discing and fragmentation.
The remainder of the paper is structured as follows. In Section 2, this paper starts by reviewing the CoM inequality, the OUQ approach and the corresponding probability bounds used for purposes of UQ. Section 3 proceeds to investigate the effects of material uncertainties on the ballistic performance of AZ31B Mg alloys subject to high-speed impact. A conclusion with a summary and short discussion is given in Section 4.

Methodology
For the sake of completeness and convenience, this section briefly summarizes the CoM and OUQ theories for rigorous UQ and conservative safe design. Additional details can be found in Refs. [26,32,43].

Concentration-of-Measure (CoM) Inequality
This paper considers a system characterized by N real-valued random variables X ≡ (X 1 , . . . , X N ) ∈ R N and a single real-valued performance measure Y ∈ R. The values of the input random variables lie within intervals I ≡ (I 1 , . . . , I N ), i.e., x 1 ∈ I 1 , . . . x N ∈ I N . This paper begins by supposing that the system performance can be described by a deterministic response function F : R N → R through either experiments or an exact model. It is further assumed that the system fails if Y ≥ Y c where Y c is a threshold value for the safe operation of the system, and that the expected system performance E[Y] is known exactly. Then, a direct application of McDiarmid's concentration-of-measure (CoM) inequality [24] provides an upper bound on the probability-of-failure of the system [32] where is the system diameter and x + := max(x, 0). In Equation (2), D i denotes the sub-diameter corresponding to the input variable X i , which is calculated through the optimization problem The preceding methodology can be extended to the case in which the exact mean performance E[Y] is not available and the mean performance must be estimated instead. To this end, suppose that n evaluations of the model F(X) are conducted based on unbiased sampling of the random input variables, resulting in predicted performance measures y 1 , y 2 , . . . , y n . Then an empirical mean performance can be defined as Lucas et al. [43] showed that the probability of failure P[Y ≥ Y c ] can be determined to within confidence intervals by considering the randomness of the estimated mean Y , with the result where denotes a pre-specified tolerance for the mean estimation and D is the same as in Equation (2). In Equation (6), α characterizes the effect of estimating mean performance, which has a form of Another equivalent expression of Equation (6) is that, with a probability that is greater which also supplies an upper bound on the probability of failure for the scenario of estimating mean performance. With these identifications, a conservative and rigorous design criterion can be achieved through requiring that this upper bound on the probability of failure less than a tolerance, with the result where M = Y c − Y − α + measures the design margin and U = D provides an unambiguous definition and measure of uncertainty. The ratio CF of margin to uncertainty measures the confidence that can be placed on the design as is referred to as confidence factor. The design criterion Equation (9) simply requires that the confidence in the design, as measured by the confidence factor, be greater than a minimum value.

Optimal Uncertainty Quantification (OUQ)
The McDiarmid's approach for UQ in Equations (1) and (8) is attractive because it requires limited but tractable information on input variables (i.e., independence and intervals), response functions (i.e., sub-diameters) and performance measures (i.e., mean response). A question of theoretical and practical importance concerns whether it is possible to obtain an optimal bounds on the probability of failure using the same given information. Other related questions concerns the possibility of using other information than subdiameters and mean output. These questions have been addressed by Owhadi et al. [26], and this paper proceeds to summarize their main results for completeness. Assume that the safety of a system needs to be certified and the criterion is given by based on the information that X ≡ (X 1 , . . . , X N ), X 1 , . . . , X N are independent, X ∈ I and that sup |F( As a result, the optimal bound U (A MD ) on the probability of failure P[F(X) ≥ Y c ] is the solution of the following optimization problem where and M(I i ) denotes the set of probability measures on I i .
In practical applications, the available information does not determine a unique solution of (G, P) but instead provides an information set A. This set applies constraints on (G, P) and hence consists of all possible values of (G, P). As a result, the optimal uncertainty quantification (OUQ) (Owhadi et al. [26]) aims to find the optimal bounds on probabilities given such set of information about the uncertainties. These bounds are calculated as extreme values of well-defined optimization problems corresponding to extremizing probabilities of deviation subject to the constraints imposed by the known information. As a result, McDiarmid's concentration-of-measure approach in Equations (1) and (8) provides an upper bound on U (A MD ).
Although the optimization problem Equation (11) needs to be solved in infinitedimensional spaces of measures and functions and therefore is extremely large, under general moment and independence conditions, Owhadi et al. [26] have shown that they have finite-dimensional reductions. An application of OUQ that is relevant to the present work concerns the development of explicit and optimal concentration inequalities of the McDiarmid's type. Namely, considering the information given in Equation (12) and assuming the optimal bound is given by Owhadi et al. [26] is also referred to for detailed derivation of Equation (13). The resultant optimal bounds provide a means of improving on the simple McDiarmid's bounds that are taken as the basis for the present work. Since the bounds are optimal, further improvements inevitably require information other than or in addition to system sub-diameters and mean performance.

Numerical Experiments
This paper now proceeds to quantify the uncertainties of the constitutive and failure properties in the modeling of a ballistic impact problem, using the UQ strategies described in the foregoing. Specifically, the target is an AZ31B Mg alloy plate and the projectile is a steel ball, as shown in Figure 1a. The residual velocity of the projectile is considered as the objective of interest in the numerical experiments, which can be considered as a metric to measure the performance of the plate subject to high-speed impact. Figure 1b visualizes the system after penetration. It is assumed that all uncertainties arise from an imperfect characterization of the mechanical response of the plate. As a simple scenario, this paper further assumes that, under the conditions of interest, the plate is well-characterized by the Johnson-Cook plasticity and fracture models [38,39], but the parameters of the two models are uncertain. Specifically, they must be allowed to vary over certain ranges in order to cover the experimental data with prescribed probability. For simplicity, the projectile is assumed to be uncertainty-free.

Material Modeling
It is assumed that the constitutive behavior of the plate is characterized by an appropriately calibrated Johnson-Cook plasticity model [38], where σ is the true Mises stress, p is the equivalent plastic strain,˙ p is the plastic strain rate, and T is the temperature. The normalized plastic strain rate˙ * p is defined aṡ * where˙ p0 is a reference strain rate. The model also uses the normalized temperature where T 0 is a reference temperature and T m is the melting temperature. The model parameters are: A, the yield stress; B, the strain-hardening modulus; n, the strain-hardening exponent; C, the strengthening coefficient of strain rate; and m, the thermal-softening exponent.
It is further assumed that the failure behavior of the plate is well-described by the Johnson-Cook fracture model [39]. Specifically, the damage of an element is defined on a cumulative damage parameter where the summation is conducted over time steps and ∆ is the plastic strain increment in each time step. f denotes the strain at fracture which is given by where E 1 , E 2 , E 3 , E 4 and E 5 are material damage constants. p * and T * are defined in Equations (15) and 16, respectively. σ * is the ratio of the pressure p divided by the von-Mises equivalent stress, i.e., where σ 1 , σ 2 and σ 3 are the principle stresses. Based on this damage model, fracture takes place when the damage parameter E reaches the value of 1.
The set X ≡ (A, B, n, C, m, E 1 , E 2 , E 3 , E 4 , E 5 ) of Johnson-Cook plasticity and damage parameters are regarded as the main source of uncertainty in the analysis. The estimated values and the bounds are tabulated in Table 1. In practice, the material parameters are derived from specific data sources, e.g., experiments or sub-grid simulations. It is assumed that these data are sufficient to determine confidence intervals for each parameter. In our calculations, this paper specifically uses the AZ31B Mg alloy characterization of Hasenpouth [44], which, conveniently, includes the lower and upper bound of the 95% confidence intervals for Johnson-Cook plasticity parameters, i.e., A, B, n, C and m. Specifically, the experiments are conducted using partially hardened AZ31B sheets in the rolling direction. All the samples are of the "miniature dog-bone" shape with a gauge length of 12.5 mm, a height of 1.75 mm and a thickness of 1.6 mm. The intervals of the plasticity parameters are determined by fitting the 10 stress-strain curves at the strain rates in the range of 0.003 to 1500 s −1 and the temperatures in the range of 20 to 300 • C. In addition, regarding the Johnson-Cook damage parameters, the estimated values are provided by Feng et al. [45]. Similar to the plasticity parameters, the bounds of the damage parameters should also be determined by tuning them to bracket the failure strains at different strain rates and temperatures with prescribed confidence. It would be also expected that some parameters have a higher uncertainty, like what happens to the plasticity parameters. However, their bounds of confidence intervals are scarce in the literature. To this end, the similar level of uncertainty, i.e., 10.0%, is added to the estimated values to generate the lower and upper bounds. The level is set the same for all the damage parameters for the sake of comparison of uncertainty contribution. This uncertainty level is also very close to the average uncertainty over the five plasticity parameters, and therefore might provide a good estimation on the imperfect knowledge of damage properties. The resultant bounds are shown in Table 1.

Forward Solver
For a given realization of the system parameters, the ballstic imapct problem is solved using the explicit dynamics solver available within the commercial finite element analysis software package LS-DYNA [42]. The initial conditions of the computational model are shown in Figure 1a. The plate is resolved using 185,600 elements, while the number of elements for the projectile is 3584. All the elements are linear hex, single point integration with careful hourglass control. Elements are refined in the impact region of the plate. The size of the elements in the refined region of the plate is 0.25 mm in all the length, width and height directions. The size of the elements in the projectile is in the range of 0.17 to 0.5 mm. This number and distribution of elements are enough to make the calculations converged. The projectile is assumed to be rigid. As a result, all the solid hex elements in the projectile are turned into a rigid body. All the nodes in the projectile are given the impact velocity as the initial conditions, and the projectile has no contact with the plate before the simulations start. No constraint is applied on the plate. All simulations were ran for 30 µs before termination. This simulation duration is sufficiently long to allow for either the penetration of the projectile through the plate, or the rebound and separation of the projectile from the plate. The time-step size is adaptive and determined by the critical size of elements. Element erosion is used to characterize material failure. Additionally, the calculations are adiabatic with the initial temperature set at room temperature. The equation-of-state, which controls the volumetric response of the material, is assumed to be of the Gruneisen type. The values of the fixed parameters of the plate and the projectile used in the calculations are tabulated in Tables 2 and 3, respectively.

Ballistic Behavior
This subsection begins by evaluating the ballistic property of the plate subject to normal impact, using one representative group of the Johnson-Cook plasticity and damage parameters listed in Table 1. Figure 2 shows the residual velocity of the projectile, as a function of impact velocity ranging from 400 m/s to 1000 m/s. In order to estimate the ballistic limit, the Recht-Ipson model [50] is employed, which provides an analytical expression based on the conversation of energy and momentum. Specifically, the model has the form of where a, p and v bl are fitting parameters. Specifically, v bl represents the ballistic limit of the plate. The Recht-Ipson model is fitted to the residual perforation velocities calculated by LS-DYNA, and the values of the fitting parameters are tabulated in Table 4. For comparison, the resulting Recht-Ipson curve is also plotted in Figure 2. It is notable that the LS-DYNA results conform to the form of the Recht-Ipson model. The coefficient of determination R 2 is extremely close to 1, also showing that the Recht-Ipson predictions well fit the LS-DYNA data. The predicted value of the ballistic limit by LS-DYNA calculations is 696.9 m/s. Moreover, although the parameter p is fitted to the numerical data, its value is very close to that given in the original Recht-Ipson model (i.e., p = 2).  This subsection proceeds to examine material failure mechanisms involved in the LS-DYNA calculations. Figure 3 shows the time history of the projectile velocity and the snapshots of the impact regions at three time instances of the calculation concerned with the normal strike with the impact velocity of 1000 m/s. Specifically, Figure 3b shows level contours of maximum principal stress and temperature at the same moments in time. As may be seen from Figure 3a, the velocity of the projectile decreases smoothly during the penetration process. The penetration completes at around 20 µs, and afterwards the velocity equals to a constant, i.e., the residual velocity. In addition, from Figure 3b it is noteworthy that the LS-DYNA solver, when equipped with the element erosion criterion, is capable of capturing several well-known failure modes for materials under high-speed impact. Overall, the plate is perforated mainly by a conventional plugging mechanism. The impact energy is released by the large plastic work around the shear rupture that separates the cavity from the rest of the plate. The temperate field also peaks at the cavity boundary, resulting in thermal softening of the plate. This thermal softening in turn facilitates and promotes localization of deformation, eventually resulting in plugging formation. In addition to the shear plugging, spalling occurs in the region of high maximum principal stress near the backface of the plate at 2.8 µs, due to the interaction between two reflecting tensile waves. Then at 11.2 µs, because of the bending and stretching of the plate, several lateral cracks have appeared, which are parallel to the plane of the plate. As a result of the lateral cracks, discing happens near the backface of the plate at 30.0 µs. The calculation also shows fragmented materials, which is typical for brittle materials such as AZ31B Mg alloys.

UQ Analysis
Computing sub-diameters requires a constrained optimization over the space of input variables in order to determine the largest deviation in the performance measure. To this end, a genetic algorithm (GA) is employed, which, as a global and derivative-free optimization method, provides the greatest flexibility in applications to non-linear problems. Another advantage of the GA is its high degree of concurrency. In particular, each iteration of the solution algorithm can be evaluated independently across multiple processors. In calculations, the DAKOTA Version 6.7 software package [40] of the Sandia National Laboratories is employed. A fixed population size of 64 is chosen throughout. One seed in the initial population is generated by setting the two repeated optimization variables associated with the sub-diameter at the two limits of that parameter range, with the remaining optimization variables set at the mid-span of their respective ranges. The remaining individuals in the initial population are selected randomly. It is found that this initial setup accelerates the convergence of the GA iterations. Additionally, the crossover and the mutation rates are 0.8 and 0.25, respectively. The numbers of parents and offspring of each generation are 64 and 48, respectively. More details about the computational framework can be found at Ref. [32].
As aforementioned, the Johnson-Cook parameters A, B, n, C, m, E 1 , E 2 , E 3 , E 4 and E 5 are assumed to be uncertain and known to be within intervals only. Specifically, in the UQ analysis the worst-case scenario is investigated for all operating conditions under consideration, i.e., the largest velocity of 1000 m/s with normal impact. The residual velocity of the projectile is used as the performance measure of the system, which provides an excellent measurement for the behavior of the AZ31B Mg plate under high-speed impact. Figure 4a shows the best fitness of each generation in the GA calculations. It is notable that, in spite of high non-linearity and irregularity of the fracture model and the contact condition used in the present ballistic problem, all GA calculations start to converge to a maxima after 200 to 300 generations. For the sake of comparison, Figure 4b compares the sub-diameters computed for each of the material parameters of the Johnson-Cook plasticity and fracture models. The numerical values of the sub-diameters are also tabulated in Table 5. It is noteworthy that the sub-diameters are measured in the unit of the performance measure. Therefore, they all have the same unit. A direct consequence of this property is that the sub-diameters can be compared and rank-ordered, which in turn provides a quantitative metric of the relative contributions of the parameters to the overall uncertainty of the response. From Figure 4b, this rank-ordering is thus deduced with the parameters E 1 , E 2 and A that contribute the most to the uncertainty, B, E 5 and E 4 the least and E 3 , n, C and m intermediate. It is also notable that the contribution rank-ordering is different from the order of percentile variation of the random variables. For instance, compared to other parameters, the parameter n has the greatest percentile variation, cf. Table 1, but contributes only modestly to the total uncertainty in the system performance. This example evinces how relative uncertainties cannot be directly deduced from the variability of the input parameter in general, but also depend critically on the non-linear sensitivity of the system response to the parameters.
Another finding from the sub-diameters is the model-dependent analysis of uncertainty contributions. As mentioned before, it is assumed that the material uncertainty comes from two separate models, i.e., Johnson-Cook plasticity and fracture models. Based on the sub-diameters listed in Table 5, the corresponding total diameter is 300.97 m/s. The uncertainty contribution by the plasticity model, which includes the parameters A, B, n, C and m, is 203.91 m/s. By contrast, the fracture model, characterized by E 1 , E 2 , E 3 , E 4 and E 5 , contributes 221.37 m/s to the total uncertainty. Therefore, for the selected random parameters and their ranges under consideration, the Johnson-Cook fracture model contributes more to the uncertainty of the ballistic behavior of the AZ31B Mg alloy plate.   As already mentioned, the present approach is predicated on probability inequalities as a means of bounding uncertainties. Evidently, the tighter the bound the better the design. However, increasing tightness comes at increasing computational expense, which sets forth a trade-off between economy of design and computability. Simple probability inequalities, such as McDiarmid's [24], supply a working compromise between tightness and computational complexity. However, it is both interesting and useful to investigate the tightness of the bounds and the attendant conservativeness of the designs. To this end, Figure 5 shows comparisons between Monte Carlo (MC) sampling, concentration-of-measure (CoM) inequality and optimal uncertainty quantification (OUQ) for the projectile/plate system with impact velocity v i = 1000 m/s and normal attack. The MC sampling is performed over the ranges of the random parameters, using 1.2 × 10 4 samples with Latin hypercube scheme and uniform distribution. Specifically, Figure 5a shows the sampled distribution of the residual velocity Y. It is computed that Y = 682.433 m/s. Using this mean value, the upper bounds of the probability of failure through CoM and OUQ are shown in Figure 5b, compared with an estimate using the MC sampling. As expected, both the CoM and OUQ bounds lie uniformly above than the MC estimate, which illustrates the conservative character of the bounds and, by extension, of the corresponding designs. For the problem with the information of system sub-diameters and mean performance, OUQ provides the optimal bound. As a result, the OUQ bound is tighter than CoM bound, as also shown in Figure 5b. Moreover, Figure 5c shows the two values of the residual velocity at the bounds of the sub-diameter for each random variable. The mean, maximum and minimum values of the sampled values of the residual velocity are also shown in Figure 5c. Notably, the residual velocities associated with the sub-diameters of A, n, C, E 1 , E 2 and E 3 are beyond the scope of MC data. These random parameters also contribute significantly to the total uncertainty in the performance of the plate. Thus, due to the sharp and irregular features involved in the contact and fracture problem, the random sampling method is not able to capture those extreme cases. By contrast, our UQ framework is capable of executing tests on demand over the entire operating space of random parameters, and hence provides required data for more rigorous uncertainty quantification and safer design.

A B n C m E
The uncertainties considered in the present work come from the plasticity and fracture models, which characterize the strength and toughness of the materials, respectively. In order to explore the material regime in which the random parameters cause the largest deviation of the residual velocity, the snapshots of the perforation cavity are extracted at the two bounds of the sub-diameter for each random variable, as shown in Figure 6. Specifically, Figure 6a,b visualize the profile of the cavity for the plasticity and fracture parameters, respectively. As may be seen from Figure 6a, the plasticity parameters result in the largest output uncertainty in different material regimes. Specifically, the behavior of the plate is more sensitive to the parameters A, n, C and m when the plate is relatively brittle, whereas the parameter B has a significant effect in the regime of high toughness. Regarding the fracture parameters in Figure 6b, it is notable that the parameters E 1 , E 2 and E 4 have the largest uncertainty contribution in the range crossing both the ductile and brittle regimes. By contrast, E 3 and E 5 have a more significant contribution to the total uncertainty in the brittle regime.

Summary and Concluding Remarks
The implementation of a UQ framework has been presented to assess the effects of constitutive and fracture properties on the performance of Mg alloys subject to high speed impact. The quantification towards uncertainties is achieved by determining the largest deviation in the performance measure resulting from a finite variation in the corresponding input random variables. Both strategies of McDiarmid's CoM inequality and OUQ have been harnessed to calculate conservative upper bounds on the probability of failure, and then the safety of the system can be certified and designed rigorously using the QMU strategy. The McDiarmid's CoM inequality has a relatively simple formulation and hence can be easily employed, whereas the optimal UQ is able to provide the optimal bounds by leveraging all known information of uncertainties and measurements in spite of its complex equations. The uncertainties of the constitutive and fracture properties in the present work arise from the partial information of model parameters, determined either by experiments or by experience of experts. Several significant findings afforded by the calculations are noteworthy. For our specific ballistic impact of AZ31B and the given intervals of uncertainties, the sub-diameters of the Johnson-Cook constitutive and fracture parameters are ordered as E 1 > E 2 > A > E 3 > n > C > m > B > E 5 > E 4 . Obviously, this ordering emphasizes the specific parameters where improvements are best targeted. Such a clear set of relationships also characterizes the constitutive and failure regime of interest, especially when compared against other boundary and/or initial conditions. For example, regarding the fracture models, the parameters related to the quasistatic behaviors, i.e., E 1 , E 2 and E 3 , contribute more significantly to the ballistic performance of AZ31B compared to the strain-raterelated parameter E 4 and temperate-related parameter E 5 . Therefore, the corresponding microscopic properties determining E 1 , E 2 and E 3 need to be adjusted in order to greatly improve the ballistic performance of the material. On the other hand, if a new fracture model needs to be developed for AZ31B, the part that affects its quasistatic behavior should gain more attention. Moreover, the material failure mechanisms for AZ31B involved in the impact tests bear emphasis, including spalling, plugging, discing and fragmentation, and our ballistic tests also conform to the form of the Recht-Ipson model. This paper is closed with a comparison between the adopted UQ approach and Bayesian strategies. Bayesian inference, based on the Bayes' theorem, has been introduced as one of the main tools for UQ of the computational models, mainly due to the relative simplicity of implementation and the rigor of the resulting Bayesian analysis. Nevertheless, the quantification of uncertainties is conducted via calculating high-dimensional integrals that are very intractable or even impossible to evaluate analytically through conventional integration techniques, let alone those significantly complicated response functions that are achieved by existing open-source codes or commercial software. One numerical way to solve those integrals is Monte-Carlo sampling, which can become impractical if the probability of failure is small, i.e., if failure is a rare event, and if one-call of forward calculation is costly, as shown in the numerical tests in this paper. By way of contrast, in the adopted method the response function can effectively be regarded as a black-box and the effort required for the computation of the uncertainty bounds only depends on the solution of sub-diameters and therefore is independent of the size of the probability of failure. In addition, the most commonly used priors in engineering problems are uniform and normal distributions. However, the strong influence of priors on the outcome of the inference process is also one of the most significant criticisms of Bayesian frameworks. By contrast, the CoM and OUQ approaches only requires the intervals of uncertain parameters and then provides rigorous bounds on the output uncertainties that bracket all the possible results led by all the probability measures in such intervals. If more information of uncertainties and system, such as moments of random variables, is given, the OUQ approach is able to obtain tighter bounds by leveraging such information.
Funding: The support of the University of Kentucky through the faculty startup fund.

Informed Consent Statement: Not applicable.
Data Availability Statement: Not applicable.