1. Introduction
The Burr III distribution, renowned for its adaptability, has been widely applied across various domains, including reliability engineering and survival analysis [
1,
2,
3,
4]. It was first introduced by Burr in 1942 [
5], and has undergone substantial enhancements to augment its modeling capabilities. Such advancements have expanded the distribution’s utility, allowing it to encompass a broader spectrum of data configurations and demonstrating its significant practical value [
6,
7,
8]. Recent progress in the field has spurred the development of innovative regression models and analytical tools predicated on the structural nuances of the Burr III distribution. Based on tuning parameters and modifying functional structures, the tools can be utilized for fitting and make predictions when given various types of data. These tools have found widespread application in data analysis, underpinning informed decision-making across a variety of scientific disciplines. Contrasting with the Weibull and gamma distributions, the Burr Types III and XII display a broader spectrum of skewness and kurtosis [
2], with Burr III being particularly versatile. In this paper, we therefore focus on the extension of the Burr III distribution.
In reliability modeling, the odds ratio serves as a pivotal statistical metric. It provides profound insights into the relationships between exposures and outcomes, proving indispensable in epidemiology and public health. The odds ratio has been effectively employed in assessing medical interventions and identifying behavioral risk factors, thereby shaping healthcare policies and preventive measures [
9,
10,
11].
In recent years, numerous methodologies for constructing generalized continuous probability distributions have been proposed [
12,
13,
14,
15,
16]. Contributions to this burgeoning field include the modified slash distributions [
17], modified-X family of distributions [
18], the new arcsine-generator distribution [
19], enhanced version of the generalized Weibull distribution [
20], McDonald Generalized Power Weibull distributions [
21], the exponentiated XLindley distribution [
22], the Pareto–Poisson distribution [
23], the Kumaraswamy Generalized Inverse Lomax distribution [
24], and others that have significantly advanced the statistical modeling landscape.
In Chen et al. [
25], we explored the exponentiated odds ratio generator to find the general form of distribution in terms of the odds ratio. The mathematical construct is defined as follows:
where
represents the cumulative distribution function (cdf) of the transformer, respectively.
, and
denote the cdf and survival function, respectively, of any baseline distribution associated with a random variable
x.
refers to the baseline distribution’s vector of parameters.
The development of statistical distributions for bathtub-shaped datasets is crucial for accurately modeling various real-life phenomena where the hazard rate initially decreases, then stabilizes, and, finally, increases over time. This characteristic shape is observed in numerous applications, particularly in reliability engineering and biomedical fields. For example, the mortality rate of humans typically follows a bathtub curve: high during infancy, low during most of adulthood, and high again in old age. This pattern is also prevalent in the failure rates of mechanical and electronic components, where early failures (infant mortality), a period of reliable operation (useful life), and wear-out failures toward the end of the life cycle are common. Drugs also exhibit similar patterns where efficacy might vary significantly across different age groups, with higher failure rates in children and the elderly compared to middle-aged individuals. By adding additional parameters to existing distributions, statisticians can improve the precision of reliability assessments and risk evaluations, which are critical for industries like insurance, engineering, and healthcare.
Due to the inherent complexity of bathtub-shaped data, developing new distributions using traditional methods, such as those used for uniform, exponential, gamma, and Weibull distributions, is not feasible. However, integrating established general structures with simpler distributions can yield new families of distributions capable of modeling bathtub shapes. These new distributions also perform well with other dataset shapes, enhancing their versatility and applicability across various fields. This approach allows for the creation of models that capture the unique characteristics of bathtub-shaped hazard functions while maintaining the simplicity and robustness of traditional distributions.
This paper introduces a revised Burr III distribution that integrates the odds ratio generator, aimed at improving the modeling of real-world data. The motivation for developing this new family of distributions is to enhance the flexibility of the basic function
, especially for single-parameter baseline distributions. We aim to show that, by adding additional parameters, simple distributions with limited variability can be transformed to display a wide range of shapes and skewness, as explored in
Section 3. Moreover, the flexibility and usefulness of this new family of distributions are demonstrated through four real-life examples. Notably, as illustrated in the first two examples of
Section 6, even with the simplest parent distribution, the uniform distribution, the B-SIOR-Uniform provides a robust bathtub shape for modeling real-life datasets. As shown in Figures 11 and 13, well-known distributions such as the gamma, Weibull, and generalized exponential distributions all fail to model bathtub-shaped datasets. Complex models, like the Burr III [
5] and those related to the Weibull and exponential distributions, fail to model the bathtub-shaped dataset as well. These models include the Weibull generalized exponential distribution [
26], the Type-2 Gumbel [
27], the Lomax Gumbel Type-2 [
28], and the Exponentiated Generalized Gumbel Type-2 distributions [
29]. These observations further support the novelty and significance of the proposed model. Recent literature has introduced several univariate distributions, such as those in [
30,
31,
32], which often possess complex structures that can complicate their practical use. In contrast, the B-SIOR-G model, described in Equation (
2), maintains a relatively simple structure. This simplicity aids in the ease of computing its properties and performing parameter inference, offering advantages over many more generalized distributions.
The structure of the paper is as follows:
Section 2 delineates the new distribution family and examines its key sub-model, the B-SIOR-G family of distributions. In
Section 3, we highlight several special cases with illustrative examples of probability density functions, hazard rate functions, and plots of skewness and kurtosis.
Section 4 is dedicated to exploring the statistical properties of the B-SIOR-G family of distributions, covering aspects such as hazard rate functions, quantile functions, moments, and more.
Section 5 presents various estimation methods along with simulation results. Finally,
Section 6 demonstrates the proposed model’s flexibility through real-life applications.
4. Statistical Properties of the Burr III Scaled Inverse Odds Ratio–G Distribution
Given the extensive formulation of the probability density function in Equation (
3), computing the statistical properties of this new family of distributions could involve intricate processes, and deriving closed forms for complex parent distributions may be challenging. To expedite these calculations, we will first expand the pdf of the B-SIOR-G into the well-known exponentiated-G distribution, for which the statistical properties have been established both theoretically and numerically. This expansion simplifies the computation of all related statistical measures, including moments, the moment-generating function, incomplete and conditional moments, moment of residual life and reversed residual life, the Rényi entropy, order statistics, stochastic ordering, and probability-weighted moments. Due to the complexity of this paper, comprehensive proofs and detailed derivations of these properties are provided in the
Supplementary Information, accessible at
https://github.com/shusenpu/B-SIOR-G/blob/main/Supplementary_Info.pdf (accessed on 2 May 2024).
4.1. Basic Properties of the B-SIOR-G Distribution
A series of expansions for the probability density function allows for simplifying calculations in properties’ derivations, approximating functions during simulation processes, and facilitating the analytical manipulation of the pdf.
Theorem 1. The pdf of the B-SIOR-G distribution can be expressed as a linear combination of the exponentiated generalized distribution as follows:where the coefficients are given byand the term is defined asrepresenting the exponentiated generalized distribution’s pdf given parameter . Proof. We consider the general form of binomial series expansion:
Thus, the pdf of the B-SIOR-G distribution can be expanded as
We note that, using the definition of odds ratio, the function can be generalized as
Therefore, we have the pdf as
where
and
which is the pdf of the exponentiated generalized distribution with parameter
. □
The pdf is represented as an infinite series, which implies that the function can be approximated by sufficiently summing many terms of the series to approximate its true value. The quality of the approximation typically improves as more terms are included. The coefficients modify the contribution of each term in the series. These coefficients can be interpreted as weights that adjust the influence of each term based on the distribution. It can be easily demonstrated that the series converges because it forms a geometric sequence with a common ratio . The expansion provides an intuitive and easier way to calculate the statistical properties of the distribution, while it is relatively difficult to perform derivations like integration on the original pdf with odds ratios.
The hrf defines the instantaneous rate at which events occur, given that no prior event has happened. A higher hazard rate at a certain time indicates a greater risk of the event happening. The hazard rate does not represent the probability of an event occurring at a specific time but the rate or intensity of occurrence at an infinitesimal interval around that time. In their seminal work, ref. [
33] demonstrated the equivalent behavior of hrf, reverse hazard function and the mean residual life function.
Remark 1. The hrf of the distribution can be derived as follows:and the reverse hrf is The quantile function, which can also be termed as the inverse cdf, serves as a method to map probabilities back to values in the domain of a random variable. For a given probability p (where ), the quantile function provides the value x such that the probability of the random variable being less than or equal to x is p.
Remark 2. The quantile function for the B-SIOR-G distribution is defined aswhere and Generally, the quantile function can be used to find percentiles and could be used for inferences. For example, gives the median of the distribution, the value below which of the data falls. The simple form of the quantile function q provides a direct way to calculate the probability since it can be solved directly by plugging in the quantile values.
4.2. Moments, Incomplete Moments, and Generating Functions
Moments serve as statistical indicators that succinctly capture the essence of probability distributions and datasets.
Remark 3. For a random variable Y∼, the rth moment of the B-SIOR-G distribution iswhere follows the exponentiated generalized distribution, which is obtained by raising the cdf of a baseline distribution to a certain power, with as defined in Theorem 1. The convergence of the rth moment can be demonstrated by examining the expansion of the pdf of B-SIOR-G, where the pdf can be expressed as a coefficient multiplied by a certain power of the parent cdf, . On one end, as i and j increase, the tails of the power series approach zero since . On the other end, the powers of can be sorted while calculating the rth moment of . To establish an upper bound for , we replace all coefficients with their maximum value and rank the according to the power of , forming a geometric series with a common ratio of . Therefore, we conclude that converges. A similar approach can be applied to demonstrate that the infinite series converges in all remaining conclusions and remarks.
An incomplete moment refers to the moment of a portion of a distribution. It is defined as the expected value of a given function of a random variable over a specified range, which is in contrast to the complete moment that takes into account the entire distribution.
Remark 4. The incomplete moment for the distribution is formulated aswhere is the incomplete moments of the exponentiated-G distribution. Here, we can examine the moment in the designated range from 0 to z as specified by the range, which is especially useful when dealing with truncated proportions of distribution, such as censoring in survival analysis. The moment generating function (mgf) encapsulates all the moments of the probability distribution. By evaluating its derivatives at zero, the mgf provides a systematic way to calculate the mean, variance, skewness, and other moments of the distribution.
Remark 5. The mgf can be found usingwhere denotes the mgf of the exponentiated generalized distribution with parameter . Thus, the overall distribution is a mixture or combination of multiple distributions, each with a different parameter . Adjusted by the coefficient weights , each contributes to the overall , which can be used for finding the n-th moment of the distribution by taking the n-th derivative of the resulting function.
4.3. Moment of Residual Life and Reversed Residual Life
In survival analysis and reliability engineering, the moment of residual life and moment of reversed residual life are essential for analyzing the distribution of time-to-event data. The moment of residual life at the time t quantifies the expected remaining lifetime given that an event has not occurred by t. Conversely, the moment of reversed residual life at the time t reflects the expected past duration given that an event occurred before t.
Lemma 1. The moment of residual life, denoted as , is formally defined aswhere X is the lifetime random variable, and m is the moment order. The moment of residual life of the distribution can be derived from Equation (10):where is the pdf of exponentiated-G with parameter . Following a similar calculation, we can find the mth moment of reversed residual life as follows.
Remark 6. The moment of reversed residual life, denoted as , can be derived as 4.4. Skewness and Kurtosis Analysis
Skewness quantifies the asymmetry of the distribution, while kurtosis measures its tail heaviness, with positive skewness indicating a right-skewed distribution and kurtosis indicating peakedness.
Lemma 1. Given , the coefficient of skewness () for Y∼ isand the coefficient of kurtosis () is 4.5. Rényi Entropy
Rényi entropy is a generalization of the Shannon entropy and provides a measure of the diversity, uncertainty, or randomness of a probability distribution.
Theorem 2. The Rényi entropy for the B-SIOR-G distribution is calculated aswhere , indicating the diversity of values the distribution can take, with being the Rényi entropy for the exponentiated generalized distribution parameterized by . Proof. The Rényi entropy of the B-SIOR-G distribution is given by
where
and
. By applying the same expansion technique for the pdf, we obtain
Using the definition of odds ratio, we have
Therefore, the further expansion of the Rényi entropy can be generalized as
where
is the Rényi entropy of the exponentiated generalized distribution with parameter
. □
The Rényi entropy for the B-SIOR-G distribution captures the diversity of the distribution’s outcomes weighted by the order . Higher values of indicate greater uncertainty or randomness within the distribution for the given order .
4.6. Order Statistics and Stochastic Ordering
Theorem 3. For as i.i.d. random variables from the B-SIOR-G distribution, the pdf of the jth order statistic isexpressing as a linear combination of B-SIOR-G with parameters , where . Proof. Let
be independent identically distributed random variables distributed by the B-SIOR-G distribution. The pdf of the
jth order statistic
is given by
□
Theorem 4. Given ∼ and ∼, the likelihood ratio Λ isindicating the relative likelihood of outcomes from two distributions based on their parameters. Earlier order statistics (smaller j) tend to have “lighter” tails as they are biased towards smaller values, whereas later order statistics (larger j) have “heavier” tails, reflecting larger sample values. The parameter modification implies that higher-order statistics or terms in the sum increasingly stretch or scale the distribution, accounting for the more extreme values expected in higher-order statistics.
4.7. Probability Weighted Moments
Probability Weighted Moments (PWMs) are derived from our probability distribution, which weights moments by the probabilities themselves. They provide a way to summarize the shape of a probability distribution and are particularly useful for characterizing the tails of the distribution.
Lemma 4. The rth PWM for the distribution can be calculated as follows:which simplifies tofor non-negative integers p, q, and r. For practical applications, often, ; thus, the PWM simplifies further, toA common example used for illustration is the first PWM (), expressed asfor and , which simplifies the evaluation process for specific cases. Commonly, in practice, particularly for hydrological models where one might want to understand the mean behavior rather than moments of higher order, r is set to 0, simplifying the PWM to only consider the probability weights without raising x to any power.
5. Methods of Estimation
In this section, we discuss five estimation methods crucial for discerning the parameters of proposed B-SIOR-G distributions. The exploration of these methods is the key to achieving accurate approximations that align B-SIOR-G models with observed data. Then, we apply the Monte Carlo simulation to test the convergence of approximated parameters as the sample sizes grow to determine its reliability. The estimation methods include the Maximum Likelihood Estimation (MLE), Least Square (LS) and Weighted Least Square (WLS) Estimation, Maximum Product Spacing Estimation (MPS), the Cramér–von Mises Estimation (CVM), and the Anderson and Darling Estimation (ADE).
Given an independent random sample
from the B-SIOR-G distribution with parameter vector
, the likelihood function
is formulated as
MLE seeks to find the set of parameters that maximize the likelihood function, making it a powerful method for parameter estimation by leveraging the full probability model, as shown in Equation (
11).
The LS technique in Equation (
12) minimizes the sum of the squared differences between observed and theoretical values, offering a straightforward approach for fitting models to data by emphasizing overall error reduction. An extension of LS, from Equ ation (
13), WLS, assigns weights to data points, prioritizing certain observations over others, thus refining the fitting process, especially when dealing with heteroscedastic data. The LS and WLS methods are expressed as [
34]
and
Both methods aim to find the best parameters (
) that align the theoretical distribution specified by the B-SIOR-G model as closely as possible to the observed data.
The MPS is valuable for complex distributions, as highlighted by Cheng and Amin [
35]. The MPS is maximized by optimizing
where
represents spacing functions and estimators are found by solving
. MPS in Equation (
14) focuses on maximizing the product of the spacings between ordered observations and their estimated distribution, providing an alternative to MLE that is less sensitive to outliers.
Estimators are derived by minimizing the Cramér–von Mises criterion,
, with respect to
, where
The Cramér–von Mises distance between continuous distribution functions is one of the distinguished measures of deviation between distributions, as shown in Equation (
15).
As an advancement of the Cramér–von Mises criterion, the Anderson–Darling approach in Equation (
16) places more weight on the tails of the distribution, making it particularly useful for detecting discrepancies in the distribution’s extremities. By minimizing the formula below, we can obtain the Anderson–Darling estimators:
This method emphasizes the tails of the distribution, making it particularly useful for distributions with significant tail behaviors.
To determine the maximum or minimum values of each estimator, we differentiate the specified objective functions and use numerical methods such as iteratively reweighted least squares (IRLS) or Newton–Raphson to locate their extreme values. For a practical demonstration of how the numerical simulation operates, we employ B-SIOR-E as an illustrative example. The detailed steps of the simulation process are outlined in Algorithm 1. We conducted a comprehensive simulation study using Monte Carlo simulations to estimate the parameters of the B-SIOR-E distribution. Then, we picked
,
,
, and
as the initial values for each parameter. Please note that the model is highly nonlinear and sensitive to the initial parameter values. Then,
, 80, 160, 320, and 640 were selected as sample sizes to generate random samples, with each experiment replicated 500 times to ensure statistical reliability. The resulting data were then analyzed to compute both bias and mean squared error (MSE) for each dataset. As illustrated in
Figure 10, the MSE demonstrated convergence towards 0 with increasing
N, affirming the stability and reliability of the estimations across all cases.
Algorithm 1 Monte Carlo Simulation for Parameter Estimation |
- 1:
Input: Randomly select a set of parameters as the true value - 2:
Initialize: Initiation of each parameter for optimization, - 3:
Set: Sample sizes , number of simulations - 4:
Define: BEORE function as per the model specifics - 5:
for each n in do - 6:
Initialize data frame error to store parameters - 7:
for to do - 8:
Generate uniform random numbers of size n - 9:
Initialize vectors: - 10:
for each i in 1 to n do - 11:
- 12:
Store r in x - 13:
Calculate indices for estimation methods including MLE, LS, WLS, etc. - 14:
end for - 15:
Sort x - 16:
- 17:
Apply estimation methods and store results - 18:
Store errors in error data frame - 19:
end for - 20:
Compute mean parameter estimates and MSE from error - 21:
end for - 22:
Output: Final parameter estimates and MSE for each sample size
|
6. Application
In this section, we transition from theoretical discussions to practical examinations, highlighting the applicability of our model through the analysis of real-world datasets. This exploration is designed to validate the practical utility of the newly introduced B-SIOR-U and B-SIOR-E distributions by showcasing their effectiveness in real data-driven scenarios. Some simple and well-known distributions such as gamma, generalized exponential, Log-logistic, and Weibull distributions are included in the comparison. Given that the B-SIOR-E model derives from the Burr III and exponential distributions, we aim to benchmark it against other models based on Burr III [
5], as well as those related to the Weibull and exponential distributions—Weibull generalized exponential (WGE) distribution [
26], known for its applicability in complex scenarios. Furthermore, in light of Chen’s work [
25] on Type-2 Gumbel distributions, we have also incorporated comparisons with Type-2 Gumbel (T2G) [
27], Lomax Gumbel Type-2 (LGT) [
28], and Exponentiated Generalized Gumbel Type-2 (EGG2) distributions [
29] into our analysis. It should be noted that both LGT and EGG2 have four parameters each. When these models are compared to B-SIOR-E and B-SIOR-U, the significance of the proposed model is evident, even though it shares the same number of parameters with its counterparts.
To thoroughly assess and compare the efficacy of statistical models, several goodness-of-fit metrics were employed, each evaluating distinct facets of model performance. The metrics include the following:
The -2 Log-Likelihood Statistic [
36], which quantifies the fit of a model by summarizing the discrepancies between observed and expected values under the model. A lower statistic suggests a better fit and this metric underpins various other statistical tests.
Cramér–von Mises Statistic (
) [
37], which measures how closely a theoretical cdf matches the empirical cdf by integrating the squared differences across all values. A lower
value indicates a better fit, as it means there is less deviation between the theoretical and empirical cdfs. This provides a thorough assessment of the deviation between the modeled and observed data.
Anderson–Darling Statistic (
) [
38], akin to
but placing greater emphasis on the tails of the distribution. This makes it particularly sensitive to extremities in data, valuable for analyses where tail behavior is crucial. A lower
value indicates a better fit, especially when the tails of the distribution are well modeled.
Akaike Information Criterion (
) [
36], which balances model fit against the number of parameters, penalizing unnecessary complexity. Derived from information entropy, it seeks to minimize information loss, preferring models with lower
values. A lower
value indicates a better fit, as it suggests that the model achieves a good balance between accuracy and simplicity.
Bayesian Information Criterion (
) [
39], similar to
but imposing a stronger penalty on the number of parameters. Based on Bayesian probability, it is useful for selecting among a finite set of models, favoring simplicity unless a more complex model significantly enhances fit. A lower
value indicates a better fit, favoring models that are simpler and have fewer parameters unless a more complex model significantly improves the fit.
Consistent Akaike Information Criterion (
) [
40], an augmentation of
that incorporates an extra penalty for parameter count, making it more conservative and particularly apt for larger datasets where overfitting is a concern. A lower
value indicates a better fit, particularly in larger datasets where it helps avoid overfitting.
Hannan–Quinn Criterion (
) [
41], which, like
and
, employs a logarithmically growing penalty term with sample size. It offers a compromise between the propensity of
to overfit and the strict penalties of
. A lower
value indicates a better fit, balancing between the risk of overfitting and underfitting.
Kolmogorov–Smirnov Test Statistic (
) [
42], which identifies the maximum divergence between the empirical distribution function of the sample and the cumulative distribution function of the reference distribution. The corresponding p-value assists in recognizing statistically significant deviations, with a smaller
statistic indicating a more accurate fit between observed and modeled distributions, as it suggests minimal divergence.
6.1. Lifetime Data
The first dataset contains the lifetime data of 50 devices, which was provided in [
43]. The goodness-of-fit statistics are shown in
Table 1 and
Figure 11 presents the histogram of observed data and the density plots of fitted distributions. The B-SIOR-U distribution emerges as the best model for this bathtub dataset since it has the optimal goodness-of-fit metrics and the highest K-S test p-value.
Figure 12 illustrates the Kaplan–Meier (K-M) survival curve and both theoretical and empirical cdfs. The theoretical predictions align well with actual data, highlighting that the B-SIOR-U distribution is effective in modeling the data with a bathtub shape.
6.2. Failure Data
This dataset contains the failure and running times of a sample of 30 devices given by Meeker and Escobar [
44]. The model results are shown in
Table 2.
Figure 13 presents the histogram of observed data and the density plots of fitted distributions. The B-SIOR-U distribution has the best performance for this dataset since it has the optimal value of all goodness-of-fit metrics.
Figure 14 presents the Kaplan–Meier (K-M) survival curve and the theoretical and empirical cdfs plot. It is obvious that the B-SIOR-U distribution can approximate the true data well.
6.3. Lung Cancer Data
This dataset contains the survival time of 128 patients with advanced lung cancer from the North Central Cancer Treatment Group. A summary of parameter estimates and goodness-of-fit metrics is presented in
Table 3, while
Figure 15 illustrates the comparison between the histogram of observed data and the density plots of fitted distributions. The B-SIOR-E distribution emerges as the superior model for this dataset, evidenced by its better goodness-of-fit metrics and the highest K-S test
p-value, as detailed in
Table 3.
Figure 16 presents a suite of plots including the Kaplan–Meier (K-M) survival curve, both theoretical and empirical cdfs, and a Total Time on Test (TTT) plot adjusted for scaling. The empirical consistency observed between theoretical predictions and actual data underscores the B-SIOR-E distribution’s adeptness at modeling the data with a monotonic hazard rate structure.
6.4. Bladder Cancer Data
This dataset represents the recorded remission times given in months from bladder cancer patients, reported by Lee and Wang [
45]. A summary of parameter estimates and goodness-of-fit metrics is presented in
Table 4, while
Figure 17 illustrates the comparison between the histogram of observed data and the density plots of fitted distributions. The B-SIOR-E distribution emerges as the superior model for this dataset, evidenced by its better goodness-of-fit metrics and the highest K-S test
p-value, as detailed in
Table 4.
Figure 18 presents a suite of plots including the K-M survival curve, both theoretical and empirical cdfs, and a TTT plot adjusted for scaling. The empirical consistency observed between theoretical predictions and actual data underscores the B-SIOR-E distribution’s adeptness at modeling the data with a monotonic hazard rate structure.
These results from real-world datasets do more than just confirm the statistical robustness of the B-SIOR-G model; they also amplify its practical relevance in the field of biomedical science. By accurately capturing the dynamics of the specified cancer datasets, this method supports predictive maintenance and ensures the quality of materials essential for advanced applications. The successful correlation of theoretical models with empirical evidence highlights the model’s capability to enhance decision-making in healthcare. The primary contribution of the proposed family of distributions lies in its exceptional performance in modeling bathtub-shaped data. While traditional distributions such as gamma, generalized exponential, Log-logistic, and Weibull may achieve a better Bayesian Information Criterion (BIC) for simpler shapes like unimodal data due to their fewer parameters, the B-SIOR-G family excels overall. Across diverse datasets, the B-SIOR-G distributions consistently deliver superior fits, making them highly effective for complex data structures.
7. Conclusions
This paper presented the modified Burr III Odds Ratio–G distribution, a comprehensive generalization aimed at enhancing data modeling through the incorporation of Burr III and the odds ratio. The extensive examination of its subfamilies, particularly the Burr III Scaled Inverse Odds Ratio–G distribution, revealed its versatility and efficiency in fitting cancer data. The theoretical underpinnings were rigorously explored, providing insights into its statistical properties and potential applications. Through simulation studies and real-life data analyses, the B-SIOR-G distribution demonstrated superior performance in modeling and prediction accuracy over several well-known distributions, underlining its significance and utility in statistical modeling. Notably, even the simplest model of B-SIOR-G, the B-SIOR–Uniform distribution, is capable of generating bathtub-shaped density and hazard rate functions. As demonstrated in the application section, the B-SIOR-G distribution flexibly models both complex bathtub shapes and skewed data. This broad applicability highlights the model’s practical utility for various types of data. With the key statistical properties thoroughly examined in this paper, this new model can be readily applied to data analysis and statistical modeling. However, we acknowledge that no single model is universally the best. In line with George E. P. Box’s observation, “All models are wrong, but some are useful”, we have transparently presented instances where our model may not be the optimal fit, to provide an honest and balanced view. Future research could further explore the application of this model in more diverse datasets and compare its performance with other recent distributions, potentially opening new avenues for statistical analysis and modeling across various scientific domains. Given the limited information available on distributions with bathtub-shaped probability density functions and hazard rate functions, a systematic literature review of all models capable of generating these shapes would be particularly valuable. Furthermore, our ongoing work aims to assess the robustness and accuracy of various statistical distributions in producing bathtub shapes, enabling a comprehensive evaluation of their efficacy in modeling diverse real-world data.