Next Article in Journal
A Dynamic Regional-Aggregation-Based Heterogeneous Graph Neural Network for Traffic Prediction
Previous Article in Journal
Hybrid Deep Neural Network with Domain Knowledge for Text Sentiment Analysis
Previous Article in Special Issue
A New Class of Bivariate Distributions: Properties, Estimation, and Modeling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detecting Clinical Risk Shift Through log–logistic Hazard Change-Point Model

by
Shobhana Selvaraj Nadar
,
Vasudha Upadhyay
and
Savitri Joshi
*
Department of Applied Sciences, Indian Institute of Information Technology Allahabad, Prayagraj 211015, India
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2025, 13(9), 1457; https://doi.org/10.3390/math13091457
Submission received: 27 March 2025 / Revised: 20 April 2025 / Accepted: 23 April 2025 / Published: 29 April 2025
(This article belongs to the Special Issue Advances in Statistical Methods with Applications)

Abstract

:
The change–point problem is about identifying when a pattern or trend shifts in time–ordered data. In survival analysis, change–point detection focuses on identifying alterations in the distribution of time–to–event data, which may be subject to censoring or truncation. In this paper, we introduce a change–point in the hazard rate of the log–logistic distribution. The log–logistic distribution is a flexible probability distribution used in survival analysis, reliability engineering, and economics. It is particularly useful for modeling time–to–event data exhibiting decreasing hazard rates. We estimate the parameters of the proposed change–point model using profile maximum likelihood estimation. We also carry out a simulation study and Bayesian analysis using the Metropolis–Hastings algorithm to study the properties of the proposed estimators. The proposed log–logistic change–point model is applied to survival data from kidney catheter patients and acute myeloid leukemia (AML) cases. A late change–point with a decreasing scale parameter in the catheter data reflects an abrupt increase in risk due to delayed complications, whereas an early change–point with an increasing scale parameter in AML indicates high early mortality followed by slower hazard progression in survivors. We find that the log–logistic change–point model performs better in comparison to the existing change–point models.

1. Introduction and Background

In survival analysis, the log–logistic distribution serves as an alternative to the more commonly used Weibull and log–normal distributions, offering an advantage when the data display a decreasing hazard rate. The probability density function (PDF) and the hazard function of the log–logistic distribution are, respectively, given by the following:
f ( x ; α ) = α ( α + x ) 2 , x > 0
h ( x ; α ) = α α + x , x > 0
where α > 0 .
The log–logistic distribution is widely used in survival analysis due to its ability to model decreasing hazard rates, making it suitable where the risk of an event declines over time. Several studies have contributed to the application of the log–logistic model. To mention a few, Gupta et al. (1999) [1] considered the log–logistic model to analyze lung cancer data and evaluated the confidence intervals for both parameters and the critical points. They also computed the mean residual life function and maximum likelihood estimators, which were found to be unique for both parameters. Bennett (1983) [2] considered a log–logistic regression model by equating one parameter of the distribution with a linear function of the covariates, which was a special case of the proportional odds model, and tested it on patients with advanced inoperable lung cancer. He fitted the data using the Generalized Linear Interactive Modeling software and found that the model captured the general shape of the data very well. Muse et al. (2021) [3] comprehensively reviewed the log–logistic distribution and its generalizations, highlighting its mathematical properties, applications, and advancements in statistical modeling.
Apart from analyzing survival data, the log–logistic distribution has gained increasing traction in various fields such as economics, medical sciences, hydrology, and social sciences. Lemonte (2014) [4] incorporated the beta distribution as a weight function, allowing for greater adaptability in fitting real–world data. Ashkar and Mahdi (2006) [5] investigated the use of the generalized moment (GM) method for estimating parameters and quantiles in the two–parameter log–logistic (LL2) distribution. They compared the GM method with other common estimation techniques, including the generalized probability–weighted moments and the maximum likelihood method, and observed that the GM method provided more accurate parameter estimates in the LL2 case when moment orders were appropriately selected. Shoukri, Mian, and Tracy (1988) [6] examined the sampling properties of the estimators for the log–logistic distribution, focusing on its application to Canadian precipitation data. They compared the following two estimation methods: probability–weighted moments and maximum likelihood estimation. Their findings confirmed that the log–logistic distribution provided an excellent fit for various meteorological datasets, making it a valuable tool in hydrology and climatology. De Santana et al. (2012) [7] introduced the Kumaraswamy–log–logistic distribution, a generalization of the log–logistic distribution designed to provide greater flexibility in modeling lifetime and survival data. Ramos et al. (2013) [8] introduced the Zografos–Balakrishnan–log–logistic distribution, an extension of the traditional log–logistic model. This distribution accommodated various hazard function shapes, enhancing flexibility in survival analysis. Alfaer et al. (2021) [9] introduced the extended log–logistic (Ex–LL) distribution, a flexible extension of the traditional log–logistic model designed to better capture actuarial and engineering risk data, particularly those with heavy–tailed distributions. Felipe et al. (2023) [10] proposed a robust estimation approach for the log–logistic distribution using minimum density power divergence estimators (MDPDEs) to address the limitations of traditional methods in the presence of outliers. The study demonstrated, through simulation and real data, that the MDPDEs offered a favorable balance between efficiency and robustness, making them suitable for practical survival analysis. Gaire and Gurung (2024) [11] introduced a new four–parameter log–logistic distribution based on the Rayleigh distribution. They carried out two methods of estimation and derived the statistical properties of the estimators, such as moments, survival function, etc. They applied it on two real–world datasets and found that the proposed model performed better than some select probability models.

Change–Point Problems in Survival Analysis

The problem of change–points in survival analysis has been a topic of interest for researchers in the past few decades. Matthews and Farewell (1982) [12] were among the first to consider change–point detection in hazard functions. Chang, Chen, and Hsiung (1994) [13] developed estimation methods for change–point hazard rate models under random censorship. Gijbels and Gürler (2003) [14] proposed a model incorporating a hazard jump at a specific time point and estimated the unknown parameters using maximum likelihood estimation, a least squares estimator, and the cumulative hazard–based approach introduced by Chang, Chen and Hsiung. Goodman, Li, and Tiwari (2011) [15] developed methods to detect multiple change–points in piecewise constant hazard functions. Williams and Kim (2013) [16] developed a Weibull hazard model considering Type–I censoring and staggered entry and studied the applicability of the model through a data on chronic granulomatous disease. Joshi et al. (2017) [17] proposed the Lindley hazard change–point model, estimating that the risk of death from a bone marrow transplant in patients with acute lymphoblastic leukemia significantly decreases after certain time. Palmeros et al. (2018) [18] considered a Weibull hazard regression model in the presence of covariates and censored observations. They estimated the parameters using maximum likelihood estimation and Monte Carlo simulation and applied the model on data regarding chronic granulomatous disease. Joshi and Rattihalli (2019) [19] proposed a general hazard regression change–point model where the effect of covariates on lifetime begins after the change–point. The Exponential–Lindley hazard change–point model was introduced by Joshi and Rattihalli (2020), ref. [20] where the failure rate remained constant before and after the change–point, following a Lindley distribution failure rate. Gierz and Park (2022) [21] proposed a sequential testing approach for detecting multiple change–points in a Weibull accelerated failure time (AFT) model and conducted a simulation study to show that the proposed method detected change–points and estimated the model quite accurately.
Despite the log–logistic model being a popular survival model, its scope in the field of change–point analysis is unattempted. Thus, in this paper, we aim to achieve the following:
  • Introduce a novel hazard change–point model, namely the log–logistic hazard change–point model;
  • Estimate the model parameters using the profile maximum likelihood estimation method and the Bayesian method, as well as compare the results for varying sample sizes and censoring levels;
  • Show the suitability of the model to detect a change–point in two real–world datasets.
The article is organized as follows: After an introduction and background in Section 1, the proposed change–point model is introduced in Section 2. In Section 3, parameter estimation using the profile maximum likelihood estimation procedure is shown followed by a simulation study in Section 4. We also evaluate the performance of the estimators through a Bayesian analysis, carried out in Section 5. In Section 6, a real–life data analysis is performed using the proposed change–point model. Lastly, Section 7 discusses the findings and future works.

2. log–logistic Hazard Change–Point Model for Survival Analysis

Let T be a survival time variable with a hazard function given by Equation (1). Suppose after a certain point in time, say τ , α 1 changes to α 2 . The hazard model under this setup is now given by the following equation:
h ( t ) = 1 α 1 + t , 0 t τ , 1 α 2 + t t > τ ,
where α 1 > 0 and α 2 + τ > 0 . It can be seen that if α 1 < (>) α 2 , then the hazard function beyond the change–point is at lower (higher) level than the hazard function below the change–point. It can also be noted that both the hazard functions are decreasing in X (Figure 1).
The PDF and the survival function corresponding to the above model are as follows
f ( t ) = α 1 ( α 1 + t ) 2 , 0 t τ , α 1 ( α 2 + τ ) ( α 1 + τ ) ( α 2 + t ) 2 t > τ ,
and
S ( t ) = α 1 α 1 + t , 0 t τ , α 1 ( α 2 + τ ) ( α 1 + τ ) ( α 2 + t ) t > τ .
where we have three parameters— α 1 , α 2 , and τ —to be estimated. In the next section, we discuss the profile maximum likelihood estimation (PMLE) method.

3. Parameter Estimation Using PMLE

To estimate the parameters, we opt for PMLE; however, one may use any command based approach of multi parameter estimation in R or any other suitable platform.
We shall consider random right censoring, i.e., instead of observing the survival time T i , we observe a pair ( T i , C i ) ( C i is the right censored observations) and a censoring indicator ϵ i , where ϵ i = 1 for an uncensored observation, and 0 for a censored observation. The study variable now becomes X i = min ( T i , C i ) .
With the assumption of independent censoring, the likelihood function for the pair ( X i , ϵ i ) is expressed as follows:
L ( α 1 , α 2 , τ | x ̲ ) i = 1 n ( f ( x i ) ) ϵ i i = 1 n ( S ( x i ) ) ( 1 ϵ i ) .
where n is the total number of observations in the data. In the presence of a change–point, the likelihood changes to the following:
L ( α 1 , α 2 , τ | x ̲ ) i = 1 n ( f ( x i ) ) I ( X i τ ) ϵ i i = 1 n ( S ( x i ) ) I ( X i τ ) ( 1 ϵ i ) i = 1 n ( f ( x i ) ) I ( X i > τ ) ϵ i i = 1 n ( S ( x i ) ) I ( X i > τ ) ( 1 ϵ i ) ,
where I [ 0 , τ ] ( X i ) is the indicator function. Substituting the values of f ( x ) and S ( x ) for the proposed log–logistic model, we obtain the likelihood function as follows:
L ( α 1 , α 2 , τ | x ̲ ) i = 1 n α 1 ( α 1 + x i ) 2 I ( X i τ ) ϵ i i = 1 n α 1 α 1 + x i I ( X i τ ) ( 1 ϵ i ) i = 1 n α 1 ( α 2 + τ ) ( α 1 + τ ) ( α 2 + x i ) 2 I ( X i > τ ) ϵ i i = 1 n α 1 ( α 2 + τ ) ( α 1 + τ ) ( α 2 + x i ) I ( X i > τ ) ( 1 ϵ i ) .
The log-likelihood function corresponding to the above expression is as follows:
log L ( α 1 , α 2 , τ | x ̲ ) i = 1 n ϵ i I ( X i τ ) log α 1 ( α 1 + x i ) 2 + i = 1 n ( 1 ϵ i ) I ( X i τ ) log α 1 α 1 + x i + i = 1 n ϵ i I ( X i > τ ) log α 1 ( α 2 + τ ) ( α 1 + τ ) ( α 2 + x i ) 2 + i = 1 n ( 1 ϵ i ) I ( X i > τ ) log α 1 ( α 2 + τ ) ( α 1 + τ ) ( α 2 + x i ) .
Let n u be the total number of uncensored observations in the data. Let u ( τ ) = i = 1 n I ( X i τ ) ϵ i and c ( τ ) = i = 1 n I ( T i τ ) ( 1 ϵ i ) be the number of uncensored and censored observations by time τ , respectively. The log-likelihood function can be further simplified as follows:
log L ( α 1 , α 2 , τ | x ̲ ) n log α 1 2 i = 1 n ϵ i I ( X i τ ) log ( x i + α 1 ) i = 1 n ( 1 ϵ i ) I ( X i τ ) log ( x i + α 1 ) ( n u ( τ ) c ( τ ) ) log ( α 1 + τ ) log ( α 2 + τ ) 2 i = 1 n ϵ i I ( X i > τ ) log ( x i + α 2 ) i = 1 n ( 1 ϵ i ) I ( X i > τ ) log ( x i + α 2 )
Let X ( i ) be the order statistics corresponding to X i , and ϵ ( i ) be the corresponding censoring indicator. For estimating α 1 , α 2 and τ , we assume that τ ( = τ ˜ ) is known. We consider τ ˜ to be the mid-point of each interval in all the ( n 1 ) intervals [ X ( j ) , X ( j + 1 ) ] , j = 1 , 2 , , ( n 1 ) . The confined log-likelihood is given as follows:
log L ( α 1 , α 2 , τ ˜ | x ̲ ) n log α 1 2 i = 1 n ϵ ( i ) I ( X ( i ) τ ˜ ) log ( x ( i ) + α 1 ) i = 1 n ( 1 ϵ ( i ) ) I ( X ( i ) τ ˜ ) log ( x ( i ) + α 1 ) ( n u ( τ ˜ ) c ( τ ˜ ) ) log ( α 1 + τ ˜ ) log ( α 2 + τ ˜ ) 2 i = 1 n ϵ ( i ) I ( X ( i ) > τ ˜ ) log ( x ( i ) + α 2 ) i = 1 n ( 1 ϵ ( i ) ) I ( X ( i ) > τ ˜ ) log ( x ( i ) + α 2 )
By applying the principle of maximum likelihood (ML) estimation and taking the partial derivatives of the simplified log-likelihood with respect to α 1 and α 2 , we get the following equations:
log L ( α 1 , α 2 , τ | x ̲ ) α 1 = n α 1 2 i = 1 n ϵ i I ( X i τ ) α 1 + x i i = 1 n ( 1 ϵ i ) I ( X i τ ) α 1 + x i n u ( τ ) c ( τ ) α 1 + τ ,
and
log L ( α 1 , α 2 , τ | x ̲ ) α 2 = n u u ( τ ) c ( τ ) α 2 + τ 2 i = 1 n ϵ i I ( X i > τ ) α 2 + x i i = 1 n ( 1 ϵ i ) I ( X i > τ ) α 2 + x i .
The above likelihood equations cannot be solved analytically for both α 1 and α 2 , therefore we employ numerical methods of estimation using the nlm() and the optimize() function in R [22]. To estimate α 1 , we consider α 2 to be known (nuisance parameter). The initial value of α 2 is taken to be the one used in the data simulation. By maximizing Equation (3) with respect to α 1 , using the nlm() function and the same seed value of α 1 as that used in the data simulation, we obtain the PML estimate of α 1 for the j t h interval, denoted as α 1 ^ j . Then, by replacing α 1 with α 1 ^ j in (3) and maximizing it for α 2 using the optimize() function, as well as considering the interval for α 2 such that the estimated value falls in the given interval, (0,3) yields the PML estimate of α 2 for the j t h interval, denoted as α 2 ^ j . Substituting the values of α 1 ^ j and α 2 ^ j obtained in (3) and maximizing the function over [ X ( 1 ) , X ( n ) ] will give us the estimate τ ^ j . Proceeding iteratively, we obtain the estimates α 1 ^ j , α 2 ^ j , and τ ^ j for all the intervals.
We compute the log-likelihood log L j ( α 1 ^ j , α 2 ^ j , τ ^ j ) , j = 1 , 2 , , ( n 1 ) . Suppose the maximum value of the likelihood is obtained at index j ¯ . The estimates of ( α 1 ^ j , α 2 ^ j , τ ^ j ) corresponding to j ¯ will provide us with the PML estimates of the parameters, i.e., α 1 ^ j ¯ , α 2 ^ j ¯ , τ ^ j ¯ .

4. Simulation Study

In order to simulate data for our analysis, we use the inverse cumulative distribution function (CDF) technique. For this purpose, we obtain a random sample from a standard rectangular distribution U ( 0 , 1 ) of the same size as our data. If our distribution function is denoted by F ( x ) , equating it with the random uniform sample F ( x ) = u previously generated will yield the required data x = F 1 ( u ) for a fixed value of ( α 1 , α 2 , τ ) .
The distribution function for the model proposed in Equation (2) is as follows:
F ( x ) = 1 S ( x ) = x α 1 + x , 0 x τ , ( α 1 + τ ) x + ( α 2 α 1 ) τ ( α 1 + τ ) ( α 2 + x ) , x > τ .
For the change–point τ , the corresponding uniform number will be u 0 = F ( τ ) = τ α 1 + τ . The generated sample is thus given by the following:
x = F 1 ( u ) = u α 1 1 u , 0 u u 0 , 1 1 u α 2 + τ α 1 + τ α 1 α 2 + α 2 u , u 0 < u 1 .
To check how well the proposed estimators work, we use PMLE. We use the following two sets of parameters for the estimation: ( α 1 = 1, α 2 = 2, τ = 1) and ( α 1 = 2, α 2 = 1, τ = 3). The estimation procedure is repeated 1000 times to ensure the reliability of the results. For each simulation run, we generate random samples from the log–logistic distribution with predefined parameters. We test for different sample sizes (n), 20, 50, 200, 500, and 1000, to see how the estimators perform under four levels of right-censoring, including 0%, 10%, 20%, and 50%. To assess the performance of the proposed estimators, we calculate bias and mean square error (MSE) and present them in Table 1 and Table 2.

Conclusion of the Simulation Study

From Table 1 and Table 2, we observe the following:
  • The estimators for α 1 , α 2 , and τ are consistently close to the true values across all sample sizes and censoring levels, with both bias and MSE decreasing as the sample size increases. This indicates that the estimators are approximately unbiased and consistent, even under high censoring. However, an increase in censoring percentage increases the MSE, because it introduces uncertainty due to missing information in the likelihood.
  • We plot bee-swarm plots to visualize the estimates of all the parameters (Figure A1, Figure A2, Figure A3 and Figure A7, Figure A8, Figure A9), and we observe that all the plots show similar observations, which indicates that the estimators are asymptotically unbiased and consistent.
  • We also plot the kernel density estimates to assess the normality. From Figure A4, Figure A5, Figure A6 and Figure A10, Figure A11, Figure A12, we see that the estimators are asymptotically normal. Here, in the case of 50% censoring when α 2 = 1 , we find that the kernel density estimator does not converge to normality at sample size 1000; however, the convergence can be achieved with a larger sample size.

5. Bayesian Analysis

We also establish the properties of the estimators obtained in Section 3 using a Bayesian analysis. We apply the Metropolis–Hastings (M-H) algorithm [23] and a Markov Chain Monte Carlo (MCMC) technique to estimate α 1 , α 2 and τ . We choose a normal distribution for all our three parameters (with an optimal choice of variance for all three parameters) as our proposal distribution.
If α ˜ 1 , α ˜ 2 , and τ ˜ denote, respectively, the random variables α 1 , α 2 , and τ , they are then expressed as follows:
α ˜ 1 N ( α 1 , 0.5 2 ) , α ˜ 2 N ( α 2 , 0.3 2 ) and τ ˜ N ( τ , 0.7 2 )
Under a normal prior, the log of the posterior likelihood is given by the following:
log L ( α ˜ 1 , α ˜ 2 , τ ˜ | α 1 , α 2 , τ , x ̲ ) n log α ˜ 1 2 i = 1 n ϵ i I ( X i τ ˜ ) log ( x i + α ˜ 1 ) i = 1 n ( 1 ϵ i ) I ( X i τ ˜ ) log ( x i + α ˜ 1 ) ( n u ( τ ˜ ) c ( τ ˜ ) ) log ( α ˜ 1 + τ ˜ ) log ( α ˜ 2 + τ ˜ ) 2 i = 1 n ϵ i I ( X i > τ ˜ ) log ( x i + α ˜ 2 ) i = 1 n ( 1 ϵ i ) I ( X i > τ ˜ ) log ( x i + α ˜ 2 ) log ( 2 π 0.5 ) 1 2 α ˜ 1 α 1 0.5 2 log ( 2 π 0.3 ) 1 2 α ˜ 2 α 2 0.3 2 log ( 2 π 0.7 ) 1 2 τ ˜ τ 0.7 2
We run the algorithm 10,000 times. At each iteration i, a proposed value is generated from θ ̲ ( i ) N ( θ ̲ ( i 1 ) , σ 2 ) , where θ ̲ ( i 1 ) = ( α ˜ 1 ( i 1 ) , α ˜ 2 ( i 1 ) , τ ˜ ( i 1 ) ) denotes the value of the parameters at the ( i 1 ) t h iteration. The M-H algorithm is described as follows:
Algorithm:
  • First, generate a proposal θ * = θ ( i ) from the proposal distribution N ( θ ̲ ( i 1 ) , σ 2 ) .
  • Next, compute the acceptance ratio, which is given as follows:
    r i = min 1 , L ( θ ( i ) | α 1 , α 2 , τ , x ̲ ) L ( θ ̲ ( i 1 ) | α 1 , α 2 , τ , x ̲ )
  • Then, draw a random variable u i U [ 0 , 1 ] .
  • If u i r i , accept the proposal and set θ * = θ ( i ) ; otherwise, reject the proposal and set θ * = θ ( i 1 ) .
We consider sample sizes of 20, 50, 200, 500, and 1000 under 0%, 10%, 20%, and 50%, censoring for the set of parameters α 1 = 2, α 2 = 1, and τ = 3. We do not discard the burn-in samples; we retain all 10,000 samples.
We plot trace plots for all parameters (shown in Figure A13, Figure A14, Figure A15) and tabulate the mean, bias, MSE, and the acceptance rate (AR), which is defined as the proportion of accepted samples out of the 10,000 simulated samples. The results are shown in Table 3.
From Table 3 and from the trace plots, we observe the following:
  • With an increase in sample size, there is a reduction in the bias and the MSE for all three parameters.
  • We observe that the AR increases with an increase in sample size. We also see that the AR is lower for the same sample size when we increase the censoring percentage. Hence, we can infer that censoring has a slight effect on the AR of the samples.
  • From the traceplots, we see that there is no trend or fluctuation observed in any of the three parameters, indicating a good mix of the chain.
Comparing the Bayesian analysis with the PMLE, we find that the bias and MSE are significantly lower for the Bayesian analysis compared to the PMLE; this could be due to the greater number of iterations in the Bayesian scenario, as well as the use of normal priors with small variance as the proposal distribution. Overall, the simulation results and the Bayesian analysis indicate that the proposed estimators are asymptotically unbiased, consistent, and asymptotically normal.

6. Data Analysis

In this section, we apply the proposed log–logistic hazard change–point model (LLHM) to two real-life datasets and evaluate its performance by comparing it with the exponential hazard change–point model (EHM), the Lindley hazard change–point model [LHM], and the Weibull hazard change–point model [WHM].
  • Model Evaluation Criteria and Goodness-of-Fit Measures
The AIC and BIC are defined as follows: [24]
AIC = −2 ln L + 2k, BIC = −2 ln L + k ln n
where ln L is the maximum value of the log-likelihood function for the estimated model, k is number of estimated parameters, and n is the sample size. Further, to assess the accuracy of the proposed model, the goodness-of-fit metrics—Manhattan distance ( L 1 norm), Euclidean distance ( L 2 norm), and Kolmogorov–Smirnov ( K S ) statistic—are defined as follows:
L 1 norm , d 1 ( a , b ) = j = 1 n | a ( j ) b ( j ) |
L 2 norm , d 2 ( a , b ) = j = 1 n ( a ( j ) b ( j ) ) 2
K S statistic , d 3 ( a , b ) = max 1 k n j = 1 k a ( j ) j = 1 k b ( j )
The Anderson–Darling (A-D) test is employed to determine whether the distribution of the model-generated data is statistically similar to that of the observed data. This test evaluates the null hypothesis H 0 , which states that both samples are drawn from the same underlying distribution.
The A-D statistic is given by the following:
A 2 = n 1 n i = 1 n ( 2 i 1 ) ln F ( X ( i ) ) + ln 1 F ( X ( n + 1 i ) )
where n is the sample size, X ( i ) are the ordered sample values, and F is the cumulative distribution function of the reference distribution (in this case, the empirical distribution of the observed data).
A high p-value (typically p > 0.05 ) indicates that there is no statistically significant difference between the model and observed distributions, supporting the similarity assumption and indicating a good fit of the model to the data.

6.1. Kidney Catheter Data

The kidney dataset in the survival package is based on a medical study that investigated repeated kidney infection in patients following catheter treatment. A total of 76 patients were monitored over time to record the duration until a recurrence of infection, out of which 18 were censored. The dataset included information such as the time (in days) until the event of interest occurred, such as catheter failure or removal, censoring status, treatment type, and other related clinical details. The study was originally described in McGilchrist and Aisbett (1991).
Early catheter failures are frequently associated with acute complications, such as improper placement, immediate post-insertion infection, or suboptimal initial care. However, patients who avoid these early risks tend to reach a period of relative stability, during which the catheter can function effectively for an extended duration [25].
The change–point models are applied to the selected dataset. Model adequacy is assessed using the standard selection criteria, and their performance is evaluated through a range of predefined similarity and goodness-of-fit measures. The results of these evaluations are summarized in Table 4 and Table 5, with graphical representations of the cumulative distribution functions (CDFs) (Figure 2) illustrating how well each model fits the observed data.
From Table 5, it is observed that, among all the competing models, the proposed LLHM yields the smallest values across all distance-based evaluation metrics, indicating a better fit to the data. Also from the A-D test, the asymptotic p-value indicates that the proposed model is not significant at 5% level of significance and is therefore a good fit. Based on the analysis, from Table 4, a change–point is identified at 201 days and the scale parameter decreases from 0.0170 to 0.0155, i.e., there is an abrupt increase in the hazard rate after around 201 days, after which the infection increases at a faster rate, signaling catheter removal or failure. This finding also aligns with the fact that the median duration of catheter dialysis is 190 days [25].

6.2. Acute Myeloid Leukemia

The analysis in this study is based on the LeukSurv dataset, which contains detailed survival data for 1043 patients diagnosed with acute myeloid leukemia (AML). It includes key variables such as time (in days) to event (death), event status (indicating death or censoring), age, sex, and white blood cell count (wbc) at diagnosis. For AML, the hazard rate plays a critical role in understanding disease progression and evaluating treatment outcomes. AML typically exhibits a high early hazard, especially during induction therapy, due to treatment-related toxicity and complications such as infections or bleeding [26].
The change–point models are applied to the selected dataset. Model adequacy is assessed using the standard selection criteria, and their performance is evaluated through a range of predefined similarity and goodness-of-fit measures. The results of these evaluations are summarized in Table 6 and Table 7, with graphical representations of the cumulative distribution functions (CDFs) (Figure 3) illustrating how well each model fits the observed data.
From the above-mentioned results (Table 6 and Table 7), the EHM shows lower A I C and B I C values, which might be due to the simplicity of the model as compared to LLHM. However, other evaluation measures such as the goodness-of-fit tests show higher values, indicating a poor fit of EHM to the data. This means that although EHM looks good based on A I C and B I C , it does not capture the actual pattern of the data well. Apart from A I C and B I C , the LLHM performs best across multiple evaluation criteria. It shows the smallest values in the L 1 norm, L 2 norm, and K S statistic, indicating that its CDF curve is closest to the observed data. Additionally, the A - D test yields a p-value that is not significant at 5% level of significance, further supporting that the model fits the data well.
The WHM, LLHM, and EHM estimate delayed change-points ranging from approximately 479 to 535 days. Among these, the LLHM identifies a change–point at 11 days, indicating a sharp increase in hazard very early in the disease course. From Table 6, a small change–point indicates that the risk of failure (e.g., death or relapse) is initially high during the early weeks after diagnosis. However, the increase in the scale parameter after this early change–point suggests that patients who survive this critical early period tend to have a slower progression of risk over time. This finding of a change–point at 11 days is consistent with the fact that the maximum risk of death due to AML occurs within 3–4 weeks after start of the treatment [26].

7. Conclusions and Future Work

In this paper, we introduce a log–logistic hazard model in the presence of a change–point. We derive expressions for the PDF and the survival function. We then estimate parameters using the PMLE and study the estimators obtained using different statistical measures and plots. The proposed change–point model is applied to both the Kidney Catheter and the Acute Myeloid Leukemia datasets and compared with the EHM, LLHM, and WHM. In both cases, LLHM provides the best fit, as indicated by lower values in the distance-based metrics (e.g., K S statistic, L 1 and L 2 norms) and the A-D test asymptotic p-value.
Interestingly, the change–point occurred earlier in the AML dataset, and the scale parameter increased after the change–point, suggesting that risk was highest in the early phase, but patients who survived this period tended to have slower hazard progression. In contrast, the Kidney Catheter dataset showed a later change–point with a decrease in the scale parameter, indicating that the patients were initially relatively stable, but hazard increased more sharply beyond a certain time, aligning with the possibility of delayed complications such as infections. These findings demonstrate the flexibility and clinical relevance of the LLHM in capturing disease-specific survival dynamics.
We may further extend the log–logistic change–point model to test for the presence of multiple change-points in a dataset, as well as the presence of covariates, so as to enhance the applicability of this model to diverse fields.

Author Contributions

Study Conceptualization, S.S.N. and S.J.; Methodology, S.S.N. and V.U.; Software, S.S.N. and V.U.; Validation, S.S.N., V.U., and S.J.; Formal Analysis, S.S.N. and V.U.; Data Curation, S.S.N. and V.U.; Writing—Original Draft Preparation, S.S.N. and V.U.; Writing—Review and Editing, S.S.N., V.U., and S.J.; Supervision, S.J.; Funding Acquisition, V.U. and S.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The Kidney Catheter dataset used in this study is publicly available in the survival package in R and can be accessed using the kidney dataset object. The Acute Myeloid Leukemia dataset, referred to as LeukSurv, is available through the SurvSet package in R. Both datasets are open access and can be obtained directly from the respective R packages version 4.4.2.

Acknowledgments

The first author acknowledges the cooperation of the Administration of the Guru Nanak Khalsa College of Arts, Science & Commerce, Mumbai. The second author acknowledges the financial assistance received in the form of a research fellowship from the Indian Institute of Information Technology Allahabad (IIITA) for pursuing this research. The third author acknowledges the SEED grant received from IIITA to carry out this research.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AMLAcute Myeloid Leukemia
PDFProbability Density Function
GMGeneralized Moments
LL22-parameter log–logistic
Ex-LLExtended log–logistic
MDPDEsMinimum Density Power Divergence Estimators
AFTAccelerated Failure Time
PMLEProfile Maximum Likelihood Estimation
MLMaximum Likelihood
CDFCumulative Distribution Function
MSEMean Square Error
MEVMean Estimated Value
M-HMetropolis–Hastings
MCMCMarkov Chain Monte Carlo
ARAcceptance Rate
LLHMlog–logistic Hazard Change-Point Model
EHMExponential Hazard Change-Point Model
LHMLindley Hazard Change-Point Model
WHMWeibull Hazard Change-Point Model
AICAkaike Information Criterion
BICBayesian Information Criterion
K S Kolmogrov–Smirnov
A - D Anderson–Darling

Appendix A

Figure A1. Bee–swarm plot for the estimates of α 1 = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A1. Bee–swarm plot for the estimates of α 1 = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a1
Figure A2. Bee–swarm plot for the estimates of α 2 = 2 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A2. Bee–swarm plot for the estimates of α 2 = 2 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a2
Figure A3. Bee–swarm plot for the estimates of τ = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A3. Bee–swarm plot for the estimates of τ = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a3
Figure A4. Gaussian kernel density estimates of α 1 = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A4. Gaussian kernel density estimates of α 1 = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a4
Figure A5. Gaussian kernel density estimates of α 2 = 2 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A5. Gaussian kernel density estimates of α 2 = 2 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a5
Figure A6. Gaussian kernel density estimates of τ = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A6. Gaussian kernel density estimates of τ = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a6
Figure A7. Bee–swarm plot for the estimates of α 1 = 2 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A7. Bee–swarm plot for the estimates of α 1 = 2 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a7
Figure A8. Bee–swarm plot for the estimates of α 2 = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A8. Bee–swarm plot for the estimates of α 2 = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a8
Figure A9. Bee–swarm plot for the estimates of τ = 3 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A9. Bee–swarm plot for the estimates of τ = 3 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a9
Figure A10. Gaussian kernel density estimates of α 1 = 2 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A10. Gaussian kernel density estimates of α 1 = 2 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a10
Figure A11. Gaussian kernel density estimates of α 2 = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A11. Gaussian kernel density estimates of α 2 = 1 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a11
Figure A12. Gaussian kernel density estimates of τ = 3 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Figure A12. Gaussian kernel density estimates of τ = 3 at 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000, respectively.
Mathematics 13 01457 g0a12
Figure A13. Trace plots for the estimates of α 1 = 2 in the presence of 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000.
Figure A13. Trace plots for the estimates of α 1 = 2 in the presence of 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000.
Mathematics 13 01457 g0a13
Figure A14. Trace plots for the estimates of α 2 = 1 in the presence of 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000.
Figure A14. Trace plots for the estimates of α 2 = 1 in the presence of 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000.
Mathematics 13 01457 g0a14
Figure A15. Trace plots for the estimates of τ = 3 in the presence of 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000.
Figure A15. Trace plots for the estimates of τ = 3 in the presence of 0%, 10%, 20%, and 50%, censoring for sample sizes 20, 50, 200, 500, and 1000.
Mathematics 13 01457 g0a15

References

  1. Gupta, R.C.; Olcay, A.; Sergey, L. A study of log-logistic model in survival analysis. Biom. J. J. Math. Methods Biosci. 1999, 41, 431–443. [Google Scholar] [CrossRef]
  2. Steve, B. log–logistic regression models for survival data. J. R. Stat. Soc. Ser. C Appl. Stat. 1983, 32, 165–171. [Google Scholar]
  3. Muse, A.H.; Mwalili, S.; Oscar, N. On the log–logistic distribution and its generalizations: A survey. Int. J. Stat. Probab. 2021, 10, 93. [Google Scholar] [CrossRef]
  4. Lemonte, A.J. The beta log–logistic distribution. Braz. J. Probab. Stat. 2014, 28, 313–332. [Google Scholar] [CrossRef]
  5. Fahim, A.; Mahdi, S. Fitting the log–logistic distribution by generalized moments. J. Hydrol. 2006, 328, 694–703. [Google Scholar]
  6. Shoukri, M.M.; Mian, I.U.H.; Tracy, D.S. Sampling properties of estimators of the log-logistic distribution with application to Canadian precipitation data. Can. J. Stat. 1988, 16, 223–236. [Google Scholar] [CrossRef]
  7. De Santana, T.V.F.; Ortega, E.M.; Cordeiro, G.M.; Silva, G.O. The Kumaraswamy-log–logistic distribution. J. Stat. Theory Appl. 2012, 11, 265–291. [Google Scholar]
  8. Ramos, P.L.; Balakrishnan, N.; Zografos, K. The Zografos-Balakrishnan log–logistic distribution: Properties and applications. J. Stat. Theory Appl. 2013, 12, 275–290. [Google Scholar]
  9. Alfaer, N.M.; Gemeay, A.M.; Aljohani, H.M.; Afify, A.Z. The extended log–logistic distribution: Inference and actuarial applications. Mathematics 2021, 9, 1386. [Google Scholar] [CrossRef]
  10. Felipe, A.; Jaenada, M.; Miranda, P.; Pardo, L. Robust parameter estimation of the log–logistic distribution based on density power divergence estimators. arXiv 2023, arXiv:2312.02662. [Google Scholar] [CrossRef]
  11. Gaire, A.K.; Gurung, Y.B. Rayleigh generated log–logistic distribution properties and performance analysis. J. Turk. Stat. Assoc. 2024, 15, 13–28. [Google Scholar]
  12. Matthews, D.E.; Farewell, V.T. On Testing for a Constant Hazard against a Change-Point Alternative. Biometrics 1982, 38, 463–468. [Google Scholar] [CrossRef]
  13. Chang, I.S.; Chen, C.H.; Hsiung, C.A. Estimation in Change-Point Hazard Rate Models with Random Censorship. Lect. Notes-Monogr. Ser. 1994, 23, 78–92. [Google Scholar]
  14. Gijbels, I.; Gürle, U. Estimation of a Change-Point in a Hazard Function Based on Censored Data. Lifetime Data Anal. 2003, 9, 395–411. [Google Scholar] [CrossRef] [PubMed]
  15. Goodman, M.S.; Li, Y.; Tiwari, R.C. Detecting Multiple Change-Points in piece-wise Constant Hazard Functions. J. Appl. Stat. 2011, 38, 2523–2532. [Google Scholar] [CrossRef]
  16. Williams, M.R.; Kim, D.Y. A test for an Abrupt Change in Weibull Hazard Functions with Staggered Entry and Type-I Censoring. Commun. Stat. Theory Methods 2013, 42, 1922–1933. [Google Scholar] [CrossRef]
  17. Joshi, S.; Jose, K.K.; Bhati, D. Estimation of a Change-Point in the Hazard Rate of Lindley Model under Right Censoring. Commun. Stat. Simul. Comput. 2017, 46, 3563–3574. [Google Scholar] [CrossRef]
  18. Palmeros, O.; Villaseñor, J.A.; González, E. On computing estimates of a change–point in the Weibull regression hazard model. J. Appl. Stat. 2018, 45, 642–648. [Google Scholar] [CrossRef]
  19. Joshi, S.; Rattihalli, R.N. Estimation of parameters in a general hazard regression change–point model. J. Indian Stat. Assoc. 2019, 57, 19–40. [Google Scholar]
  20. Joshi, S.; Rattihalli, R.N. Estimation of Parameters in the Exponential-Lindley Hazard Change-Point Model. Adv. Intell. Syst. Comput. 2020, 1169, 345–356. [Google Scholar]
  21. Gierz, K.; Park, K. Detection of multiple change points in a Weibull accelerated failure time model using sequential testing. Biom. J. 2022, 64, 617–634. [Google Scholar] [CrossRef] [PubMed]
  22. Matloff, N. The Art of R Programming: A Tour of Statistical Software Design; No Starch Press: San Francisco, CA, USA, 2011. [Google Scholar]
  23. Lynch, S.M. Introduction to Applied Bayesian Statistics and Estimation for Social Scientists; Springer: New York, NY, USA, 2007; Volume 1. [Google Scholar]
  24. Akaike, H. A new look at the statistical model identification. IEEE Trans. Automat. Contr. 1974, 19, 716–723. [Google Scholar] [CrossRef]
  25. Griffith, I.R.; Newsome, B.B.; Leung, G.; Block, G.A.; Herbert, R.J.; Danese, M.D. Impact of Hemodialysis Catheter Dysfunction on Dialysis and Other Medical Services: An Observational Cohort Study. Int. J. Nephrol. 2012, 8, 1179–1187. [Google Scholar] [CrossRef]
  26. Walter, R.B.; Othus, M.; Borthakur, G.; Ravandi, F.; Cortes, J.E.; Pierce, S.A.; Appelbaum, F.R.; Kantarjian, H.A.; Estey, E.H. Prediction of early death after induction therapy for newly diagnosed acute myeloid leukemia with pretreatment risk scores: A novel paradigm for treatment assignment. J. Clin. Oncol. 2011, 29, 4417–4423. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Hazard function of the log–logistic distribution at the change–point.
Figure 1. Hazard function of the log–logistic distribution at the change–point.
Mathematics 13 01457 g001
Figure 2. CDF Curves of kidney catheter data set under different models.
Figure 2. CDF Curves of kidney catheter data set under different models.
Mathematics 13 01457 g002
Figure 3. CDF Curves of acute myeloid leukemia data set under different models.
Figure 3. CDF Curves of acute myeloid leukemia data set under different models.
Mathematics 13 01457 g003
Table 1. Mean estimated value (MEV), bias, and MSE of proposed model based on 1000 iterations for α 1 = 1, α 2 = 2, τ = 1.
Table 1. Mean estimated value (MEV), bias, and MSE of proposed model based on 1000 iterations for α 1 = 1, α 2 = 2, τ = 1.
α 1 α 2 τ
Censoring n MEV Bias MSE MEV Bias MSE MEV Bias MSE
0%201.11660.11660.34262.19870.19871.26130.9383−0.06160.0225
501.00760.00760.07692.17700.17700.78370.9565−0.04340.0156
2000.9972−0.00280.01792.10790.10790.28900.9887−0.01120.0105
5000.9973−0.00270.00671.9134−0.08650.12900.9922−0.00770.0088
10000.9981−0.00180.00312.06050.06050.06840.9960−0.00390.0062
10%201.11750.11750.40222.16090.16091.29230.9388−0.06110.0222
501.03120.03120.08352.15250.15250.81020.9640−0.03590.0140
2000.9953−0.00460.01792.12940.12940.30240.9859−0.01400.0113
5000.9984−0.00150.00722.06450.06450.14770.9950−0.00500.0089
10000.9995−0.00040.00342.03090.03090.06860.9966−0.00310.0063
20%201.14210.14210.43102.20000.20001.36330.9374−0.06250.0239
501.02860.02860.08712.19370.19370.89940.9611−0.03880.0146
2000.9971−0.00280.01802.13830.13830.33710.9823−0.01760.0113
5000.9990−0.00090.00722.09720.09720.15220.9921−0.00780.0088
10000.9995−0.00040.00312.00150.00150.08431.00000.00000.0068
50%201.18630.18630.50012.23890.23891.86970.9289−0.07100.0296
501.04130.04130.10632.22630.22631.39490.9712−0.02880.0165
2000.9973−0.00270.01932.19040.19040.66040.9873−0.01260.0126
5001.01070.01070.00782.16280.16280.33690.9925−0.00740.0097
10000.9992−0.00080.00381.9302−0.06970.21240.9976−0.00230.0072
Table 2. Mean estimated value (MEV), bias and MSE of proposed model based on 1000 iterations for α 1 = 2, α 2 = 1, τ = 3.
Table 2. Mean estimated value (MEV), bias and MSE of proposed model based on 1000 iterations for α 1 = 2, α 2 = 1, τ = 3.
α 1 α 2 τ
Censoring n MEV Bias MSE MEV Bias MSE MEV Bias MSE
0%202.28310.28311.01491.28530.28531.99943.04500.04500.0397
502.08770.08770.30951.09130.09131.50683.03170.03170.0244
2002.02220.02220.06660.9105−0.08950.72333.03130.03130.0156
5002.01390.01390.02670.9184−0.08160.27823.02920.02920.0149
10002.00980.00980.01290.9682−0.03180.12723.01350.01350.0120
10%202.28860.28861.14771.29940.29942.00883.04850.04850.0413
502.10160.10160.33390.9163−0.08371.50013.03110.03110.0259
2002.03260.03260.07300.9180−0.08200.72563.02760.02760.0161
5002.02150.02150.02720.9453−0.05470.30603.02280.02280.0140
10002.00950.00950.01141.02380.02380.13773.01990.01990.0124
20%202.24380.24381.21291.24140.24142.03813.04710.04710.0424
502.11640.11640.36230.8776−0.12241.67153.03160.03160.0252
2002.03350.03350.07061.11760.11760.88123.02900.02900.0157
5002.01460.01460.02790.9209−0.07910.35973.02700.02700.0140
10002.01410.01410.01350.9915−0.00850.17423.02330.02330.0138
50%202.31930.31931.98431.53870.53872.49753.02670.02670.0523
502.09390.09390.38051.30160.30162.23893.02200.02200.0411
2002.03650.03650.07901.14060.14061.88723.01300.01300.0167
5002.02270.02270.03250.9288−0.07121.40623.00530.00530.0153
10002.00840.00840.01491.05510.05510.91913.00040.00040.0150
Table 3. Mean, bias, MSE and AR of the proposed log–logistic hazard change–point model based on 10,000 iterations of the Metropolis–Hastings algorithm for α 1 = 2, α 2 = 1, τ = 3.
Table 3. Mean, bias, MSE and AR of the proposed log–logistic hazard change–point model based on 10,000 iterations of the Metropolis–Hastings algorithm for α 1 = 2, α 2 = 1, τ = 3.
α 1 α 2 τ AR
Censoring n Mean Bias MSE Mean Bias MSE Mean Bias MSE
0%201.9973−0.00270.06400.9986−0.00140.00812.9879−0.01210.24650.9861
501.9979−0.00210.06371.00100.00100.00812.9895−0.01050.24440.9906
2002.00170.00170.06221.00080.00080.00802.9929−0.00710.24420.9948
5001.9994−0.00060.06171.00060.00060.00802.9932−0.00680.24300.9968
10002.00050.00050.06120.9999−0.00010.00792.9954−0.00460.24270.9976
10%201.9937−0.00630.06430.9986−0.00140.00822.9903−0.00970.24260.9776
501.9969−0.00310.06420.9989−0.00110.00822.9909−0.00910.24060.9827
2002.00310.00310.06341.00030.00030.00812.9952−0.00480.23910.9954
5001.9977−0.00230.06271.00010.00010.00802.9959−0.00410.23880.9961
10001.9983−0.00170.06250.9999−0.00010.00802.9972−0.00280.23700.9968
20%201.9964−0.00360.06310.9992−0.00080.00822.9913−0.00870.24510.9823
501.9980−0.00200.06280.9995−0.00050.00812.9933−0.00670.24080.9902
2001.998−0.0020.06261.00040.00040.0082.9950−0.00500.24040.9941
5001.9979−0.00210.06221.00010.00010.00803.00380.00380.23610.9956
10002.00010.00010.06211.00000.00000.00803.00170.00170.23470.9965
50%201.9939−0.00610.06240.9976−0.00240.00832.9773−0.02270.24210.9745
501.9947−0.00530.06300.9990−0.00100.00822.9912−0.00880.24130.983
2002.00050.00050.06301.00100.00100.00812.9919−0.00810.23910.9949
5001.9979−0.00210.06290.9999−0.00010.00793.00490.00490.23860.9955
10001.998−0.0020.06221.00000.00000.00782.9986−0.00140.23610.9965
Table 4. Parameter estimates of the change–point models when applied to data on recurrence times to infection of kidney patients.
Table 4. Parameter estimates of the change–point models when applied to data on recurrence times to infection of kidney patients.
ParameterEHMLHMWHMLLHM
α 1 0.00920.023962.21490.0170
α 2 0.00550.0088143.70170.0155
τ 96.000030.0000107.9997201.0000
Table 5. Different measures to assess the goodness of fit of the change–point models for Kidney Catheter data set.
Table 5. Different measures to assess the goodness of fit of the change–point models for Kidney Catheter data set.
MetricsEHMLHMWHMLLHM
AIC684.5601696.1884681.1359677.3659
BIC691.5523703.1806688.1281684.3581
L 1 -norm7.477116.42116.42725.1790
L 2 -norm1.00322.08900.85410.6724
K S statistic0.21740.36200.18100.1351
A - D test2.60511.071.8911.658
A - D asymptotic p-value0.042790.00000.10410.1413
Table 6. Parameter estimates of change–point models when applied to data on duration of acute myeloid leukemia in patients.
Table 6. Parameter estimates of change–point models when applied to data on duration of acute myeloid leukemia in patients.
ParameterEHMLHMWHMLLHM
α 1 0.00280.00660.0181262.7735
α 2 0.00040.00080.0085232.6557
τ 527.0006479.0017534.999711.0000
Table 7. Different measures to assess the goodness of fit of the change–point models for Acute Myeloid Leukemia data set.
Table 7. Different measures to assess the goodness of fit of the change–point models for Acute Myeloid Leukemia data set.
MetricsEHMLHMWHMLLHM
AIC10,783.4413,692.6712,224.0512,180.56
BIC10,798.2913,707.5212,238.912,195.41
L 1 -norm71.4478114.033122.548511.7279
L 2 -norm2.74994.49370.97090.4706
K S statistic0.15310.25860.06710.0397
A - D asymptotic p-value0.00000.00000.00150.1225
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nadar, S.S.; Upadhyay, V.; Joshi, S. Detecting Clinical Risk Shift Through log–logistic Hazard Change-Point Model. Mathematics 2025, 13, 1457. https://doi.org/10.3390/math13091457

AMA Style

Nadar SS, Upadhyay V, Joshi S. Detecting Clinical Risk Shift Through log–logistic Hazard Change-Point Model. Mathematics. 2025; 13(9):1457. https://doi.org/10.3390/math13091457

Chicago/Turabian Style

Nadar, Shobhana Selvaraj, Vasudha Upadhyay, and Savitri Joshi. 2025. "Detecting Clinical Risk Shift Through log–logistic Hazard Change-Point Model" Mathematics 13, no. 9: 1457. https://doi.org/10.3390/math13091457

APA Style

Nadar, S. S., Upadhyay, V., & Joshi, S. (2025). Detecting Clinical Risk Shift Through log–logistic Hazard Change-Point Model. Mathematics, 13(9), 1457. https://doi.org/10.3390/math13091457

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop