Next Article in Journal
CFD-Simulation-Based Multi-Axial Differential Mixing Enhancement Study for High-Viscosity Adhesives: From the Perspective of Breaking the Symmetry of the Flow Field
Previous Article in Journal
A Symmetric Multiscale Feature Fusion Architecture Based on CNN and GNN for Hyperspectral Image Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Parameter Estimation of MSNBurr-Based Hidden Markov Model: A Simulation Study

Department of Statistics, Institut Teknologi Sepuluh Nopember, Surabaya 60111, Indonesia
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(11), 1931; https://doi.org/10.3390/sym17111931
Submission received: 7 October 2025 / Revised: 29 October 2025 / Accepted: 6 November 2025 / Published: 11 November 2025

Abstract

Hidden Markov Model (HMM) is a well-known probabilistic framework for representing sequential phenomena governed by doubly stochastic processes. Specifically, it features a Markov chain with hidden (unobserved) states, where each state emits observable values through a state-conditioned emission distribution at every time step. In this framework, selecting an appropriate emission distribution is essential because an unsuitable choice may prevent the HMM from accurately representing the observed phenomenon. To accommodate emission phenomena with situational symmetry, we propose an HMM framework with an adaptive emission distribution, named MSNBurr-HMM. This method is based on the MSNBurr distribution, which can effectively represent symmetric, right-skewed, and left-skewed emission patterns. We also provide its parameter estimation algorithm using the Baum–Welch algorithm. For model validation, we conduct fitting simulations across diverse scenarios and compare the findings against Gaussian-HMM and Fernández–Steel Skew Normal-HMM using log-likelihood, the Akaike Information Criterion (AIC), the corrected AIC (AICc), and the Bayesian Information Criterion (BIC). The results demonstrate that the algorithm can effectively estimate the target parameters accurately in all tested scenarios. In terms of performance, MSNBurr-HMM generally outperforms other models with strong dominance in various aspects across all evaluation metrics, confirming the promising and excellent results of this proposed method.

1. Introduction

The Hidden Markov Model (HMM) is a probabilistic modeling framework that represents a doubly stochastic process, where one stochastic process underlies another stochastic process [1]. The underlying process in HMM is a Markov chain characterized by a set of hidden states and transition probabilities [2]. The second stochastic process generates an observable emission variable whose value depends on the hidden state at each time point [3]. This framework has been widely implemented in various research domains [4], including speech recognition [5], econometrics [6], ecology [7], bioinformatics [8], and predictive maintenance [9]. Despite its widespread use, there are important points to note in HMM modeling. A key feature of HMMs is that observed values are emitted according to a particular probability distribution, known as the emission distribution, conditioned on the hidden state. This distribution must accurately represent the data’s properties including whether it is continuous or discrete, as well as its characteristics, such as skewness, kurtosis, or modality. If not appropriate, this will lead to poor fitting results and affect the overall model performance [10]. Several studies have taken different approaches when dealing with the skew phenomenon. For instance, Sowan et al. [11] address skewness and instability in traditional lift values by proposing four adjusted lift formulations, namely Smoothed, Weighted, Log, and Thresholded lift. On the other hand, Dash et al. [12] address skewed data by removing outliers. In the context of HMM, accommodation of skew patterns can be carried out by selecting an emission distribution that is adaptive to the skew pattern. Notably, Gaussian distributions are frequently employed for continuous data [13], but their symmetry assumption often mismatches real-world data. Other continuous distributions, such as Gamma, Log-Gaussian, Log-Gamma, or Weibull distribution, may seem tempting for modeling asymmetric emissions. However, these distributions cannot easily accommodate varying skew patterns (both right and left) or nearly symmetric patterns. Moreover, their restricted domain suitability further limits flexibility.
Based on these findings, we develop an adaptive HMM framework capable of accommodating both symmetric and asymmetric emission distributions. Specifically, we integrate the HMM with the Modified-to-be-Stable-to-Normal from Burr (MSNBurr) distribution proposed by Iriawan [14]. Several factors have motivated us in this integration. First, MSNBurr distribution was formulated to effectively capture symmetric, right-skewed, and left-skewed patterns. Choir et al. [15] also explain that this distribution can perfectly fit the symmetric pattern of the Gaussian distribution while exhibiting heavier and longer tails, making it more representative for data with extreme events. Another advantage of the MSNBurr distribution is the stability of its mode with respect to the location parameter. This contrasts with other distributions such as Azzalini’s Skew-Normal [16] and Skew-t [17], where changes in parameter values cause the mode to shift away from the location parameter. The stability of the mode is important because it ensures that the central position of the distribution remains consistently identifiable when other parameters change. This understanding is crucial in statistics, especially for large real-world datasets, as it helps identify where most of the data is concentrated [18]. Due to its adaptive properties, the use of MSNBurr distribution led to improvements in several methods, including mixture models [19], Markov switching autoregressive models [20], and the Ising model [21].
This integration yields a new framework, which we name MSNBurr-HMM. We present its parameter estimation algorithm based on an Expectation-Maximization scheme, familiarly known as the Baum–Welch Algorithm [22]. It is an effective and widely used method for estimating HMM parameters when hidden states are unknown. Performance validation via simulations will be conducted in various scenarios that represent different emission symmetry conditions. We compare our proposed method with a well-known Gaussian-HMM and Fernández–Steel Skew-Normal-based HMM (FSSN-HMM) through evaluation metrics such as Log-Likelihood, Akaike Information Criterion (AIC) [23], corrected AIC (AICc) [24], and Bayesian Information Criterion (BIC) [25]. Besides Gaussian-HMM, FSSN-HMM was also chosen because the FSSN distribution has similar characteristics to the MSNBurr distribution, particularly its flexibility in modeling both symmetric and asymmetric fat-tailed patterns [26]. In addition, the mode of the FSSN distribution is also stable at the location parameter [19]. Several studies have confirmed the superiority of the FSSN distribution compared to the Gaussian distribution, such as in the modeling data of volcano height [27], magnetic resonance imaging brain image intensity data [19], and the Scotland lip cancer dataset [28]. Based on these considerations, FSSN-HMM serves as an additional ideal benchmark for the MSNBurr-based method proposed here, alongside the symmetric Gaussian-HMM. The results will demonstrate the advantages of the MSNBurr-HMM and highlight its potential for further HMM development.

2. Methods

2.1. Hidden Markov Model

A Hidden Markov Model (HMM) is a doubly stochastic process in which the hidden process follows a Markov chain, and each hidden state generates an observation based on a specific probability distribution. Formally, an HMM consists of two sequences, namely the hidden state sequence q 1 : T = q 1 , q 2 , q T representing the underlying Markov process and the sequence of observations O 1 : T = O 1 , O 2 , O T , where T denotes the sequence length. Each hidden state q t for t = 1 , 2 , , T takes values from a finite set S 1 , S 2 , , S K , with K is the number of possible hidden states. The observation O t for t = 1 , 2 , , T is generated from a conditional emission distribution p O t | q t = S i , and is conditionally independent, meaning that each observation depends only on its corresponding hidden state and not on other observations when the state is known.
HMMs are parameterized by a set of parameters, λ , which consists of three components [29]: the initial state distribution ( π ), the transition probability matrix ( A ), and the emission distribution parameters Θ = θ 1 , θ 2 , θ K corresponding for K hidden states. The definitions of π and A , along with their constraints, are provided in Equations (1)–(8). Regarding Θ , each θ i defines the parameters of the emission distribution for state S i ( i = 1 , 2 , , K ). If the emission probability p O t | q t = S i follows a Gaussian distribution, then θ i will consist of the mean and standard deviation parameters of that distribution. Using the Markov property of the hidden sequence and the conditional independence assumption of O t given q t , we can obtain the complete-data likelihood p O 1 : T , q 1 : T | λ in Equation (9) [30]. Equation (9) can also be rewritten as Equation (10) by decomposing λ into its explicit components π , A , Θ , where each parameter governs a different part of the model.
π = π 1 , π 2 , , π K
π i = p q 1 = S i ,       i = 1 ,   2 , , K
π i 0 ,       i = 1 ,   2 , , K
i = 1 K π i = 1
A = a 11 a 12 a 1 K a 21 a 22 a 2 K a K 1 a K 2 a K K
a i j = p q t = S j | q t 1 = S i ,       i , j = 1 ,   2 , , K ,       t = 2 ,   3 , , T
a i j 0 ,         i , j = 1,2 , , K
j = 1 K a i j = 1 ,       i = 1,2 , , K
p O 1 : T , q 1 : T | λ = p q 1 | λ   t = 2 T p q t | q t 1 , λ   t = 1 T p O t | q t , λ
p O 1 : T , q 1 : T | π , A , Θ = p q 1 | π   t = 2 T p q t | q t 1 , A   t = 1 T p O t | q t , Θ

2.2. MSNBurr-Based HMM

We introduce a modified HMM with an emission distribution following the MSNBurr distribution [14]. Derived from the Burr Type II distribution, the MSNBurr distribution accommodates both symmetric and asymmetric patterns while ensuring mode stability with respect to its location parameter [14]. Additionally, it can also accommodates long-tailed data [19]. We formulate an MSNBurr-based HMM, abbreviated as MSNBurr-HMM, by defining the emission probability p O t q t = S i to follow the MSNBurr distribution. Each state’s emission distribution is characterized by parameters θ i = μ i , ϕ i , γ i , where μ i , ϕ i , and γ i are the location, scale, and shape parameters, respectively. The complete specification of this distribution is provided in Equation (11), with < O t < , < μ i < , ϕ i > 0 , γ i > 0 , and κ i is a function of γ i as shown in Equation (12).
p O t | q t = S i , θ i = κ i ϕ i exp κ i O t μ i ϕ i 1 + exp κ i O t μ i ϕ i γ i γ i + 1
κ i = 1 2 π 1 + 1 γ i γ i + 1
To illustrate the adaptability of this distribution, we present the probability density function (pdf) plots for several different parameter values in Figure 1. We can observe that symmetry is achieved when γ = 1 . A left-skewed pattern begins to emerge when γ < 1 , while γ > 1 is associated with right-skewness.

2.3. Parameter Estimation of MSNBurr-HMM

In this section, we present the parameter estimation of the MSNBurr-HMM using an EM-based approach, commonly known as the Baum–Welch Algorithm. At this point, we will apply Equation (10) to enter the EM framework, where in the E-Step, we compute the expected log-likelihood of the complete data. In the M-Step, it is going to be maximized with respect to the model parameter λ . The expected log-likelihood of the complete data is expressed as Q λ , λ o l d in Equation (13), where λ o l d is the current parameter estimate obtained from the previous iteration. To solve this, we adopt the Baum–Welch algorithm steps where we need a forward variable α t i , backward variable β t i , and two other quantities calculated from α t i and β t i , namely δ t ( i ) and ξ t ( i , j ) [22,29,31].
Q λ , λ o l d = E q 1 : T | O 1 : T , λ o l d log p O 1 : T , q 1 : T | λ
The forward variable, denoted by α t i , satisfies the definition in Equation (14). At t = 1 , this quantity is calculated using Equation (15), where p O 1 | q t = S i , θ i follows the form given in Equation (11). For t = 2 , 3 , , T , α t i is computed recursively via Equation (16). We next define the backward variable β t i in Equation (17). In contrast to α t i , which is computed forward from t = 1 , β t i is computed backward from t = T using Equation (18), then recursively for t = T 1 , T 2 , , 1 via Equation (19).
α t i = p O 1 : t   , q t = S i | λ
α 1 i = π i   p O 1 | q 1 = S i , θ i
α t i = j = 1 K α t 1 j   a j i p O t | q t = S i , θ i
β t i = p O t + 1 : T | q t = S i , λ
β T i = 1
β t i = j = 1 K a i j   p O t + 1 | q t + 1 = S j , θ j   β t + 1 j
Once α t i and β t i are obtained for each t = 1 , 2 , , T dan i = 1 , 2 , , K , we compute δ t i and ξ t i , j using the formal definitions given in Equations (20) and (21), respectively. We refer to δ t i as the expected state occupancy count and ξ t i , j as the expected state transition count. Both become the basis for updating the parameters λ in the next stage.
δ t i = p q t = S i | O 1 : T , λ = α t i   β t i l = 1 K α t l   β t l
ξ t i , j = p q t = S i ,   q t + 1 = S j | O 1 : T , λ = α t i   a i j   p O t + 1 | q t + 1 = S j , θ j   β t + 1 j i = 1 K j = 1 K α t i   a i j   p O t + 1 | q t + 1 = S j , θ j   β t + 1 j
The updating process of λ is operated independently for each component ( π , A , and Θ ). The update of π i ( i = 1 , 2 , . . . , K ) is calculated using Equation (22), while a i j ( i , j = 1,2 , . . . , K ) is obtained by Equation (23). Regarding Θ , the update is performed on each hidden state emission θ i using Equation (24). The δ t i and ξ t i , j values used in all three equations are calculated using the current λ . The output of Equations (22)–(24) gives us the updated parameter estimates λ n e w = ( π n e w , A n e w , Θ n e w ) . Since δ t ( i ) and ξ t ( i , j ) directly depend on these parameters, their values will consequently be updated as well.
π i n e w = δ 1 i
a i j n e w = t = 1 T 1 ξ t i , j t = 1 T 1 δ t i
θ i n e w = arg max θ i t = 1 T δ t i log p O t | q t = S i , θ i
The next step is to repeat the process of calculating the forward–backward variable applying the updated parameter λ n e w . This process should be repeated until it reaches M iterations or satisfies the specified convergence criteria. Such a criterion can be defined, for example, by requiring that the changes in all parameter components become sufficiently small, as formulated in Equation (25), where m denotes the iteration index and ϵ is a convergence tolerance.
max π m + 1 π m ,   A m + 1 A m , Θ m + 1 Θ m < ϵ
Furthermore, we suggest readers to see Yang et al. [32] for a detailed discussion of the properties and convergence guarantees of the Baum–Welch algorithm. We finally summarize the procedure in Algorithm 1. Notice that an initial parameter value λ 0 is required to start the iteration, which can be obtained using a reasonable data-driven strategy. For instance, λ 0 may be assigned by using values derived from simple statistics, such as the sample mean and standard deviation of the observable values.
Algorithm 1. Baum–Welch algorithm for MSNBurr-HMM.
Input: O 1 : T ; Initialization λ 0 = π 0 , A 0 , Θ 0 ; convergence tolerance ϵ
Output: λ m *
Initialization:
Set the iteration index m = 0
Repeat:
E-Step:
Compute α t i and β t i using Equations (15), (16), (18) and (19).
Compute δ t i and ξ t ( i , j ) using Equations (20) and (21).
t = 1,2 , , T , i , j = 1,2 , , K
M-Step:
Update π using Equation (22) to obtain π m + 1
Update A using Equation (23) to obtain A m + 1
Update Θ using Equation (24) to obtain Θ m + 1
Convergence Check:
Evaluate the parameter change using Equation (25)
If the criterion in Equation (25) is satisfied, then stop the iteration and set m * = m + 1 ;
Else set m m + 1 and repeat the process.
End Repeat

2.4. Simulation Design

In this study, we focus on the case of two hidden states with a fixed sequence length of T = 200. We examine three main simulation scenarios, each representing a different skewness in the emission distributions. Scenario 1 consists of two symmetric distributions. In Scenario 2, one emission distribution remains symmetric, while the other exhibits a skewed (asymmetric) behavior. Scenario 3 extends this further, with both emissions demonstrating skewness. Each of these scenarios is then divided into two sub-scenarios based on the overlap between the emission distributions. In sub-scenario (a), the two emission distributions are completely separated, whereas in sub-scenario (b), they partially overlap. As a result, we analyze a total of six distinct cases, labeled as Scen1a, Scen1b, Scen2a, Scen2b, Scen3a, and Scen3b.
The data generation process utilizes an equal initial state distribution, π = 0.5 , 0.5 , and a fixed transition probability matrix A = a 11 a 12 a 21 a 22 = 0.7 0.3 0.2 0.8 for all scenarios. The emission distribution parameters vary across scenarios according to their respective criteria with full details provided in Table 1. We also replicate each scenario 100 times to obtain a comprehensive view of the results. Finally, the R code of this simulation is publicly available at https://github.com/dbunggul/msnburrHMM-bwa (accessed on 20 June 2025) for reproducibility.

2.5. Evaluation Criteria

The evaluation procedure will be conducted by analyzing the log-likelihood and three information criteria, namely Akaike’s Information Criterion (AIC) [23], its bias-corrected version (AICc) [24], and Bayesian Information Criterion (BIC) [25]. The log-likelihood, denoted as log p O 1 : T | λ , can be efficiently computed using the forward variable at the final time step within the Baum–Welch algorithm. The formula of log-likelihood is given by Equation (26) below. Furthermore, we also consider AIC, AICc, and BIC as they account for the number of parameters ( d ). This process of penalizing complexity offers a clearer perspective when comparing models with different numbers of parameters. AIC is defined by Equation (27), while AICc is given in Equation (28), where n represents the number of observations. Lastly, BIC is written in Equation (29). Models with higher log-likelihood are preferred [33]. As for AIC, AICc, and BIC, we prefer models with lower values.
log p O 1 : T | λ = log i = 1 K α T i
A I C = 2 log p O 1 : T | λ + 2 d
A I C c = A I C + 2 d d + 1 n d 1
B I C = 2 log p O 1 : T | λ + d ln n

3. Results

3.1. Simulation Results

We begin by describing the data generation results from 100 replications per scenario. Their suitability to the desired conditions can be observed through the median of skewness values in Table 1. We can observe from Table 1 that emissions generated from the MSNBurr distribution with γ = 1 had medians of empirical skewness close to zero. In contrast, the emission distribution generated with γ 1 successfully simulated asymmetric conditions, with the median of skewness deviating from zero. As an illustration, Figure 2 visualizes the data generation output and the corresponding emission histogram, specific to one of the replications in Scen3b.
In general, the parameter estimation procedure constructed based on Algorithm 1 was executed well. The convergence speed in each simulation varied per scenario. To summarize this, we provide Table 2, which contains the number of replications that converged within a certain iteration range. The model was deemed converged when the difference between each parameter’s value in consecutive iterations was smaller than a set threshold, which was chosen to be ϵ = 0.0001 for this study. The first key insight is that sub-scenario (b) required more iterations to achieve convergence than sub-scenario (a). This pattern holds when comparing Scen1a-Scen1b, Scen2a-Scen2b, and Scen3a-Scen3b. The majority of replications in sub-scenario (a) converged before the eleventh iteration. In contrast, sub-scenario (b) took longer to converge (11–50 iterations or even more). This result was expected, given that sub-scenario (a) emissions were conditioned to be fully separated, whereas sub-scenario (b) involved two overlapping emissions and thus required more time to converge.
Our next finding indicates that scenario 1 (two symmetrical emissions) converged more slowly than scenario 2 (one skew emission and one symmetrical) and scenario 3 (two skew emissions). Specifically, up to 81% of replications in Scen1a converged before the eleventh iteration. This percentage was lower than those observed in Scen2a (95%) and Scen3a (100%). Similar patterns were evident when comparing Scen1b, Scen2b, and Scen3b. In Scen1b, only 5% of the replications converged in fewer than 26 iterations. This was markedly different from Scen2b and Scen3b, where the corresponding percentages are 35% and 54%, respectively.
We now turn to the results of parameter estimation and their closeness to the target parameters in each scenario. To summarize the results, we present Table 3, which shows the measures of central tendency for the estimated parameter values. For the parameters in π and A , we computed the mean of the 100 replication estimates as the measure of central tendency, which is shown in the ‘Est.’ column of Table 3. On the other hand, the median was preferred for the emission parameters, given the potential deviation of the mean due to extreme values. This issue is less likely to occur in π and A because their values are restricted to the range [0, 1]. In general, the parameter estimation performed well, as the central tendency estimates closely approximated the true values. This demonstrated the model’s reliable estimation performance.
We next compare the performance of MSNBurr-HMM against Gaussian-HMM and FSSN-HMM as the benchmark models. The model fitting results across all scenarios are presented in Figure 3, Figure 4, Figure 5 and Figure 6. Figure 3 illustrates the violin plots of the log-likelihood values, while Figure 4, Figure 5 and Figure 6 correspond to the AIC, AICc, and BIC values, respectively. These visualizations reveal key information about model performance.
Firstly, the spread of log-likelihood, AIC, AICc, and BIC values for three models appeared fairly wide across all scenarios. However, MSNBurr-HMM typically displayed a narrower spread compared to the Gaussian-HMM and FSSN-HMM, particularly in Scen1b, Scen2a, and Scen3a. This was further confirmed by the Interquartile Range (IQR) presented in Table 4. We can see that the IQRs for the log-likelihood, AIC, AICc, and BIC values of MSNBurr-HMM were all lower than those for Gaussian-HMM. We also encountered a similar case when comparing it with the FSSN-HMM, except in Scen2b. At this point, it can be said that MSNBurr-HMM exhibited more consistent performance overall.
Secondly, across all six scenarios, the log-likelihood values for the MSNBurr-HMM were also visually higher than those for the Gaussian model, indicating that MSNBurr-HMM generally provided a better fit to the data. When switching to the FSSN-HMM, we observe that in certain scenarios such as Scen1b, Scen2b, and Scen3b, the log-likelihood violin plot of the MSNBurr-HMM is visually located higher with a narrower spread, indicating better performance.
This was consistent for the AIC, AICc, and BIC metrics, where visually, MSNBurr-HMM had lower values compared to Gaussian-HMM and FSSN-HMM, especially in scenarios Scen2a, Scen2b, Scen3a, and Scen3b, which involve asymmetric emission conditions. This was supported by the median values of the four metrics in Table 4. The median log-likelihood of MSNBurr-HMM outperformed Gaussian-HMM and FSSN-HMM in all scenarios. For AIC, AICc, and BIC, the superiority of MSNBurr-HMM was also evident in five of the six scenarios, marked by blue colour in Table 4.
However, it should be noted that the first and second points do not account for pairwise comparisons between models and instead only provide an overview of their performance across 100 replications. As a result, the superiority of MSNBurr-HMM still requires further investigation. To address this issue, for each scenario, we conducted four separate one-sided Wilcoxon signed-rank tests (one for each evaluation metric) on paired replications. More precisely, we compared MSNBurr-HMM against Gaussian-HMM and MSNBurr-HMM against FSSN-HMM under each metric, resulting in two pairwise tests per metric for each scenario. We then applied the Bonferroni correction to mitigate the increased risk of Type I errors that arises when multiple hypothesis tests are performed [34].
In the case of the log-likelihood, the null hypothesis stated that the median difference between the log-likelihoods of MSNBurr-HMM and a comparator model was less than or equal to zero (i.e., the log-likelihood of MSNBurr-HMM was not greater than that of Gaussian-HMM). The alternative hypothesis, therefore, was that the median of the difference was greater than zero. A significance level of 0.05 was adopted and subsequently adjusted using the Bonferroni correction. If the p-value was below Bonferroni-adjusted significance level of 0.025 (reflecting two pairwise comparisons: MSNBurr-HMM versus Gaussian-HMM and MSNBurr-HMM versus FSSN-HMM), then the null hypothesis was rejected. This provided statistically significant evidence that the log-likelihood values of MSNBurr-HMM were greater than those of both Gaussian-HMM and FSSN-HMM. On the other hand, we reversed the hypotheses defined for the AIC, AICc, and BIC cases. For both metrics, we defined the null hypothesis as the median of the difference between the AIC (or AICc or BIC) values of MSNBurr-HMM and a comparator model being zero or greater. Consequently, the alternative hypothesis was that the median of the difference was less than zero. This approach was reasonable because the nature of AIC, AICc, and BIC (where smaller values indicate better performance) is opposite to that of log-likelihood.
The results reveal that, for the MSNBurr-HMM versus FSSN-HMM comparison, the null hypothesis was rejected in all scenarios and across all evaluation metrics, confirming that MSNBurr-HMM consistently delivered significantly better performance. For the MSNBurr-HMM versus Gaussian-HMM comparison, the null hypothesis was not rejected in three specific cases: AICc under Scen1b (V = 2071, p-value = 0.059, one-sided 95% CI bounded above at 0.050), BIC under Scen1a (V = 4201, p-value = 1.000, one-sided 95% CI bounded above at 5.318), and BIC under Scen1b (V = 4318, p-value = 1.000, one-sided 95% CI bounded above at 6.283). In these scenarios, the one-sided confidence intervals crossed zero and the p-values exceeded the Bonferroni-adjusted significance level of 0.025, indicating insufficient statistical evidence that MSNBurr-HMM provided a lower AICc or BIC than Gaussian-HMM. In all other scenarios, however, the null hypothesis was decisively rejected. These outcomes are reasonable from a model selection perspective. In Scen1a and Scen1b, both emission distributions were intentionally constructed to be symmetric, meaning that Gaussian-HMM, which assumes symmetric emission behavior, is already compatible with the data-generating process. In contrast, MSNBurr-HMM is more flexible and contains additional parameters to accommodate skewness and heavy-tailed characteristics. Since AICc and BIC explicitly penalize model complexity, this additional flexibility does not translate into a meaningful advantage when the true emission structure is already symmetric. However, in all other scenarios, where at least one emission distribution exhibits asymmetry, the superiority of MSNBurr-HMM becomes evident. This confirms that when the data deviates from symmetry, MSNBurr-HMM is clearly preferable over Gaussian-HMM due to its enhanced ability to capture skewed behavior.
If we shift the analysis to count the number of replications where MSNBurr-HMM outperformed Gaussian-HMM and FSSN-HMM, the results remain largely consistent. Figure 7, Figure 8, Figure 9 and Figure 10 display the number of replications won by each model based on log-likelihood, AIC, AICc, and BIC respectively. Key information can be drawn from these figures. For instance, the top-left panel of Figure 7 exhibits a bar chart showing values of 78 for MSNBurr-HMM, 22 for FSSN-HMM, and 0 for Gaussian-HMM. This indicated that in Scen1a, MSNBurr-HMM achieved highest log-likelihood values in 78 out of 100 replications while the FSSN-HMM claims the best performance in 22 replications. Shortly, MSNBurr-HMM showed better log-likelihood-based performance in most replications. Figure 7 further revealed that the proportion of MSNBurr-HMM’s superior performance was consistent in scenarios 2 and 3. Moreover, as asymmetric emission behavior emerges in Scenarios 2 and 3, Gaussian-HMM becomes increasingly less competitive.
We observed a trend where the dominance of MSNBurr-HMM in Figure 8 becomes increasingly pronounced as the emission structure shifts from symmetric to asymmetric. In Scen1a and Scen1b (both emissions are symmetric), MSNBurr-HMM achieved lowest AIC values in 65 and 45 replications, respectively. These numbers increased to 72 and 64 replications in Scen2a and Scen2b (one asymmetric emission). The trend continued with 78 and 74 replications in Scen3a and Scen3b (both emissions are asymmetric). In contrast, the number of replications won by Gaussian-HMM consistently decreased as the emission setting transitioned from symmetric to asymmetric. A very similar behavior is observed for AICc in Figure 9. Gaussian-HMM still remained relatively competitive under fully symmetric emission conditions, especially under the overlapping case (Scen1b). However, MSNBurr-HMM took over the dominance in the remaining scenarios. A comparable pattern is also found for BIC. In Figure 10, Gaussian-HMM exhibited stronger performance in Scen1a and Scen1b. The situation began to reverse in Scen2a and Scen2b, where MSNBurr-HMM started to outperform the competing models more frequently. Finally, in Scen3a and Scen3b, which involve fully asymmetric emissions, more than 70% of the replications indicated that MSNBurr-HMM attained the lowest BIC.

3.2. Real-Data Example

This section demonstrates the application of the MSNBurr-HMM in modeling the daily closing price movements of the Jakarta Composite Index (JCI) during 2023–2024. This case was selected based on the distributional characteristics of the emissions, as visualized in Figure 11. First, the histogram clearly exhibits a consistent bimodal pattern, indicating the presence of two distinct components that are naturally modeled as two emissions in an HMM framework. Second, each mode displays a non-symmetric (skewed) shape, suggesting that the symmetric Gaussian assumption may be inadequate. The data will be modeled using the MSNBurr-HMM, and the results will be compared with the performance of the FSSN-HMM and the Gaussian-HMM.
The initial values were set to π = ( 0.4 ,   0.6 ) and A = 0.6 0.4 0.4 0.6 . We chose a slightly larger initial-state probability for the second emission ( π 2 0 = 0.6 ) to reflect its visually wider spread around the higher-valued mode, as inferred from the histogram. The diagonal entries of A were set to 0.6 to encode the expectation that remaining in the same state is more likely than switching. These values were kept close to 0.5 to maintain a conservative, minimally biased initialization. For emission parameter initialization, we use the sample mean and standard deviation as references for the location and scale. Specifically, μ 1 0 = O ¯ 0.5   s d O , μ 2 0 = O ¯ + 0.5   s d O , ϕ 1 0 = ϕ 2 0 = s d O , and γ 1 0 = γ 2 0 = 0.1 , where O ¯ and s d O denote the mean and standard deviation of the observable data. The convergence criterion is set to ϵ = 0.0001 . With this setup, the MSNBurr-HMM reached convergence at the ninth iteration, and the resulting parameter estimates are given in Table 5. Figure 12 is also provided to illustrate the trajectory of the parameter estimates up to the point of convergence.
The estimated parameters indicate a clear tendency for the process to start in state 1. Furthermore, the estimated transition probabilities highlight that the process tends to stay in the same state rather than switch to another. Regarding the emission distributions, the first emission exhibits its mode at 6879.568, whereas the second emission attains its mode at 7229.954. The second emission has a wider spread than the first, as reflected by the larger ϕ ^ 2 relative to ϕ ^ 1 . In addition, the two emission distributions differ in their skewness characteristics. The first emission shows left skewness ( γ ^ 1 < 1 ), whereas the second emission displays right skewness ( γ ^ 2 > 1 ). We can observe how the estimated emission distributions align with the data pattern in Figure 13.
The figure confirms that the overall data distribution is well captured by the emission from MSNBurr-HMM. As a further analysis, we present the model performance measures including Log-Likelihood, AIC, AICc, and BIC for the MSNBurr-HMM, FSSN-HMM, and Gaussian-HMM in Table 6. It can be seen that the MSNBurr-HMM consistently outperforms both benchmark models across all evaluated metrics. It achieves the highest log-likelihood of −3060.60, outperforming FSSN-HMM (−3066.02) and Gaussian-HMM (−3083.43). This improvement is also reflected in the lowest AIC value of 6139.19, compared to 6150.04 for FSSN-HMM and 6180.86 for Gaussian-HMM. In terms of AICc, MSNBurr-HMM similarly achieves the smallest value of 6139.57, whereas FSSN-HMM and Gaussian-HMM yield 6150.42 and 6181.10, respectively. The lowest BIC value of 6176.72 further validates the advantage of MSNBurr-HMM in balancing model fit and complexity.

4. Discussion

Several points can be discussed from the findings presented in the previous section. MSNBurr-HMM most clearly outperforms competing models when the underlying emission distributions exhibit asymmetric behavior. In Scenarios 2 and 3, where at least one emission distribution is skewed, MSNBurr-HMM consistently achieves the best performance across log-likelihood, AIC, AICc, and BIC. This demonstrates that the model is particularly effective when the data deviate from Gaussian assumptions, especially in the presence of skewness.
In terms of computational considerations, MSNBurr-HMM indeed involves more parameters than Gaussian-HMM. However, the superiority of MSNBurr-HMM under AIC, AICc, and BIC, all of which incorporate explicit penalties for model complexity, confirms that its improved performance is not simply a consequence of overfitting. When compared to FSSN-HMM, which has the same number of emission parameters, MSNBurr-HMM consistently delivers better results, highlighting its modeling capability. This advantage is further supported in the real-data application, where MSNBurr-HMM not only maintains superior performance but also offers clear interpretability. In particular, the estimated emission parameters provide meaningful insights regarding the mode position, dispersion, and skewness direction of the observed process.
Despite the promising results, this study has several limitations that also present opportunities for future research. First, the simulation design and estimation algorithm are currently restricted to a predetermined number of hidden states. Further exploration is needed on how the MSNBurr-HMM algorithm can be designed to automatically identify the optimal number of hidden states, which would reduce potential bias and enhance the objectivity of the results. Second, this article focuses solely on parameter estimation and has not yet been extended to decode hidden states for each observation. Developing a decoding algorithm and evaluating the accuracy of the MSNBurr-HMM in performing this task would be a valuable direction for future research.

5. Conclusions

We can draw several conclusions from this study. First, MSNBurr-HMM can be effectively estimated using the Baum–Welch Algorithm. Second, the model shows its strongest advantage when the underlying emission distribution is asymmetric. Together, these findings highlight MSNBurr-HMM’s adaptability and effectiveness, confirming its practical potential across diverse modeling contexts.

Author Contributions

Conceptualization, D.B.U.; methodology, D.B.U., N.I. and I.I.; software, D.B.U.; validation, D.B.U., N.I. and I.I.; formal analysis, D.B.U. and N.I.; investigation, D.B.U. and N.I.; resources, D.B.U. and N.I.; data curation, D.B.U.; writing—original draft preparation, D.B.U.; writing—review and editing, N.I. and I.I.; visualization, D.B.U.; supervision, N.I. and I.I.; project administration, D.B.U. and N.I.; funding acquisition, D.B.U. and N.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Lembaga Pengelola Dana Pendidikan (LPDP) through a Doctoral Scholarship (Award Letter No. LOG-19050/LPDP.3/2024).

Data Availability Statement

The data used in the simulation study were generated through simulation procedures described in the paper. All codes required to reproduce the results and regenerate the data are openly available at https://github.com/dbunggul/msnburrHMM-bwa (accessed on 20 June 2025). The daily closing price data for the JCI is available at https://id.investing.com/indices/idx-composite-historical-data (accessed on 15 August 2025).

Acknowledgments

The authors gratefully acknowledge the financial support provided by the Lembaga Pengelola Dana Pendidikan (LPDP). We also thank Department of Statistics Institut Teknologi Sepuluh Nopember for its support throughout this research.

Conflicts of Interest

The authors declare no conflicts of interest.

Nomenclature

The following notations are used in this manuscript:
T Total number of observation periods
K Number of hidden states in the MSNBurr-HMM
O t ;   O 1 : T Observable value at time t , for t = 1,2 , , T ; Full observation sequence O 1 , O 2 , , O T
q t ; q 1 : T Hidden state at time t , for t = 1,2 , , T ; Full hidden state sequence q 1 , q 2 , , q T
λ Full parameter set of MSNBurr-HMM
π Initial state probability
A Transition probability matrix
Θ Emission parameter set
θ i Emission parameter set for i-th hidden state, θ i = μ i , ϕ i , γ i , for i = 1,2 , , K
μ i Location parameter of the emission distribution in the i-th hidden state
ϕ i Scale parameter of the emission distribution in the i-th hidden state
γ i Shape parameter of the emission distribution in the i-th hidden state
α t i Forward variable, for t = 1,2 , , T and i = 1,2 , , K
β t i Backward variable, for t = 1,2 , , T and i = 1,2 , , K
δ t i Expected state occupancy, for t = 1,2 , , T and i = 1,2 , , K
ξ t i , j Expected state transition, for t = 1,2 , , T and i , j = 1,2 , , K
m Iteration index of the BWA
ϵ Convergence tolerance

References

  1. Yu, S.-Z. Hidden Semi-Markov Models. Artif. Intell. 2010, 174, 215–243. [Google Scholar] [CrossRef]
  2. Kouemou, G.L. History and Theoretical Basics of Hidden Markov Models. In Hidden Markov Models, Theory and Applications; Dymarski, P., Ed.; InTech: Houston, TX, USA, 2011; ISBN 978-953-307-208-1. [Google Scholar]
  3. Bouarada, O.; Azam, M.; Amayri, M.; Bouguila, N. Hidden Markov Models with Multivariate Bounded Asymmetric Student’s t-Mixture Model Emissions. Pattern Anal. Appl. 2024, 27, 117. [Google Scholar] [CrossRef]
  4. Mor, B.; Garhwal, S.; Kumar, A. A Systematic Review of Hidden Markov Models and Their Applications. Arch. Comput. Methods Eng. 2021, 28, 1429–1448. [Google Scholar] [CrossRef]
  5. Srivastava, R.K.; Shree, R.; Shukla, A.K.; Pandey, R.P.; Shukla, V.; Pandey, D. A Feature Based Classification and Analysis of Hidden Markov Model in Speech Recognition. In Cyber Intelligence and Information Retrieval; Tavares, J.M.R.S., Dutta, P., Dutta, S., Samanta, D., Eds.; Lecture Notes in Networks and Systems; Springer: Singapore, 2022; Volume 291, pp. 365–379. ISBN 978-981-16-4283-8. [Google Scholar]
  6. Bhar, R.; Hamori, S. Hidden Markov Models: Applications to Financial Economics; Advanced Studies in Theoretical and Applied Econometrics; Kluwer Academic Publishers: Boston, MA, USA, 2004; Volume 40, ISBN 978-1-4020-7899-6. [Google Scholar]
  7. Glennie, R.; Adam, T.; Leos-Barajas, V.; Michelot, T.; Photopoulou, T.; McClintock, B.T. Hidden Markov Models: Pitfalls and Opportunities in Ecology. Methods Ecol. Evol. 2023, 14, 43–56. [Google Scholar] [CrossRef]
  8. Yoon, B.-J. Hidden Markov Models and Their Applications in Biological Sequence Analysis. Curr. Genom. 2009, 10, 402–415. [Google Scholar] [CrossRef]
  9. Ramezani, S.B.; Killen, B.; Cummins, L.; Rahimi, S.; Amirlatifi, A.; Seale, M. A Survey of HMM-Based Algorithms in Machinery Fault Prediction. In Proceedings of the 2021 IEEE Symposium Series on Computational Intelligence (SSCI), Orlando, FL, USA, 5–7 December 2021; IEEE: Orlando, FL, USA, 2021; pp. 1–9. [Google Scholar]
  10. Volant, S.; Bérard, C.; Martin-Magniette, M.-L.; Robin, S. Hidden Markov Models with Mixtures as Emission Distributions. Stat. Comput. 2014, 24, 493–504. [Google Scholar] [CrossRef]
  11. Sowan, B.; Zhang, L.; Matar, N.; Zraqou, J.; Omar, F.; Alnatsheh, A. A Novel Lift Adjustment Methodology for Improving Association Rule Interpretation. Decis. Anal. J. 2025, 15, 100582. [Google Scholar] [CrossRef]
  12. Dash, C.S.K.; Behera, A.K.; Dehuri, S.; Ghosh, A. An Outliers Detection and Elimination Framework in Classification Task of Data Mining. Decis. Anal. J. 2023, 6, 100164. [Google Scholar] [CrossRef]
  13. Murphy, K.P. Machine Learning: A Probabilistic Perspective; Adaptive Computation and Machine Learning Series; MIT Press: Cambridge, MA, USA, 2012; ISBN 978-0-262-01802-9. [Google Scholar]
  14. Iriawan, N. Computationally Intensive Approaches to Inference in Neo-Normal Linear Models. Ph.D. Thesis, Curtin University of Technology, Perth, Australia, 2000. [Google Scholar]
  15. Choir, A.S.; Iriawan, N.; Ulama, B.S.S.; Dokhi, M. MSEPBurr Distribution: Properties and Parameter Estimation. Pak. J. Stat. Oper. Res. 2019, 15, 179–193. [Google Scholar] [CrossRef]
  16. Azzalini, A. A Class of Distributions Which Includes the Normal Ones. Scand. J. Stat. 1985, 12, 171–178. [Google Scholar]
  17. Azzalini, A.; Capitanio, A. Distributions Generated by Perturbation of Symmetry with Emphasis on a Multivariate Skew T-Distribution. J. R. Stat. Soc. Ser. B Stat. Methodol. 2003, 65, 367–389. [Google Scholar] [CrossRef]
  18. Alruwaili, B. The Modality of Skew T-Distribution. Stat. Pap. 2023, 64, 497–507. [Google Scholar] [CrossRef]
  19. Pravitasari, A.A.; Iriawan, N.; Fithriasari, K.; Purnami, S.W.; Irhamah; Ferriastuti, W. A Bayesian Neo-Normal Mixture Model (Nenomimo) for MRI-Based Brain Tumor Segmentation. Appl. Sci. 2020, 10, 4892. [Google Scholar] [CrossRef]
  20. Rasyid, D.A.; Iriawan, N.; Mashuri, M. On the Flexible Neo-Normal MSAR MSN-Burr Control Chart in Air Quality Monitoring. Air Soil Water Res. 2024, 17, 11786221241272391. [Google Scholar] [CrossRef]
  21. Nuraini, U.S.; Iriawan, N.; Fithriasari, K.; Hidayat, T. Ising Neo-Normal Model (INNM) for Segmentation of Cardiac Ultrasound Imaging. In Proceedings of the 2024 Beyond Technology Summit on Informatics International Conference (BTS-I2C), Jember, Indonesia, 19 December 2024; IEEE: Jember, Indonesia, 2024; pp. 298–303. [Google Scholar]
  22. Welch, L.R. Hidden Markov Models and the Baum-Welch Algorithm. IEEE Inf. Theory Soc. Newsl. 2003, 53, 10–13. [Google Scholar]
  23. Akaike, H. A New Look at the Statistical Model Identification. IEEE Trans. Autom. Control 1974, 19, 716–723. [Google Scholar] [CrossRef]
  24. Hurvich, C.M.; Tsai, C.-L. Regression and Time Series Model Selection in Small Samples. Biometrika 1989, 76, 297–307. [Google Scholar] [CrossRef]
  25. Schwarz, G. Estimating the Dimension of a Model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  26. Fernández, C.; Steel, M.F.J. On Bayesian Modeling of Fat Tails and Skewness. J. Am. Stat. Assoc. 1998, 93, 359–371. [Google Scholar] [CrossRef]
  27. Castillo, N.O.; Gómez, H.W.; Leiva, V.; Sanhueza, A. On the Fernández–Steel Distribution: Inference and Application. Comput. Stat. Data Anal. 2011, 55, 2951–2961. [Google Scholar] [CrossRef]
  28. Rantini, D.; Iriawan, N.; Irhamah. Fernandez–Steel Skew Normal Conditional Autoregressive (FSSN CAR) Model in Stan for Spatial Data. Symmetry 2021, 13, 545. [Google Scholar] [CrossRef]
  29. Rabiner, L.R. A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition. Proc. IEEE 1989, 77, 257–286. [Google Scholar] [CrossRef]
  30. Chen, Z.; Li, Y.; Xia, T.; Pan, E. Hidden Markov Model with Auto-Correlated Observations for Remaining Useful Life Prediction and Optimal Maintenance Policy. Reliab. Eng. Syst. Saf. 2019, 184, 123–136. [Google Scholar] [CrossRef]
  31. Baum, L.E.; Petrie, T.; Soules, G.; Weiss, N. A Maximization Technique Occurring in the Statistical Analysis of Probabilistic Functions of Markov Chains. Ann. Math. Stat. 1970, 41, 164–171. [Google Scholar] [CrossRef]
  32. Yang, F.; Balakrishnan, S.; Wainwright, M.J. Statistical and Computational Guarantees for the Baum-Welch Algorithm. J. Mach. Learn. Res. 2017, 18, 1–53. [Google Scholar]
  33. Nirmal, A.; Jayaswal, D.; Kachare, P.H. A Hybrid Bald Eagle-Crow Search Algorithm for Gaussian Mixture Model Optimisation in the Speaker Verification Framework. Decis. Anal. J. 2024, 10, 100385. [Google Scholar] [CrossRef]
  34. Armstrong, R.A. When to Use the Bonferroni Correction. Ophthalmic Physiol. Opt. 2014, 34, 502–508. [Google Scholar] [CrossRef]
Figure 1. Density of the MSNBurr distribution for μ = 0 , ϕ = 1 , and various values of γ : (a) γ = 1 ; (b) γ = 0.5 ; (c) γ = 2 ; (d) γ = 0.2 ; (e) γ = 5 ; (f) γ = 0.1 ; (g) γ = 10 .
Figure 1. Density of the MSNBurr distribution for μ = 0 , ϕ = 1 , and various values of γ : (a) γ = 1 ; (b) γ = 0.5 ; (c) γ = 2 ; (d) γ = 0.2 ; (e) γ = 5 ; (f) γ = 0.1 ; (g) γ = 10 .
Symmetry 17 01931 g001
Figure 2. Illustration of generated data and emission histogram in the Scen3b.
Figure 2. Illustration of generated data and emission histogram in the Scen3b.
Symmetry 17 01931 g002
Figure 3. Violin plots of log-likelihood values from 100 replications per scenario.
Figure 3. Violin plots of log-likelihood values from 100 replications per scenario.
Symmetry 17 01931 g003
Figure 4. Violin plots of AIC values from 100 replications per scenario.
Figure 4. Violin plots of AIC values from 100 replications per scenario.
Symmetry 17 01931 g004
Figure 5. Violin plots of AICc values from 100 replications per scenario.
Figure 5. Violin plots of AICc values from 100 replications per scenario.
Symmetry 17 01931 g005
Figure 6. Violin plots of BIC values from 100 replications per scenario.
Figure 6. Violin plots of BIC values from 100 replications per scenario.
Symmetry 17 01931 g006
Figure 7. The number of replications where each HMM outperforms the others based on the log-likelihood.
Figure 7. The number of replications where each HMM outperforms the others based on the log-likelihood.
Symmetry 17 01931 g007
Figure 8. The number of replications where each HMM outperforms the others based on the AIC.
Figure 8. The number of replications where each HMM outperforms the others based on the AIC.
Symmetry 17 01931 g008
Figure 9. The number of replications where each HMM outperforms the others based on the AICc.
Figure 9. The number of replications where each HMM outperforms the others based on the AICc.
Symmetry 17 01931 g009
Figure 10. The number of replications where each HMM outperforms the others based on the BIC.
Figure 10. The number of replications where each HMM outperforms the others based on the BIC.
Symmetry 17 01931 g010
Figure 11. Observable movement of daily JCI closing prices from 2023 to 2024 along with its histogram.
Figure 11. Observable movement of daily JCI closing prices from 2023 to 2024 along with its histogram.
Symmetry 17 01931 g011
Figure 12. Trajectory of the parameter estimates up to the point of convergence indicated by the red dotted line.
Figure 12. Trajectory of the parameter estimates up to the point of convergence indicated by the red dotted line.
Symmetry 17 01931 g012
Figure 13. Histogram of the Daily JCI closing price with two estimated emission density functions.
Figure 13. Histogram of the Daily JCI closing price with two estimated emission density functions.
Symmetry 17 01931 g013
Table 1. Characteristics summary of generated data for all replications per scenario.
Table 1. Characteristics summary of generated data for all replications per scenario.
ScenarioTarget of Emission DistributionMedian of Empirical Skewness
Hidden State 1Hidden State 2Hidden State 1Hidden State 2
Scen1aMSNBurr(2, 1, 1)MSNBurr(10, 1, 1)−0.01590.0584
Scen1bMSNBurr(4, 1, 1)MSNBurr(7, 1, 1)−0.0110−0.0834
Scen2aMSNBurr(2, 1, 1)MSNBurr(10, 1, 5)−0.04660.8220
Scen2bMSNBurr(4, 1, 1)MSNBurr(7, 1, 5)−0.09290.8865
Scen3aMSNBurr(2, 1, 0.5)MSNBurr(10, 1, 5)−0.64430.8058
Scen3bMSNBurr(4, 1, 0.5)MSNBurr(7, 1, 5)−0.76810.8264
Table 2. Number of replications converging within specific iteration ranges for each scenario.
Table 2. Number of replications converging within specific iteration ranges for each scenario.
Iterations to ConvergenceScen1aScen1bScen2aScen2bScen3aScen3b
Below 11 Iterations8109511000
11–25 Iterations195434054
26–50 Iterations064148041
Above 50 Iterations03101705
Table 3. Measure of central tendency for parameter estimate values.
Table 3. Measure of central tendency for parameter estimate values.
ParameterScen1aScen1bScen2aScen2bScen3aScen3b
TargetEst.TargetEst.TargetEst.TargetEst.TargetEst.TargetEst.
π 1 0.500.470.500.530.500.470.500.500.500.470.500.52
π 2 0.500.530.500.470.500.530.500.500.500.530.500.48
a 11 0.700.680.700.710.700.680.700.700.700.690.700.71
a 12 0.300.320.300.290.300.320.300.300.300.310.300.29
a 21 0.200.210.200.210.200.210.200.200.200.210.200.20
a 22 0.800.790.800.790.800.790.800.800.800.790.800.80
μ 1 2.001.984.003.992.002.014.004.022.001.974.004.02
μ 2 10.009.987.006.9610.009.997.006.9810.0010.017.007.00
ϕ 1 1.000.981.000.991.000.981.000.971.001.001.000.98
ϕ 2 1.001.031.000.951.000.991.000.991.001.001.000.98
γ 1 1.001.061.001.061.000.941.000.940.500.530.500.48
γ 2 1.001.071.001.165.005.825.006.685.005.995.006.15
Table 4. Comparison of the medians and IQRs of performance metrics between MSNBurr-HMM, Gaussian-HMM, and FSSN-HMM.
Table 4. Comparison of the medians and IQRs of performance metrics between MSNBurr-HMM, Gaussian-HMM, and FSSN-HMM.
ScenarioModelLog-LikelihoodAICAICcBIC
MedianIQRMedianIQRMedianIQRMedianIQR
Scen1aMSNBurr-HMM−416.9318.77851.8637.54852.8137.54881.5537.54
FSSN-HMM−417.5818.91853.1637.81854.1137.81882.8537.81
Gaussian-HMM−418.8018.80851.5937.60852.1837.60874.6837.60
Scen1bMSNBurr-HMM−377.6114.68773.2229.36774.1729.36802.9129.36
FSSN-HMM−389.2617.44796.5234.88797.4734.88826.2134.88
Gaussian-HMM−381.2317.51776.4635.02777.0435.02799.5435.02
Scen2aMSNBurr-HMM−409.6819.91837.3739.83838.3239.83867.0539.83
FSSN-HMM−412.3221.94842.6543.89843.6043.89872.3343.89
Gaussian-HMM−418.8922.74851.7945.47852.3745.47874.8845.47
Scen2bMSNBurr-HMM−387.6620.52793.3241.04794.2741.04823.0041.04
FSSN-HMM−393.5319.35805.0638.69806.0138.69834.7538.69
Gaussian-HMM−395.1021.37804.1942.75804.7842.75827.2842.75
Scen3aMSNBurr-HMM−417.5017.04852.9934.09853.9434.09882.6834.09
FSSN-HMM−419.7619.65857.5239.29858.4739.29887.2139.29
Gaussian-HMM−429.2021.37872.4142.75872.9942.75895.4942.75
Scen3bMSNBurr-HMM−400.6718.10819.3436.19820.2936.19849.0236.19
FSSN-HMM−404.8018.43827.5936.85828.5436.85857.2836.85
Gaussian-HMM−408.7018.64831.4137.28831.9937.28854.5037.28
The blue-highlighted cells in the median column indicated the best model (higher log-likelihood or lower AIC, AICc, BIC), while those in the IQR column represented smaller measures of dispersion.
Table 5. Parameter estimates of the MSNBurr-HMM for the daily JCI closing prices data from 2023 to 2024.
Table 5. Parameter estimates of the MSNBurr-HMM for the daily JCI closing prices data from 2023 to 2024.
ParameterEstimated Value
π 1 1
π 2 1 × 10−12
a 11 0.991
a 12 0.009
a 21 0.004
a 22 0.996
μ 1 6879.568
μ 2 7229.954
ϕ 1 88.677
ϕ 2 170.374
γ 1 0.348
γ 2 27,871.600
Table 6. Performance of the three models based on Log-Likelihood, AIC, AICc, and BIC.
Table 6. Performance of the three models based on Log-Likelihood, AIC, AICc, and BIC.
ModelLog-LikelihoodAICAICcBIC
MSNBurr-HMM−3060.606139.196139.576176.72
FSSN-HMM−3066.026150.046150.426187.56
Gaussian-HMM−3083.436180.866181.106210.05
The blue-highlighted cells indicate the best performance (the highest log-likelihood or the lowest AIC, AICc, and BIC).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Unggul, D.B.; Iriawan, N.; Irhamah, I. Parameter Estimation of MSNBurr-Based Hidden Markov Model: A Simulation Study. Symmetry 2025, 17, 1931. https://doi.org/10.3390/sym17111931

AMA Style

Unggul DB, Iriawan N, Irhamah I. Parameter Estimation of MSNBurr-Based Hidden Markov Model: A Simulation Study. Symmetry. 2025; 17(11):1931. https://doi.org/10.3390/sym17111931

Chicago/Turabian Style

Unggul, Didik Bani, Nur Iriawan, and Irhamah Irhamah. 2025. "Parameter Estimation of MSNBurr-Based Hidden Markov Model: A Simulation Study" Symmetry 17, no. 11: 1931. https://doi.org/10.3390/sym17111931

APA Style

Unggul, D. B., Iriawan, N., & Irhamah, I. (2025). Parameter Estimation of MSNBurr-Based Hidden Markov Model: A Simulation Study. Symmetry, 17(11), 1931. https://doi.org/10.3390/sym17111931

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop