Next Article in Journal
Joint Event Density and Curvature Within Spatio-Temporal Neighborhoods-Based Event Camera Noise Reduction and Pose Estimation Method for Underground Coal Mine
Previous Article in Journal
Geometric Properties of a General Kohn–Nirenberg Domain in ℂn
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feedforward Factorial Hidden Markov Model

School of Computer Information Engineering, Hanshan Normal University, Chaozhou 521041, China
*
Authors to whom correspondence should be addressed.
Mathematics 2025, 13(7), 1201; https://doi.org/10.3390/math13071201
Submission received: 12 November 2024 / Revised: 1 April 2025 / Accepted: 2 April 2025 / Published: 5 April 2025

Abstract

:
This paper introduces a novel kind of factorial hidden Markov model (FHMM), specifically the feedforward FHMM (FFHMM). In contrast to traditional FHMMs, the FFHMM is capable of directly utilizing supplementary information from observations through predefined states, which are derived using an automatic feature filter (AFF). We investigate two variations of FFHMM models that integrate predefined states with the FHMM: the direct FFHMM and the embedded FFHMM. In the direct FFHMM, alterations to one sub-hidden Markov model (HMM) do not affect the others, enabling individual improvements in HMM estimation. On the other hand, the sub-HMM chains within the embedded FFHMM are interconnected, suggesting that adjustments to one HMM chain may enhance the estimations of other HMM chains. Consequently, we propose two algorithms for these FFHMM models to estimate their respective hidden states. Ultimately, experiments conducted on two real-world datasets validate the efficacy of the proposed models and algorithms.

1. Introduction

The factorial hidden Markov model (FHMM) represents a variant of the Hidden Markov Model (HMM) that incorporates distributed state representations [1]. The latent state space encompasses multiple chains that evolve concurrently, collectively influencing the single observation at any given moment [2]. Due to its ability to model outcomes influenced by several independent, time-varying processes, the FHMM is suitable for a wide range of applications. For instance, the FHMM has been successfully utilized in various fields, including visual tracking [3], non-intrusive load monitoring (NILM) [4], consumers’ choice models [5], natural language processing [6], estimating tumor clonal composition [7], fault diagnosis [8], online skill rating [9], random telegraph noise [10], forecasting financial portfolios with value at risk [11], and pitch and formant tracking [12].
According to [13], an additive FHMM model is defined as
Model 1. 
x 1 ( i ) Cat ( ϕ ( i ) ) ,
x t ( i ) | x t 1 ( i ) Cat ( P x t 1 ( i ) x t ( i ) ( i ) ) ,
y ¯ t | x t ( 1 : N ) N i = 1 N μ x t ( i ) ( i ) , Σ ,
In this model, N Z + is the number of HMM chains; m i Z + is the number of states in the i-th HMM; x t ( i ) { 1 , , m i } denotes the state of the i-th HMM at time t; ϕ ( i ) [ 0 , 1 ] m i is the initial state distribution for the i-th HMM, while ϕ j i ( i ) [ 0 , 1 ] is the initial state distribution for the i-th HMM at state j; P ( i ) [ 0 , 1 ] m i × m i is the transition matrix for the i-th HMM, and its element P x t 1 ( i ) x t ( i ) ( i ) [ 0 , 1 ] represents the probability of the i-th HMM being transferred from state x t 1 ( i ) to state x t ( i ) ; y ¯ t R N is the aggregated observation; μ j i ( i ) R N is the mean of the i-th HMM for state j i ; and Σ R N × N is the covariance matrix. More specifically, Equation (3) can be explicitly written as
Pr ( y ¯ t | x t ) = 1 ( 2 π ) N det ( Σ ) exp 1 2 ( y ¯ t μ t ) T Σ 1 ( y ¯ t μ t ) ,
where μ t = i = 1 N μ x t ( i ) ( i ) and x t = [ x t ( 1 ) , , x t ( N ) ] T . A scheme of the FHMM model is depicted in Figure 1a. Then, the task of the FHMM is to estimate the states of all HMMs, which is illustrated in Figure 1b.
To achieve optimal estimates, the Viterbi algorithm as described in [14] for Hidden Markov Models (HMMs) can be adapted for use with FHMMs, resulting in the EXACT algorithm. Nonetheless, the EXACT algorithm necessitates the enumeration of an exponential number of states, leading to significant computational demands and highlighting the need for approximate algorithms. Even particle approximations via particle filters [15] suffer from the curse of dimensionality in high-dimensional FHMMs. As the dimensionality increases, the state space becomes increasingly sparse, and the number of particles required to accurately represent the posterior distribution grows exponentially. A possible solution can be found in localised particle filters [16] and localised forward–backward algorithms [17], which trade-off accuracy to overcome the curse of dimensionality. As an alternative,  Ref. [13] introduces an approximate approach termed the additive factorial approximate MAP (AFAMAP), designed to globally optimize state estimates within FHMMs. AFAMAP offers three advantages over other approximation techniques: (1) it accounts for the variations in observations between successive time points, (2) it employs robust component accounting to manage unmodeled observations effectively, and (3) it transforms the problem into a convex optimization task with a guaranteed global optimum.
Nevertheless, in numerous instances, the FHMM struggles to detect hidden state variables that exhibit notable patterns but possess small magnitudes. In such scenarios, during specific time intervals, state variables with distinct patterns can be readily identified using alternative methods. In this paper, we define a pattern as significant if it is readily discernible from observations, such as the square-wave electricity consumption of a refrigerator, illustrated in Figure 2a between 1000 and 1200 min. To address these limitations of the FHMM, this paper proposes a feedforward FHMM (FFHMM) that employs predefined states via an automatic feature filter (AFF). The term “feedforward” is derived from feedforward control systems [18], referring to the predefined states in our context. In this study, we only utilize the AFF to extract information from testing data. Nonetheless, other forms of prior knowledge can also be employed to determine the predefined states. Consequently, our FFHMM framework can improve the performance of the FHMM by incorporating various additional pieces of information.
The remainder of this paper is structured as follows. In Section 2, we first introduce an AFF designed to identify significant patterns and treat them as predefined states. Subsequently, we present our key contributions: the direct FFHMM and embedded FFHMM. We also outline the exact and approximate algorithms associated with these models. The efficacy of the FFHMM models is evaluated in Section 3 through rigorous experimentation on two distinct datasets. Lastly, Section 4 summarizes the key points and findings of this paper.

2. Feedforward FHMM

In this section, we first introduce an automatic feature filter (AFF) aimed at extracting notable patterns from the aggregated observations—patterns that are discernible yet not accurately identified by the FHMM. The AFF subsequently facilitates the determination of certain predefined states. We then propose a novel model, the feedforward FHMM (FFHMM), which integrates the FHMM with these predefined states. Depending on the method of integration, we discuss two variations: the direct FFHMM and the embedded FFHMM.

2.1. Automatic Feature Filter

As previously discussed, the FHMM tends to overlook numerous significant patterns present in the observations, which could otherwise be utilized to refine its estimates for individual HMM chains. To address this issue, we have developed the following automatic feature filter:
  • If for all j 1 , j 2 = 1 , , k i such that j 1 j 2 , we have g j 1 ( i ) ( y ¯ ( t ) ) = g j 2 ( i ) ( y ¯ ( t ) ) , then
    x ˜ t ( i ) = g j 1 ( i ) ( y ¯ ( t ) ) ,
  • Otherwise,
    x ˜ t ( i ) = 0 ,
where x ˜ t ( i ) is the estimated status from the AFF for the i-th HMM chain at time t. g j ( i ) ( y ¯ ( t ) ) is a selection function, whose output is the corresponding state of the i-th HMM chain if its j-th feature has been detected from the aggregated observation at time t. k i is the number of features we used for detecting the i-th HMM chain.
Figure 3 presents an illustration of the AFF method. Here, we employ green-colored operational windows with specified width w and height h as the features for detection purposes. The filter automatically retrieves the most similar pairs of w and h values. Following the clustering process, the selection function g j ( i ) ( y ¯ ( t ) ) is responsible for determining the predefined state at time t.

2.2. Direct FFHMM

Now, we examine two distinct approaches to integrating predefined states with the FHMM: first, a simple concatenation method, followed by a more intricate technique. Alongside these, we also introduce the EXACT algorithms and their derivatives, the AFAMAP algorithms.
Firstly, the straightforward method involves the direct fusion of estimates from the FHMM and AFF. Specifically, this entails substituting the non-zero estimates from the AFF into the corresponding positions within the FHMM. This approach gives rise to what we term the direct FFHMM model:
Model 2. 
During t [ 1 , T ] , every individual HMM chain i { 1 , , N } satisfies
x ^ t ( i ) = x ˜ t ( i ) , if   x ˜ t ( i ) > 0 , x t ( i ) , otherwise ,
x 1 ( i ) Cat ( ϕ ( i ) ) ,
x t ( i ) | x t 1 ( i ) Cat ( P x t 1 ( i ) x t ( i ) ( i ) ) ,
y ¯ t | x t ( 1 : N ) N i = 1 N μ x t ( i ) ( i ) , Σ ,
where x ^ t ( i ) { 1 , , m i } denotes the output estimation provided by the direct FFHMM for the i-th HMM at time t; x ˜ t ( i ) { 0 , 1 , , m i } represents the output from the AFF for the i-th HMM at time t, with  x ˜ t ( i ) = 0 indicating that no state has been retrieved by the AFF from the aggregated observation for the i-th HMM at time t; and finally, x t ( i ) { 1 , , m i } signifies the output of the FHMM as described in Model 1, which corresponds to the optimal estimation for the i-th HMM at time t.
Figure 4 presents a schematic representation of the direct FFHMM approach. According to this figure, one can readily discern that the estimation for the individual HMM, denoted as x ^ t ( i ) , is chiefly reliant on the FHMM’s corresponding estimate x t * ( i ) . Nevertheless, there are instances at particular time points where the AFF yields non-zero estimations; in such cases, these AFF-derived estimates supersede those originating from the FHMM. Then, an EXACT algorithm for the direct FFHMM is proposed in Algorithm 1.
Algorithm 1: EXACT algorithm for direct FFHMM
1Initialization:
δ 1 ( j ) = i = 1 N ϕ j i ( i ) Pr ( y ¯ 1 | x 1 ( j ) ) ,
where j = [ j 1 , , j N ] T , j i { 1 , , m i } , 1 i N ,
x 1 ( j ) = [ x 1 ( 1 ) = j 1 , , x 1 ( N ) = j N ] T .
2Recursion:
δ t ( j ) = max k i = 1 N δ t 1 ( k ) P k i j i ( i ) Pr ( y ¯ t | x t ( j ) ) ,
ψ t ( j ) = arg max k i = 1 N δ t 1 ( k ) P k i j i ( i ) ,
where 2 t T , k = [ k 1 , , k N ] T , k i { 1 , , m i } , 1 i N ,
x t ( j ) = [ x t ( 1 ) = j 1 , , x t ( N ) = j N ] T .
3Termination:
q T * = arg max k i = 1 N δ t 1 ( k ) P k i j i ( i ) .
4Path backtracking:
q t * = ψ t + 1 ( q t + 1 * ) , t = T 1 , , 1 .
5Merging:
q ^ t ( i ) = q ˜ t ( i ) , if   q ˜ t ( i ) > 0 , q t * ( i ) , otherwise .
In Algorithm 1, q ^ t ( i ) { 1 , , m i } is the index of the estimation of the FFHMM for the i-th HMM at time t; q ˜ t ( i ) { 0 , 1 , , m i } is the index of the estimation of the AFF for the i-th HMM at time t, where q ˜ t ( i ) = 0 means no state has been retrieved by the AFF from the aggregated observation for the i-th HMM at time t; and q t * ( i ) { 1 , , m i } is the index of the optimal estimation for the i-th HMM at time t by the FHMM.
Evidently, Algorithm 1 employs the Viterbi algorithm within the proposed direct FFHMM. In this algorithm, all possible states are traced through the application of the dynamic programming algorithm approach. However, this method necessitates the enumeration of an exponentially large number of states, which in turn gives rise to substantial computational demands. In light of these computational challenges, we introduce an approximation, specifically Algorithm 2, with the aim of mitigating the heavy computational burden imposed by Algorithm 1.
Algorithm 2: AFAMAP algorithm for direct FFHMM
1Input:  x ˜ 1 : T ( 1 : N ) , y ¯ 1 : T , μ 1 : m i ( 1 : N ) , Σ 1 , Σ 2 , λ 1 and λ 2 .
2Minimization:
min S L O , z 1 : T = f ( S , z 1 : T )
3Output:  y ^ 1 : T ( 1 : N ) , predicted the output of individual HMM
y ^ t ( i ) = μ x ˜ t ( i ) ( i ) , if   x ˜ t ( i ) > 0 , l = 1 m i μ l ( i ) Q ( x t * ( i ) ) l , otherwise .
In Algorithm 2, y ¯ 1 : T is the aggregated observation; μ 1 : m i ( 1 : N ) represents state means for N HMMs; Σ 1 and Σ 2 are the covariance matrices for the additive FHMM and difference FHMM, respectively; and λ 1 and λ 2 are regularization parameters. The cost function f in (17) is defined as
f ( S , z 1 : T ) = 1 2 t | | y ¯ t Σ 1 2 z t i , j μ j ( i ) Q ( x t ( i ) ) j | | Σ 1 2 + 1 2 t , i , j , k j | | Δ y ¯ t Δ μ k , j ( i ) | | Σ 1 2 Q ( x t 1 ( i ) , x t ( i ) ) j , k + 1 2 t D ( Σ 1 2 Δ y ¯ t , λ 2 ) 1 i , j Q ( x t 1 ( i ) , x t ( i ) ) j , j + t , i , j , k Q ( x t 1 ( i ) , x t ( i ) ) j , k ( log P k , j ( i ) ) + λ 1 t z t z t 1 1 ,
where Δ y ¯ t and Δ μ j , k ( i ) represents the changes in observations and states, given by
Δ y ¯ t | x t ( 1 : N ) , x t 1 ( 1 : N ) N i = 1 N μ x t ( i ) μ x t 1 ( i ) , Σ ,
and
Δ μ j , k ( i ) = μ j ( i ) μ k ( i ) .
Meanwhile, the set S of indicators is defined as
S = i { 1 , , N } t { 1 , , T } Q ( x t ( i ) ) { 0 , 1 } m i Q ( x t 1 ( i ) , x t ( i ) ) { 0 , 1 } m i × m i
in which Q ( x t ( i ) ) j = 1 x t ( i ) = j , Q ( x t 1 ( i ) , x t ( i ) ) j , k = 1 x t 1 ( i ) = j , x t ( i ) = k , and t { 1 , , T } ; i { 1 , , N } ; j , k { 1 , , m i } . The minimization constraints are
L = S : j Q ( x t ( i ) ) j = 1 k Q ( x t 1 ( i ) , x t ( i ) ) j , k = Q ( x t 1 ( i ) ) j k Q ( x t 1 ( i ) , x t ( i ) ) k , j = Q ( x t ( i ) ) j Q ( x t ( i ) ) j , Q ( x t 1 ( i ) , x t ( i ) ) k , j 0 .
In addition, the one-at-a-time condition leads to an extra constraint as
O = S : i , j , k j Q ( x t 1 ( i ) , x t ( i ) ) j , k 1 ,
which means the off-diagonal elements of Q ( x t 1 ( i : N ) , x t ( i : N ) ) sum to at most one. The last term z t z t 1 1 in the cost function is the robust component. It is used to handle unmodeled observations.
Clearly, the direct FFHMM possesses two notable advantages:
  • Firstly, it simplifies the control over outcomes for a designated HMM chain. Specifically, as the AFF’s estimations directly supersede those of the FHMM during particular timeframes, the accuracy of the FFHMM for the target HMM chain is unlikely to decline, assuming the AFF’s accuracy is maintained with sufficient conservatism. Secondly, an aggressive AFF can be employed to enhance the performance of the target HMM chain, though this approach carries a risk of potentially exacerbating outcomes.
  • Secondly, alterations in accuracy for the estimation of one HMM chain are independent of changes in others. This attribute facilitates the independent improvement in accuracy for each HMM chain. Additionally, customized AFFs can be applied to specific HMM chains, thereby achieving superior results.

2.3. Embedded FFHMM

In this subsection, we employ an alternative method to integrate predefined states with the FHMM, whereby the non-zero estimations from the AFF take the place of the hidden states within the embedded HMMs of the FHMM model. The structure of the embedded FFHMM can be described as follows:
Model 3. 
1. 
The initial state:
(a) 
If x ˜ 1 ( i ) > 0 , then
x 1 ( i ) = x ˜ 1 ( i ) ,
(b) 
Otherwise,
x 1 ( i ) Cat ( ϕ ( i ) )
2. 
The state transition:
(a) 
If x ˜ t 1 ( i ) > 0 , then
x t ( i ) = x ˜ t ( i ) ,
(b) 
Otherwise,
x t ( i ) | x t 1 ( i ) Cat ( P x t 1 ( i ) x t ( i ) ( i ) )
3. 
The observation probability distribution:
y ¯ t | x t ( 1 : N ) N i = 1 N μ x t ( i ) ( i ) , Σ ,
The notations used in this model can be understood in reference to the preceding discussions. Figure 5 presents a diagram of the embedded FFHMM, where a hidden state x t ( i ) from the FHMM is substituted by the non-zero estimation x ˜ t i from the AFF for the i-th HMM at time t. The FFHMM’s output x ^ t ( i ) is the optimal estimation, determined by the observation probability distribution, as described in (29). Subsequently, we introduce the EXACT and AFAMAP algorithms designed for the embedded FFHMM, as outlined in Algorithms 3 and Algorithm 4, respectively.
Algorithm 3: EXACT algorithm for embedded FFHMM
1Initialization:  j = [ j 1 , , j N ] T , 1 i N ,
  • if x ˜ 1 ( i ) > 0 , then
    j i = x ˜ 1 ( i ) ,
  • otherwise,
    j i { 1 , , m i } .
Then, we initialize δ 1 ( j ) by (11), which is the same as Algorithm 1,
2Recursion:  2 t T , k = [ k 1 , , k N ] T , k i { 1 , , m i } , 1 i N ,
  • if x ˜ t ( i ) > 0 , then
    j i = x ˜ t ( i ) ,
    P k i j i ( i ) = 1 ,
  • if x ˜ t ( i ) = 0 but x ˜ t 1 ( i ) > 0 , then
    k i = x ˜ t 1 ( i ) ,
  • otherwise,
    j i { 1 , , m i } , k i { 1 , , m i } .
Then, we recur δ t ( j ) and ψ t ( j ) using (12) and (13), which is the same as Algorithm 1.
3Termination:
q T * = arg max k i = 1 N δ t 1 ( k ) P k i j i ( i ) .
4Path backtracking:
q t * = ψ t + 1 ( q t + 1 * ) , t = T 1 , , 1 .
In Algorithm 4, constraint L ^ is defined as
L ^ = S : Q ( x t ( i ) ) x ˜ t ( i ) = 1   for   all   x ˜ t ( i ) > 0 j Q ( x t ( i ) ) j = 1 k Q ( x t 1 ( i ) , x t ( i ) ) j , k = Q ( x t 1 ( i ) ) j k Q ( x t 1 ( i ) , x t ( i ) ) k , j = Q ( x t ( i ) ) j Q ( x t ( i ) ) j , Q ( x t 1 ( i ) , x t ( i ) ) k , j 0 .
Algorithm 4: AFAMAP algorithm for embedded FFHMM
1Input:  x ˜ 1 : T ( 1 : N ) , y ¯ 1 : T , μ 1 : m i ( 1 : N ) , Σ 1 , Σ 2 , λ 1 and λ 2 .
2Minimization:
min S L ^ O , z 1 : T = f ( S , z 1 : T )

Output:  y 1 : T * ( 1 : N ) , predicted individual HMM output
y t * ( i ) = l = 1 m i μ l ( i ) Q ( x t ( i ) ) l
As previously discussed in the earlier subsection, the estimations of the direct FFHMM for the target HMM chain are independent of other HMM chains. In contrast, the estimations of the embedded FFHMM for the target HMM chain are interconnected with other HMM chains via the observation probability distribution described in (29). Due to these interdependencies, enhancing the accuracy of the AFF for one HMM can indirectly influence the estimation accuracies of other HMMs. This represents a significant advantage of the embedded FFHMM approach. However, this also necessitates greater care in managing the embedded FFHMM, as these correlations can lead to unforeseen changes in the estimations of other HMM chains.
In summary, given their distinct characteristics, the following guidelines can be followed when selecting between the direct FFHMM and embedded FFHMM: Firstly, opt for the direct FFHMM if the goal is to independently improve the estimation of a single HMM chain. Secondly, choose the embedded FFHMM if the aim is to enhance not only the estimation of an individual HMM chain but also the estimations across multiple HMM chains simultaneously.
The computational complexity of the Viterbi algorithm applied to the FHMM is
O i = 1 N m i 2 T ,
where N is the number of HMMs, m i is the number of states in the i-th HMM, and T is the number of time steps. This exponential complexity makes the Viterbi algorithm intractable for large FHMMs. When contrasted with the exponential complexity, the cost associated with implementing the AFF is so minimal that it can be considered negligible. As a result, the computational complexities of the EXACT algorithms in Algorithms 1 and 3 are identical to those expressed in Equation (40), both of which exhibit an exponential-in-dimension computational cost.
Similarly, as per the findings in [13], the computational complexity of AFAMAP for the additive FHMM is approximately linear with respect to the number of HMMs. Consequently, the AFAMAP algorithms for the FFHMM presented in Algorithms 2 and 4 share the same computational complexity.

3. Results

In this section, we implement our models and algorithms to address the non-intrusive load monitoring (NILM) problem, showcasing their advantages over the traditional FHMM. The evaluations are conducted using the Pecan Street dataset [19] and the REDD [20] dataset.
The Pecan dataset comprises daily electricity consumption data from households located in the green-built Mueller community in Austin, Texas. This dataset encompasses both aggregated and appliance-level consumption data, collected at a sampling rate of 1 / 60 Hz, meaning one sample is recorded per minute. Specifically, the Pecan dataset used in our experiment contains energy consumption data from 10 households over a one-week period, with one sample recorded every minute. The entire dataset consists of 60 min / h × 24 h / day × 7 days = 10,080 sampled points. The first half of these sampled points, i.e., 5040 points, are employed for training, while the remaining 5040 points are reserved for testing purposes.
Meanwhile, the power consumption data of house 1 from the REDD dataset are also utilized to validate our algorithms. The sampling rate of these data is 1 / 3 Hz. We selected house 1 because it contains the most comprehensive information and features 18 devices [21].

3.1. Automatic Pattern Recognition Using AFF

In this subsection, we employ the AFF to identify similar patterns of refrigerator usage from aggregated observations. The effectiveness of the AFF is assessed using the metrics “Rate of Detection” and “precision”, which are defined as follows:
Rate   of   Detection = t p + f p t p + f p + f n ,
Precision = t p t p + f p ,
where t p , f p , and f n denote “true positive”, “false positive”, and “false negative”, respectively. These terms represent correct identifications, erroneous detections, and missed identifications. It is important to note that our definition of “Rate of Detection” differs from the conventional definition of “recall”, as outlined in [22]:
Recall = t p t p + f n .
In the definition of “Rate of Detection”, we use “tp+fp+fn” to signify the total number of time points under evaluation, and “tp+fp” to denote the number of time points that have been identified by the AFF.
We then apply the AFF to the Pecan dataset, with the results presented in Table 1, evaluated based on the metrics of “Rate of Detection” and “precision”. These findings are derived from a moderate configuration of the AFF parameters. As discussed in the previous section, relaxing the variances of the filter window’s width w and height h can enhance the “Rate of Detection” but may lead to a decrease in “precision”. Conversely, tightening these variances can boost “precision” at the cost of a reduced “Rate of Detection”.
Assuming the refrigerator operates in two states, “turn on” (cooling) or “turn off” (standby), the AFF’s task is to identify the time periods during which these states can be classified as “on” or “off” by the AFF. To illustrate the detailed outcomes of our AFF, Figure 6 displays the estimated operational status of the refrigerator in the Pecan dataset. In this figure, higher red lines represent estimations of the “turn on” state, while lower red lines indicate the “turn off” state. When comparing these estimated states with the aggregated observations (depicted by the blue lines), we observe that the AFF has successfully captured most of the significant patterns of the refrigerator’s operation.
In the subsequent subsection, the retrieved statuses will serve as the predefined states within the models of both the direct FFHMM and the embedded FFHMM.

3.2. Experiments Using Direct FFHMM and Embedded FFHMM

In order to assess the performance of various models and algorithms, we utilize the metric known as “total energy correctly assigned” [20], which is defined as follows:
acc = 1 t = 1 T i = 1 n | y ^ t ( i ) y t ( i ) | 2 t = 1 T y ¯ t ,
where y ¯ t represents the observed total power consumption at time t, while y t ( i ) denotes the actual power consumed by the i-th appliance at time t, with its corresponding estimate being y ^ t ( i ) . In our experimental setup, we apply both the EXACT and AFAMAP algorithms to both the direct FFHMM model and the embedded FFHMM model.
Based on the design presented in Figure 4, the direct FFHMM model can enhance—and is specifically designed to enhance—the performance of appliances with predefined states. In this experiment, the refrigerator serves as the target appliance. When compared to the FHMM, the direct FFHMM shows this enhancement. Therefore, we only need to compare the performances of the direct FFHMM and FHMM on the refrigerator, as illustrated in Figure 7. It is evident that, regardless of whether the EXACT algorithm or the AFAMAP algorithm is employed, the direct FFHMM invariably attains a higher level of accuracy compared to the FHMM. This is attributable to the fact that the AFF within the direct FFHMM model takes into account additional information that the FHMM overlooks.
When using the EXACT algorithm, the mean of the difference in accuracy between the direct FFHMM and FHMM is 0.0788 , with a standard deviation of 0.0246 . The calculated T-statistic is 10.1158 . These statistical results clearly indicate that there is a highly significant difference in the performance of these two models. Additionally, the 95 % confidence interval for this difference is [ 0.0612 , 0.0964 ] , further validating the advantage of the proposed direct FFHMM over the FHMM.
When the AFAMAP algorithm is utilized, the mean of the difference in accuracy between the direct FFHMM and the FHMM amounts to 0.0759 . The standard deviation associated with this difference is 0.0407 . A T-statistic of 5.8959 is computed. Such a T-statistic value strongly suggests that there exists a statistically significant difference in the performance of the two models. Moreover, the 95 % confidence interval for this performance difference is precisely determined to be in the range of [ 0.0468 , 0.1050 ] , which further bolsters the conclusion that the performance of the direct FFHMM is better than that of the FHMM.
The estimation results by the EXACT and AFAMAP algorithms for the Pecan dataset using the embedded FFHMM model are presented in Table 2. The third column, labeled as “Ratio”, indicates the proportion of the cumulative consumption of each appliance relative to the total consumption of the entire household.
The bolded figures in Table 2 emphasize the best accuracies, which were all obtained by either the EXACT or AFAMAP algorithm. It is evident that the embedded FFHMM model has the potential to boost the performance of several additional appliances, albeit it may also lead to a decrease in performance for some.
These findings substantiate the assertions made in the preceding section, indicating that improvements for appliances are independent in the direct FFHMM model but become interdependent in the embedded model.
For ease of comparison, Figure 8 illustrates the enhancements in disaggregation accuracy for the refrigerator after substituting the FFHMM models for the FHMM model. Within the figure, the term “Improvement of AFAMAP by direct FFHMM” denotes the disparity between the accuracy achieved by the AFAMAP algorithm when applied to the direct FFHMM model and that of the FHMM model. A positive difference indicates that the accuracy derived from the FFHMM-based algorithm surpasses that of the FHMM-based algorithm, while a negative difference signifies the converse. The remaining three lines are derived in an analogous manner. Consequently, as all lines in the figure are above zero, it can be inferred that the FFHMM models yield superior disaggregation results across all households compared to the FHMM models.
Figure 9 presents a detailed disaggregation analysis of the refrigerator. Upon examining the figure, it becomes evident that both the EXACT and AFAMAP algorithms significantly enhance the disaggregation results when applied to either the direct FFHMM or embedded FFHMM models. For instance, between 3600 and 4000 min, the operational status of the refrigerator is accurately estimated by the algorithms employing FFHMM models. In contrast, both the EXACT and AFAMAP algorithms based on the FHMM model fail to detect such significant patterns.
Meanwhile, we apply FFHMM models to the REDD dataset. To demonstrate that the efficacy of our models is not confined to a single metric, we employ “precision”, “recall”, and “F-measure” to evaluate the performance of FFHMM models on the REDD dataset. The F-measure is defined as per [22]:
F = 2 · precision · recall precision + recall
where the precision and recall are defined in (42) and (43), respectively.
On the REDD dataset, when the direct FFHMM model replaces the FHMM model, the enhancements in precision, recall, and F-measure of the AFAMAP algorithm for the refrigerator are 0.0149 , 0.2668 , and 0.2020 respectively. Meanwhile, the findings for the embedded FFHMM are presented in Table 3. Further detailed illustrations about the direct FFHMM and embedded FFHMM are provided in Figure 10. These outcomes bear resemblance to those obtained from experiments conducted on the Pecan dataset, thereby validating the effectiveness of the proposed FFHMM models.
In this subsection, the proposed direct FFHMM and embedded FFHMM are demonstrated by their application to NILM. However, it is evident that these two models are not limited to NILM. As stated in the introduction, similar to the FHMM model, they can also be applied to a wide range of other domains. These include consumers’ choice models, estimating tumor clonal composition, fault diagnosis, online skill rating, and forecasting financial portfolios with value at risk.

4. Discussion

In order to leverage the information embedded within the aggregated observations, which is often significant yet overlooked by the FHMM, we employ an empirical automatic feature filter (AFF) to extract pertinent patterns designated as predefined states. The AFF utilized in this study is sufficient to showcase the efficacy of both the direct FFHMM and embedded FFHMM; however, more specialized approaches could be explored in future research to refine the identification of these predefined states with enhanced precision. Subsequently, we introduce the FFHMM model, which integrates the FHMM with these predefined states. The FFHMM can be classified into two distinct variants: the direct FFHMM and embedded FFHMM. In the former, the predefined states directly substitute the estimated outcomes of the FHMM, allowing for the independent enhancement of each individual HMM chain’s estimation. Conversely, in the embedded FFHMM, the hidden state variables of each HMM chain are replaced by the predefined states, which, in conjunction with the states of other HMM chains, determine the final outcome of the embedded FFHMM. As a result, within the embedded FFHMM framework, the estimation of multiple HMM chains can potentially be improved by merely defining some states for a single HMM chain. To validate the effectiveness of the FFHMM models, we conducted experiments on two real-world datasets, employing both the EXACT and AFAMAP algorithms.

Author Contributions

Conceptualization, Z.P.; methodology, Z.P.; software, Z.P.; validation, Z.P., W.H. and Y.Z.; formal analysis, Z.P.; investigation, Z.P.; resources, W.H. and Y.Z.; data curation, Z.P.; writing—original draft preparation, Z.P.; writing—review and editing, Z.P., W.H. and Y.Z.; visualization, Z.P.; supervision, W.H. and Y.Z.; project administration, W.H. and Y.Z.; funding acquisition, Z.P., W.H. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Discipline Construction and Promotion Project of Guangdong Province (No. 2022ZDJS065), Key Field Special Project of Scientific Research of Education Department of Guangdong Province (No. 2023ZDZX1014), Key Research Project of Guangdong Province Undergraduate Colleges and Universities Online Open Courses Guiding Committee (No. 2022ZXKC310), Scientific Research Project of Hanshan Normal University (No. XD202304 and No. XSZ202402), and Education and Teaching Reform Project of Hanshan Normal University (No. PX-161241546).

Data Availability Statement

The data presented in this study are available upon request from the corresponding authors.

Acknowledgments

The authors would like to thank the referees and the editors for their very useful suggestions, which significantly improved this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ghahramani, Z.; Jordan, M. Factorial hidden Markov models. Mach. Learn. 1997, 29, 245–273. [Google Scholar] [CrossRef]
  2. Schweiger, R.; Erlich, Y.; Carmi, S. FactorialHMM: Fast and exact inference in factorial hidden Markov models. Bioinformatics 2019, 35, 2162–2164. [Google Scholar] [CrossRef] [PubMed]
  3. Paeng, J.W.; Kwon, J. Visual tracking using interactive factorial hidden Markov models. IET Signal Process. 2021, 15, 365–374. [Google Scholar] [CrossRef]
  4. Kumar, P.; Abhyankar, A.R. A Time Efficient Factorial Hidden Markov Model-Based Approach for Non-Intrusive Load Monitoring. IEEE Trans. Smart Grid 2023, 14, 3627–3639. [Google Scholar] [CrossRef]
  5. Kani, A.; DeSarbo, W.S.; Fong, D.K. A factorial hidden Markov model for the analysis of temporal change in choice models. Cust. Needs Solut. 2018, 5, 162–177. [Google Scholar]
  6. Nepal, A.; Yates, A. Factorial Hidden Markov Models for Learning Representations of Natural Language. arXiv 2013, arXiv:1312.6168. [Google Scholar]
  7. Rahman, M.S.; Nicholson, A.E.; Haffari, G. HetFHMM: A Novel Approach to Infer Tumor Heterogeneity Using Factorial Hidden Markov Models. J. Comput. Biol. 2018, 25, 182–193. [Google Scholar] [CrossRef]
  8. Zhou, D.; Zhao, Y.; Wang, Z.; He, X.; Gao, M. Review on diagnosis techniques for intermittent faults in dynamic systems. IEEE Trans. Ind. Electron. 2019, 67, 2337–2347. [Google Scholar] [CrossRef]
  9. Duffield, S.; Power, S.; Rimella, L. A state-space perspective on modelling and inference for online skill rating. J. R. Stat. Soc. Ser. C Appl. Stat. 2024, 73, 1262–1282. [Google Scholar] [CrossRef]
  10. Puglisi, F.M.; Pavan, P. Factorial hidden Markov model analysis of random telegraph noise in resistive random access memories. ECTI Trans. Electr. Eng. Electron. Commun. 2014, 12, 24–29. [Google Scholar] [CrossRef]
  11. Saidane, M. Forecasting portfolio-Value-at-Risk with mixed factorial hidden Markov models. Croat. Oper. Res. Rev. 2019, 10, 241–255. [Google Scholar] [CrossRef]
  12. Durrieu, J.L.; Thiran, J.P. Source/filter factorial hidden markov model, with application to pitch and formant tracking. IEEE Trans. Audio Speech Lang. Process. 2013, 21, 2541–2553. [Google Scholar]
  13. Kolter, J.; Jaakkola, T. Approximate inference in additive factorial hmms with application to energy disaggregation. In Proceedings of the International Conference on Artificial Intelligence and Statistics, La Palma, Spain, 21–23 April 2012. [Google Scholar]
  14. Rabiner, L. A tutorial on Hidden Markov Models and selected applications in speech recognition. Proc. IEEE 1989, 77, 257–286. [Google Scholar]
  15. Chopin, N.; Papaspiliopoulos, O. An Introduction to Sequential Monte Carlo; Springer: Berlin/Heidelberg, Germany, 2020; Volume 4. [Google Scholar]
  16. Rebeschini, P.; Van Handel, R. Can local particle filters beat the curse of dimensionality? Ann. Appl. Probab. 2015, 25, 2809–2866. [Google Scholar] [CrossRef]
  17. Rimella, L.; Whiteley, N. Exploiting locality in high-dimensional Factorial hidden Markov models. J. Mach. Learn. Res. 2022, 23, 1–34. [Google Scholar]
  18. Franklin, G.F.; Powell, J.D.; Emami-Naeini, A. Feedback Control of Dynamic Systems, 8th ed.; Pearson: London, UK, 2018. [Google Scholar]
  19. Holcomb, C. Pecan Street Inc.: A test-bed for nilm. In Proceedings of the International Workshop on Non-Intrusive Load Monitoring, Pittsburgh, PA, USA, 7 May 2012. [Google Scholar]
  20. Kolter, J.; Johnson, M. REDD: A public data set for energy disaggregation research. In Proceedings of the Workshop on Data Mining Applications in Sustainability (SIGKDD), San Diego, CA, USA, 21 August 2011. [Google Scholar]
  21. Shao, H.; Marwah, M.; Ramakrishnan, N. A Temporal Motif Mining Approach to Unsupervised Energy Disaggregation: Applications to Residential and Commercial Buildings. In Proceedings of the Twenty-Seventh Conference on Artificial Intelligence (AAAI’13), Bellevue, WA, USA, 14–18 July 2013. [Google Scholar]
  22. Olson, D.; Delen, D. Advanced Data Mining Techniques, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
Figure 1. Additive FHMM and its output of the optimal estimations for the individual HMMs. (a) Depicts the model of the FHMM [13]. (b) Depicts the optimal estimation for the individual HMMs.
Figure 1. Additive FHMM and its output of the optimal estimations for the individual HMMs. (a) Depicts the model of the FHMM [13]. (b) Depicts the optimal estimation for the individual HMMs.
Mathematics 13 01201 g001
Figure 2. An illustrative example reveals that both the exact Viterbi algorithm and the approximate AFAMAP algorithm, when applied to the FHMM, struggle to detect the electricity consumption of the refrigerator during the interval between 1000 and 1200 min. In this figure, the blue lines denote the actual consumption (ground truth), while the red lines represent the estimated consumption. The sub-figures are as follows: (a) displays the aggregated consumption (observation) for the entire household; (b) compares the actual consumption of air conditioner 1 with its estimation derived from the EXACT algorithm; (c) contrasts the actual consumption of air conditioner 1 with its approximation obtained through the AFAMAP algorithm; (d) shows the true consumption pattern of the refrigerator alongside its EXACT algorithm estimate; and (e) presents the actual consumption of the refrigerator with its approximation derived from the AFAMAP algorithm.
Figure 2. An illustrative example reveals that both the exact Viterbi algorithm and the approximate AFAMAP algorithm, when applied to the FHMM, struggle to detect the electricity consumption of the refrigerator during the interval between 1000 and 1200 min. In this figure, the blue lines denote the actual consumption (ground truth), while the red lines represent the estimated consumption. The sub-figures are as follows: (a) displays the aggregated consumption (observation) for the entire household; (b) compares the actual consumption of air conditioner 1 with its estimation derived from the EXACT algorithm; (c) contrasts the actual consumption of air conditioner 1 with its approximation obtained through the AFAMAP algorithm; (d) shows the true consumption pattern of the refrigerator alongside its EXACT algorithm estimate; and (e) presents the actual consumption of the refrigerator with its approximation derived from the AFAMAP algorithm.
Mathematics 13 01201 g002
Figure 3. Example of automatic feature filter (AFF). (a) Illustrates the aggregated observation, with a segment expanded to form sub-figure (b). Each green window in (b) encapsulates a similar pattern identified during the first step of the AFF process. In the second step of AFF, depicted in (c), the dimensions—width w and height h—of the windows are employed to classify the recognized patterns.
Figure 3. Example of automatic feature filter (AFF). (a) Illustrates the aggregated observation, with a segment expanded to form sub-figure (b). Each green window in (b) encapsulates a similar pattern identified during the first step of the AFF process. In the second step of AFF, depicted in (c), the dimensions—width w and height h—of the windows are employed to classify the recognized patterns.
Mathematics 13 01201 g003
Figure 4. Scheme of the direct FFHMM. (a) Incorporates the FHMM model. (b) Presents the estimation of the AFF, where each orange square denotes a state that has been identified by the AFF at a specific moment in time. In forming the output estimation of the FFHMM as shown in (c), the majority of state estimations are derived from the FHMM, with exceptions at certain time points where the AFF provides non-zero estimations.
Figure 4. Scheme of the direct FFHMM. (a) Incorporates the FHMM model. (b) Presents the estimation of the AFF, where each orange square denotes a state that has been identified by the AFF at a specific moment in time. In forming the output estimation of the FFHMM as shown in (c), the majority of state estimations are derived from the FHMM, with exceptions at certain time points where the AFF provides non-zero estimations.
Mathematics 13 01201 g004
Figure 5. Diagram of the embedded FFHMM. (a) Illustrates the amalgamation of the FHMM and the AFF estimations, with the orange square boxes representing the AFF replacing the hidden states of the FHMM. (b) Shows the estimation output of the FFHMM, where each orange square box signifies an estimation made by the AFF at the respective time.
Figure 5. Diagram of the embedded FFHMM. (a) Illustrates the amalgamation of the FHMM and the AFF estimations, with the orange square boxes representing the AFF replacing the hidden states of the FHMM. (b) Shows the estimation output of the FFHMM, where each orange square box signifies an estimation made by the AFF at the respective time.
Mathematics 13 01201 g005
Figure 6. The estimated status of the refrigerator in the Pecan dataset as determined by the AFF. The blue lines represent the aggregated energy consumption of the entire household, whereas the red lines depict the estimated status of the refrigerator by the AFF. When the red lines are at a height of 2, it indicates that the refrigerator is operating in the “On” state during the corresponding time intervals, as estimated by the AFF. Conversely, the lower red lines at a height of 1 signify the estimated “Off” state of the refrigerator.
Figure 6. The estimated status of the refrigerator in the Pecan dataset as determined by the AFF. The blue lines represent the aggregated energy consumption of the entire household, whereas the red lines depict the estimated status of the refrigerator by the AFF. When the red lines are at a height of 2, it indicates that the refrigerator is operating in the “On” state during the corresponding time intervals, as estimated by the AFF. Conversely, the lower red lines at a height of 1 signify the estimated “Off” state of the refrigerator.
Mathematics 13 01201 g006
Figure 7. Accuracy of energy disaggregation for refrigerator employing the FHMM and direct FFHMM models on the Pecan dataset. (Left) Results obtained with the EXACT algorithm. (Right) Results obtained with the AFAMAP algorihtm.
Figure 7. Accuracy of energy disaggregation for refrigerator employing the FHMM and direct FFHMM models on the Pecan dataset. (Left) Results obtained with the EXACT algorithm. (Right) Results obtained with the AFAMAP algorihtm.
Mathematics 13 01201 g007
Figure 8. The enhancement in accuracy observed for the refrigerator following the substitution of FFHMM models for the FHMM.
Figure 8. The enhancement in accuracy observed for the refrigerator following the substitution of FFHMM models for the FHMM.
Mathematics 13 01201 g008
Figure 9. Disaggregation results for the refrigerator within the Pecan dataset. The blue lines indicate the actual consumption, whereas the red lines display the disaggregation outcomes generated by the respective algorithms: (a) aggregated observation, (b) AFAMAP algorithm utilizing the FHMM, (c) EXACT algorithm with the FHMM, (d) AFAMAP algorithm utilizing the direct FFHMM, (e) EXACT algorithm employing the direct FFHMM, (f) AFAMAP algorithm utilizing the embedded FFHMM, and (g) EXACT algorithm with the embedded FFHMM.
Figure 9. Disaggregation results for the refrigerator within the Pecan dataset. The blue lines indicate the actual consumption, whereas the red lines display the disaggregation outcomes generated by the respective algorithms: (a) aggregated observation, (b) AFAMAP algorithm utilizing the FHMM, (c) EXACT algorithm with the FHMM, (d) AFAMAP algorithm utilizing the direct FFHMM, (e) EXACT algorithm employing the direct FFHMM, (f) AFAMAP algorithm utilizing the embedded FFHMM, and (g) EXACT algorithm with the embedded FFHMM.
Mathematics 13 01201 g009
Figure 10. Applying the AFAMAP algorithms for energy disaggregation of the refrigerator to the REDD dataset. (a) Presents the aggregated consumption. (b) Displays the estimation result based on the FHMM model (real consumption in blue and estimation result in red). (c) Illustrates the estimation result derived from the direct FFHMM model. (d) Shows the estimation result obtained through the embedded FFHMM model.
Figure 10. Applying the AFAMAP algorithms for energy disaggregation of the refrigerator to the REDD dataset. (a) Presents the aggregated consumption. (b) Displays the estimation result based on the FHMM model (real consumption in blue and estimation result in red). (c) Illustrates the estimation result derived from the direct FFHMM model. (d) Shows the estimation result obtained through the embedded FFHMM model.
Mathematics 13 01201 g010
Table 1. Performance of AFF on Pecan dataset.
Table 1. Performance of AFF on Pecan dataset.
No. of HomeRate of DetectionPrecision
130.52%76.01%
227.54%86.02%
317.94%98.78%
420.87%75.29%
537.92%95.36%
636.09%93.95%
726.81%79.94%
836.85%88.53%
936.79%85.22%
1017.92%74.97%
Table 2. Accuracy of energy disaggregation on Pecan dataset employing the FHMM and embedded FFHMM models with the EXACT and the AFAMAP algorithms.
Table 2. Accuracy of energy disaggregation on Pecan dataset employing the FHMM and embedded FFHMM models with the EXACT and the AFAMAP algorithms.
HomesAppliancesRatioFHMMEmbedded FFHMM
AFAMAPEXACTAFAMAPEXACT
1Refrigerator5.31%0.62620.63360.67490.7012
Air conditioner 170.99%0.84170.94760.93870.9475
Furnace13.51%0.75800.73720.76450.7542
2Refrigerator3.42%0.53150.54290.63440.5864
Air conditioner 150.35%0.80960.92080.81060.9158
Furnace 127.29%0.96940.95890.96980.9608
Electrical dryer 16.74%0.52510.49550.52430.4008
Living room 15.47%0.29590.69700.30940.6491
3Refrigerator8.84%0.55110.67060.60030.6831
Air conditioner 145.60%0.89230.89750.89410.8988
Furnace21.22%0.82980.83600.84150.8368
Electrical Car17.30%0.70650.81520.69940.8152
4Refrigerator4.78%0.59860.53040.60480.5716
Air conditioner 163.63%0.94370.94320.94360.9433
Furnace22.10%0.70890.80830.69960.8102
5Refrigerator4.85%0.90650.85070.93090.9313
Air conditioner 149.13%0.83550.92920.83630.9287
Furnace7.79%0.71690.71340.71710.7297
Electrical car34.30%0.85640.95680.85480.9591
6Refrigerator5.16%0.52010.59920.61630.6963
Air conditioner 169.27%0.81980.92370.81500.9213
Air conditioner 213.67%0.53840.85790.52050.8526
7Refrigerator4.15%0.57740.55240.66450.6638
Air conditioner 157.32%0.88690.89780.88830.9003
Furnace16.10%0.81830.84540.82210.8451
Electrical dryer 18.23%0.77710.82250.77800.8219
Family room 17.34%0.73010.76470.72080.7626
8Refrigerator2.39%0.57950.55100.60650.6109
Air conditioner 140.61%0.81290.87400.81560.8622
Air conditioner 218.02%0.64020.76090.65000.7084
Furnace17.89%0.81680.87890.81710.8662
Electrical dryer 15.87%0.76770.82480.76970.8173
Living room 17.03%0.65100.64630.64380.6812
9Refrigerator4.33%0.61930.62820.68390.6992
Air conditioner 159.86%0.94370.94580.94220.9437
Furnace15.65%0.79210.80270.79280.8168
Theater13.50%0.66930.82220.67070.8580
10Refrigerator4.03%0.41360.46610.49530.5845
Air conditioner 166.92%0.92810.95140.93270.9502
Furnace15.75%0.79100.81730.80800.8266
Living room 15.08%0.65250.65410.64610.7529
Table 3. REDD dataset using the FHMM and embedded FFHMM models with the AFAMAP algorithm.
Table 3. REDD dataset using the FHMM and embedded FFHMM models with the AFAMAP algorithm.
AppliancesNo. of
Status
FHMMEmbedded FFHMM
Precision Recall F-Measure Precision Recall F-Measure
refrigerator20.97450.50430.66470.98390.77430.8666
dishwasher50.98710.87980.93040.98730.88400.9328
microwave20.98600.63650.77360.99830.68750.8143
bathroomfi20.98380.99980.99170.99210.99990.9960
kOutlet 220.99060.30080.46150.99450.35570.5240
kOutlets 330.45460.69330.54910.47800.67820.5607
light 330.71030.39440.50720.78030.38260.5135
light 120.99970.14550.25410.99940.16840.2883
light 220.96890.24510.39130.93680.33400.4924
washdryer 221.00000.95380.97640.99990.98160.9907
washdryer 130.97350.77830.86500.98140.74770.8488
kOutlet 120.99660.77880.87430.99520.72260.8373
oven 1& 221.00000.97280.98620.99250.81040.8923
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, Z.; Huang, W.; Zhu, Y. Feedforward Factorial Hidden Markov Model. Mathematics 2025, 13, 1201. https://doi.org/10.3390/math13071201

AMA Style

Peng Z, Huang W, Zhu Y. Feedforward Factorial Hidden Markov Model. Mathematics. 2025; 13(7):1201. https://doi.org/10.3390/math13071201

Chicago/Turabian Style

Peng, Zhongxing, Wei Huang, and Yinghui Zhu. 2025. "Feedforward Factorial Hidden Markov Model" Mathematics 13, no. 7: 1201. https://doi.org/10.3390/math13071201

APA Style

Peng, Z., Huang, W., & Zhu, Y. (2025). Feedforward Factorial Hidden Markov Model. Mathematics, 13(7), 1201. https://doi.org/10.3390/math13071201

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop