Next Article in Journal
Lexicons of Key Terms in Scholarly Texts and Their Disciplinary Differences: From Quantum Semantics Construction to Relative-Entropy-Based Comparisons
Next Article in Special Issue
Information Theoretic Methods for Variable Selection—A Review
Previous Article in Journal
Financial Network Analysis on the Performance of Companies Using Integrated Entropy–DEMATEL–TOPSIS Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multivariate Time Series Imputation: An Approach Based on Dictionary Learning

by
Xiaomeng Zheng
1,
Bogdan Dumitrescu
2,
Jiamou Liu
3 and
Ciprian Doru Giurcăneanu
1,*
1
Department of Statistics, University of Auckland, Auckland 1142, New Zealand
2
Department of Automatic Control and Computers, University Politehnica of Bucharest, 060042 Bucharest, Romania
3
School of Computer Science, University of Auckland, Auckland 1142, New Zealand
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(8), 1057; https://doi.org/10.3390/e24081057
Submission received: 17 June 2022 / Revised: 25 July 2022 / Accepted: 29 July 2022 / Published: 31 July 2022

Abstract

:
The problem addressed by dictionary learning (DL) is the representation of data as a sparse linear combination of columns of a matrix called dictionary. Both the dictionary and the sparse representations are learned from the data. We show how DL can be employed in the imputation of multivariate time series. We use a structured dictionary, which is comprised of one block for each time series and a common block for all the time series. The size of each block and the sparsity level of the representation are selected by using information theoretic criteria. The objective function used in learning is designed to minimize either the sum of the squared errors or the sum of the magnitudes of the errors. We propose dimensionality reduction techniques for the case of high-dimensional time series. For demonstrating how the new algorithms can be used in practical applications, we conduct a large set of experiments on five real-life data sets. The missing data (MD) are simulated according to various scenarios where both the percentage of MD and the length of the sequences of MD are considered. This allows us to identify the situations in which the novel DL-based methods are superior to the existing methods.

1. Introduction

1.1. Background

It is well-known from the classical literature on time series that a multivariate time series data set is obtained by measuring K > 1 variables at time points 1 , , T . The observations are stored in a matrix with T rows and K columns. For ease of writing, we use the notation Z for this matrix. In our notation, the bold letters are used for vectors and matrices. Due to recent technological advances, both T and K are very large for the data sets that are collected nowadays. The massive amount of data poses difficulties for both the storage and the processing. Another challenge comes from the fact that some of the entries of the big matrix Z are missing. The data are incomplete for various reasons: malfunction of the sensors, problems with the transmission of the measurements between devices, or the fact that it is practically impossible to collect all the data (for example, this happens often in astronomy).
The conventional approach is to estimate the missing data and then to use the resulting complete data set in statistical inference. The estimation methods span a wide range from the simple ones that perform imputation for each time series individually by considering the mean or the median, or employ the last value carried forward or the next value carried backward, to the more advanced ones that involve the evaluation of the (Gaussian) likelihood, see for example [1].
In here we do not discuss the imputation methods that can be easily found in the time series textbooks, but we briefly present the newer methods that have been compared in [2]. For instance, we consider the method DynaMMO from [3], which is closer to the traditional methods in the sense that it uses a technique akin to the Kalman filter (see again [1]) to estimate the missing values. The other methods that we outline below are based on the decomposition of the matrix Z ; hence, it is not surprising that they have a certain relationship with the singular value decomposition (SVD). In fact, SVDImp from [4] explicitly uses the SVD factorization U Σ V of the matrix Z , after replacing the missing values with the mean computed for the corresponding row. Note that the symbol ( · ) denotes transposition. The most significant κ rows of V are selected, where the value of κ is chosen empirically. Then the linear regression is used to express each row of Z as a linear combination of the most significant κ rows of V . In this way, new estimates are obtained for the missing data, and the procedure above is applied again to the matrix Z that contains these imputed values. The iterations are continued until the change of the magnitudes of the imputed values between two consecutive iterations is smaller than a threshold selected by the user.
In [5], the imputation problem is formulated as a matrix completion problem. The obtained estimate Z ^ has the same entries as Z at the locations for which the measurements are available. The matrix Z ^ is found by solving a penalized least-squares problem whose expression contains a coefficient λ that balances the two terms involved: (i) half of the sum of the squares of the approximation errors, or equivalently 1 2 | | Z ^ Z | | F 2 (the notation | | · | | F stands for the Frobenius norm), and (ii) the sum of the singular values of Z ^ , or equivalently the nuclear norm of Z ^ . The solution is given by a soft-thresholded SVD of Z , where the soft threshold is λ . This suggests the name Soft-Impute of the method for which we use the acronym SoftImp. In SoftImp, the solutions are obtained in a computationally efficient manner for all the values of λ on a predefined grid. Another algorithm that solves the same penalized least-squares problem by computing the soft-thresholded SVD at each iteration is the Singular Value Thresholding (SVT) from [6]. A particular attribute of SVT is that it automatically finds the optimal rank of the estimated matrix.
An approximation of SVD, which is called centroid decomposition (CD), is employed in [7,8] for representing the matrix Z as Z = L R , where L R T × K and R R K × K . Either interpolation or extrapolation is applied for estimating the missing entries of Z and then a vector s { 1 , 1 } T is found such that the Euclidean norm | | Z s | | 2 is maximized. The vector c = Z s is obtained and is further used to obtain the first column of R , R : 1 = c / | | c | | 2 , and the first column of L , L : 1 = Z R : 1 . For a better understanding of why the method is named CD, we mention that c is the first centroid vector. The “new” matrix Z is taken to be Z L : 1 R : 1 , and the algorithm continues until all K columns of L and R are obtained. Only the first κ columns are used to obtain the approximation of Z given by i = 1 κ L : i R : i , and this approximation yields the estimates for the missing values. It is interesting that the selection of κ is performed by an entropy-based criterion. When CD is employed in time series recovery we call it CDRec (as in [2]).
Another imputation method is dubbed Grassmannian Rank-One Update Subspace Estimation (GROUSE), see for example [9,10]. The name comes from the set of all subspaces of R T of dimension κ that is called Grassmannian. It is evident that an element of the Grassmanian can be represented by any matrix U R T × κ with the property that U U = I , where I denotes the identity matrix of appropriate dimension. GROUSE finds the matrix U that minimizes the objective function j = 1 K | | Δ j ( Z : j U U Z : j ) | | 2 2 , where Δ j R T × T is a diagonal matrix which has on the main diagonal ones on the locations corresponding to the data that are available for the jth column of Z and zeros otherwise. The presence of Δ j in the formula above shows that the algorithm can work directly with the columns of Z that have missing data. In fact, GROUSE optimizes the cost function by considering one column of Z at a time, and at each such step, the matrix U is updated by adding a rank-one matrix to the matrix U obtained at the previous step. Once the “final” U is found at the last step, the incomplete columns are projected onto the low-rank subspace that was identified to complete the matrix.
In robust principal component analysis (RPCA), the data matrix (which is supposed to be complete) is represented as a low-rank matrix plus a sparse matrix [11]. Because the recovery of the low-rank matrix is computationally intensive, an efficient algorithm called Robust Orthogonal Subspace Learning (ROSL) was proposed in [12]. The algorithm was altered in [2] to be applied to an incomplete data matrix for estimating the missing values. We use the acronym ROSL for the version of the algorithm from [2].
The imputation method from [13] relies on the nonnegative matrix factorization (NMF) technique and, to be suitable for electricity consumption, uses temporal aggregates. The optimization problem solved by the algorithm proposed in [13] takes into consideration the correlation between time series. As in [2], we call this method temporal NMF (TeNMF). The matrix factorization is also used in [14], but in contrast to TeNMF, the entries of the two factor matrices are not constrained to be nonnegative. An important feature of the method is that the regularization term of the objective function takes explicitly into consideration the temporal structures, and this is why the method is termed Temporal Regularized Matrix Factorization (TRMF).
Another matrix factorization that can be instrumental in time series imputation is the one generated by dictionary learning (DL).

1.2. Organization of the Paper and the Main Contributions

In this article, we extend the DL-based solution for time series imputation which we proposed in our earlier work [15]. Our previous results from [15] consist of two imputation methods: DLU (for univariate time series) and DLM (for multivariate time series). However, because of the computational complexity, the original version of DLM can be utilized only when the number of the time series involved is very small. In contrast, the variants of DLM that we introduce in this study can be applied to data sets that contain tens of time series.
DLM is presented in Section 2.2 after briefly discussing the DL optimization problem in Section 2.1. An important characteristic of DLM is that it solves the optimization problem by minimizing the Frobenius norm of the errors. In many practical situations, the imputation should be performed to minimize the 1 -norm of the errors and not the sum of the squared errors. In Section 3, we demonstrate how the optimization problem can be solved when the Frobenius norm is replaced with the sum of the magnitudes of errors. The method that involves the 1 -norm is dubbed DLM 1 . Another characteristic of DLM (which is also inherited by DLM 1 ) is the use of a structured dictionary. In Section 4, we present the expressions of the IT criteria that are employed to select the size of the structured dictionary, the size for each of its blocks as well as the sparsity level of the representation. Section 5 is focused on the techniques that we propose for dimensionality reduction. It allows us to apply DLM and DLM 1 to data sets that comprise tens of time series with thousands of measurements. For demonstrating how the new algorithms can be used in practice, we conduct a large set of experiments with real-life data. The experimental settings are presented in Section 6, and the empirical results are discussed in Section 7. Section 8 concludes the paper.
Hence, the main contributions of this work are the following:
  • A flexible approach that allows the user to choose the norm of the errors (Frobenius norm or 1 -norm) minimized in the optimization problem.
  • An automatic method for selecting the sparsity as well as the size for each block of the dictionary.
  • The exemplification of two techniques for dimensionality reduction that enable DLM to impute values on multivariate time series for which K is large.
  • An extensive empirical study which compares DLM with nine other imputation methods on data sets with various characteristics. On many of these data sets, DLM has the best performance among the considered methods when the missing data are simulated by sampling without replacement.

2. Dictionary Learning for Data Sets with Missing Values

2.1. Preliminaries

The DL problem is formulated as follows. Given N signals of length m that are grouped into the matrix Y R m × N , we approximate Y by the product D X , where D R m × n is the dictionary, and its columns are usually named atoms. The Euclidean norm of each atom equals one. The matrix X R n × N is sparse in the sense that each of its columns contains at most s non-zero entries, the parameter s being named sparsity level. We emphasize that both D and X are learned from the signals by solving the following optimization problem [16]:
minimize D , X Y DX F subject to | | X : | | 0 s , { 1 , , N } | | D : j | | 2 = 1 , j { 1 , , n }
the -th column of X is denoted X : , and the j-th column of D is denoted D : j . The symbol · 0 represents the number of the non-zero entries for the vector in the argument.
The algorithm that solves the optimization problem in (1) is initialized with a dictionary D , which is generally randomly generated. The user selects the number of iterations, and the following steps are executed at each iteration:
(i)
The current dictionary D is used to find the matrix X , which provides a representation for the signals in Y . This goal is achieved by employing the Orthogonal Matching Pursuit (OMP).
(ii)
The dictionary D is updated by using the current sparse representation X . This is performed by using the Approximate K-Singular Value Decomposition (AK-SVD) algorithm.
The two steps of the main algorithm are presented in [17].
There are other ways of posing the DL problem. For example, one may add a sparsity enhancing term to the objective and thus impose sparsity globally, not on each representation; thus, one can obtain a matrix X that has around s N nonzeros without the explicit constraint that each of its columns has s nonzeros. A representative of this approach is [18]. Convolutional DL [19] does not split the time series into signals of size m, but works with a single long signal that is approximated as a linear combination of atoms of length m that may be placed at any position; the same atom can be used repeatedly. These approaches may provide more flexibility, but they require more fine tuning of the parameters and adaptation of dictionary size criteria. We prefer AK-SVD because it is one of the fastest DL-algorithms that have been proposed in the previous literature. Another important feature of the algorithm is its conceptual simplicity, which allows it to be easily modified for solving particular formulations of DL that appear in the context of imputation.

2.2. Optimization Problem for Incomplete Data: Formulation, Solution and Applications

When some of the entries of the matrix Y are missing, the optimization problem in (1) becomes [20]:
minimize D , X M ( Y DX ) F subject to | | X : | | 0 s , { 1 , , N } | | D : j | | 2 = 1 , j { 1 , , n }
where M is a mask matrix with the same size as Y . Its entries are equal to zero for the positions in Y that correspond to the missing data. All other entries of M are equal to one. The operator ⊙ is the element-wise product. The role of M is to guarantee that only the available data are used in learning. Note that the missing data are replaced with zeros in Y . For the optimization problem in (2), a specialized version of the AK-SVD algorithm is applied [16] [Section 5.9]; the representation matrix X is found using OMP by ignoring the missing samples and working only with the present ones; the atom update formulas are simple adaptations of AK-SVD rules to the incomplete data case.
An important application of (2) consists of filling the gaps of an incomplete image and is called image inpainting. In the case of this application, the matrix Y is generated as follows (see, for example, [16] [Section 2.3.1]). A patch of pixels of size m × m is selected from a random location in the image. Then its columns are stacked to generate the signal y , which is a column of Y . The procedure continues until all N signals are produced. Obviously, it is not allowed to select the same patch twice, but it is highly recommended to select patches that overlap. In what concerns the sizes of the patches, the value m = 8 is often used. Values such as m = 12 and m = 16 have also been used, but they are not commonly employed because of the increased computational burden. Once the dictionary is learned, the product D X yields an estimate Y ^ of Y . Any missing pixel is obtained by averaging its values from all entries of Y ^ where it appears. More details about image inpainting can be found in [21,22,23,24]. The use of the inpainting was extended from images to audio signals in [20,25].
Our main goal is to show how DL can be employed for estimating the values of the missing data in multivariate time series. As we have already pointed out, the use of DL in the imputation of time series was discussed in [15]. The approach adopted in the multivariate case should take into consideration the dynamic interrelationships between K > 1 variables whose measurements collected at time points 1 , , T are stored in matrix Z . Suppose that some of the entries of Z are missing. As the positions of the missing data are not necessarily the same for all the time series, we use the symbol Ψ k to denote the indexes of the measurements that are available for the k-th time series, where 1 k K . Obviously, the set of indexes of missing data for the k-th time series is Ψ ¯ k = { 1 , , T } Ψ k .
Let z be one of the columns of the data matrix Z in which the missing data indexed by Ψ ¯ are replaced with zeros. We define a matrix Y as follows:
Y = z 1 : m z 1 + h : m + h z 1 + q h : m + q h .
For an arbitrary vector v , v a : b denotes the entries of the vector whose indexes belong to the set { a , a + 1 , , b } , where a < b . The number of rows of the matrix Y is m, and its choice depends on the sampling period. The parameter h is called signal shift and controls the overlapping between the columns of Y , and the value of q is given by ( T m ) / h . It follows that the number of columns of the matrix Y is N = q + 1 . Herein we take h = 1 , which leads to q = T m and N = T m + 1 .
If t Ψ ¯ , then z t is a missing value, and this will be represented as a zero-entry in Y . As there is an overlap between the columns of Y , the missing value z t leads to several zero-entries in Y . We collect all the values of these entries from Y ^ = D X and compute an estimate for z t by averaging them.
For example, suppose 4 Ψ ¯ , which means that z 4 is missing. Then the matrix Y is given by:
Y = z 1 z 2 z 3 0 z T m + 1 z 2 z 3 0 z 5 z T m + 2 z 3 0 z 5 z 6 z T m + 3 0 z 5 z 6 z 7 z T m + 4 z 5 z 6 z 7 z 8 z T m + 5 z m z m + 1 z m + 2 z m + 3 z T .
In addition, the entries of the matrix Y ^ are:
Y ^ = y ^ 1 , 1 y ^ 1 , 2 y ^ 1 , 3 y ^ 1 , 4 y ^ 1 , N y ^ 2 , 1 y ^ 2 , 2 y ^ 2 , 3 y ^ 2 , 4 y ^ 2 , N y ^ 3 , 1 y ^ 3 , 2 y ^ 3 , 3 y ^ 3 , 4 y ^ 3 , N y ^ 4 , 1 y ^ 4 , 2 y ^ 4 , 3 y ^ 4 , 4 y ^ 4 , N y ^ m , 1 y ^ m , 2 y ^ m , 3 y ^ m , 4 y ^ m , N ,
where N = T m + 1 . It results that the estimate of the missing value z 4 is computed by averaging the red-colored entries of Y ^ , whose positions are the same as the positions of z 4 in Y . Thus, we obtain z ^ 4 as:
z ^ 4 = 1 4 ( y ^ 4 , 1 + y ^ 3 , 2 + y ^ 2 , 3 + y ^ 1 , 4 ) .
This imputation method was introduced in [15]. As it can be easily seen from the description above, the method is suitable for univariate time series, and for this reason it is named DLU (see again Section 1.2). In the multivariate case, the data matrix is
Y = [ Y 1 Y K ] ,
where Y i R m × ( T m + 1 ) is made of data measured for the i-th time series for i { 1 , , K } . In [15], it was pointed out that in the DLM algorithm designed for the multivariate case, a structured dictionary should be used:
D = [ D 1 D K D K + 1 ] ,
where D 1 , , D K R m × n d and D K + 1 R m × n K + 1 . The dictionary D i is dedicated to the representation of the i-th time series, while D K + 1 is common for all time series. It follows that the number of atoms used in the representation of each time series is n u = n d + n K + 1 . The values of n u and n K + 1 as well as the sparsity level s are selected by using information theoretic (IT) criteria [26]. The main advantage is that the procedure for choosing the triple ( n u , n K + 1 , s ) does not rely on prior knowledge. We take the sizes of dictionaries D 1 , , D K to be equal to simplify the decision process, but also to use similar representation power for all times series (or groups of time series, as we will see later).
We mention that there are many DL algorithms that choose the size of the dictionary. Most of them are based on heuristics, such as, for example, growing a small dictionary [27] or removing atoms from a large one [28], with the general purpose of achieving parsimony. Other approaches are more principled, using Bayesian learning [29] or an Indian Buffet Process [30]. We have used IT criteria in [26], where we also presented a more detailed view of the topic, including bibliographic references. IT criteria offer a sound evaluation of the trade-off between dictionary size and representation error.
Next we show how the new algorithm DLM 1 can be devised.

3. DLM 1 : DL-Algorithm for Incomplete Data (with 1 -Norm)

We solve the 1 -norm version of (2):
minimize D , X M ( Y DX ) 1 , 1 subject to | | X : | | 0 s , { 1 , , N } | | D : j | | 2 = 1 , j { 1 , , n }
where for a matrix G R m × N we denote G 1 , 1 = i = 1 m = 1 N | g i | , the 1 -equivalent of the Frobenius norm. So, the aim of (5) is to optimize the sparse 1 -norm representation of the signals, whereas (2) targets the 2 -norm.
We modify the algorithm proposed in [31] such that it is suitable for the missing data case. The algorithm is an adaptation of the AK-SVD [17] idea and consists of iterations containing the usual two steps, sparse representation and dictionary update, as described in Section 2.1. We will next discuss these steps in detail.

3.1. 1 -Norm OMP with Missing Data

We present a 1 -norm version of the greedy approach whose most prominent representative is OMP [32]. It is enough to consider a single signal y R m , for which we have to minimize m ( y Dx ) 1 , where D R m × n is the given dictionary, x R n must have at most s nonzero elements, and m R m is the mask whose entries are zeros and ones.
We denote y ¯ R μ ( μ m ) the vector that results from y by keeping only the elements that correspond to nonzero values in m . Similarly, D ¯ R μ × n is the matrix obtained from D by keeping only the rows corresponding to nonzero values in m . We are thus left with the problem y ¯ D ¯ x 1 , which is a usual 1 -norm sparse representation problem for which the algorithm was described in [31].
For the sake of completeness, we revisit the main operations here. The algorithm has s steps. Denoting x ˜ the representation at the beginning of the current step and r = y ¯ D ¯ x ˜ the current residual, the next selected atom d ¯ j is that for which
min j { 1 , , n } min ξ r ξ d ¯ j 1
is attained. Thus, we follow the idea of finding the atom with the best projection on the residual.
The problem min ξ r ξ d 1 (we lighten the notation for the sake of simplicity) can be easily solved. It is not only convex, but its solution can be found by inspection [33]. Denote c i = r i / d i , for i { 1 , , μ } . Denote c ˜ the vector containing the elements of c sorted increasingly and π ( · ) the permutation for which c ˜ i = c π ( i ) . Denote d ˜ i = d π ( i ) . The desired minimum is ξ = c π ( k ) , where the index k is the largest for which
i = 1 k 1 | d ˜ j | i = k μ | d ˜ j | .
So, finding the solution essentially requires only a sort operation. Moreover, in solving (6), some atoms can be ignored if their scalar product (usual orthogonal projection) with the residual is small.
Once the current atom has been found, it is added to the support, and the optimal 1 -norm representation with that support is computed. This is a convex problem and can be solved by several nonlinear optimization algorithms (we have used a few coordinate descent iterations). Moreover, a good initialization is available in the representation at the previous step. (In OMP, these operations correspond to finding a least-squares solution.)

3.2. Dictionary Update with Missing Data in the 1 -Norm

The update stage of AK-SVD optimizes the atoms one by one, also updating the coefficients of the corresponding representations. Denote d j the current atom and I j the set of signals where this atom contributes to the representation. Denote E = Y DX the current residual and
R = Y i j d i x i I j = E + d j x j I j ,
the error without the contribution of d j , keeping only the columns where d j appears in the representation.
With lighter notation, namely d for the current atom, x for the vector of its nonzero representation coefficients and M for the mask (even though the signals where d is not used are removed), the atom update problem becomes
min d M ( R dx ) 1 , 1 .
Denoting M N 2 the indexes of available data, the problem can be written as
min d ( i , j ) M | r i j d i x j | .
The minimization can be performed on each d i separately and has the form min d i r ^ d i x ^ 1 , where r ^ and x ^ are vectors that can be easily built. We thus end up with a problem similar to that described after (6).
Keeping the updated atom fixed, we can now optimize the associated representation coefficients by solving
min x M ( R >dx ) 1 , 1 ,
which can be written
min x ( i , j ) M | r i j d i x j |
like (7) was written as (8). The problem is separable on each x j and can be solved as above.
We note that we use the same basic algorithm in the 1 -norm OMP, atom and representation update. The approach can be extended to p -norms, with p 1 , transforming the p -norm AK-SVD from [31] to the missing data case, similarly to the transformations described in this section.

4. Information Theoretic Criteria

We have already mentioned in Section 2.2 that we employ IT criteria for selecting the triple ( n u , n K + 1 , s ) . More precisely, the criteria that we use are derived from the well-known Bayesian information criterion (BIC) [34]. For evaluating the complexity of the model, we need to calculate the number of parameters (NoP). We have that NoP = s N + ( m 1 ) n . The first term is given by the number of the non-zero entries for the representation matrix X ^ , which is estimated from the available data by solving the optimization problem (2). The second term is equal to the number of the entries of the estimated dictionary D ^ ; for each column of D ^ , we count m 1 entries (and not m entries) because each column is constrained to have the Euclidean norm equal to one. Furthermore, we define the matrix of residuals U ^ = M ( Y D ^ X ^ ) . Note that the residuals located at the positions corresponding to the missing data are forced to be zero. With the understanding that η is the number of the entries of M that are equal to one, the expression of the first IT criterion that we employ is:
BIC ( Y ; n u , n K + 1 , s ) = η 2 ln | | U ^ | | F 2 η + NoP 2 ln η ,
where ln ( · ) denotes the natural logarithm. This criterion was proposed in [15], where its “extended” variant was also used (see [26,35]):
EBIC ( Y ; n u , n K + 1 , s ) = BIC ( Y ; n u , n K + 1 , s ) + N ln n s .
However, because of the way in which they have been derived, the formulas in (9) and (10) can only be used when D ^ and X ^ are outputs of the DLM algorithm. When the estimation is performed by applying the DLM 1 algorithm from Section 3, we employ the following formula for BIC:
BIC ( Y ; n u , n K + 1 , s ) = η 2 ln 2 | | U ^ | | 1 , 1 2 η 2 + NoP 2 ln ( 2 η ) .
The expression above is based on the criterion obtained in [31] for signals in additive Laplacian noise and is altered to be suitable for the missing data case. Its “extended” variant is easily constructed by adding the term N ln n s to the expression above (see again [31]).
For clarifying the notation, we mention that when we write DLM+BIC, it means that the criterion in (9) is applied for model selection. At the same time, DLM1+BIC involves the criterion from (11). A similar convention is used for the “extended” criteria.
Whenever DLM is applied to multivariate time series with missing values, 10 random initializations of the dictionary are considered. For each initialization, 50 iterations of the two-step algorithm that involves OMP and AK-SVD for incomplete data are executed for each possible combination of n u , n K + 1 and s. Then the triple ( n u , n K + 1 , s) which minimizes BIC/EBIC is selected. The procedure is the same when DLM 1 is used instead of DLM.

5. Dimensionality Reduction

5.1. Dimensionality Reduction via Clustering

When K is large, we group the time series to reduce the dimensionality. For exemplification, we refer to the particular case in which K is an even number, and the time series are clustered into two groups such that each group contains K / 2 columns of Z . The set of the indexes of the time series that belong to the first group is Φ = { ϕ 1 , , ϕ K / 2 } , whereas the set of the indexes corresponding to the second group is Φ ¯ = { ϕ ¯ 1 , , ϕ ¯ K / 2 } . It is evident that Φ Φ ¯ = { 1 , , K } and Φ Φ ¯ = . Furthermore, we re-arrange the columns of the matrix Z R T × K into the matrix Z ( c ) as follows:
Z ( c ) = z ϕ 1 z ϕ ¯ 1 z ϕ K / 2 z ϕ ¯ K / 2 .
The newly obtained matrix Z ( c ) is regarded as a multivariate time series that contains K ( c ) = 2 time series observed at time points 1 , , T ( c ) , where T ( c ) = ( K / 2 ) × T . Hence, the data matrix Y used by the DL algorithm (see (3)) has the expression Y = [ Y 1 Y 2 ] where, for i { 1 , 2 } , Y i R m × ( T ( c ) m + 1 ) is constructed from the entries of the i-th column of Z ( c ) as in Section 2.2. According to the convention from (4), the structure of the dictionary D is given by D = [ D 1 D 2 D 3 ] . It follows that the block D 1 is for the time series in group Φ , the block D 2 is for the time series in group Φ ¯ and the block D 3 is common for all the time series. The estimation of the missing values is performed as it was described in the previous sections.

5.2. Time Series Grouping

When deciding what time series to assign to the group Φ , we should take into consideration that all these time series are represented by using atoms from the same blocks: D 1 and D 3 . Hence, it is desirable (i) to minimize the overlap of the sequences of missing data for the columns of Z that belong to Φ , and (ii) to maximize the linear dependence between any two time series in Φ . Because it is non-trivial to combine the two requirements, we first focused on the condition (i). After some preliminary experiments, we came to the conclusion that the approach does not lead to good results. Then we investigated more carefully the condition (ii). The result of this investigation is the heuristic for cluster selection that we present below and which is based on the evaluation of the absolute value of the Pearson correlation between the pairs of columns of the matrix Z .
For ease of exposition, we introduce the following notation. If W R p × p is symmetric, then α ( W ) = 1 i < j p | w i j | , where | w i j | is the magnitude of the entry of W located at the intersection of i-th row and the j-th column. Let C R K × K be the matrix of the pairwise correlations between the columns of Z . For a subset Φ ( g ) { 1 , , K } whose cardinality equals K / 2 , we take C ( g ) to be the block of C that corresponds to the rows and the columns indexed by Φ ( g ) . Then we select Φ as follows:
Φ = argmin Φ ( g ) α C ( g ) 1 2 α C .
We note that the cardinality of Φ ¯ is also K / 2 . The formula implies that we find Φ with the property that the sum of the absolute pairwise correlations of the time series in cluster Φ is as close as possible to the sum of the absolute pairwise correlations of the time series in cluster Φ ¯ plus the sum of the absolute correlations of the pairs that contain a time series from Φ and a time series from Φ ¯ . Remark that, in the particular case when the correlations for all the pairs that contain a time series from Φ and a time series from Φ ¯ are zero, the sum of the absolute pairwise correlations of the time series in cluster Φ are approximately equal to the sum of the absolute pairwise correlations of the time series in cluster Φ ¯ . This approach has two limitations: (i) it can be applied only when the number of groups is two, and (ii) the computational burden is too high when K is large.
An alternative solution is a greedy algorithm which can be employed when the number of groups, K ( c ) , is greater than two. For simplicity, we assume that K is a multiple integer of K ( c ) . The algorithm constructs the groups as follows. Initially, the two time series that have the largest absolute correlation are included in the first group Φ 1 . In other words, we take
( ϕ 1 , ϕ 2 ) = argmax 1 i < j K | c i j | ,
and then Φ 1 = { ϕ 1 , ϕ 2 } . At the next step, it is included in Φ 1 the time series that increases the sum of the absolute correlations in the group the most:
ϕ 3 = argmax i { 1 , , K } Φ 1 | c i ϕ 1 | + | c i ϕ 2 | ,
and Φ 1 becomes Φ 1 = { ϕ 1 , ϕ 2 , ϕ 3 } . In general, after it was decided that Φ 1 = { ϕ 1 , , ϕ r } , where 2 r < K / K ( c ) , the ( r + 1 ) th time series is selected as follows:
ϕ r + 1 = argmax i { 1 , , K } Φ 1 q = 1 r | c i ϕ q | .
Once the first cluster is built, the second one is initialized with the two time series from { 1 , , K } Φ 1 that have the largest absolute correlation, and the steps described above are applied for obtaining the second cluster. The procedure continues until K ( c ) 1 groups are produced. The last group results automatically.

6. Experimental Settings

6.1. Simulation of the Missing Data

When we conduct experiments on a real-life multivariate time series Z R T × K , we randomly select the positions of the missing data. The selection is performed such that the number of missing data, M miss , is the same for each of the K time series. Hence, all the sets Ψ 1 , , Ψ K have the same cardinality. It follows that the percentage of missing data is
ρ = 100 M miss T
for each time series. In our experiments, we consider ρ = 5 % , ρ = 10 % , ρ = 15 % and ρ = 20 % .
The indexes of the missing data for a particular time series are independent of the positions of the missing data in the other time series from the same data set. They are selected by either sampling without replacement M miss integers from the set { 1 , , T } or by using the Polya urn model (with finite memory), which was introduced in [36]. The Polya urn model is well-known, and it was employed in various applications. Some of these applications are: modeling of the communication channels [36,37,38,39], image segmentation [40] and modeling of epidemics on networks [41]. In [15], we have proposed the use of the Polya urn model for simulating the missing data in time series.
In the Polya model, the urn initially contains R red balls and S black balls ( R < S ). At each time moment t 1 , a ball is drawn from the urn and, after each draw, ( 1 + Δ ) balls of the same color as the drawn ball are returned to the urn. We take Δ > 0 . More details about the selection of Δ are provided below after the presentation of the most important properties of the model. Since we want the model to have finite memory, the experiment is performed as described above only for 1 t M , where the parameter M is a positive integer (see the discussion in [36]). At each time moment t > M , a ball is drawn from the urn and, after each draw, two operations are executed: (i) ( 1 + Δ ) balls of the same color as the drawn ball are returned to the urn and (ii) Δ balls of the same color as the ball picked at time t M are removed from the urn.
A sequence of random variables { Ξ t } 1 t T is defined as follows: Ξ t = 1 if the ball drawn at time t is red and Ξ t = 0 if the ball drawn at time t is black. It was proven in [36,39] that the sequence { Ξ t } is a Markov process of order M. For t > M , let S ̲ t denote the state ( Ξ t M , , Ξ t 1 ) . The Polya urn model has the remarkable property that the probability of having Ξ t = 1 after the state S ̲ t was observed depends on the number of ones in ( Ξ t M , , Ξ t 1 ) , but not on their locations. We mention that M = 5 in our settings.
The indexes of the missing data correspond to the positions of ones in the sequence { Ξ t } 1 t T . It is known that P ( Ξ t = 1 ) = R R + S , where the symbol P ( · ) denotes probability. In our simulations, R and S are chosen such that P ( Ξ t = 1 ) = M miss T . According to [39], the correlation Corr ( Ξ t , Ξ t i ) is equal to δ 1 + δ , where 0 < i < M and δ = Δ R + S . This property allows us to simulate bursts of missing data by taking δ = 1 . Obviously, this is different from the situation when the sampling without replacement is applied and when it is more likely to have isolated missing data. At the same time, the simulation of the missing data by using the Polya urn model is different from the approach in [2], where blocks of missing data are considered.

6.2. Data Pre-Processing

After the missing values are simulated, each time series is decomposed into trend, seasonal component and remainder. Then the DLM imputation method is applied on the T × K matrix of the remainder components. For each time series, both the trend and the seasonal components are added to the estimates produced by DLM to obtain the estimates for the missing data.
The decomposition uses the implementation for the R package imputeTS [42,43], which is available at https://github.com/SteffenMoritz/imputeTS/blob/master/R/na_seadec.R (accessed on 28 February 2022). The implementation returns a specific output when it cannot detect a seasonal pattern for the analyzed time series. From the package imputeTS, we only use the decomposition technique and not the imputation methods because all the imputation methods are designed for univariate time series; thus they are sub-optimal for the multivariate case.

6.3. Performance Evaluation

Let z be a column of the data matrix Z R T × K . We collect in the vector z Ψ ¯ the entries of the time series z that are indexed by the elements of the set Ψ ¯ . With the convention that z ^ Ψ ¯ is the vector of estimates produced by DLM for the missing values of z , we calculate the following normalized error:
E = | | z Ψ ¯ z ^ Ψ ¯ | | 2 | | z Ψ ¯ | | 2 .
The normalized errors are computed similarly for the estimates yielded by the imputation methods from [2]. To rank the methods, we calculate scores as follows. For each time series z , the imputation method that achieves the minimum normalized error yields two points, the method that leads to the second smallest normalized error yields one point, and all other methods yield zero points. The number of points accumulated by each method from the experiment with all time series in Z are divided by 2 K to ensure that the scores take values in the interval [ 0 , 1 ] .
When the imputation is performed by using DLM 1 , the expression in (14) is replaced with
E 1 = | | z Ψ ¯ z ^ Ψ ¯ | | 1 | | z Ψ ¯ | | 1 ,
and the scores are calculated as explained above.
In the empirical comparison of the methods, we have used the code available at https://github.com/eXascaleInfolab/bench-vldb20.git (accessed on 3 October 2021) for the imputation methods that have been assessed in [2]. Short descriptions of these methods have been given in Section 1.1.
In the next section, we present the results obtained by DLM on five data sets that have been also used in [2]. For the sake of conciseness, we report the scores for DLM 1 only for three data sets. The experimental results can be reproduced by using the Matlab code available at https://www.stat.auckland.ac.nz/%7Ecgiu216/PUBLICATIONS.htm (accessed on 17 June 2022).

7. Experimental Results

7.1. Climate Time Series (K = 10, T = 5000)

The data set comprises monthly climate measurements that have been recorded at various locations in North America from 1990 to 2002. We do not transform the time series with the method from Section 6.2 because it does not improve the quality of the imputation. As K is relatively small, we cluster the time series into K ( c ) = 2 groups that are found by using (12): Φ = { 5 , 6 , 7 , 8 , 10 } and Φ ¯ = { 1 , 2 , 3 , 4 , 9 } . It is interesting that the rule in (12) yields the same grouping for all percentages of the missing data, for both sampling without replacement and for the Polya model.
Since the data are sampled monthly, it is natural to take the signal length m = 12 . We have n u { 5 × 2 m , 5 × 3 m , 5 × 4 m } , and for each value of n u , we take n 3 { 5 × m , 5 × 2 m , , n u 5 × m } . Observe that n 3 = n K ( c ) + 1 , and it denotes the size of the block of the structured dictionary which contains atoms that are common for all time series. The sparsity level s is selected from the set { 2 , 3 , 4 } . It follows that the total number of triples ( n u , n 3 , s ) that we consider is 18.
We compute the normalized errors (see Table A1, Table A2, Table A3, Table A4, Table A5, Table A6, Table A7 and Table A8 in Appendix A.1.1), which lead to the scores shown in Figure 1. From the plot in the left panel of the figure, it is evident that both DLM+BIC and DLM+EBIC work better than other methods when the missing data are simulated by sampling without replacement. The method DLM+BIC is slightly superior to DLM+EBIC for all missing percentages, except for ρ = 10 % . In the right panel of the figure, where the Polya urn model is employed for simulating the missing data, we observe the following: the method DynaMMo is ranked the best for all missing percentages, except for ρ = 20 % , where DLM+BIC works better than DynaMMo. In Figure 2, we can see that DLM1+BIC and DLM1+EBIC are also very good when sampling without replacement is employed, but their performance diminishes in the case of the Polya model (for more details, see Table A9, Table A10, Table A11, Table A12, Table A13, Table A14, Table A15 and Table A16 in Appendix A.1.2).

7.2. Meteoswiss Time Series (K = 10, T = 10,000)

The measurements represent weather data collected in Swiss cities from 1980 to 2018. After simulating the missing data, we remove both the trend and the seasonal components for each time series (see Section 6.2). The transformed time series are clustered into K ( c ) = 2 groups as follows: Φ = { 1 , 2 , 3 , 4 , 5 } and Φ ¯ = { 6 , 7 , 8 , 9 , 10 } (see (12)).
As the time interval between successive observations of these time series is 10 min, the most suitable value for m would be 6 × 24 = 144 (which corresponds to 24 h), but this makes the computational complexity too high. For keeping the computational burden at a reasonable level, we conduct experiments for three different values of m: m = 6 × 4 = 24 , m = 6 × 6 = 36 and m = 6 × 8 = 48 . Note that each value of m corresponds to a time interval (in hours) that is a divisor of 24. For each value of m, n u is selected from the set { 5 × 2 m , 5 × 3 m , 5 × 4 m } and for each value of n u , we have that n 3 { 5 × m , 5 × 2 m , , n u 5 × m } . The sparsity level s is chosen from the set { 3 , 4 , 6 } . We use these settings for the case when the missing data are generated by sampling without replacement and ρ = 5 % . The results can be found in Table A17. It can be easily noticed that both DLM+BIC and DLM+EBIC yield very good results for all values of m.
Taking into consideration these results, we further conduct the experiments for all cases of missing data simulations by setting m = 24 and s = 3 . Hence, we have six candidates for ( n u , n 3 , s ) . The grouping for which Φ = { 1 , 2 , 3 , 4 , 5 } is employed. The results are reported in Table A18, Table A19, Table A20, Table A21, Table A22, Table A23, Table A24 and Table A25, in Appendix A.2.1, and in Figure 3. Both DLM+BIC and DLM+EBIC have outstanding performance for sampling without replacement, and they are very good when the Polya model is used. In the latter case, DLM+BIC is less successful when ρ = 20 % . According to the scores shown in Figure 4, which are based on Table A26, Table A27, Table A28, Table A29, Table A30, Table A31, Table A32 and Table A33 (see Appendix A.2.2), the ranking of the imputation methods does not change significantly when the algorithm DLM is replaced with DLM 1 .

7.3. BAFU Time Series (K = 10, T = 50,000)

These are water discharge time series recorded by BAFU (BundesAmt Für Umwelt - Swiss Federal Office for the Environment) on Swiss rivers from 2010 to 2015. We do not remove the seasonality from the simulated incomplete time series because the method from [42] does not detect seasonal components for 3 out of 10 time series. Hence, the imputation is performed on the original time series. The grouping Φ = { 1 , 3 , 6 , 7 , 9 } and Φ ¯ = { 2 , 4 , 5 , 8 , 10 } is used to cluster the time series into K ( c ) = 2 groups when the missing data are simulated by sampling without replacement (for all missing percentages) or by the Polya model with ρ = 15 % . For the other three cases of missing data simulations, the grouping Φ = { 2 , 3 , 6 , 7 , 9 } and Φ ¯ = { 1 , 4 , 5 , 8 , 10 } is applied.
Relying on some of the empirical observations that we made for the MeteoSwiss time series and taking into consideration the fact that the sampling period for the BAFU time series is equal to 30 min, we set m = 12 (which corresponds to a time interval of 6 h). The possible values for n u and n 3 are calculated by using the same formulas as in the experiments with climate and MeteoSwiss time series. In those experiments, we have noticed that the small values are preferred for the sparsity level, hence we take s = 3 . This implies that the number of candidates for ( n u , n 3 , s ) is the same as in the case of MeteoSwiss time series.
The normalized errors for this data set are given in Appendix A.3 (see Table A34, Table A35, Table A36, Table A37, Table A38, Table A39, Table A40 and Table A41). In Figure 5, we show the scores computed for various imputation methods. From the left panel of the figure, it is clear that both DLM+BIC and DLM+EBIC are the best for ρ = 5 % , 10 % , 20 % , where their scores are approximately one. The imputation method DLM+BIC does not work as well as DLM+EBIC when the percentage of the missing data is ρ = 15 % . In the right panel of the figure, we observe that both DLM+BIC and DLM+EBIC have modest performance when the missing data are simulated by the Polya model. In this case, DynaMMo and ROSL are the winners.

7.4. Temperature Time Series (K = 50, T = 5000)

The data set contains average daily temperatures recorded at various sites in China from 1960 to 2012. After removing the trend and the seasonal components from all 50 time series with simulated missing data, we cluster them by applying the greedy algorithm which was introduced in Section 5.2. The resulting groups of time series are listed in Table A42 (see Appendix A.4). Note that the total number of groups is K ( c ) = 5 , and there are 10 time series in each group. It is interesting that the way in which the time series are assigned to the clusters depends very much on the method employed for simulating the missing data and on the percentage of missing data.
Mainly based on the lessons learned from the experiments with the data sets that have been analyzed in the previous sections, we have decided to take the signal length m = 12 . We opt for relatively small values for n u , thus n u is selected from the set 10 × 1 2 m , 10 × m , 10 × 3 2 m . For n u = 10 × 1 2 m , we have that n 6 = n u , which means that all the atoms are common for all the time series, and atoms that are specific for a certain group do not exist. For the other two values of n u , we allow n 6 to be selected from the set { n u 5 m , n u } . The sparsity level s is selected from the set { 2 , 3 } . Simple calculations lead to the conclusion that the total number of triples ( n u , n 6 , s ) equals 10.
The normalized errors from Table A43, Table A44, Table A45, Table A46, Table A47, Table A48, Table A49 and Table A50 (see Appendix A.4) lead to the scores displayed in Figure 6. For the case of sampling without replacement, DLM+BIC, DLM+EBIC and ROSL have similar performance, and they are followed by CDRec. It is interesting that although SVT has a very modest rank when the percentage of the missing data is small, it becomes superior to all other methods when ρ = 20 % . As we have already observed for the BAFU time series, DLM+BIC and DLM+EBIC are not well ranked when the data are simulated by the Polya model. The difference compared with the results for the BAFU time series comes from the fact that ROSL and DynaMMo are not clear winners. This time, ROSL and CDRec are better than the other competitors. Similar to the case when the missing data are simulated by sampling without replacement, SVT outperforms all other imputation methods when ρ = 20 % .

7.5. Air Time Series (K = 10, T = 1000)

The data set comprises hourly sampled air quality measurements that have been recorded at monitoring stations in China from 2014 to 2015. For the case when the missing data are simulated by sampling without replacement (with ρ = 5 % , 10 % , 15 % ) and the Polya model (with ρ = 5 % , 10 % , 20 % ), the groups produced by the criterion in (12) are Φ = { 2 , 3 , 5 , 9 , 10 } and Φ ¯ = { 1 , 4 , 6 , 7 , 8 } . For the other two cases of missing data simulations, the clustering is: Φ = { 1 , 2 , 4 , 5 , 9 } and Φ ¯ = { 3 , 6 , 7 , 8 , 10 } . We do not remove the trend and seasonal components since, for each missing data simulation, the seasonal patterns are not detected in more than half of the time series.
We have applied DLM and DLM 1 algorithms on this data set. In both cases, we take the signal length m = 24 because we analyze hourly data. The parameter n u is selected from the set { 5 × 2 m , 5 × 3 m , 5 × 4 m } . For each value of n u , we have that n 3 { 5 × m , 5 × 2 m , , n u 5 × m } . The sparsity level s is selected from the set { 2 , 3 , 4 } .
In Figure 7, we show the scores obtained by various imputation methods when the DLM algorithm is applied, and the normalized errors are computed with the formula from (14), see also Table A51, Table A52, Table A53, Table A54, Table A55, Table A56, Table A57 and Table A58 in Appendix A.5.1. It is evident that both DLM+BIC and DLM+EBIC have better performance when sampling without replacement is employed to simulate the missing data. SVT is clearly the best method for both cases of missing data simulations and for all percentages ρ . The ranking of the imputation methods is the same in Figure 8, where the scores have been calculated by using (15), see Table A59, Table A60, Table A61, Table A62, Table A63, Table A64, Table A65 and Table A66 in Appendix A.5.2. An important difference between the two figures is that DLM 1 and the corresponding IT criteria are employed in Figure 8.

8. Final Remarks

In this work, we have exemplified how two techniques for dimensionality reduction can be employed for extending the use of the DLM algorithm to data sets that contain tens of time series. It remains to be investigated by the future research how the previous results on time series clustering can be utilized for grouping the time series and finding the number of groups (see, for example [44], for the definition of a metric that evaluates the dissimilarity between two time series). We have also derived the novel DLM 1 algorithm and the corresponding BIC and EBIC criteria. We have conducted a large set of experiments for comparing DLM and DLM 1 with nine algorithms that represent the state-of-the-art in the multivariate time series imputation.
Our empirical study confirms a fact already observed in the previous literature: There is no imputation method which always yields the best results. Although not always the best, our method clearly outperforms the other studied methods for some missing data models. So, it appears to be a useful addition to the existing methods. DLM tends to work better when the missing data are isolated (see the results for simulation by sampling without replacement) than in the case when the sequences of missing data are long (see the results for simulation by using the Polya model).
Both BIC and EBIC are effective in selecting the best structure of the dictionary and the sparsity. It is interesting that small values are chosen for the sparsity s, and this supports the idea that sparse models are appropriate for the multivariate time series. The values selected for n u are also relatively small. Recall that n u equals the number of atoms used in the representation of a specific time series (or a group of time series) plus the number of atoms that are common for all the time series in the data set.
Our imputation method can also be applied when the percentage of the missing data is not the same for all the time series in the data set. Based on the experimental results, it is easy to see that the percentage of missing data does not have an important influence on how DLM is ranked in comparison with other methods; thus we expect the same to be true when the number of missing data varies across the time series.

Author Contributions

Conceptualization, X.Z., B.D. and C.D.G.; methodology, X.Z., B.D. and C.D.G.; software, X.Z., B.D. and C.D.G.; validation, X.Z.; investigation, X.Z.; data curation, X.Z.; writing—original draft preparation, X.Z., B.D. and C.D.G.; writing—review and editing, X.Z., B.D., J.L. and C.D.G.; supervision, J.L. and C.D.G.; project administration, C.D.G. All authors have read and agreed to the published version of the manuscript.

Funding

B.D. was supported in part by a grant of the Ministry of Research, Innovation and Digitization, CNCS-UEFISCDI, project number PN-III-P4-PCE-2021-0154, within PNCDI III.

Data Availability Statement

Not applicable.

Acknowledgments

X.Z. would like to thank to Mourad Khayati from Department of Computer Science of the University of Fribourg, Switzerland for numerous hints on how to use the data and the code from https://github.com/eXascaleInfolab/bench-vldb20.git (accessed on 3 October 2021).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The tables below contain the normalized errors (see (14) and (15)) for the imputation methods that are assessed in this work. Note that the errors are computed for each time series in the data sets used in the empirical study. For each row in the tables, the best result is written in red and the second best result is written with bold font. For the methods DLM and DLM 1 , we also show the values of the triple ( n u , n K + 1 , s) selected by BIC (see (9) and (11)) and EBIC (10). In the caption of each table, we provide information about the missing data: how they have been simulated and what is the value of ρ (13). The only exception is Table A42 in Appendix A.4, which does not contain normalized errors, but shows the clustering of the time series.

Appendix A.1. Climate Time Series: Numerical Results

Appendix A.1.1. Results Obtained by Using DLM

Table A1. Climate time series: Sampling without replacement ( ρ = 5 % ).
Table A1. Climate time series: Sampling without replacement ( ρ = 5 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 60 n 3 = 180
TS10.2206 0.2378 0.90760.92061.13680.87910.89680.77940.97390.54960.8835
TS20.3859 0.4220 0.64360.76920.79930.62250.63570.61400.69310.54370.6271
TS30.4527 0.4836 0.90990.93460.85710.87320.88110.78240.87890.72860.8764
TS40.3806 0.4138 0.99011.52101.06131.00051.01040.84150.95800.86241.0043
TS50.59970.64530.59790.7721 0.5650 0.58620.59180.55970.57230.60480.5894
TS60.4955 0.5227 0.62120.75660.58560.60370.60570.68120.66900.55290.6027
TS70.1928 0.2078 0.69140.88580.75100.66830.67460.73311.02710.52010.6696
TS80.70100.72250.63460.71270.63690.62600.62520.67480.64240.6557 0.6253
TS9 0.4412 0.42320.90461.07130.91320.93050.94320.86291.01280.79830.9340
TS10 0.5064 0.49630.73000.93660.84760.72160.72880.72191.12010.65850.7245
Table A2. Climate time series: Sampling without replacement ( ρ = 10 % ).
Table A2. Climate time series: Sampling without replacement ( ρ = 10 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 60 n 3 = 180
TS1 0.3141 0.30790.88540.95011.16320.86250.87930.77620.99420.56580.8676
TS20.3910 0.4051 0.64540.88260.80760.61520.63120.61440.77610.53340.6226
TS30.4310 0.4611 1.01240.94820.84840.94830.96360.79130.91590.75690.9514
TS40.3739 0.4019 0.94421.21651.13010.95570.96600.83991.13050.81200.9574
TS50.62850.65360.56300.75620.56550.55520.55750.58680.56590.5763 0.5556
TS6 0.5302 0.52950.65701.05710.61880.64300.64810.69100.68380.57870.6439
TS70.1716 0.1882 0.70240.87660.71070.67600.68660.72251.04630.50450.6772
TS80.74430.7381 0.6309 0.71410.63950.62990.63440.64530.65170.64340.6321
TS9 0.4504 0.44780.90881.09340.90060.95150.96590.85611.00370.78340.9546
TS10 0.5064 0.48630.78620.95470.79820.77440.78030.77351.25190.70480.7760
Table A3. Climate time series: Sampling without replacement ( ρ = 15 % ).
Table A3. Climate time series: Sampling without replacement ( ρ = 15 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 60 n 3 = 180
TS1 0.2762 0.26360.96451.00271.15020.88500.91420.78590.93770.55330.9007
TS20.4363 0.4590 0.71131.08610.81710.64710.66970.66350.73960.56120.6638
TS30.4786 0.5362 1.12010.95920.88380.98891.01800.80240.93620.77711.0032
TS40.4216 0.4314 0.99391.60141.08381.00611.01980.85201.06600.82951.0139
TS50.67330.69310.66060.99570.63020.63770.65170.62370.6289 0.6256 0.6497
TS60.4903 0.5191 0.67710.96160.59940.64070.65430.68310.68650.54650.6496
TS70.2039 0.2113 0.78410.87740.71930.70260.72390.72470.97320.51250.7212
TS80.74030.76230.67060.88880.62650.66280.6738 0.6560 0.68140.65980.6732
TS9 0.5364 0.50810.93381.06570.91340.95370.97370.85871.04720.80680.9665
TS100.4992 0.5110 0.77611.18710.82770.75260.76040.77041.19420.66690.7602
Table A4. Climate time series: Sampling without replacement ( ρ = 20 % ).
Table A4. Climate time series: Sampling without replacement ( ρ = 20 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 60 n 3 = 180
TS10.2684 0.2989 1.03921.22511.14570.88580.93790.80190.98230.58300.9257
TS20.4676 0.4778 0.89301.41690.85770.71340.77750.68470.84360.58460.7721
TS30.4978 0.5197 1.01420.97210.84610.91770.94220.82070.93070.73490.9261
TS40.4291 0.4642 1.04041.62391.09290.96721.01130.85161.25700.78921.0085
TS50.65790.64850.58180.89840.60540.56890.57590.62250.59110.5892 0.5719
TS60.5604 0.5611 0.66220.92140.60530.63930.64980.70080.69200.56740.6425
TS70.2014 0.2298 0.74250.84860.74100.70110.72370.74421.08990.51370.7074
TS80.76870.7896 0.6668 0.77660.66050.66990.67600.70120.69390.67440.6722
TS9 0.5123 0.50260.90511.15120.91040.92470.94700.86881.03670.77910.9312
TS10 0.5210 0.51970.71731.09140.85340.71050.71720.76881.31760.64970.7123
Table A5. Climate time series: Polya model ( ρ = 5 % ).
Table A5. Climate time series: Polya model ( ρ = 5 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 60 n 3 = 180
TS11.53390.59991.08750.90791.08491.00331.04020.74380.9397 0.6961 1.0238
TS20.51480.54520.58470.87600.84090.58700.60500.64490.7186 0.5397 0.5936
TS32.46900.82700.94460.98940.90880.87060.88250.74420.9026 0.7754 0.8759
TS41.58780.72570.94491.55921.07530.96220.97360.86601.0507 0.8517 0.9706
TS50.90750.80440.48670.66710.54740.47910.48210.52270.50370.5441 0.4805
TS62.82310.80230.63501.01700.6313 0.6155 0.62140.66380.64350.59850.6175
TS72.4865 0.6010 0.64830.86250.74720.63800.64430.72061.02700.52240.6376
TS82.68430.79760.58150.7659 0.5559 0.60090.61450.54030.61350.58300.6093
TS91.20030.76490.87881.00360.94310.90640.91860.84771.0531 0.8149 0.9097
TS102.12810.72430.57340.80830.82920.58650.58450.73460.9960 0.5809 0.5842
Table A6. Climate time series: Polya model ( ρ = 10 % ).
Table A6. Climate time series: Polya model ( ρ = 10 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 60 n 3 = 180
TS10.49061.06680.97040.87990.96000.93150.95230.77680.9242 0.6934 0.9396
TS20.80931.20250.67500.89250.85610.64090.6653 0.6263 0.75400.55690.6490
TS3 0.7706 1.18660.99410.97250.92340.93000.94710.80020.91330.76740.9396
TS40.74881.35680.97281.33941.06300.98831.0031 0.8413 1.11030.86060.9951
TS50.78620.78850.57660.75560.59990.56760.57320.59180.57030.5985 0.5689
TS60.71620.73360.66310.8485 0.6282 0.64250.64960.68430.64640.58550.6447
TS7 0.5634 0.57560.70290.75930.78230.67180.68450.71461.04170.52540.6798
TS80.84540.82140.67380.81090.6639 0.6713 0.67380.68640.69290.68910.6725
TS9 0.8455 1.17030.87230.90820.89470.88630.89750.85351.07710.80480.8900
TS100.5941 0.6240 0.71830.99150.80480.70610.71500.73081.23040.65460.7091
Table A7. Climate time series: Polya model ( ρ = 15 % ).
Table A7. Climate time series: Polya model ( ρ = 15 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 60 n 3 = 180
TS10.66231.05461.07981.04941.08910.94970.99250.79891.0618 0.7023 0.9922
TS20.72680.93310.84101.05430.83700.69700.7368 0.6759 0.80740.61270.7424
TS3 0.7846 0.79511.09280.96450.93320.94530.99200.81660.91720.75200.9678
TS40.72571.02850.94721.58681.06320.91270.92890.86201.1940 0.7932 0.9260
TS50.85050.88420.68701.10350.6246 0.6239 0.65020.61930.62710.65030.6464
TS60.71780.99690.69511.2827 0.6276 0.64760.66890.71630.70320.59530.6624
TS7 0.6162 0.67480.85780.97290.83380.78130.83340.73501.23790.59200.8106
TS80.83030.93810.73261.06160.7029 0.6996 0.72420.70870.74110.69860.7206
TS90.78071.03600.89640.96120.91820.91160.93210.86511.0274 0.8242 0.9203
TS100.73650.84990.90791.70800.95400.83170.8692 0.7640 1.56980.78000.8616
Table A8. Climate time series: Polya model ( ρ = 20 % ).
Table A8. Climate time series: Polya model ( ρ = 20 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 180
s = 2s = 2
n 3 = 60 n 3 = 120
TS10.57730.68770.96611.26871.29440.88260.91850.80510.9959 0.6391 0.9124
TS20.77050.80070.75181.06780.7904 0.6526 0.68090.67650.80300.59220.6834
TS30.75020.85091.02672.23260.92220.91880.94340.81870.9178 0.7690 0.9332
TS40.7871 0.8449 0.99171.18061.06390.96610.98100.86951.09510.85490.9832
TS50.86350.88030.68190.73450.6824 0.6583 0.68230.64640.69040.66370.6724
TS60.70280.69490.66831.0010 0.6070 0.64460.65760.72050.72340.57780.6490
TS70.78770.78430.75330.78380.8042 0.6952 0.71620.75371.02470.56830.7089
TS80.89870.92510.72120.8163 0.6920 0.71610.73720.68170.73750.70800.7283
TS90.7284 0.8182 0.97671.04950.93170.99891.03300.87461.06630.85151.0227
TS100.6367 0.6441 0.78930.82290.82190.78690.79530.81081.22280.74140.7905

Appendix A.1.2. Results obtained by using DLM1

Table A9. Climate time series: Sampling without replacement ( ρ = 5 % ).
Table A9. Climate time series: Sampling without replacement ( ρ = 5 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 60 n 3 = 180
TS10.2103 0.2245 0.85940.86441.09380.83450.85040.77080.93750.54000.8382
TS20.3741 0.4063 0.57770.74920.77300.56380.57230.57890.68260.47950.5654
TS30.4323 0.4799 0.88210.91860.80960.83560.84560.74950.82330.69600.8412
TS40.3703 0.4176 0.99661.43681.07051.00741.02060.83640.94420.86541.0132
TS50.60570.62580.53980.71320.51910.52870.53240.5410 0.5253 0.56830.5305
TS60.5005 0.5038 0.60400.76900.56850.58520.58670.67050.65140.52470.5839
TS7 0.2160 0.20310.66270.88860.72820.64540.65090.72301.01340.50680.6463
TS80.61250.58180.55160.64890.5281 0.5484 0.55110.58430.57150.55830.5494
TS90.3788 0.3953 0.93111.05290.84860.96350.98310.80330.96730.80360.9730
TS100.4958 0.5023 0.62110.83620.78750.61400.62100.64901.06910.56680.6161
Table A10. Climate time series: Sampling without replacement ( ρ = 10 % ).
Table A10. Climate time series: Sampling without replacement ( ρ = 10 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 180
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.2338 0.2356 0.81800.82091.09480.79420.81120.75990.93900.53530.7995
TS20.4073 0.4233 0.59870.73770.74710.56920.58400.58030.71190.48180.5768
TS30.5076 0.5258 0.96330.92810.78970.91060.92250.77230.88240.73900.9121
TS40.4399 0.4535 0.94021.16801.12880.95350.96590.83361.07580.80720.9561
TS50.62270.61390.50270.67440.5100 0.4950 0.49560.55660.52350.51890.4943
TS60.5269 0.5277 0.62721.02200.59460.61310.61750.67200.65910.54820.6135
TS70.2757 0.2761 0.64330.89410.64380.62800.63530.70570.98700.47180.6277
TS80.65250.66110.59040.67580.55980.59500.60430.58090.6216 0.5782 0.6003
TS9 0.4141 0.40510.91991.08250.83040.97190.99410.80540.98610.77640.9798
TS100.4810 0.4832 0.71950.86630.73120.70810.71300.72631.15060.63640.7088
Table A11. Climate time series: Sampling without replacement ( ρ = 15 % ).
Table A11. Climate time series: Sampling without replacement ( ρ = 15 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 180
s = 2s = 2
n 3 = 60 n 3 = 120
TS1 0.2964 0.28650.87640.86901.03590.81440.84000.77220.88330.53550.8261
TS20.4662 0.4741 0.65150.96660.76510.60760.62770.62940.70320.50480.6193
TS30.5636 0.5786 1.02680.94380.81170.93070.95310.78610.89720.75040.9419
TS40.5144 0.5263 1.00451.52891.07571.01821.03130.85101.03830.83651.0273
TS50.69990.68770.58470.76220.5742 0.5709 0.57880.59880.57080.58270.5765
TS6 0.5305 0.53330.62950.76170.56810.60490.61380.66570.64790.52290.6079
TS7 0.3366 0.32260.70210.84380.68140.65770.67170.70410.92640.48360.6658
TS80.66940.66220.60480.71210.53270.61220.6244 0.5860 0.63360.59220.6205
TS9 0.5185 0.49770.94111.07140.84460.97871.00350.80601.03950.80330.9936
TS100.5308 0.5367 0.69110.88910.77100.67710.68320.72061.13440.60400.6810
Table A12. Climate time series: Sampling without replacement ( ρ = 20 % ).
Table A12. Climate time series: Sampling without replacement ( ρ = 20 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 120
s = 2s = 2
n 3 = 60 n 3 = 60
TS10.3706 0.3792 0.91860.95621.04490.83150.86810.78950.91110.56240.8532
TS20.5608 0.5602 0.73671.18440.81740.65240.68990.64250.78930.52660.6810
TS3 0.6398 0.63700.95440.96600.77080.88300.90260.80130.90300.71580.8892
TS4 0.5846 0.57040.98301.48341.10160.96050.99260.83931.12600.77080.9832
TS50.69470.67420.53890.81710.55840.52840.53230.59590.55130.5458 0.5298
TS60.59790.58660.63630.9182 0.5795 0.61840.62910.68390.66400.54240.6215
TS70.4053 0.4338 0.69130.85040.67900.65640.67740.72571.00440.47280.6636
TS80.70120.69350.61200.73070.57720.62270.63160.61680.6502 0.5890 0.6263
TS90.5611 0.5855 0.90421.11700.83620.93510.96030.82471.02190.76580.9448
TS100.57850.55340.63260.97460.80480.62670.63310.71931.1550 0.5768 0.6282
Table A13. Climate time series: Polya model ( ρ = 5 % ).
Table A13. Climate time series: Polya model ( ρ = 5 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 60 n 3 = 180
TS10.6019 0.6286 1.07810.89161.03370.99561.03340.73900.90450.70501.0084
TS20.67860.69400.54190.82470.7931 0.5352 0.54830.61190.66800.48380.5357
TS30.83490.84300.90590.97630.85770.84130.85160.73930.8804 0.7465 0.8461
TS40.7421 0.7546 0.92451.56981.07300.94820.96190.83431.05100.83390.9562
TS50.80350.79820.45000.61950.51070.44180.44400.50610.45600.5069 0.4429
TS60.79180.79280.58040.94590.5751 0.5601 0.56410.62650.58830.55660.5611
TS70.63430.64290.60500.84940.7045 0.6001 0.60810.69681.00630.49720.6016
TS80.83250.82720.48690.64520.43590.51420.5324 0.4633 0.52740.46490.5260
TS90.77960.75970.90501.03760.86290.94090.9614 0.7744 1.06030.81680.9501
TS100.78980.80200.51990.71900.77840.53650.53330.70950.9513 0.5317 0.5329
Table A14. Climate time series: Polya model ( ρ = 10 % ).
Table A14. Climate time series: Polya model ( ρ = 10 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 180
s = 2s = 2
n 3 = 60 n 3 = 120
TS10.65260.66670.88600.78670.86750.85640.87290.75410.8396 0.6533 0.8603
TS20.79420.78600.62250.85600.82060.59600.6150 0.5911 0.71800.52240.6037
TS30.79630.80490.93410.95760.86660.87950.8919 0.7843 0.89150.73890.8856
TS40.7353 0.7373 0.97881.32971.05480.99581.01310.83111.04910.86771.0024
TS50.82480.84480.51090.65930.53020.50230.50630.56220.50960.5459 0.5027
TS60.75170.76680.63970.8052 0.5986 0.62100.62700.65820.61150.55510.6222
TS70.67500.66890.66130.74680.7378 0.6317 0.64150.70310.96580.49180.6358
TS80.82060.82130.61680.76870.58700.62200.6312 0.6109 0.65850.61540.6268
TS9 0.7840 0.77740.88550.89030.83880.89760.91320.81161.04990.80270.9048
TS100.77200.76260.65880.88600.7420 0.6468 0.65420.69211.09630.60300.6491
Table A15. Climate time series: Polya model ( ρ = 15 % ).
Table A15. Climate time series: Polya model ( ρ = 15 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 120
s = 2s = 2
n 3 = 120 n 3 = 60
TS10.7120 0.6969 0.98350.95281.01510.89080.92840.78580.99980.67920.9244
TS20.78400.77750.73190.96680.78520.65100.6813 0.6460 0.75570.55770.6814
TS30.83210.83180.99700.94840.84740.88270.9195 0.7970 0.87730.70770.9008
TS40.7928 0.7898 0.90361.55731.05800.88310.89570.84171.02990.77040.8934
TS50.85500.84770.59460.75830.57910.56420.57270.59280.57680.6134 0.5725
TS60.81860.81290.65680.87050.6329 0.6323 0.64150.70220.69940.58840.6376
TS70.72320.72610.77340.84490.77040.72670.7591 0.7190 1.14030.55310.7433
TS80.87040.85610.64970.82610.62030.64390.6582 0.6273 0.69710.63180.6553
TS90.7777 0.7823 0.90760.96810.87120.93320.95920.81601.03500.82000.9459
TS100.81410.80740.80271.15250.84680.76010.7805 0.7270 1.35610.72330.7741
Table A16. Climate time series: Polya model ( ρ = 20 % ).
Table A16. Climate time series: Polya model ( ρ = 20 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 120
s = 2s = 2
n 3 = 120 n 3 = 60
TS1 0.7124 0.71680.87741.21021.15760.81150.84450.79370.92420.60610.8375
TS20.78340.79550.67681.04940.7561 0.6077 0.63180.65080.75830.53940.6310
TS30.84300.85030.94312.13510.85140.87450.8913 0.8084 0.89480.74530.8830
TS40.8271 0.8349 0.99371.18971.06520.97560.99020.86131.06800.85870.9892
TS50.87810.87970.58660.68170.60610.57050.58050.61010.62390.6047 0.5762
TS60.82100.82310.63940.9726 0.5960 0.62250.62860.70400.69700.55800.6232
TS70.79460.78330.69310.77210.7533 0.6589 0.67440.74380.97630.54100.6670
TS80.88550.88760.63340.78650.60140.63610.6523 0.6141 0.66930.63050.6455
TS90.8000 0.8022 0.97651.04900.85751.02511.06840.82871.09790.84701.0535
TS100.82670.8275 0.6779 0.73240.79510.67940.68420.76091.13940.64440.6807

Appendix A.2. Meteoswiss Time Series: Numerical Results

Appendix A.2.1. Results obtained by using DLM

Table A17. MeteoSwiss time series: Sampling without replacement ( ρ = 5 % ). Comparison of the results obtained for three different values of m.
Table A17. MeteoSwiss time series: Sampling without replacement ( ρ = 5 % ). Comparison of the results obtained for three different values of m.
TimeDLM+BICDLM+EBICDLM+BICDLM+EBICDLM+BICDLM+EBIC
Seriesm = 24m = 24m = 36m = 36m = 48m = 48
n u = 480 n u = 480 n u = 720 n u = 540 n u = 960 n u = 720
s = 3s = 3s = 3s = 3s = 3s = 3
n 3 = 360 n 3 = 360 n 3 = 540 n 3 = 360 n 3 = 720 n 3 = 480
TS1 0.0131 0.0131 0.01270.01270.01320.0132
TS20.00720.0072 0.0074 0.00750.00770.0079
TS30.02210.0221 0.0217 0.02140.02260.0224
TS40.01680.0168 0.0158 0.01560.01620.0166
TS50.01750.01750.0167 0.0170 0.01780.0179
TS60.03070.03070.0294 0.0278 0.02760.0291
TS70.08370.08370.08360.08360.0831 0.0834
TS8 0.0796 0.0796 0.0791 0.0796 0.07980.0809
TS90.08110.0811 0.0791 0.07790.08110.0811
TS100.08920.08920.08660.0855 0.0853 0.0849
Table A18. MeteoSwiss time series: Sampling without replacement ( ρ = 5 % ).
Table A18. MeteoSwiss time series: Sampling without replacement ( ρ = 5 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 480
s = 3s = 3
n 3 = 360 n 3 = 360
TS10.01310.01310.06320.08420.05680.06000.06000.05430.0589 0.0538 0.0589
TS20.00720.00720.04430.08470.04280.04290.04300.04190.0642 0.0407 0.0426
TS30.02210.02210.09320.11550.07340.08960.09100.08490.0864 0.0718 0.0873
TS40.01680.01680.10010.21010.10440.09990.10020.10030.0960 0.0809 0.0975
TS50.01750.01750.07470.15720.06260.07140.07180.06290.0654 0.0588 0.0695
TS60.03070.03070.08980.12700.09000.08920.08990.08460.0933 0.0841 0.0883
TS70.08370.08370.21100.18610.13270.15770.16040.14980.1531 0.1302 0.1511
TS80.07960.07960.22660.29840.14650.19670.20520.18270.1841 0.1360 0.1809
TS90.08110.08110.15550.29630.13350.15460.15650.15610.1516 0.1302 0.1514
TS100.08920.08920.45080.50710.10880.34510.64890.17780.5598 0.0914 0.2171
Table A19. MeteoSwiss time series: Sampling without replacement ( ρ = 10 % ).
Table A19. MeteoSwiss time series: Sampling without replacement ( ρ = 10 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 360
s = 3s = 3
n 3 = 360 n 3 = 240
TS1 0.0137 0.01330.05780.10570.05350.05700.05690.05240.06780.05080.0564
TS20.0068 0.0069 0.04320.15850.04170.04240.04240.04150.06930.03970.0421
TS30.0212 0.0213 0.10350.11450.08220.10060.10380.09031.09910.07970.0975
TS4 0.0161 0.01590.09800.28480.09810.09900.10020.09240.09720.07940.0965
TS50.0249 0.0254 0.08350.16370.06770.08080.08180.07130.08160.06650.0787
TS6 0.0217 0.02100.08170.13280.08540.08220.08260.07950.09230.07710.0814
TS7 0.0698 0.06920.21440.21440.12480.16770.17430.16950.18190.13040.1588
TS80.0741 0.0757 0.26030.31240.14130.21380.23050.18140.21840.12990.1897
TS90.0756 0.0770 0.16170.32640.13270.16000.16310.16520.17390.13450.1555
TS100.0818 0.0825 0.47010.55960.11680.35410.61480.20730.21400.08830.2412
Table A20. MeteoSwiss time series: Sampling without replacement ( ρ = 15 % ).
Table A20. MeteoSwiss time series: Sampling without replacement ( ρ = 15 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 360
s = 3s = 3
n 3 = 360 n 3 = 240
TS10.0140 0.0141 0.06000.11260.05430.05950.06000.05260.07060.05040.0580
TS20.00790.00790.04400.19010.04180.04370.04390.04200.0741 0.0412 0.0433
TS30.0273 0.0278 0.10590.11440.07750.10100.10410.09041.36980.07810.0973
TS40.0178 0.0181 0.09680.30180.09220.09620.09720.08880.09880.07980.0945
TS5 0.0202 0.01970.07890.18200.06390.07710.07810.06580.08080.06160.0745
TS60.0236 0.0237 0.08070.13930.08380.07980.08040.07670.08620.07260.0785
TS7 0.0821 0.08140.27980.23110.13780.21810.23350.18280.19380.14320.1915
TS80.0698 0.0703 0.24730.32420.12950.21500.24950.17070.20610.11810.1802
TS90.0763 0.0766 0.15990.33660.12350.15430.16090.15200.16930.12470.1451
TS100.0801 0.0808 0.41000.54570.11050.31440.46660.19970.20290.09050.2465
Table A21. MeteoSwiss time series: Sampling without replacement ( ρ = 20 % ).
Table A21. MeteoSwiss time series: Sampling without replacement ( ρ = 20 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 3
n 3 = 360 n 3 = 120
TS1 0.0150 0.01480.05990.15250.05500.05860.05870.05300.05830.05010.0581
TS2 0.0082 0.00780.04730.15130.04470.04670.04680.04480.06340.04320.0464
TS3 0.0255 0.02540.09780.10470.06750.09390.09810.08240.08720.07220.0899
TS4 0.0185 0.01780.10710.26600.10380.10780.10900.09830.09980.08230.1060
TS50.0262 0.0264 0.08050.22200.06770.07780.07870.06860.07230.06290.0763
TS6 0.0262 0.02580.08380.19160.08170.08210.08310.07780.08820.07610.0808
TS7 0.0857 0.08370.26120.21600.13220.19140.19650.17100.23180.13660.1801
TS8 0.0753 0.07490.27580.31500.14090.22040.25430.18440.26950.13340.1846
TS9 0.0751 0.07470.16700.30360.12970.15760.16400.15810.21400.13110.1530
TS10 0.0849 0.08410.42390.61820.11520.31960.51810.19890.43890.09160.2393
Table A22. MeteoSwiss time series: Polya model ( ρ = 5 % ).
Table A22. MeteoSwiss time series: Polya model ( ρ = 5 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 480
s = 3s = 3
n 3 = 360 n 3 = 360
TS10.06620.06620.05960.0761 0.0504 0.05810.05880.04680.05310.05120.0572
TS20.03390.03390.03420.06980.03400.03360.0338 0.0334 0.05790.03300.0335
TS30.03430.03430.07690.08160.08060.07740.0803 0.0640 0.06960.06690.0755
TS40.04750.04750.10820.21010.11030.10700.10820.10420.1037 0.0951 0.1056
TS50.05030.05030.07600.13040.07220.07580.0769 0.0620 0.07060.07190.0753
TS60.02870.02870.05260.09770.04710.05140.0520 0.0463 0.05920.04850.0511
TS70.14350.14350.27880.2602 0.1652 0.21820.22080.20270.18090.17510.2101
TS80.13020.13020.18160.20660.10010.16370.17760.11990.1436 0.1175 0.1511
TS90.11190.11190.16120.2597 0.1435 0.16300.16450.15890.16520.14830.1611
TS100.13880.13880.38560.57160.19540.32840.50850.21590.5496 0.1697 0.2791
Table A23. MeteoSwiss time series: Polya model ( ρ = 10 % ).
Table A23. MeteoSwiss time series: Polya model ( ρ = 10 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 360
s = 3s = 3
n 3 = 360 n 3 = 240
TS1 0.0494 0.04690.06870.09670.05740.06400.06600.06020.09060.05950.0643
TS2 0.0106 0.00980.03630.08180.03220.03530.03570.03460.07610.03400.0356
TS30.0414 0.0436 0.08190.08410.06870.07440.08430.06641.10170.06530.0758
TS40.0371 0.0408 0.11350.20030.11370.10850.11180.10280.11640.09510.1088
TS50.0503 0.0535 0.08900.19230.07350.08600.08470.07630.10580.07500.0835
TS60.0250 0.0252 0.06790.10270.05980.06280.06730.05900.08880.06010.0650
TS70.10530.18480.22670.1806 0.1294 0.16210.17570.14390.14610.14020.1632
TS8 0.0999 0.09760.20890.25030.13400.18450.20610.15770.16270.13160.1765
TS90.0948 0.0970 0.13970.25860.11470.15600.13500.13840.14000.11650.1303
TS100.1283 0.1371 0.47320.48460.21420.35640.61500.17640.17810.16110.3046
Table A24. MeteoSwiss time series: model ( ρ = 15 % ).
Table A24. MeteoSwiss time series: model ( ρ = 15 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 3
n 3 = 360 n 3 = 120
TS10.06360.10290.06540.10850.05910.06450.06450.05570.0621 0.0558 0.0622
TS20.0206 0.0268 0.05120.11200.05010.05100.05100.05040.06050.04770.0503
TS30.0562 0.0689 0.09710.10970.07500.09540.09860.08300.08700.07860.0915
TS40.0445 0.0497 0.10570.18760.09550.11470.11200.09090.09660.09150.1108
TS50.0579 0.0662 0.08460.19010.07650.08460.08380.07870.07590.07080.0813
TS6 0.0560 0.05280.09030.12750.09660.08900.08880.08810.09080.08480.0880
TS70.14390.14600.22240.2122 0.1387 0.20020.19480.16600.18110.13370.1670
TS80.18210.16280.34140.31650.18440.23140.29470.20590.2376 0.1684 0.2199
TS90.11760.14240.15140.25770.13390.15210.15170.15150.1647 0.1232 0.1389
TS100.11840.19760.42390.55780.16030.30770.49590.18650.4601 0.1557 0.2561
Table A25. MeteoSwiss time series: Polya model ( ρ = 20 % ).
Table A25. MeteoSwiss time series: Polya model ( ρ = 20 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 3
n 3 = 360 n 3 = 120
TS10.46470.12280.06490.24410.05790.06680.0660 0.0574 0.06350.05630.0653
TS20.36740.03540.04210.1161 0.0392 0.04290.04220.03970.05590.03940.0425
TS30.18170.06490.10710.11510.08860.10330.10510.09120.0915 0.0826 0.1002
TS42.27410.25790.10950.36470.11230.11230.11340.1075 0.1065 0.10070.1111
TS50.29960.03720.07070.21830.06510.07120.07200.06180.0641 0.0588 0.0692
TS60.0497 0.0541 0.08020.26650.08090.07890.07980.07470.08750.07560.0789
TS70.14910.15530.32520.21870.13120.26840.23540.16970.2019 0.1418 0.2120
TS8 0.1292 0.12830.32170.27580.14820.20130.20850.16140.23030.14460.1768
TS90.15750.16320.20790.2847 0.1510 0.20020.20700.17040.19720.14740.1772
TS10 0.1196 0.11840.38750.56140.17410.29570.51010.17410.51370.13650.2499

Appendix A.2.2. Results Obtained by Using DLM1

Table A26. MeteoSwiss time series: Sampling without replacement ( ρ = 5 % ).
Table A26. MeteoSwiss time series: Sampling without replacement ( ρ = 5 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 480
s = 3s = 3
n 3 = 360 n 3 = 360
TS10.00890.00890.04430.06200.03560.04260.0429 0.0338 0.04100.03650.0416
TS20.00590.00590.02900.06350.02560.02700.0272 0.0249 0.04240.02510.0266
TS30.01420.01420.07030.0697 0.0446 0.06610.06740.05540.05930.05080.0640
TS40.00990.00990.06410.14690.05730.06610.06690.05720.0627 0.0533 0.0644
TS50.01170.01170.05520.1079 0.0402 0.05260.05310.04100.04610.04150.0509
TS60.01310.01310.05880.0956 0.0502 0.05830.05880.05170.06350.05370.0576
TS70.06830.06830.17430.1398 0.1045 0.13100.13360.11600.12610.10810.1253
TS80.06880.06880.18470.2337 0.1099 0.16090.16930.13880.14750.11040.1472
TS90.06880.06880.12720.2200 0.1075 0.12900.13120.11960.12460.10780.1263
TS100.07130.07130.38770.40320.09060.29340.54710.13960.4733 0.0768 0.1862
Table A27. MeteoSwiss time series: Sampling without replacement ( ρ = 10 % ).
Table A27. MeteoSwiss time series: Sampling without replacement ( ρ = 10 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 3
n 3 = 360 n 3 = 120
TS1 0.0093 0.00910.04090.07090.03340.04040.04060.03250.05030.03510.0399
TS2 0.0058 0.00570.02910.07140.02550.02770.02780.02550.04980.02590.0274
TS3 0.0141 0.01350.07840.06830.04730.07480.07740.05781.17170.05640.0720
TS4 0.0112 0.01110.07000.15090.05720.07310.07400.05700.07160.05750.0715
TS5 0.0151 0.01460.06080.12050.04400.05980.06080.04700.05970.04760.0581
TS60.0142 0.0143 0.05540.09360.04800.05570.05600.04880.06330.05130.0552
TS70.0620 0.0628 0.17300.15250.10170.13740.14220.12430.13690.10740.1304
TS8 0.0665 0.06610.22010.23360.11110.18050.19530.13790.16280.10970.1601
TS9 0.0714 0.07010.13430.23020.10700.13380.13670.12750.13480.11100.1297
TS10 0.0696 0.06920.38950.42860.09250.29770.51140.14900.16100.07370.2037
Table A28. MeteoSwiss time series: Sampling without replacement ( ρ = 15 % ).
Table A28. MeteoSwiss time series: Sampling without replacement ( ρ = 15 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 3
n 3 = 360 n 3 = 120
TS1 0.0105 0.01000.04190.07350.03380.04160.04210.03250.05130.03470.0406
TS2 0.0068 0.00630.02900.07770.02500.02830.02850.02540.05490.02650.0280
TS3 0.0185 0.01770.07880.06760.04440.07510.07760.05741.32540.05540.0721
TS4 0.0136 0.01300.06660.14300.05140.06670.06790.05330.06900.05420.0656
TS5 0.0161 0.01570.05650.11860.04180.05500.05610.04300.05690.04300.0532
TS6 0.0158 0.01560.05570.09100.04800.05480.05550.04830.06410.04920.0541
TS7 0.0735 0.07260.22440.16220.10960.17900.18920.13170.14600.11470.1577
TS8 0.0684 0.06760.20360.22410.10190.18030.20690.13310.15380.09900.1527
TS9 0.0674 0.06670.12950.21480.09790.12570.13080.11720.12970.10290.1196
TS10 0.0720 0.07070.33170.39930.08590.25810.38190.14120.14550.07480.2023
Table A29. MeteoSwiss time series: Sampling without replacement ( ρ = 20 % ).
Table A29. MeteoSwiss time series: Sampling without replacement ( ρ = 20 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 3
n 3 = 360 n 3 = 120
TS1 0.0137 0.01350.04220.08120.03490.04170.04170.03370.04160.03490.0414
TS20.00890.00890.03100.0864 0.0272 0.03020.03000.02750.04430.02770.0301
TS3 0.0217 0.02110.07290.06310.04150.06920.07100.05370.05810.05150.0668
TS4 0.0183 0.01790.07090.16170.05820.07350.07520.05710.06520.05580.0708
TS5 0.0209 0.02020.05840.13850.04270.05620.05690.04410.04950.04390.0547
TS6 0.0192 0.01900.05770.10640.04640.05600.05700.04840.06230.05130.0551
TS7 0.0781 0.07760.21310.15410.10540.15930.16160.12780.16970.11150.1504
TS80.0782 0.0789 0.22670.23180.10720.18320.21120.13900.19580.10800.1516
TS9 0.0744 0.07380.13560.21530.10060.12920.13490.12010.16410.10670.1251
TS100.0816 0.0811 0.34910.43970.09220.26560.43210.14160.36710.07640.1985
Table A30. MeteoSwiss time series: Polya model ( ρ = 5 % ).
Table A30. MeteoSwiss time series: Polya model ( ρ = 5 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 360
s = 3s = 3
n 3 = 360 n 3 = 240
TS10.0248 0.0254 0.04370.05730.03280.04180.04250.03010.03810.03570.0410
TS2 0.0118 0.01150.02320.05180.02020.02240.02270.02070.03720.02130.0222
TS30.0293 0.0296 0.06100.05070.05020.06070.06360.04310.04920.05170.0590
TS4 0.0441 0.04360.07130.14850.06090.07200.07340.05930.06830.06300.0709
TS50.0360 0.0366 0.06270.10780.05140.06180.06290.04660.05520.05610.0610
TS6 0.0191 0.01890.03880.07090.02770.03820.03890.03110.04700.03700.0381
TS7 0.1404 0.14150.23310.18640.13450.18530.18750.15150.15350.14590.1781
TS80.0889 0.0893 0.15450.17570.08890.14220.15340.10440.11890.10010.1318
TS90.1064 0.1081 0.13790.20480.11820.13940.14050.12860.14120.12510.1376
TS100.1496 0.1482 0.32410.43290.15470.28390.42760.17040.45470.13570.2366
Table A31. MeteoSwiss time series: Polya model ( ρ = 10 % ).
Table A31. MeteoSwiss time series: Polya model ( ρ = 10 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 3
n 3 = 360 n 3 = 120
TS10.0352 0.0355 0.05150.07510.03870.04700.04910.03980.07430.04310.0475
TS20.0129 0.0131 0.02390.05740.01900.02250.02310.02070.05620.02210.0229
TS3 0.0331 0.03270.06320.05250.04400.05760.06490.04551.18590.05020.0589
TS40.0392 0.0394 0.07450.13980.05960.07660.07550.05730.08520.06330.0731
TS50.0535 0.0534 0.06970.15030.05300.06770.06660.05470.08840.05830.0650
TS60.0205 0.0210 0.04530.07330.03130.04090.04500.03510.06640.03890.0433
TS70.10390.10570.18960.1357 0.1040 0.12480.14580.10970.11400.10800.1346
TS80.11330.11320.17400.19010.10350.14910.17200.11630.1220 0.1086 0.1473
TS90.09230.09390.11390.1891 0.0932 0.12880.11190.10520.10850.09600.1079
TS10 0.1203 0.12000.39140.37700.14870.29340.51500.13660.14330.13050.2606
Table A32. MeteoSwiss time series: Polya model ( ρ = 15 % ).
Table A32. MeteoSwiss time series: Polya model ( ρ = 15 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 3
n 3 = 360 n 3 = 120
TS10.0334 0.0337 0.04760.07620.03980.04630.04690.03640.04680.04100.0455
TS20.0225 0.0229 0.03220.07600.02860.03140.03130.02940.04120.02940.0310
TS3 0.0459 0.04550.07110.06600.04670.07060.07220.05480.06050.05830.0684
TS40.0365 0.0369 0.07010.12520.05320.07500.07540.05360.06340.05960.0734
TS5 0.0451 0.04450.06100.12970.04900.06020.05960.05070.05200.04960.0582
TS60.0381 0.0382 0.05850.09200.05240.05790.05770.05200.06180.05370.0573
TS70.11630.11820.18310.1488 0.1118 0.16700.15800.12040.14140.11020.1394
TS80.13580.13800.28070.2418 0.1337 0.18430.23870.14620.17850.13330.1777
TS90.10990.10940.12100.1935 0.1031 0.12080.12060.11490.12760.09850.1120
TS100.1073 0.1084 0.33070.37870.12540.24110.38980.12570.38480.11550.1974
Table A33. MeteoSwiss time series: Polya model ( ρ = 20 % ).
Table A33. MeteoSwiss time series: Polya model ( ρ = 20 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 3
n 3 = 360 n 3 = 120
TS10.03660.03660.04790.09480.04040.04910.0495 0.0385 0.04740.04220.0487
TS20.0215 0.0217 0.02780.07640.02480.02740.02720.02540.04060.02620.0272
TS30.05260.05340.07830.0704 0.0532 0.07460.07700.05980.06260.06140.0734
TS40.0442 0.0448 0.06970.16480.06170.07640.07850.06230.06970.06570.0760
TS50.0346 0.0352 0.05160.12580.04210.05230.05270.04130.04540.04320.0510
TS60.0343 0.0345 0.05540.10680.04500.05050.05210.04610.06130.05150.0510
TS70.12110.12280.26280.14840.10670.20760.18270.12310.1556 0.1139 0.1657
TS80.1178 0.1189 0.25810.21860.11950.17070.17570.12640.17890.12420.1522
TS90.12290.12540.15770.2102 0.1204 0.17000.17550.12650.15540.11840.1511
TS10 0.1092 0.10950.30140.36580.12840.23540.39750.12520.40100.10720.1996

Appendix A.3. Bafu Time Series: Numerical Results

Table A34. BAFU time series: Sampling without replacement ( ρ = 5 % ).
Table A34. BAFU time series: Sampling without replacement ( ρ = 5 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 3s = 3
n 3 = 180 n 3 = 180
TS10.17040.17040.33040.35720.28720.32690.33420.37090.3345 0.2342 0.3344
TS20.08250.08250.52050.58910.49240.48010.49120.49150.4937 0.3400 0.4895
TS30.11970.11970.37760.39630.34140.37250.37470.36290.3761 0.3226 0.3743
TS40.09420.09420.36280.6117 0.2387 0.36530.36780.45740.37840.27040.3731
TS50.14690.14690.46210.53630.43000.42540.44270.45170.4463 0.3087 0.4421
TS60.10790.10790.25300.34230.23670.24600.24830.26030.2508 0.1954 0.2478
TS70.10680.10680.27960.36010.23280.27650.27960.28860.2780 0.2192 0.2791
TS80.07300.07300.68120.74310.62060.70730.74890.69900.7790 0.3866 0.7501
TS90.11790.11790.38260.3724 0.2617 0.36070.36510.38920.36500.27090.3656
TS100.08940.08940.59591.92420.44470.51320.52620.53340.5299 0.3764 0.5273
Table A35. BAFU time series: Sampling without replacement ( ρ = 10 % ).
Table A35. BAFU time series: Sampling without replacement ( ρ = 10 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 3s = 3
n 3 = 180 n 3 = 180
TS10.10760.10760.33390.55450.28960.32540.33060.39080.3388 0.2224 0.3375
TS20.03640.03640.51890.65460.46490.47290.48820.49530.4872 0.3321 0.4880
TS30.03780.03780.40190.42550.34660.39260.39640.37650.3959 0.3400 0.3966
TS40.07000.07000.42140.8424 0.2420 0.41470.42400.46780.44250.29820.4358
TS50.10110.10110.43660.66210.40930.40720.41860.45750.4251 0.2809 0.4208
TS60.03270.03270.27370.35700.21730.25410.25980.27430.2575 0.1986 0.2606
TS70.06060.06060.32180.4167 0.2261 0.30630.31470.31370.30490.23320.3154
TS80.08700.08700.66190.67530.59220.68470.73080.68600.7701 0.3650 0.7368
TS90.05490.05490.40810.6734 0.2851 0.37980.38700.39950.39060.28820.3911
TS100.07200.07200.59440.75940.42150.51880.52760.55930.5409 0.3822 0.5362
Table A36. BAFU time series: Sampling without replacement ( ρ = 15 % ).
Table A36. BAFU time series: Sampling without replacement ( ρ = 15 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 180
s = 3s = 3
n 3 = 120 n 3 = 120
TS1 0.1329 0.10580.36010.44730.27700.33500.34760.39840.35210.21720.3547
TS2 0.0775 0.06450.53291.26610.47230.47440.48900.50380.49450.33410.4872
TS3 0.1090 0.10540.40810.85850.34720.40260.40390.40630.40720.35030.4034
TS40.0945 0.1008 0.44532.27110.23650.42620.43930.47950.46120.30350.4550
TS5 0.1286 0.12580.50290.60970.42940.44020.46250.46990.47020.30170.4683
TS6 0.0653 0.05710.27291.41820.21430.26010.26200.29190.26750.20000.2616
TS7 0.0810 0.07010.29800.82050.21290.29520.29770.32400.29790.22510.2971
TS8 0.0965 0.08050.65111.34040.56880.68310.74660.67850.79070.35650.7574
TS9 0.1024 0.08540.43460.46470.28510.39320.40410.42160.40720.29280.4090
TS10 0.0810 0.07470.63142.24950.43080.53050.54700.55850.56950.37980.5579
Table A37. BAFU time series: Sampling without replacement ( ρ = 20 % ). Because the values of the errors computed for GROUSE are very large, we prefer to replace them with Inf.
Table A37. BAFU time series: Sampling without replacement ( ρ = 20 % ). Because the values of the errors computed for GROUSE are very large, we prefer to replace them with Inf.
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 180 n u = 180
s = 3s = 3
n 3 = 120 n 3 = 120
TS10.08480.08480.3876Inf0.26840.35130.37140.41120.4034 0.2159 0.3830
TS20.06220.06220.5501Inf0.47670.49230.51190.51430.5237 0.3430 0.5095
TS30.08230.08230.4079Inf 0.3397 0.40340.40670.40832.41340.34720.4060
TS40.09300.09300.4864Inf 0.2145 0.43600.46410.48560.57330.29820.4955
TS50.12140.12140.5083Inf0.41200.43650.46350.47330.5047 0.2838 0.4737
TS60.06230.06230.2780Inf0.21650.26950.27450.30420.3623 0.2028 0.2735
TS70.06460.06460.3186Inf 0.2205 0.30920.31700.33380.39330.23520.3164
TS80.08980.08980.7019Inf0.58430.71520.79610.69440.8595 0.3701 0.8199
TS90.06720.06720.4495Inf 0.2745 0.39460.41130.43090.40000.28420.4207
TS100.06760.06760.6526Inf0.43360.52970.55370.56220.6377 0.3724 0.5729
Table A38. BAFU time series: Polya model ( ρ = 5 % ).
Table A38. BAFU time series: Polya model ( ρ = 5 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 120 n u = 120
s = 3s = 3
n 3 = 60 n 3 = 60
TS11.88131.88130.27250.31160.33840.2649 0.2568 0.37870.25770.18820.2573
TS20.55240.55240.45410.61880.4407 0.4321 0.44080.46440.45140.33810.4395
TS31.90941.90940.36800.37040.30860.36000.35890.36580.3621 0.3227 0.3588
TS40.71250.71250.38860.67160.22870.39390.40030.44430.4057 0.3049 0.4012
TS51.71621.71620.44060.48700.4358 0.4070 0.42390.44510.42830.34960.4241
TS60.62640.62640.27840.30860.24460.26970.27580.25720.2755 0.2463 0.2753
TS70.61970.61970.30210.38150.25070.29460.29560.33420.2965 0.2657 0.2952
TS80.39840.39840.82630.83330.76370.85570.89920.80560.9287 0.6502 0.8988
TS91.27891.27890.41550.63290.34480.36540.3937 0.3321 0.39630.33080.3915
TS100.37200.37200.59510.6249 0.4378 0.54560.55060.56920.55380.45320.5499
Table A39. BAFU time series: Polya model ( ρ = 10 % ).
Table A39. BAFU time series: Polya model ( ρ = 10 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 120 n u = 120
s = 3s = 3
n 3 = 60 n 3 = 60
TS14.21004.21000.30180.35060.30160.2970 0.2954 0.38710.29880.25380.3006
TS20.48610.48610.46570.5445 0.4192 0.45080.46650.47580.44970.36340.4647
TS34.57394.57390.44040.47950.39130.44210.44360.42460.4451 0.4093 0.4435
TS40.42850.42850.35800.62930.24010.36320.36650.47520.3752 0.2898 0.3703
TS50.49760.49760.45150.5132 0.3945 0.41140.43920.42240.43840.33700.4417
TS64.92494.92490.28700.34420.23100.27070.27390.27890.2800 0.2448 0.2735
TS74.53684.53680.33700.41550.22690.33680.34280.33360.3398 0.2930 0.3424
TS8 0.5594 0.5594 0.64410.70980.58250.66510.70580.70080.73520.47400.7090
TS94.39444.39440.37860.39100.25290.35200.35770.39710.3601 0.2972 0.3617
TS10 0.4510 0.4510 0.63881.96240.47450.52180.55150.51970.55640.43640.5498
Table A40. BAFU time series: Polya model ( ρ = 15 % ).
Table A40. BAFU time series: Polya model ( ρ = 15 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 120 n u = 120
s = 3s = 3
n 3 = 60 n 3 = 60
TS112.381112.38110.32360.3763 0.3152 0.33760.32870.43150.32470.23980.3261
TS24.12114.12110.54300.5602 0.4583 0.50790.53230.50800.52780.40730.5299
TS313.289813.28980.38490.42210.32860.38670.38630.39980.3899 0.3580 0.3862
TS43.52733.52730.48350.82930.24890.45340.47620.48790.4877 0.3812 0.4977
TS53.71083.71080.42630.53970.4142 0.4016 0.40680.46860.40830.30860.4053
TS613.138113.13810.27070.51570.21180.26930.27640.27720.2752 0.2313 0.2759
TS711.131411.13140.30650.38230.23860.30570.31250.31610.3099 0.2633 0.3114
TS82.93382.93380.67090.8033 0.5345 0.69810.80710.64870.86340.49090.8371
TS912.430112.43010.40070.49400.23370.36220.36840.41050.3715 0.2962 0.3711
TS104.25524.25520.62451.79830.39750.51310.53680.55330.5498 0.4139 0.5506
Table A41. BAFU time series: Polya model ( ρ = 20 % ).
Table A41. BAFU time series: Polya model ( ρ = 20 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 180 n u = 180
s = 3s = 3
n 3 = 120 n 3 = 120
TS12.34742.34740.36072.0702 0.2869 0.33330.33840.41700.35430.25090.3502
TS21.55741.55740.56591.5948 0.4837 0.52230.54760.52410.54390.42180.5455
TS31.69921.69920.38510.64710.31710.37980.38120.40100.3847 0.3442 0.3819
TS41.35671.35670.40244.25990.22540.38770.39870.47110.4118 0.3025 0.4122
TS51.72021.72020.51403.7538 0.4206 0.43920.46200.47740.47900.33780.4749
TS61.12941.12940.28153.0014 0.2368 0.27550.27730.31030.27960.23380.2767
TS71.83911.83910.27990.37890.20080.27990.27770.33610.2775 0.2354 0.2784
TS82.04992.04990.66960.7806 0.6095 0.69330.74210.71310.77220.48870.7541
TS92.35532.35530.41580.57420.26950.39500.39940.44130.4073 0.3210 0.4099
TS102.11972.11970.61104.8646 0.4281 0.52200.53540.56110.54330.42260.5419

Appendix A.4. Temperature Time Series: Numerical Results

Table A42. Temperature time series: The clustering of the time series when the missing data are simulated by sampling without replacement (SWP) or by the Polya model (Polya).
Table A42. Temperature time series: The clustering of the time series when the missing data are simulated by sampling without replacement (SWP) or by the Polya model (Polya).
Simulation
Method
Percentage
Missing Data
Group1Group2Group3Group4Group5
SWP ρ = 5 % {7, 10, 11, 12, 13,
15, 18, 19, 20, 25}
{1, 14, 16, 17, 21,
22, 23, 24, 26, 27}
{2, 3, 9, 31, 39,
45, 46, 47, 48, 49}
{28, 29, 30, 32, 33,
34, 36, 37, 38, 41}
{4, 5, 6, 8, 35,
40, 42, 43, 44, 50}
SWP ρ = 10 % {7, 9, 11, 12, 13,
15, 18, 19, 20, 25}
{3, 4, 6, 14, 16,
17, 23, 24, 26, 27}
{1, 2, 10, 21, 22,
45, 46, 47, 48, 49}
{29, 30, 32, 33, 34,
35, 36, 37, 38, 41}
{5, 8, 28, 31, 39,
40, 42, 43, 44, 50}
SWP ρ = 15 % {3, 7, 9, 10, 11,
15, 18, 19, 20, 25}
{4, 6, 14, 16, 17,
21, 23, 24, 26, 27}
{1, 2, 12, 22, 43,
45, 46, 47, 48, 49}
{28, 29, 30, 32, 33,
34, 36, 37, 38, 41}
{5, 8, 13, 31, 35,
39, 40, 42, 44, 50}
SWP ρ = 20 % {7, 9, 10, 11, 13,
18, 19, 20, 22, 25}
{3, 4, 6, 14, 16,
17, 23, 24, 26, 27}
{1, 2, 12, 15, 43,
45, 46, 47, 48, 49}
{29, 30, 32, 33, 34,
35, 36, 37, 38, 41}
{5, 8, 21, 28, 31,
39, 40, 42, 44, 50}
Polya ρ = 5 % {3, 7, 9, 11, 12,
13, 18, 19, 22, 25}
{6, 14, 16, 23, 43,
45, 46, 47, 48, 49}
{1, 2, 10, 15, 17,
20, 21, 24, 26, 27}
{29, 30, 32, 33, 34,
35, 36, 37, 38, 41}
{4, 5, 8, 28, 31,
39, 40, 42, 44, 50}
Polya ρ = 10 % {3, 7, 9, 11, 12,
13, 15, 18, 20, 25}
{2, 10, 14, 16, 17,
21, 22, 23, 24, 26}
{1, 6, 19, 27, 39,
45, 46, 47, 48, 49}
{28, 29, 30, 31, 32,
33, 35, 37, 38, 41}
{4, 5, 8, 34, 36,
40, 42, 43, 44, 50}
Polya ρ = 15 % {6, 7, 9, 14, 16,
17, 18, 20, 22, 23}
{1, 2, 3, 10, 11,
12, 13, 15, 19, 25}
{4, 21, 24, 26, 27,
45, 46, 47, 48, 49}
{28, 29, 30, 32, 33,
34, 36, 37, 38, 41}
{5, 8, 31, 35, 39,
40, 42, 43, 44, 50}
Polya ρ = 20 % {3, 6, 7, 9, 10,
15, 16, 19, 24, 26}
{1, 11, 12, 14, 25,
43, 45, 46, 47, 49}
{2, 4, 5, 13, 18,
20, 21, 22, 23, 27}
{28, 29, 30, 31, 32,
34, 35, 36, 37, 41}
{8, 17, 33, 38, 39,
40, 42, 44, 48, 50}
Table A43. Temperature time series: Sampling without replacement ( ρ = 5 % ).
Table A43. Temperature time series: Sampling without replacement ( ρ = 5 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 180 n u = 60
s = 2s = 2
n 3 = 120 n 3 = 60
TS10.1524 0.1552 0.19890.21790.21020.20250.20060.93850.21360.19240.2003
TS2 0.1329 0.13250.15790.23260.15620.17420.17311.05130.18220.16270.1724
TS30.16170.16240.13550.18420.13830.15050.14741.09660.1627 0.1380 0.1469
TS4 0.1698 0.16890.21370.20880.21540.20770.20720.25540.19490.20960.2073
TS5 0.1618 0.16050.22410.23150.22620.21260.21250.26230.19820.21740.2126
TS6 0.1519 0.15060.18650.21660.19170.16990.16730.20570.16050.17760.1682
TS70.13490.13750.09100.14070.10030.10010.09791.06190.1150 0.0914 0.0972
TS8 0.2055 0.20400.26000.26680.25580.24940.24880.29710.23100.25530.2492
TS90.11950.1181 0.0917 0.13310.09580.09870.09691.10920.10670.09070.0963
TS100.13970.14220.11230.1650 0.1113 0.11830.11501.12670.12370.11100.1148
TS110.13070.13590.08510.15730.09080.10180.09811.13470.1173 0.0884 0.0975
TS120.15360.15100.10300.1506 0.1084 0.11740.11421.05590.1361 0.1084 0.1139
TS130.12560.12990.11050.1623 0.1137 0.13870.13171.03980.14850.11970.1315
TS140.14040.1455 0.1359 0.16840.10380.15730.15641.05130.15260.15340.1563
TS150.16360.1678 0.1194 0.15160.12210.12130.12091.13660.13090.11690.1204
TS160.14090.1424 0.1233 0.16130.08650.15440.15321.04500.15140.14710.1531
TS170.18400.1936 0.1617 0.20060.14780.17620.17550.99000.17350.17210.1755
TS180.14210.14270.10100.13930.10780.11520.11321.11520.1320 0.1051 0.1124
TS190.14400.13610.10310.15560.11390.11970.11681.14510.1436 0.1056 0.1159
TS200.17570.19050.13870.18740.14600.15230.14721.05460.1670 0.1428 0.1469
TS210.1671 0.1715 0.19190.21210.17480.21450.21460.97750.21470.20510.2141
TS220.14890.1538 0.1260 0.17210.09230.16110.15971.09430.16650.14880.1592
TS230.15530.1630 0.1386 0.16400.09870.16880.16921.11470.16510.16430.1689
TS240.17570.1784 0.1478 0.17220.10630.17810.17691.07640.17070.17290.1769
TS250.14110.14500.11060.1468 0.1133 0.12840.12551.07400.14040.11600.1248
TS260.17990.1824 0.1730 0.19140.16230.19340.19180.99280.19350.18590.1916
TS270.1487 0.1514 0.17550.18110.15690.19080.18960.98440.18810.18440.1895
TS280.17250.18050.17770.20530.17930.15970.16140.22680.17400.1660 0.1610
TS290.13020.13540.12020.17710.12850.11430.11350.20090.11960.1162 0.1136
TS300.1579 0.1580 0.18720.27440.19520.17730.17740.27200.18430.18040.1773
TS310.1769 0.1816 0.21230.23850.23140.19560.19430.27070.21670.20010.1943
TS320.1473 0.1491 0.16580.25830.17200.14950.14730.26650.16140.15300.1473
TS330.17210.17390.16630.24230.16740.1620 0.1611 0.25230.15970.1643 0.1611
TS34 0.1156 0.11550.15930.21590.16380.15480.15120.23550.14860.15270.1515
TS35 0.2026 0.19760.22140.26180.23600.21080.20600.29510.22800.21370.2067
TS36 0.1139 0.10790.13580.22250.14620.12960.12590.24290.12800.12760.1259
TS370.13410.13050.13530.20860.14150.13160.12720.24020.12920.1305 0.1277
TS380.1289 0.1336 0.14930.22410.15070.14890.14510.26590.14370.14310.1449
TS390.1706 0.1722 0.19510.25360.20720.18930.18970.28040.19530.19060.1894
TS40 0.1792 0.17750.21760.26400.23480.21340.21210.29980.22080.21330.2121
TS410.15110.14720.14870.23540.15940.1444 0.1411 0.25910.14090.14300.1412
TS42 0.1490 0.14680.17170.22190.16840.17190.17030.25310.16590.16930.1701
TS43 0.1222 0.11770.14880.23260.14450.14910.14910.18230.15980.14870.1489
TS440.1522 0.1539 0.21150.24050.21200.21150.21160.25250.21110.21110.2114
TS45 0.1266 0.13060.17710.25650.22590.13460.12430.94840.15540.13840.1243
TS460.0950 0.0964 0.20130.28810.24730.15460.14870.99220.18520.14690.1477
TS470.0937 0.0970 0.23180.26680.23820.21680.21050.99680.19270.22090.2117
TS480.1236 0.1252 0.28210.33130.28930.26080.26211.01950.23850.26770.2620
TS490.0987 0.0996 0.19620.28440.24620.14270.13940.98900.16580.13960.1377
TS50 0.1010 0.09760.16580.23440.16790.16320.16320.19420.18310.16410.1631
Table A44. Temperature time series: Sampling without replacement ( ρ = 10 % ).
Table A44. Temperature time series: Sampling without replacement ( ρ = 10 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 180 n u = 60
s = 2s = 2
n 3 = 180 n 3 = 60
TS10.1544 0.1572 0.20070.22200.20480.20940.20900.92930.21160.19040.2083
TS20.1516 0.1519 0.16510.21860.16100.17450.17441.00550.18350.16170.1739
TS3 0.1773 0.17920.18070.21510.18370.18290.18240.97060.18510.17470.1817
TS4 0.1651 0.16280.17100.22340.17010.18270.18030.90350.19060.17360.1794
TS50.1460 0.1471 0.20450.20430.20380.19630.19670.18160.22080.20390.1968
TS60.15000.1530 0.1357 0.21020.13580.14800.14460.96040.15080.13560.1440
TS70.13450.1367 0.1048 0.13820.09780.12400.12121.04730.12640.10620.1211
TS80.2213 0.2179 0.26480.27310.26290.25450.25410.17350.27540.26630.2543
TS90.13420.14310.10720.13260.10330.11970.11891.06600.1226 0.1062 0.1186
TS100.14210.14060.11710.16710.10910.12780.12741.06770.1297 0.1163 0.1271
TS110.15800.1558 0.1066 0.16160.09870.13060.12951.08100.13650.11240.1291
TS120.16410.1694 0.1269 0.16240.11860.15460.15441.01110.16200.13620.1539
TS130.13730.1401 0.1229 0.17090.11430.15270.15141.01330.15720.13420.1509
TS140.15050.14950.10210.1681 0.1024 0.11130.10660.99680.11680.10850.1061
TS150.16800.1627 0.1285 0.16610.12140.15030.15011.05450.15530.13130.1497
TS160.15680.15810.10530.16610.09780.1071 0.1015 0.97680.11000.11070.1017
TS170.16530.17550.14220.20160.13590.13950.13690.94550.14340.1377 0.1367
TS180.15660.1600 0.1297 0.16580.12280.15700.15431.02870.16060.13940.1543
TS190.15350.1550 0.1174 0.15400.11090.14580.14321.02640.15070.12660.1430
TS200.16480.1708 0.1546 0.19350.14680.17840.17901.03720.18550.16370.1784
TS21 0.1897 0.18710.19150.22700.19380.20310.20290.93200.20640.19520.2020
TS220.14790.15060.13770.15410.13440.14680.14811.05590.1507 0.1349 0.1476
TS230.16460.1700 0.1104 0.15680.10770.11640.11050.96950.11900.11950.1106
TS240.17310.17730.12420.17490.11470.1214 0.1152 0.98550.12580.13010.1155
TS250.16120.1633 0.1282 0.15990.12100.15350.15401.04740.15870.13590.1534
TS260.18440.18200.16200.20730.15940.15910.15650.94680.16260.1537 0.1561
TS270.14980.15160.15230.18330.14920.15340.14920.95490.15360.1416 0.1489
TS280.16240.16920.17480.21800.18650.1671 0.1645 0.25650.17900.17100.1647
TS290.12890.13240.13430.18300.14020.12770.12680.21990.12680.1325 0.1269
TS300.1619 0.1665 0.17680.22560.18110.17030.17090.23700.17020.17560.1708
TS310.17000.17710.18630.23760.1993 0.1722 0.17000.27790.18350.17980.1700
TS320.16500.16890.15980.26920.16880.15440.15090.27730.16220.1577 0.1511
TS330.16850.17790.17710.24130.17930.17470.17480.2393 0.1709 0.17680.1746
TS340.1182 0.1237 0.13450.19410.13620.13350.13170.21120.12940.13300.1316
TS350.17410.18190.18310.23660.1926 0.1714 0.16950.27310.18410.17780.1695
TS36 0.1200 0.11830.13460.20740.13370.13150.12940.20750.12340.13230.1294
TS37 0.1201 0.11730.13350.19830.13160.13280.12950.22200.12480.13250.1297
TS380.14560.14690.14730.20750.14670.14760.14560.22600.14080.1467 0.1455
TS390.1654 0.1679 0.20350.27640.21900.20100.20020.29350.20340.20090.1999
TS400.1832 0.1886 0.23350.26860.23170.22720.23020.30980.22710.22980.2293
TS410.14560.14380.14170.22730.14180.1374 0.1370 0.22220.13730.13970.1365
TS42 0.1509 0.14630.16470.21320.16210.16410.16440.20320.16110.16400.1640
TS43 0.1167 0.11430.15860.22990.15630.15940.15880.15370.16870.15790.1586
TS44 0.1686 0.16270.19720.22390.19690.19550.19590.24020.19800.19620.1956
TS45 0.1290 0.12730.23190.26500.24030.24380.24250.90900.24820.22890.2425
TS46 0.0966 0.09410.23510.28280.23930.24190.23990.88230.24160.22880.2401
TS470.0987 0.1042 0.22860.27650.24600.18460.17460.92780.17300.21440.1753
TS48 0.1198 0.11730.26280.32140.27410.21380.21730.95360.21140.24440.2159
TS490.0936 0.0937 0.23790.26960.25960.18380.17790.96760.17420.22110.1784
TS500.0980 0.0994 0.14890.19550.14760.14720.14710.15050.16900.14810.1471
Table A45. Temperature time series: Sampling without replacement ( ρ = 15 % ).
Table A45. Temperature time series: Sampling without replacement ( ρ = 15 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 180 n u = 60
s = 2s = 2
n 3 = 120 n 3 = 60
TS10.1549 0.1573 0.19510.21300.19420.19700.19670.88710.20260.19290.1962
TS2 0.1582 0.15760.17960.19740.17590.18440.18250.85580.18870.17940.1823
TS30.17060.16600.15150.18220.15150.15710.15460.92430.1611 0.1517 0.1546
TS40.1720 0.1722 0.20220.20510.18900.20600.20590.89230.21210.20420.2057
TS50.1443 0.1479 0.20690.21360.21010.20490.20530.17140.24160.20540.2052
TS60.1590 0.1584 0.16060.20330.15020.16570.16420.89870.16960.16160.1639
TS70.14800.15070.10140.15530.10440.10840.10460.97410.1106 0.1029 0.1046
TS8 0.2093 0.21420.25770.26890.26010.25680.25690.19080.29610.25730.2569
TS90.14670.1445 0.1010 0.14160.10500.10500.10250.93680.10770.10000.1023
TS100.14750.1525 0.1198 0.16550.12180.12210.12110.96960.12400.11770.1206
TS110.15940.15930.09530.14980.09910.10390.09870.95530.1069 0.0986 0.0989
TS120.17600.17730.11360.1587 0.1167 0.12260.11690.87580.12420.11850.1172
TS130.14150.1407 0.1094 0.17080.10430.11940.11540.94290.12700.11110.1148
TS140.15790.1528 0.1432 0.15090.12260.15080.14960.90920.15690.14770.1494
TS150.16320.16800.10270.16190.10620.10910.10590.96650.1133 0.1055 0.1056
TS160.17390.1700 0.1438 0.14850.12840.14870.14680.88290.14980.14670.1469
TS170.19440.1966 0.1820 0.21720.17120.18390.18390.84350.18600.18320.1837
TS180.17610.1741 0.1136 0.16390.11110.12290.11940.91450.13120.11850.1191
TS190.15570.1521 0.0947 0.14640.09370.10280.10000.95390.11200.09810.0993
TS200.16900.1669 0.1415 0.18050.13630.14890.14680.94950.15650.14300.1462
TS21 0.1938 0.18830.21940.22060.20170.22140.22110.84110.22380.21790.2207
TS220.15590.1524 0.1300 0.16110.13100.13610.13210.91800.13690.12890.1320
TS230.16530.1637 0.1560 0.15620.13680.16310.16010.88550.16500.15840.1602
TS240.17990.1722 0.1607 0.16150.14160.16650.16460.89910.16950.16300.1646
TS250.16390.1606 0.1083 0.16340.10520.11820.11390.93910.12600.11250.1135
TS260.20070.20290.18890.19100.17880.19040.18920.88710.1902 0.1888 0.1893
TS270.1724 0.1726 0.18560.17510.17580.18760.18660.83440.18860.18560.1866
TS280.18150.18090.18320.23300.19870.1816 0.1791 0.28130.19980.18030.1789
TS290.14350.13880.13900.20400.14510.13740.13380.22480.13460.1362 0.1339
TS300.1645 0.1719 0.17950.27400.18570.17670.17520.25440.17560.17560.1748
TS310.18970.18860.18690.25240.20660.18250.17930.28490.1987 0.1816 0.1793
TS320.15560.15690.14320.25800.15340.14220.13770.25720.15570.1420 0.1380
TS330.1749 0.1743 0.18000.24730.18380.17850.17680.24400.17400.17830.1767
TS340.1235 0.1240 0.13030.18480.12730.12930.12930.19470.12840.12850.1289
TS350.18270.18860.18570.25770.19910.1817 0.1812 0.27560.19630.18250.1809
TS360.1167 0.1183 0.12760.20900.12490.12880.12560.19730.12390.12570.1256
TS370.13090.13190.13260.2054 0.1276 0.13190.12950.20930.12170.13070.1295
TS380.14390.14360.14680.2191 0.1423 0.14750.14500.22720.13760.14460.1447
TS39 0.1635 0.16140.20320.27040.21530.20130.20120.28390.20800.19870.2006
TS40 0.2107 0.20680.23870.28030.24750.23520.23630.30350.24150.23620.2358
TS410.15640.15730.14470.22930.14370.14340.14270.23080.14400.1419 0.1423
TS42 0.1540 0.14980.16990.21410.16310.17040.17030.19830.17100.16940.1699
TS430.1283 0.1297 0.15400.32300.20870.15180.14370.88870.13570.14980.1442
TS44 0.1650 0.16410.19510.22780.19400.19500.19500.22630.19930.19450.1947
TS45 0.1292 0.13560.14380.24490.20260.14000.13460.88700.12420.14000.1348
TS460.0974 0.0997 0.15380.29430.20820.14720.14280.81330.13690.14250.1426
TS47 0.1167 0.11660.13720.24800.19570.13520.12840.85360.11960.13010.1283
TS48 0.1295 0.12280.28760.33500.29500.28770.28680.78100.28710.28430.2866
TS49 0.1134 0.10820.15880.27620.21070.15330.14880.80780.13730.14800.1485
TS50 0.1084 0.10700.16270.23340.16390.16440.16330.14910.18500.16370.1634
Table A46. Temperature time series: Sampling without replacement ( ρ = 20 % ).
Table A46. Temperature time series: Sampling without replacement ( ρ = 20 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 180 n u = 60
s = 2s = 2
n 3 = 120 n 3 = 60
TS1 0.1700 0.16610.20050.21870.19810.20570.20620.17790.20960.19580.2054
TS20.1676 0.1604 0.17090.19400.17330.18020.17880.15420.18530.16990.1782
TS30.17630.1740 0.1520 0.20200.15660.16150.16100.18150.17000.15070.1599
TS40.18610.17700.18210.2400 0.1801 0.19190.19120.20670.19860.18420.1902
TS50.1589 0.1619 0.21650.22480.22240.21140.21120.16500.23620.21790.2114
TS60.15640.15320.13090.20040.13520.14530.14300.17980.1528 0.1331 0.1420
TS70.16590.15660.13450.14830.11530.13910.1389 0.1239 0.13830.13540.1387
TS8 0.2192 0.22160.25270.25930.25720.24960.25000.18450.27800.25420.2501
TS90.15760.15480.12690.13440.11460.13330.13150.13040.1330 0.1265 0.1315
TS100.15350.14560.13010.15230.11910.13630.1344 0.1212 0.13590.13030.1344
TS110.16100.15660.12800.14720.10950.13600.1350 0.1168 0.13630.12970.1348
TS120.19280.18450.15180.16240.13220.16190.1600 0.1421 0.16300.15530.1599
TS130.16060.15450.15190.1782 0.1218 0.16250.16100.12140.16160.15560.1608
TS140.17320.16550.10310.15140.11120.12010.11540.16610.1281 0.1101 0.1148
TS150.18170.17330.13870.17590.11680.14600.1440 0.1286 0.14430.14060.1441
TS160.17400.16830.09300.13920.10250.10860.10410.17200.1174 0.1001 0.1036
TS170.19030.18730.14020.22010.14650.14520.14330.17850.1520 0.1411 0.1428
TS180.17630.16600.15850.1668 0.1320 0.16910.16770.11890.16940.16300.1676
TS190.18300.17710.14470.15750.11790.15580.1542 0.1223 0.15640.14900.1540
TS200.18800.17660.16810.1784 0.1399 0.17770.17570.12710.17720.17190.1757
TS21 0.1869 0.18600.21510.23800.21110.22190.22270.21740.22660.21610.2218
TS220.16450.15740.15070.15480.14000.15410.15290.17510.1532 0.1481 0.1527
TS230.17760.17210.11880.1668 0.1247 0.13490.12900.16180.14130.12610.1287
TS240.18750.18080.12410.1755 0.1294 0.13520.13240.16390.14390.13050.1319
TS250.16620.16210.14910.1599 0.1185 0.15830.15680.11400.15790.15180.1566
TS260.19730.1841 0.1508 0.17750.14830.15780.15520.18540.16460.14830.1545
TS270.18390.17690.15650.1632 0.1531 0.16430.16230.19600.16960.15240.1614
TS28 0.1870 0.18960.20890.26390.22740.20100.19870.17150.20580.20460.1987
TS290.13870.14210.14090.20830.15230.13620.13510.12930.13530.1382 0.1349
TS30 0.1851 0.18620.19480.25140.21040.19310.18960.16660.19180.19410.1899
TS310.1913 0.1885 0.20280.27580.22340.19530.19360.16040.20300.19840.1935
TS320.19640.19140.15600.28010.17230.14990.14740.14390.15580.1524 0.1473
TS330.1623 0.1609 0.17150.25660.17710.17050.16980.14850.16960.17110.1695
TS340.17880.17890.14220.19760.14590.14160.13870.1126 0.1344 0.14090.1389
TS350.1329 0.1333 0.19570.28250.21770.18970.18410.16770.19950.19130.1846
TS360.12000.12130.12520.21840.12790.12430.12170.1013 0.1198 0.12350.1217
TS370.13610.13450.13320.20260.13680.13200.12920.1065 0.1223 0.13230.1294
TS380.1468 0.1453 0.14890.21500.15110.14730.14610.14890.14110.14680.1457
TS39 0.1687 0.16270.19760.28180.21240.19310.19440.17850.19410.19440.1937
TS400.2134 0.2126 0.23620.28990.24960.23000.23000.20410.23180.23230.2296
TS410.15730.15680.14550.25720.15520.14440.14180.16380.14030.1435 0.1416
TS420.17220.17200.17030.25140.16660.17180.17030.1807 0.1695 0.17060.1702
TS43 0.1378 0.12930.23460.28120.22430.23110.22750.17600.22000.23210.2284
TS44 0.1670 0.16620.20360.24860.20280.20440.20340.19850.20450.20360.2033
TS45 0.1320 0.12530.23510.25990.22850.22560.22320.16990.21230.23080.2239
TS46 0.1176 0.11050.18430.26310.22650.16240.15460.24360.15100.17040.1554
TS47 0.1208 0.11390.17820.24920.21100.15390.14920.22460.14340.16410.1497
TS48 0.1376 0.13010.22170.29160.26190.19230.19200.21920.19580.20470.1909
TS49 0.1272 0.11200.17090.24650.21970.14550.13560.16840.13510.15730.1362
TS50 0.1077 0.10540.15280.22740.14970.15180.15170.13160.18060.15170.1515
Table A47. Temperature time series: Polya model ( ρ = 5 % ).
Table A47. Temperature time series: Polya model ( ρ = 5 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 180 n u = 60
s = 2s = 2
n 3 = 180 n 3 = 60
TS10.62510.79180.19430.2233 0.2009 0.21320.20640.93060.2128 0.2009 0.2069
TS20.29430.3902 0.1540 0.19740.15360.17910.17601.06270.18300.16880.1756
TS30.47320.60960.10150.1813 0.1093 0.15700.14900.96820.16610.13360.1490
TS40.26500.27180.27300.30630.27460.27040.2707 0.2674 0.27540.27270.2705
TS50.1781 0.1844 0.22080.23320.22660.20560.20360.20690.21270.21020.2039
TS60.53730.57930.17220.17550.15320.17990.17610.9048 0.1721 0.17710.1767
TS70.40280.42090.10820.2025 0.1162 0.14230.13291.15330.15170.12350.1335
TS80.20360.21490.20990.22070.21590.20210.20130.24940.20690.2056 0.2014
TS90.48970.56820.11480.1710 0.1202 0.14360.14681.12260.15700.13740.1458
TS100.47290.4579 0.1092 0.12500.08820.11460.11620.95180.11370.11760.1158
TS110.38030.37330.09430.1767 0.0971 0.12430.12031.08570.14250.10790.1196
TS120.25170.25990.12990.1827 0.1375 0.15480.15480.99630.16270.14990.1543
TS130.46820.5351 0.1179 0.19140.11390.15830.14910.98660.16870.13370.1490
TS140.42050.5867 0.1307 0.15120.11200.15830.15451.06730.15490.15130.1547
TS150.58500.6819 0.1098 0.14990.10960.11740.11361.03150.11870.11030.1134
TS160.58790.6217 0.1235 0.13660.10620.14490.14411.00780.13980.14510.1439
TS170.42420.5019 0.1849 0.23470.17110.20670.20100.99280.20150.20260.2019
TS180.45990.49190.14230.17830.15140.15500.15581.08800.1671 0.1459 0.1546
TS190.46770.60580.13030.16760.14020.13850.14131.15430.1507 0.1384 0.1401
TS200.48850.5201 0.1885 0.22860.17730.21100.21360.98730.21370.21220.2129
TS210.32630.3047 0.2210 0.23540.20650.23750.23570.89110.23490.23300.2357
TS220.45900.51990.16730.19870.1728 0.1688 0.17171.07800.17050.16960.1710
TS230.49150.45950.16760.17320.13780.17430.17561.0248 0.1674 0.18040.1757
TS240.40700.4336 0.1209 0.17140.10670.15480.16081.02630.15580.15870.1596
TS250.53230.67260.12560.1731 0.1285 0.15200.14650.98230.16130.13620.1461
TS260.64060.6595 0.1639 0.17490.15710.18500.17930.94150.18410.17820.1800
TS270.35520.3690 0.1631 0.17030.15230.17840.17410.91500.17260.17440.1746
TS280.26930.25270.21130.30840.21680.1974 0.1972 0.32470.21920.19740.1967
TS290.18790.19570.17460.21460.16800.1542 0.1517 0.21950.15570.15510.1513
TS300.28300.29540.21150.27660.21430.2033 0.1994 0.26510.21520.20170.1992
TS310.23380.22950.18870.26340.19260.17320.17380.25320.19250.1752 0.1734
TS320.23100.24880.19600.32540.19000.17600.16790.31530.19810.1759 0.1687
TS330.22970.22030.19690.24380.19590.1893 0.1875 0.27070.18150.1912 0.1875
TS340.17470.18080.14640.21080.14490.14360.14070.2297 0.1401 0.13840.1403
TS350.27050.26260.22140.32450.22150.2026 0.1998 0.29210.2242 0.1998 0.1995
TS360.22450.22090.15390.27020.15090.1498 0.1429 0.24000.13840.14490.1431
TS370.21090.19270.15910.23980.15220.1533 0.1491 0.23120.14440.15160.1493
TS380.23580.24360.18020.2506 0.1712 0.18130.17830.27810.17050.17530.1773
TS390.27110.21440.20830.31250.21630.2002 0.1970 0.32060.20650.19900.1969
TS400.25850.26260.25350.30930.25380.24420.24220.36530.25550.2413 0.2417
TS410.23840.23190.16260.30830.16470.15760.15290.26210.16270.1518 0.1523
TS420.21340.2191 0.2050 0.28900.20390.20940.20910.24790.21960.20680.2081
TS430.50720.57030.15860.20810.20540.1067 0.0991 1.02100.11180.10860.0990
TS44 0.2326 0.22700.23570.27530.23320.23590.23560.29790.24000.23550.2353
TS450.45330.49600.16630.23460.21070.13020.12470.92240.13780.1339 0.1252
TS460.57800.61420.26370.29160.28480.2214 0.2052 0.91320.20120.22160.2078
TS470.31460.28680.17540.23870.2408 0.1238 0.12920.98150.15550.11470.1264
TS480.30950.51820.26100.28650.25190.25730.26810.9740 0.2557 0.27200.2670
TS490.42150.49020.22860.29730.28020.14300.15691.09350.16790.1707 0.1547
TS50 0.1399 0.13870.17130.21880.16500.17240.17520.19450.19750.17330.1747
Table A48. Temperature time series: Polya model ( ρ = 10 % ).
Table A48. Temperature time series: Polya model ( ρ = 10 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 180 n u = 60
s = 2s = 2
n 3 = 180 n 3 = 60
TS10.36930.35920.18440.18760.19070.18110.18100.8298 0.1801 0.17880.1805
TS20.45630.4058 0.1969 0.21040.19900.20110.20170.75760.20290.19380.2011
TS30.34960.35400.13940.17440.13770.15110.14840.81020.1383 0.1382 0.1456
TS4 0.2530 0.26690.26390.26130.26470.25950.25630.16960.30090.26410.2580
TS50.2629 0.2548 0.26410.26470.26330.25780.25620.18360.44360.26410.2576
TS60.41330.39360.15340.16560.15720.1555 0.1507 0.79010.15160.14810.1509
TS70.36270.3189 0.1075 0.15670.11350.13120.12680.82130.10850.10670.1237
TS8 0.3219 0.34450.33930.34220.33670.33520.33180.21940.52290.34220.3341
TS90.37710.3532 0.1194 0.16610.12470.12700.12360.88010.12020.11250.1220
TS100.34590.3205 0.1102 0.17530.10460.14200.14580.82750.13500.12410.1441
TS110.35310.37910.09230.13960.09460.10450.10420.86250.0897 0.0898 0.1003
TS120.34140.3407 0.1134 0.17040.11360.13120.13190.86450.11240.11980.1284
TS130.29630.27170.11830.1780 0.1177 0.13080.13150.87800.11410.12070.1276
TS140.34760.3094 0.1234 0.16810.10600.15150.15340.82650.14490.13770.1531
TS150.40530.36060.11470.17700.11820.12180.11600.82940.1081 0.1095 0.1142
TS160.31840.3085 0.1285 0.15610.11190.15370.15340.85940.14850.14250.1533
TS170.31350.3322 0.1731 0.24230.16480.18820.18840.80140.18700.17760.1881
TS180.42530.36820.12510.1788 0.1235 0.12900.12880.85290.11650.12450.1251
TS190.35150.36970.11260.15630.11060.11130.11020.89150.10220.1087 0.1072
TS200.33170.3151 0.1420 0.19440.14130.14340.14630.87510.1440 0.1420 0.1426
TS210.43000.4119 0.2116 0.22740.20440.23230.23070.71580.22570.22060.2303
TS220.33420.3060 0.1189 0.16500.10250.16120.16080.84400.15050.13810.1597
TS230.38400.3771 0.1381 0.17320.12190.17950.18400.80450.17340.16000.1818
TS240.40840.3870 0.1456 0.16240.13010.16490.16460.81590.16090.15950.1651
TS250.31590.2679 0.1177 0.17820.11960.13300.12830.86820.11250.12080.1255
TS260.42660.43810.1902 0.1892 0.18130.21020.20920.83130.20770.19500.2095
TS270.31370.32270.1750 0.1700 0.16670.18250.18300.81670.18500.17700.1831
TS280.28400.28280.21730.27790.22540.20330.19740.31320.22720.2095 0.1984
TS290.20720.20650.15550.22060.15800.14810.14130.25820.15260.1511 0.1428
TS300.24080.24210.19540.31660.19160.18190.17800.25630.19280.1886 0.1788
TS310.23600.23370.19050.24320.19560.17830.17270.30000.20600.1835 0.1735
TS320.22550.21790.14980.25260.15660.14830.14060.26790.19850.1469 0.1417
TS330.25010.24960.18180.24390.18280.17620.17180.23920.17480.1778 0.1725
TS340.22950.22630.17850.28560.17880.17710.17260.26550.17450.1747 0.1730
TS350.26290.25630.21150.27690.21160.19150.18850.29480.21650.2019 0.1892
TS360.23450.22990.16230.27810.15770.1663 0.1604 0.26830.16440.16150.1608
TS370.23840.22520.17810.28270.17620.1790 0.1753 0.28240.17120.17610.1754
TS380.22300.22820.17510.2573 0.1720 0.17590.17190.26010.17710.17330.1721
TS390.28750.27630.23780.29270.24880.24150.24310.30660.2518 0.2384 0.2422
TS400.26050.26680.22790.28400.24020.22580.22170.30090.23190.2234 0.2220
TS410.20360.20200.13250.23980.13440.1317 0.1315 0.23370.1710 0.1315 0.1312
TS420.21520.21970.18410.25660.18370.18510.18310.22690.25460.1829 0.1830
TS430.17920.17250.17420.2696 0.1724 0.17390.17410.16880.22840.17390.1739
TS440.24990.23290.21360.25300.21900.2169 0.2135 0.26880.21050.21380.2138
TS450.31430.30250.23870.27760.26440.17470.15620.71940.22910.2207 0.1639
TS460.42600.43100.23100.25620.24960.21680.19800.76860.25590.2304 0.2044
TS470.26050.23450.19630.27470.21830.13870.12710.81200.19440.1758 0.1314
TS480.32400.29800.26720.34630.2655 0.2629 0.27120.81160.31380.26070.2702
TS490.23970.23190.25030.30160.27810.16790.14740.69450.25610.2203 0.1534
TS50 0.1871 0.17990.20120.22040.19880.19910.20000.20550.26920.19970.1997
Table A49. Temperature time series: Polya model ( ρ = 15 % ).
Table A49. Temperature time series: Polya model ( ρ = 15 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 180 n u = 60
s = 2s = 2
n 3 = 180 n 3 = 60
TS10.43390.5828 0.2001 0.22110.20280.21120.20800.88660.21530.19430.2068
TS20.48610.54230.17690.2345 0.1785 0.19400.19290.86550.20480.17950.1911
TS30.34990.47870.14730.1664 0.1476 0.16210.15950.88520.16110.15300.1590
TS40.34890.49050.18000.24290.18870.20450.20030.86730.2257 0.1848 0.1965
TS5 0.2042 0.20030.21900.21750.22080.20930.20940.20820.22060.21840.2105
TS60.42570.5894 0.1495 0.19670.15140.15860.15690.83910.17350.14870.1546
TS70.43600.51450.13220.1721 0.1352 0.15340.15600.97120.16090.14080.1540
TS80.30430.30500.29670.29750.2944 0.2873 0.28840.28040.31230.29490.2885
TS90.34060.4121 0.1210 0.19300.12180.13330.13090.90690.15190.11850.1271
TS100.35640.54020.12390.1607 0.1248 0.13740.13340.88770.13330.12520.1333
TS110.40850.53360.11880.1594 0.1199 0.13920.13830.94750.13950.12970.1376
TS120.37910.60140.14120.1684 0.1451 0.16550.16650.91770.16730.15490.1649
TS130.31810.50000.13170.1706 0.1355 0.15640.15880.92210.16860.14540.1562
TS140.35510.5168 0.1121 0.19880.11190.12580.11610.92650.14900.11510.1132
TS150.42760.5804 0.1269 0.14280.12580.13950.13830.89140.13390.13380.1385
TS160.40870.5777 0.1354 0.17710.13210.15740.15580.88970.14900.14450.1556
TS170.49020.55290.16140.21930.15440.16000.15480.90220.17610.1522 0.1533
TS180.39650.4782 0.1398 0.17910.13910.16980.16740.90820.17190.15450.1658
TS190.43920.5995 0.1355 0.15520.13350.15700.15390.88040.15150.14600.1538
TS200.40860.46270.19390.2333 0.1888 0.19660.19010.89630.20390.18180.1892
TS210.48270.63280.19260.22810.19910.20620.20350.81620.2202 0.1980 0.2010
TS220.32730.4612 0.1280 0.19430.12490.13750.13130.90350.15890.13150.1295
TS230.38090.44690.11870.17020.11380.12290.11890.85670.14020.1196 0.1167
TS240.52490.66590.15400.16160.15160.16300.16240.8895 0.1531 0.15770.1629
TS250.44360.59330.13670.1677 0.1375 0.16540.16220.86090.16460.14740.1611
TS260.42690.60630.16640.21270.15820.1596 0.1561 0.85500.16810.15750.1559
TS270.41370.53780.15760.1881 0.1510 0.16180.15600.85440.17070.14920.1549
TS280.25540.25530.22520.26920.2272 0.2155 0.21500.29720.22260.22080.2156
TS290.20540.20450.15950.21740.15810.1488 0.1497 0.24220.14980.15570.1502
TS300.23360.22050.18660.27870.19030.1778 0.1753 0.24680.17480.18270.1760
TS310.24600.25050.23820.28220.24090.22780.22620.30780.23290.2327 0.2271
TS320.26110.26370.19060.28690.18720.17730.17070.28400.18120.1830 0.1726
TS330.23920.23820.18820.25990.18820.18520.18280.23360.18380.1871 0.1832
TS340.19290.18670.15110.22350.14810.1488 0.1463 0.18610.14610.14840.1466
TS35 1.0000 1.0000 1.0000 0.9269 1.0000 1.0000 1.0000 1.0000 4.0651 1.0000 1.0000
TS360.18760.17790.13420.23080.13420.1376 0.1347 0.19810.13540.13490.1349
TS370.21980.21730.16080.24400.15750.1573 0.1532 0.21230.15070.15720.1537
TS380.25400.25360.17030.26500.16980.1703 0.1650 0.23840.16140.16810.1654
TS390.24750.26170.21940.26570.22700.2169 0.2163 0.29960.22070.21700.2161
TS400.29170.27820.27760.30500.2764 0.2719 0.27260.31390.26920.27410.2726
TS410.24010.23150.16820.28970.16460.1666 0.1639 0.23840.16650.16550.1638
TS420.25520.25130.20920.2774 0.2038 0.21040.20780.22920.20270.20860.2079
TS430.19330.18930.16940.2476 0.1691 0.17120.16960.15640.17660.17010.1698
TS440.23170.22390.21740.24550.21780.21790.21760.21940.22130.2176 0.2175
TS450.41290.61430.21120.27010.21650.1677 0.1638 0.83250.15030.19150.1687
TS460.35170.5053 0.2306 0.28630.25000.24140.23970.81920.21940.23690.2413
TS470.32160.41710.21670.26630.22410.2156 0.2147 0.85250.20530.21500.2157
TS480.39650.60590.26670.32300.2723 0.2060 0.21270.78040.20450.24750.2183
TS490.32980.55340.24220.27740.24620.1822 0.1752 0.90000.15120.21950.1833
TS500.1966 0.1940 0.19740.23780.19830.19770.19790.18940.21060.19720.1977
Table A50. Temperature time series: Polya model ( ρ = 20 % ).
Table A50. Temperature time series: Polya model ( ρ = 20 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 180 n u = 60
s = 2s = 2
n 3 = 180 n 3 = 60
TS10.50090.3336 0.2148 0.22960.22210.22740.22540.20660.22660.22200.2253
TS20.49380.41070.16290.2285 0.1617 0.19000.18700.15760.18720.18420.1872
TS30.42880.3421 0.1707 0.19130.16940.17670.17750.23160.17660.17820.1773
TS40.55790.45740.24150.26030.23990.22500.22700.1722 0.2159 0.23200.2273
TS50.53660.41220.23200.23740.23390.20510.20580.1667 0.1940 0.21050.2059
TS60.47750.3953 0.1579 0.19030.15330.17710.17550.20690.17420.17560.1758
TS70.44760.40290.11600.1748 0.1269 0.14750.14270.20550.15320.13480.1422
TS8 0.2467 0.24920.28600.28220.28870.26750.27040.20280.26330.27140.2700
TS90.48220.40630.11440.1629 0.1248 0.15010.14390.13430.14700.13770.1441
TS100.47720.3742 0.1203 0.18460.11760.14840.15160.15200.14690.15510.1515
TS110.52250.40840.10540.1498 0.1101 0.13190.12740.12300.13560.11950.1270
TS120.54720.41720.12930.1875 0.1355 0.17380.17020.17090.17990.15910.1690
TS130.53980.4517 0.1080 0.15820.10640.15650.14800.11660.15470.13810.1480
TS140.51400.3466 0.1187 0.14470.11320.15310.15110.14560.14710.15370.1517
TS150.49630.41220.14740.16050.13930.14500.14550.1811 0.1418 0.15220.1461
TS160.39320.3368 0.1210 0.14370.11320.14560.14410.14880.13950.14840.1447
TS170.23600.22540.23880.26200.24210.21410.21730.1764 0.2087 0.22130.2171
TS180.47460.32500.12800.1712 0.1292 0.16710.16200.13120.17060.15110.1614
TS190.45660.3406 0.1440 0.16520.13870.15010.15060.18530.15040.15390.1507
TS200.52720.43190.15360.2015 0.1565 0.18530.18450.18110.19360.17460.1832
TS210.54190.4432 0.2130 0.22640.20530.23230.23130.24540.22740.23220.2318
TS220.30990.2875 0.1183 0.14950.11110.14310.14210.14370.13850.14520.1426
TS230.27250.26020.13930.1723 0.1452 0.16100.15230.15150.16050.14620.1527
TS240.41640.3403 0.1498 0.17380.14190.16740.16640.19530.16130.17130.1672
TS250.40020.30840.12280.1740 0.1257 0.16790.16510.14190.17370.15440.1641
TS260.49010.44920.16790.1722 0.1606 0.17070.16950.15210.16700.17540.1703
TS270.47290.40260.16380.1684 0.1569 0.17130.16600.14770.16480.16930.1669
TS280.23980.22820.20020.23470.19720.18370.18210.15120.20000.1845 0.1820
TS290.26410.21950.17120.23870.17580.1671 0.1636 0.15540.16650.16740.1641
TS300.25250.25150.18660.29060.18590.17870.17940.19180.18420.1804 0.1789
TS310.26810.27170.25620.30830.25560.23120.22770.19450.25620.2286 0.2273
TS320.22690.22030.19570.31440.18940.18040.17510.19220.19310.1777 0.1755
TS330.31220.29990.22760.30110.22970.2200 0.2182 0.24340.22090.22010.2181
TS340.32530.32880.17730.23710.17310.17680.17230.1539 0.1704 0.17190.1726
TS350.2186 0.2097 0.24280.32380.25140.23180.22890.20660.24470.23290.2292
TS360.22400.20380.16670.24240.15810.16040.15810.1508 0.1562 0.15920.1582
TS370.23060.22440.15740.21720.14930.15120.14770.1301 0.1468 0.14850.1479
TS380.26220.25140.16000.2326 0.1506 0.15900.15680.14230.15210.15720.1568
TS390.28850.27470.22160.29660.22530.21240.21150.19650.2220 0.2103 0.2109
TS400.28750.28900.24520.28120.2511 0.2420 0.24410.22340.24680.24250.2433
TS410.22800.21620.16600.24930.1652 0.1647 0.16510.18790.16350.16530.1648
TS420.20420.19220.17790.23390.17250.17740.1778 0.1754 0.17700.17720.1774
TS430.51960.43500.20140.29850.23220.15890.13900.20350.16740.1327 0.1382
TS440.24590.23550.22440.2541 0.2226 0.22380.22470.20030.22270.22370.2244
TS450.54640.36670.22400.25210.2297 0.1835 0.19140.17630.19130.20490.1920
TS460.48550.35860.28470.30870.2973 0.2273 0.22950.19720.23750.24490.2306
TS470.47970.40150.21070.26330.23370.15200.14750.24460.17820.1343 0.1449
TS480.2016 0.1743 0.18130.21790.18460.18410.18380.14620.19050.18330.1838
TS490.42790.32010.24990.28720.26820.19330.18660.26120.22780.1672 0.1835
TS500.1969 0.1836 0.18630.23200.18790.18530.18470.13960.19790.18460.1847

Appendix A.5. Air Time Series: Numerical Results

Appendix A.5.1. Results obtained by using DLM

Table A51. Air time series: Sampling without replacement ( ρ = 5 % ).
Table A51. Air time series: Sampling without replacement ( ρ = 5 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 2
n 3 = 360 n 3 = 120
TS10.2548 0.2434 0.36170.47140.32340.35690.35900.20710.35880.35540.3574
TS20.34200.30630.30640.50130.34000.31050.31260.0752 0.3036 0.31830.3108
TS30.2761 0.2445 0.35600.37040.45060.36460.36550.18430.36790.33990.3636
TS40.18930.23900.31251.06170.22860.31380.3153 0.2197 0.31430.29310.3136
TS50.21840.22810.34130.42870.42870.33910.3408 0.2255 0.34570.31850.3388
TS61.00380.94971.4207 0.9757 1.14871.22121.32920.99163.54581.13041.2493
TS70.9136 0.9201 3.97681.05911.85761.77522.12611.03222.07031.51061.8435
TS81.1018 1.0660 1.38261.51271.13691.63441.90631.00192.21601.46671.6949
TS90.2104 0.1867 0.25120.43740.30020.24730.24660.06050.26390.24560.2463
TS100.2302 0.2036 0.30290.32640.43770.31570.30900.17180.31230.31710.3119
Table A52. Air time series: Sampling without replacement ( ρ = 10 % ).
Table A52. Air time series: Sampling without replacement ( ρ = 10 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 360 n u = 240
s = 2s = 2
n 3 = 240 n 3 = 120
TS1 0.2260 0.23330.35310.46010.27240.35610.35440.18230.36370.35490.3548
TS20.23350.2322 0.2316 0.45010.26880.23510.23250.09970.25910.23990.2336
TS3 0.3041 0.34730.49300.48970.53940.45640.46050.16230.45180.45500.4576
TS40.2552 0.2373 0.33580.85130.41820.34350.34440.20400.33280.32140.3422
TS5 0.2240 0.23260.30170.53120.29020.29640.29740.22020.30700.27490.2961
TS60.95661.04211.85481.01130.99431.20951.5597 0.9890 3.01171.10141.2741
TS70.98770.98122.64340.97221.15761.28562.1441 0.9749 1.12351.03341.4263
TS8 1.0215 1.04391.32341.47771.05711.33431.48641.00251.45641.27351.3716
TS90.3168 0.2966 0.31500.75110.34920.29960.29670.08520.30490.29810.2984
TS10 0.3106 0.31380.40250.42480.53560.39310.39380.16140.39970.39560.3933
Table A53. Air time series: Sampling without replacement ( ρ = 15 % ).
Table A53. Air time series: Sampling without replacement ( ρ = 15 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 360 n u = 240
s = 2s = 2
n 3 = 240 n 3 = 120
TS1 0.2789 0.29060.37830.46580.31040.36360.35990.18660.36670.36620.3612
TS2 0.2654 0.27180.33530.53070.39850.33820.34170.16030.33470.30800.3377
TS30.25600.24530.35970.37930.34670.36960.3670 0.2484 0.37460.35760.3671
TS40.2605 0.2481 0.37541.35610.37730.36070.36360.20880.37490.33660.3599
TS50.2482 0.2572 0.33510.77910.38190.33300.33120.26330.33960.29840.3293
TS60.9320 0.9828 2.25861.00041.03431.50931.87910.98925.21371.32841.5798
TS70.99930.97581.6955 0.9870 1.66471.44161.65551.00161.88351.31031.4878
TS81.00810.98911.17581.2829 0.9740 1.23171.39250.96571.51701.10051.2599
TS9 0.2195 0.22440.30670.71590.36210.30750.30410.13750.32570.28710.3035
TS100.2888 0.2849 0.38550.41840.34240.37830.37870.24190.38010.36960.3776
Table A54. Air time series: Sampling without replacement ( ρ = 20 % ).
Table A54. Air time series: Sampling without replacement ( ρ = 20 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.33600.33600.36920.45550.31360.38100.3755 0.3303 0.43840.37780.3728
TS2 0.2771 0.2771 0.30850.57130.37380.33140.32120.23950.41260.30920.3201
TS3 0.2942 0.2942 0.35810.39450.32900.36920.36950.24030.41760.35970.3632
TS40.28360.28360.43081.32300.40280.43740.4359 0.3515 0.50210.42300.4306
TS50.27350.27350.30240.50520.34230.31370.31560.30010.3826 0.2884 0.3078
TS60.85160.85162.25291.08771.00431.39002.5798 0.9602 5.44481.15121.5534
TS70.94780.94781.5730 1.0077 1.38671.36961.49321.01461.64261.31491.3537
TS80.96400.96401.09010.99550.98470.78711.09140.96541.0942 0.7891 0.8220
TS90.32110.3211 0.2708 0.34340.34770.29520.28500.21020.36770.28680.2842
TS10 0.2677 0.2677 0.33461.73450.36450.34330.34520.22930.40420.31920.3371
Table A55. Air time series: Polya model ( ρ = 5 % ).
Table A55. Air time series: Polya model ( ρ = 5 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 480 n u = 240
s = 3s = 2
n 3 = 360 n 3 = 120
TS10.38261.79710.37380.4482 0.2164 0.36000.37190.07240.35200.34070.3632
TS20.23260.26810.22230.28910.21410.20530.19420.1253 0.1511 0.19110.1987
TS30.61910.56370.29990.31880.33090.3248 0.3083 0.38820.32160.34910.3172
TS40.53181.30910.32550.5860 0.2383 0.31320.32160.10060.29680.30090.3160
TS50.2660 0.2456 0.24880.77280.26410.25490.25350.23870.26040.27670.2567
TS61.02991.00933.13210.97761.17521.57093.3088 0.9938 6.86701.23021.9523
TS70.9189 0.8498 2.80331.26680.81782.42392.71781.15462.79982.17272.5071
TS80.99881.20391.06551.00280.93451.09001.1379 0.9486 1.19831.03911.1117
TS90.2070 0.1876 0.21920.38550.29140.22710.21920.13030.22340.21510.2226
TS10 0.3934 0.45430.43030.43560.43970.42490.42340.38370.42100.43100.4245
Table A56. Air time series: Polya model ( ρ = 10 % ).
Table A56. Air time series: Polya model ( ρ = 10 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 360 n u = 240
s = 2s = 2
n 3 = 240 n 3 = 120
TS10.38280.33470.41360.5378 0.2347 0.41230.42060.22250.43890.39960.4155
TS20.37110.40100.16890.3873 0.1351 0.14400.15460.09030.20160.15660.1454
TS30.82150.8423 0.2970 0.41600.39400.31530.30760.28930.29860.35790.3122
TS40.37070.39250.34671.1146 0.3237 0.35330.34670.15810.34840.35330.3501
TS50.20890.22710.27820.39100.27390.27920.2819 0.2224 0.31000.23860.2776
TS61.01100.99531.34881.01940.98581.00081.0256 0.9915 2.42241.00591.0280
TS70.97651.13352.30911.01131.69211.94912.3484 1.0056 2.08941.48802.0728
TS81.07201.10291.0466 0.9774 0.95681.24911.35820.98381.58951.12681.2811
TS90.31360.32090.25390.43610.32180.2571 0.2537 0.07000.26420.27290.2551
TS100.3709 0.3264 0.33021.97920.46220.34640.34050.29660.33500.34380.3430
Table A57. Air time series: Polya model ( ρ = 15 % ).
Table A57. Air time series: Polya model ( ρ = 15 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 360 n u = 240
s = 2s = 2
n 3 = 240 n 3 = 120
TS10.51920.4733 0.3916 0.54310.50830.43240.41940.18580.41590.44690.4312
TS20.39790.36470.35720.71370.40510.3642 0.3495 0.15430.35590.35610.3632
TS3 0.3484 0.36570.43260.86000.43190.42950.42860.32500.43260.40460.4272
TS40.31550.30360.34510.7049 0.3009 0.34070.36240.18370.34780.31810.3473
TS50.41660.41840.24690.55650.28350.24480.22520.23530.2449 0.2305 0.2434
TS60.9531 0.9484 6.99850.84751.00142.80752.97210.98746.32402.25253.1954
TS70.98520.98130.79940.96760.87120.87601.16770.9842 0.8289 0.91830.8507
TS8 0.9797 1.04711.04360.98490.97461.60154.57120.99451.23591.29371.5801
TS90.57200.54210.20660.60700.27620.2091 0.2048 0.12890.20550.22390.2115
TS100.27190.31830.38790.56830.34520.39150.3932 0.2944 0.39430.36550.3891
Table A58. Air time series: Polya model ( ρ = 20 % ).
Table A58. Air time series: Polya model ( ρ = 20 % ).
TimeDLM+BICDLM+EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.43800.43800.33780.46600.36780.36010.34330.20870.3450 0.3356 0.3592
TS20.45800.45800.23100.36910.24720.21810.22480.22870.22990.2450 0.2213
TS30.50150.50150.49770.48870.49000.49480.49570.35420.4848 0.4744 0.4946
TS40.56110.56110.43981.79640.42640.36780.38010.22050.38150.3907 0.3601
TS50.38490.38490.27091.16700.24780.2388 0.2321 0.27160.26630.21810.2554
TS61.01901.01901.97521.00821.22821.62922.7796 1.0096 4.49911.38561.7661
TS70.98100.98101.66071.00101.0619 0.9491 1.56450.95501.88310.91201.4091
TS8 1.0111 1.0111 2.86301.05110.99311.12621.84241.02921.83841.10411.3968
TS90.27040.27040.18170.64380.18600.17140.17480.21950.1746 0.1711 0.1705
TS100.59760.59760.42490.66850.46880.41170.41710.33800.4139 0.3972 0.4119

Appendix A.5.2. Results obtained by using DLM1

Table A59. Air time series: Sampling without replacement ( ρ = 5 % ).
Table A59. Air time series: Sampling without replacement ( ρ = 5 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.18060.18060.25060.3348 0.1741 0.24300.24250.10680.24440.25260.2417
TS2 0.1401 0.1401 0.19250.26690.17800.18310.19360.06170.18230.19220.1874
TS3 0.1751 0.1751 0.23800.25860.21800.27970.26960.12000.26960.26130.2740
TS40.13600.13600.21940.74780.10940.22060.2201 0.1213 0.21500.21870.2185
TS50.16620.16620.28140.32200.32080.28550.2950 0.1998 0.29510.27350.2889
TS6 0.9998 0.9998 1.68421.00941.20061.27151.43540.96675.17281.11701.3166
TS70.89900.89903.5650 0.9825 1.77111.63731.97321.01331.61221.40271.7003
TS80.96980.96981.35271.58561.15731.51151.7568 1.0206 1.94461.39341.5456
TS9 0.1176 0.1176 0.17470.32680.19380.17050.17510.04970.18220.17210.1720
TS10 0.1425 0.1425 0.21720.22860.26030.24280.23180.11740.23240.23720.2371
Table A60. Air time series: Sampling without replacement ( ρ = 10 % ).
Table A60. Air time series: Sampling without replacement ( ρ = 10 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.17250.17250.25670.3463 0.1418 0.25250.24890.12220.25400.25650.2499
TS2 0.1333 0.1333 0.16380.30360.15340.16270.16970.08030.18180.16890.1654
TS3 0.1730 0.1730 0.28890.30660.30300.30530.29700.11510.28580.29500.3005
TS4 0.1836 0.1836 0.24350.61810.20550.24120.23610.12250.22830.23310.2374
TS50.14290.14290.23890.37460.22150.24000.2451 0.1958 0.24070.22770.2419
TS6 1.0393 1.0393 2.62921.09421.06021.53422.05740.99594.66361.30771.6549
TS70.93810.93813.19590.99851.27451.52012.5258 0.9903 1.38181.19821.7120
TS8 0.9889 0.9889 1.25211.36671.05641.34371.46760.98511.54701.27371.3722
TS9 0.1533 0.1533 0.22120.53670.21410.20930.21630.06760.22360.21330.2129
TS10 0.1817 0.1817 0.27710.28420.31620.28530.27850.11770.27750.28750.2818
Table A61. Air time series: Sampling without replacement ( ρ = 15 % ).
Table A61. Air time series: Sampling without replacement ( ρ = 15 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.20000.20000.26270.3320 0.1606 0.25900.25380.13800.25420.25540.2557
TS2 0.1489 0.1489 0.21350.33970.23550.20880.21760.10460.21170.19660.2112
TS30.16620.16620.24630.27700.20040.27760.2679 0.1765 0.26740.26570.2725
TS40.19210.19210.25620.8070 0.1819 0.25230.25410.14720.25970.23540.2508
TS50.16640.16640.22680.47700.24260.22550.2258 0.2159 0.22480.21660.2247
TS60.99740.99742.10180.95491.02881.41521.6952 0.9765 4.48631.25721.4719
TS70.98140.98141.7166 0.9840 1.62691.45361.70081.00071.77771.30891.5167
TS80.96560.96561.37301.42430.99091.47011.7077 0.9724 1.85211.29541.5243
TS9 0.1342 0.1342 0.20000.44810.21630.19850.20020.08590.21140.19200.1977
TS100.15790.15790.25870.27650.18760.27450.2652 0.1660 0.26550.26740.2700
Table A62. Air time series: Sampling without replacement ( ρ = 20 % ).
Table A62. Air time series: Sampling without replacement ( ρ = 20 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.20640.20640.25420.33720.15980.26360.2546 0.1995 0.27260.25510.2556
TS2 0.1789 0.1789 0.20500.37120.23000.21460.21830.14160.23700.21200.2120
TS30.20940.20940.23820.2942 0.1803 0.27320.26560.17450.26750.25960.2669
TS40.17890.17890.30560.9343 0.1834 0.31340.31340.21020.34420.29800.3084
TS50.18480.18480.22610.38500.24110.23200.23810.24430.2447 0.2161 0.2304
TS60.97280.97282.82001.01551.04701.69873.1953 0.9899 7.14081.42691.9390
TS70.98660.98661.71511.00161.44461.34351.6010 1.0005 1.61191.27411.3968
TS8 0.9862 0.9862 1.30781.09121.02881.06381.34450.97851.37711.05091.0954
TS90.17820.1782 0.1681 0.24860.19910.17800.18090.11440.19760.17900.1750
TS10 0.1732 0.1732 0.22161.08290.17330.25130.24260.15610.26140.23730.2443
Table A63. Air time series: Polya model ( ρ = 5 % ).
Table A63. Air time series: Polya model ( ρ = 5 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.50880.50880.30540.3556 0.1543 0.28770.30270.06610.29170.28540.2929
TS20.75190.75190.12130.15980.10520.10390.1157 0.0932 0.08440.09880.1083
TS30.82210.82210.16040.19340.19300.1994 0.1776 0.25170.18120.22530.1903
TS40.44110.44110.23150.3886 0.1253 0.21820.23200.09640.19960.21820.2256
TS50.60200.6020 0.1900 0.38070.19110.19860.20800.18890.19120.22890.2073
TS61.00711.00712.8668 1.0065 1.14111.46422.97970.98495.71041.17921.7468
TS70.93190.93191.98801.0857 1.0463 1.63611.77841.06261.77131.51851.6580
TS81.00611.00611.07861.0624 0.9665 1.12981.16880.92441.22901.07931.1483
TS90.42220.4222 0.1433 0.27600.16770.14900.14960.10750.14520.14660.1494
TS100.63030.63030.29910.30520.30370.31210.30180.2534 0.2981 0.32710.3069
Table A64. Air time series: Polya model ( ρ = 10 % ).
Table A64. Air time series: Polya model ( ρ = 10 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.52830.52830.29680.4329 0.1385 0.30130.30330.13550.31750.29550.3009
TS20.73490.73490.10480.2555 0.0912 0.0912 0.10560.06370.10490.10820.0966
TS30.59050.59050.20800.3072 0.2064 0.24570.23160.19170.21520.27130.2399
TS40.71750.71750.28120.8631 0.2248 0.29100.28120.12890.27030.28650.2866
TS50.49160.49160.20030.26160.20090.20260.20750.18390.2072 0.1920 0.2033
TS61.00231.00231.78051.24200.95801.15541.1975 0.9724 5.60811.11321.1913
TS7 1.0023 1.0023 2.28201.00691.63191.99252.33580.99562.03781.54212.1045
TS80.96310.96311.13101.0313 0.9756 1.35461.50010.98631.84401.20051.4007
TS90.52980.52980.20510.38880.2201 0.2000 0.20420.05150.20860.21240.2010
TS100.63100.63100.23231.1563 0.2154 0.27110.25680.20260.24240.26980.2640
Table A65. Air time series: Polya model ( ρ = 15 % ).
Table A65. Air time series: Polya model ( ρ = 15 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.56220.56220.29280.4109 0.2667 0.31370.30770.13060.31120.31860.3143
TS20.68020.68020.25020.44140.26630.2522 0.2493 0.11970.25350.25640.2544
TS30.51890.51890.25280.47430.21740.28580.2791 0.2225 0.27490.27800.2800
TS40.71110.71110.22160.4279 0.1437 0.21770.23120.12140.21650.20580.2180
TS50.81080.81080.18510.39390.19360.18380.17290.20670.1865 0.1796 0.1839
TS60.99860.99865.61540.97931.07232.19262.8821 0.9895 4.29391.76932.4950
TS71.00171.00171.2842 0.9989 1.27491.07581.29040.98891.27211.10441.0767
TS8 0.9836 0.9836 1.12751.07650.98131.79294.96080.99571.38691.45141.7940
TS90.82540.82540.15220.33640.1829 0.1508 0.15190.09310.15270.15900.1542
TS100.67310.67310.26000.39100.17600.29180.2849 0.2078 0.28490.28070.2873
Table A66. Air time series: Polya model ( ρ = 20 % ).
Table A66. Air time series: Polya model ( ρ = 20 % ).
TimeDLM 1 +BICDLM 1 +EBICCDRecGROUSEROSLSoftImpSVDImpSVTTeNMFDynaMMoTRMF
Series n u = 240 n u = 240
s = 2s = 2
n 3 = 120 n 3 = 120
TS10.76280.76280.22910.3083 0.2214 0.25330.22850.17160.22340.23400.2431
TS20.63560.63560.13770.20890.13770.12960.13460.14620.15510.1462 0.1331
TS30.58380.58380.30790.3065 0.2938 0.32590.31040.23860.31760.32290.3204
TS40.75950.75950.22150.77330.24410.20740.21080.15420.21370.2245 0.2045
TS50.54180.54180.15480.47970.13590.14990.15860.21030.1698 0.1475 0.1534
TS6 0.9891 0.9891 2.14160.99921.20551.51872.80730.98506.82551.30071.6908
TS7 0.9971 0.9971 1.89981.03151.26261.11501.84020.99262.31241.09551.6477
TS8 1.0063 1.0063 3.10071.29661.00731.16351.98440.98852.10201.11041.5626
TS90.66720.66720.10670.49280.1051 0.1025 0.10940.13000.11190.10130.1061
TS100.63570.63570.26710.41780.27720.2694 0.2620 0.22910.27510.27710.2678

References

  1. Brockwell, P.J.; Davis, R.A. Introduction to Time Series and Forecasting; Springer: New York, NY, USA, 2002. [Google Scholar]
  2. Khayati, M.; Lerner, A.; Tymchenko, Z.; Cudré-Mauroux, P. Mind the gap: An experimental evaluation of imputation of missing values techniques in time series. VLDB Endow. 2020, 13, 768–782. [Google Scholar] [CrossRef]
  3. Li, L.; Mccann, J.; Pollard, N.; Faloutsos, C. DynaMMo: Mining and summarization of coevolving sequences with missing values. In Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Paris, France, 28 June–1 July 2009; pp. 507–516. [Google Scholar] [CrossRef]
  4. Troyanskaya, O.; Cantor, M.; Sherlock, G.; Brown, P.; Hastie, T.; Tibshirani, R.; Botstein, D.; Altman, R.B. Missing value estimation methods for DNA microarrays. Bioinformatics 2001, 17, 520–525. [Google Scholar] [CrossRef] [PubMed]
  5. Mazumder, R.; Hastie, T.; Tibshirani, R. Spectral Regularization Algorithms for Learning Large Incomplete Matrices. J. Mach. Learn. Res. 2010, 11, 2287–2322. [Google Scholar] [PubMed]
  6. Cai, J.; Candès, E.J.; Shen, Z. A Singular Value Thresholding Algorithm for Matrix Completion. SIAM J. Optim. 2010, 20, 1956–1982. [Google Scholar] [CrossRef]
  7. Khayati, M.; Böhlen, M.; Gamper, J. Memory-efficient centroid decomposition for long time series. In Proceedings of the IEEE 30th International Conference on Data Engineering, Chicago, IL, USA, 31 March–4 April 2014; pp. 100–111. [Google Scholar] [CrossRef]
  8. Khayati, M.; Cudré-Mauroux, P.; Böhlen, M.H. Scalable recovery of missing blocks in time series with high and low cross-correlations. Knowl. Inf. Syst. 2020, 62, 2257–2280. [Google Scholar] [CrossRef]
  9. Zhang, D.; Balzano, L. Global Convergence of a Grassmannian Gradient Descent Algorithm for Subspace Estimation. In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics, Limassol, Cyprus, 14–17 September 2009; Gretton, A., Robert, C.C., Eds.; PMLR: Cadiz, Spain, 2016; Volume 51, pp. 1460–1468. [Google Scholar]
  10. Balzano, L.; Chi, Y.; Lu, Y.M. Streaming PCA and Subspace Tracking: The Missing Data Case. Proc. IEEE 2018, 106, 1293–1310. [Google Scholar] [CrossRef]
  11. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? J. ACM 2011, 58, 11.1–11.37. [Google Scholar] [CrossRef]
  12. Shu, X.; Porikli, F.; Ahuja, N. Robust Orthonormal Subspace Learning: Efficient Recovery of Corrupted Low-Rank Matrices. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3874–3881. [Google Scholar] [CrossRef]
  13. Mei, J.; Castro, Y.D.; Goude, Y.; Hébrail, G. Nonnegative Matrix Factorization for Time Series Recovery From a Few Temporal Aggregates. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 2382–2390. [Google Scholar]
  14. Yu, H.; Rao, N.; Dhillon, I.S. Temporal Regularized Matrix Factorization for High-dimensional Time Series Prediction. In Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 847–855. [Google Scholar]
  15. Zheng, X.; Dumitrescu, B.; Liu, J.; Giurcăneanu, C.D. On the Use of Dictionary Learning in Time Series Imputation. In Proceedings of the 28th European Signal Processing Conference (EUSIPCO), Dublin, Ireland, 23–27 August 2021; pp. 2016–2020. [Google Scholar] [CrossRef]
  16. Dumitrescu, B.; Irofti, P. Dictionary Learning Algorithms and Applications; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar] [CrossRef]
  17. Rubinstein, R.; Zibulevsky, M.; Elad, M. Efficient Implementation of the K-SVD Algorithm Using Batch Orthogonal Matching Pursuit; Report CS-2008-08; Technion Univ.: Haifa, Israel, 2008. [Google Scholar]
  18. Sadeghi, M.; Babaie-Zadeh, M.; Jutten, C. Learning Overcomplete Dictionaries Based on Atom-by-Atom Updating. IEEE Trans. Signal Process. 2014, 62, 883–891. [Google Scholar] [CrossRef]
  19. Garcia-Cardona, C.; Wohlberg, B. Convolutional dictionary learning: A comparative review and new algorithms. IEEE Trans. Comput. Imaging 2018, 4, 366–381. [Google Scholar] [CrossRef]
  20. Guichaoua, C.; Dictionary Learning for Audio Inpainting. Report HAL Id: Dumas-00725263, HAL Archives-Ouvertes. Available online: http://dumas.ccsd.cnrs.fr/dumas-00725263 (accessed on 4 August 2020).
  21. Mairal, J.; Elad, M.; Sapiro, G. Sparse representation for color image restoration. IEEE Trans. Image Process. 2008, 17, 53–69. [Google Scholar] [CrossRef]
  22. Mairal, J.; Sapiro, G.; Elad, M. Learning multiscale sparse representations for image and video restoration. Multiscale Model. Simul. 2008, 7, 214–241. [Google Scholar] [CrossRef]
  23. Sulam, J.; Elad, M. Large inpainting of face images with Trainlets. IEEE Signal Process. Lett. 2016, 23, 1839–1843. [Google Scholar] [CrossRef]
  24. Li, S.; Cao, Q.; Chen, Y.; Hu, Y.; Luo, L.; Toumoulin, C. Dictionary learning based sinogram inpainting for CT sparse reconstruction. Optik 2014, 125, 2862–2867. [Google Scholar] [CrossRef]
  25. Jafari, M.G.; Plumbley, M.D. Fast dictionary learning for sparse representations of speech signals. IEEE J. Sel. Top. Signal Process. 2011, 5, 1025–1031. [Google Scholar] [CrossRef]
  26. Dumitrescu, B.; Giurcăneanu, C.D. Adaptive-size dictionary learning using Information Theoretic Criteria. Algorithms 2019, 12, 178. [Google Scholar] [CrossRef]
  27. Rusu, C.; Dumitrescu, B. Stagewise K-SVD to design efficient dictionaries for sparse representations. IEEE Signal Process. Lett. 2012, 19, 631–634. [Google Scholar] [CrossRef]
  28. Mazhar, R.; Gader, P.D. EK-SVD: Optimized Dictionary Design for Sparse Representations. In Proceedings of the 19th International Conference on Pattern Recognition, Tampa, FL, USA, 8–11 December 2008; pp. 1–4. [Google Scholar] [CrossRef]
  29. Zhou, M.; Chen, H.; Paisley, J.; Ren, L.; Li, L.; Xing, Z.; Dunson, D.; Sapiro, G.; Carin, L. Nonparametric Bayesian dictionary learning for analysis of noisy and incomplete images. IEEE Trans. Image Process. 2012, 21, 130–144. [Google Scholar] [CrossRef] [PubMed]
  30. Dang, H.P.; Chainais, P. Towards Dictionaries of Optimal Size: A Bayesian Non Parametric Approach. J. Signal Process. Syst. 2018, 90, 221–232. [Google Scholar] [CrossRef]
  31. Zheng, X.; Dumitrescu, B.; Liu, J.; Giurcăneanu, C.D. Dictionary learning for signals in additive noise with generalized Gaussian distribution. Signal Process. 2022, 195, 108488. [Google Scholar] [CrossRef]
  32. Pati, Y.; Rezaiifar, R.; Krishnaprasad, P. Orthogonal matching pursuit: Recursive function approximation with applications to wavelet decomposition. In Proceedings of the 27th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 1–3 November 1993; Volume 1, pp. 40–44. [Google Scholar]
  33. Zeng, W.J.; So, H.; Jiang, X. Outlier-Robust Greedy Pursuit Algorithms in p-Space for Sparse Approximation. IEEE Trans. Signal Process. 2016, 64, 60–75. [Google Scholar] [CrossRef]
  34. Schwarz, G. Estimating the Dimension of a Model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  35. Chen, J.; Chen, Z. Extended Bayesian Information Criteria for Model Selection with Large Model Spaces. Biometrika 2008, 95, 759–771. [Google Scholar] [CrossRef]
  36. Alajaji, F.; Fuja, T. A Communication Channel Modeled on Contagion. IEEE Trans. Inf. Theory 1994, 40, 2035–2041. [Google Scholar] [CrossRef]
  37. Alajaji, F.; Phamdo, N.; Farvardin, N.; Fuja, T.E. Detection of binary Markov sources over channels with additive Markov noise. IEEE Trans. Inf. Theory 1996, 42, 230–239. [Google Scholar] [CrossRef]
  38. Phamdo, N.; Alajaji, F.; Farvardin, N. Quantization of Memoryless and Gauss-Markov Sources Over Binary Markov Channels. IEEE Trans. Commun. 1997, 45, 668–675. [Google Scholar] [CrossRef]
  39. Iordache, R.; Tabus, I.; Astola, J. Robust Index Assignment Using Hadamard Transform for Vector Quantization Transmission over Finite-Memory Contagion Channels. Circuits Syst. Signal Process. 2002, 21, 485–509. [Google Scholar] [CrossRef]
  40. Banerjee, A.; Burlina, P.; Alajaji, F. Image Segmentation and Labeling Using the Polya Urn Model. IEEE Trans. Image Process. 1999, 8, 1243–1253. [Google Scholar] [PubMed]
  41. Hayhoe, M.; Alajaji, F.; Gharesifard, B. A Polya Contagion Model for Networks. IEEE Trans. Control Netw. Syst. 2018, 5, 1998–2010. [Google Scholar] [CrossRef]
  42. Moritz, S.; Bartz-Beielstein, T. imputeTS: Time Series Missing Value Imputation in R. R J. 2017, 9, 207–218. [Google Scholar] [CrossRef]
  43. Moritz, S.; Gatwa, S. Package: ImputeTS, Version 3.0. 2019. Available online: https://cran.r-project.org/web/packages/imputeTS/index.html (accessed on 16 June 2022).
  44. Mantegna, R. Hierarchical structure in financial markets. Eur. Phys. J. B 1999, 11, 193–197. [Google Scholar] [CrossRef]
Figure 1. Scores for various imputation methods applied to the Climate time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM algorithm is used.
Figure 1. Scores for various imputation methods applied to the Climate time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM algorithm is used.
Entropy 24 01057 g001
Figure 2. Scores for various imputation methods applied to the Climate time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM 1 algorithm is used.
Figure 2. Scores for various imputation methods applied to the Climate time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM 1 algorithm is used.
Entropy 24 01057 g002
Figure 3. Scores for various imputation methods applied to the MeteoSwiss time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM algorithm is used.
Figure 3. Scores for various imputation methods applied to the MeteoSwiss time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM algorithm is used.
Entropy 24 01057 g003
Figure 4. Scores for various imputation methods applied to the MeteoSwiss time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM 1 algorithm is used.
Figure 4. Scores for various imputation methods applied to the MeteoSwiss time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM 1 algorithm is used.
Entropy 24 01057 g004
Figure 5. Scores for various imputation methods applied to the BAFU time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel).
Figure 5. Scores for various imputation methods applied to the BAFU time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel).
Entropy 24 01057 g005
Figure 6. Scores for various imputation methods applied to the Temperature time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel).
Figure 6. Scores for various imputation methods applied to the Temperature time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel).
Entropy 24 01057 g006
Figure 7. Scores for various imputation methods applied to the Air time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM algorithm is used.
Figure 7. Scores for various imputation methods applied to the Air time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM algorithm is used.
Entropy 24 01057 g007
Figure 8. Scores for various imputation methods applied to the Air time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM 1 algorithm is used.
Figure 8. Scores for various imputation methods applied to the Air time series: The missing data are simulated by sampling without replacement (left panel) and by using the Polya model (right panel). Note that the DLM 1 algorithm is used.
Entropy 24 01057 g008
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zheng, X.; Dumitrescu, B.; Liu, J.; Giurcăneanu, C.D. Multivariate Time Series Imputation: An Approach Based on Dictionary Learning. Entropy 2022, 24, 1057. https://doi.org/10.3390/e24081057

AMA Style

Zheng X, Dumitrescu B, Liu J, Giurcăneanu CD. Multivariate Time Series Imputation: An Approach Based on Dictionary Learning. Entropy. 2022; 24(8):1057. https://doi.org/10.3390/e24081057

Chicago/Turabian Style

Zheng, Xiaomeng, Bogdan Dumitrescu, Jiamou Liu, and Ciprian Doru Giurcăneanu. 2022. "Multivariate Time Series Imputation: An Approach Based on Dictionary Learning" Entropy 24, no. 8: 1057. https://doi.org/10.3390/e24081057

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop