Next Article in Journal
A Mathematical Model for Dynamic Electric Vehicles: Analysis and Optimization
Previous Article in Journal
Mathematical Modeling of Two Interacting Populations’ Dynamics of Onchocerciasis Disease Spread with Nonlinear Incidence Functions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Truncated Normal Tensor Completion Method for Multi-Source Fusion Data

1
School of Computer Science and Engineering, Northwestern Polytechnical University, Xi’an 710072, China
2
School of Equipment Management and Unmanned Aerial Vehicle Engineering, Air Force Engineering University, Xi’an 710051, China
Mathematics 2024, 12(2), 223; https://doi.org/10.3390/math12020223
Submission received: 4 December 2023 / Revised: 4 January 2024 / Accepted: 7 January 2024 / Published: 9 January 2024
(This article belongs to the Section Computational and Applied Mathematics)

Abstract

:
Completing traffic data is a basic requirement for intelligent transportation systems. However, completing spatiotemporal traffic data poses a significant challenge, especially for high-dimensional data with complex missing mechanisms. Various completion methods targeting different missing mechanisms have showcased the superiority of tensor learning by effectively characterizing intricate spatiotemporal correlations. In this study, a novel tensor completion framework, known as the multi-source tensor completion method for data fusion, is proposed. This framework incorporates passenger transfer relationships between buses and subways into subway data completion, enhancing the data completion accuracy. Moreover, by combining bus transfer passenger flow data with other data dimensions, such as the different road sections, time intervals, and days, an innovative 4D low-rank tensor completion data framework was obtained. In addition, to boost the completion accuracy, a truncated l 2 , p norm optimization model was derived. This model enhances the non-convex performance of the objective function throughout the tensor completion process. The results highlight the superiority of the proposed completion method, leveraging fused subway/bus data over other completion methods that rely solely on subway data.

1. Introduction

In data-driven applications, robust and accurate data completion is of great importance [1]. Data fusion, achieved by combining data from multiple sources, can decrease the uncertainty of results [2,3,4,5]. For instance, Roy et al. improved vehicle tracking and detection performance under non-line-of-sight (NLOS) image and non-image modalities by combining multiple state-of-the-art fusion strategies [6]. Moreover, Senel et al. fused data on the object list level from distributed vehicle sensors using multiple sensors cooperatively [7]. In addition, Chen et al. fused traffic flow, occupancy, and recreational vehicle speed data and dynamically trained the fused data using a dynamic backpropagation (PB) fusion method [8]. Moreover, Lin et al. improved road traffic speed prediction by fusing traditional speed sensor data with cross-domain novel “sensor” data [9]. As for Zißner et al., they introduced the data fusion on intelligent traffic system (DataFITS), which collected four data types from seven sources over nine months and fused them in a spatiotemporal domain, leading to significant improvements in road coverage [10]. The results showed that DataFITS significantly increased the road coverage by 137% and improved the road information quality by up to 40%. Moreover, these studies demonstrated that the multi-source data could provide more information from different dimensions, thereby enhancing the quality of the data analysis.
Considering that traffic data include various multi-source and heterogeneous data, more data-fusion methods are required. Currently, classical data-fusion methods include Kalman filtering, Bayesian networks, Dempster–Shafer (D–S) evidence theory, neural networks (NNs), fuzzy theory, and voting methods. These techniques have already been applied in vehicle positioning, traffic event recognition, traffic event prediction, and traffic guidance in the transport sector. Referring to existing studies, Kalman filtering and D–S evidence theory are more commonly used. In recent years, with the advancement of artificial intelligence (AI), some machine learning methods, such as recurrent neural networks, convolutional neural networks, and graph neural networks, have also been widely applied in traffic analysis, achieving high success rates in predicting traffic flow [11].
For instance, Ji et al. encoded each feature in the spatial and temporal dimensions and considered data directly related to flow as the inputs for a graph convolutional network (GCN) [12]. Furthermore, Satish et al. proposed a multidimensional k-means++ data-fusion method, merging attribute values into a single or very few dimensions [13]. In addition, Zhang et al. introduced a data-fusion method based on Bayesian fusion rules [14]. Considering the relationships between multiple traffic factors from a system perspective, this method fuses traffic speeds from different data sources, based on prior probabilities, while using the high-order multivariate Markov model. Nevertheless, the long training time of NNs, the model-building complexity, and the dimensionality issues are still difficulties existing within fusion methods [15].
Moreover, a tensor serves as a natural mathematical representation of high-dimensional data. Drawing inspiration from compressive sensing theory, which considers sparse structures in tensors, accurate recovery of unknown information can be achieved in a small number of observations if the target tensor exhibits a low-rank structure or can be well approximated by a low-rank tensor. Recently, the tensor completion theory has shown promising applications in traffic data completion, serving as a theoretical foundation for the accurate achievement of multidimensional, multimodal, and multi-structured traffic data.
Recently, low-dimensional fusion of multimodal data has attracted great attention. Tensor-based decomposition is able to map multimodal data into a unified low-dimensional subspace, enabling clustering, classification, and other data-fusion analyses within that subspace [16]. For instance, Pan et al. utilized the tensor nuclear norm, derived from tensor singular value decomposition (t-SVD), to optimally leverage high-dimensional information, thereby improving the precision of multi-view data processing [17]. Shen et al. introduced a comprehensive method for analyzing and processing multi-source traffic data [18]. Their method incorporated data-analysis techniques rooted in spatiotemporal regression models and data-fusion methods based on confidence tensors and evidence theory. Moreover, Xing et al. established a novel tensor decomposition data-fusion framework, demonstrating that the method combining license-plate recognition (LPR) data and cellphone location (CL) data yielded significantly better results compared with the interpolation methods using only LPR data [19]. In addition, Respati et al. proposed an adaptive weight-based fusion technique (ABAFT), deploying data space coverage and quality or confidence as factors for building weights [20]. The completion methods for the different missing mechanisms demonstrated the superiority of the tensor learning method by effectively characterizing complicated spatiotemporal correlations [21]. Therefore, in this study, based on the tensor representation of high-dimensional data, subway data and bus data were fused to achieve efficient multi-source data completion.
In summary, the contributions of this study are as follows:
(1)
Subway inbound passenger flow and bus transfer time-series data were constructed, and the completion performance of single-source data containing only subway data was compared to that of multi-source data containing subway and bus transfer data;
(2)
A truncated l 2 , p norm optimization model was derived to enhance the non-convex performance of the objective function during the tensor completion process;
(3)
A multi-source data-fusion tensor completion framework was established, and ADMM was applied to achieve an effective completion algorithm.

2. Characteristics of Multi-Source Traffic Data

In practical transportation, complex traffic data from distinct sources and with different structures are present. For example, data gathered from road traffic speed, subway passenger flow, bus passenger flow, shared bike-riding, and cellphones are combined together. The data structure varies significantly among the different types of transportation,. Even for the same type of transportation, such as buses, the data structure differs due to varying data sources and collection focuses (refer to Table 1). Given that the transportation system is a complex and dynamically interactive network, additional data fusion is required to effectively capture global interactions within the system.

3. Low-Rank Completion Model of Truncated l 2 , p

3.1. Truncated l 2 , p Norm (P2TN)

As an alternative approach to solving rank-minimization problems, the truncated l 2 , p norm incorporates the advantages of the truncated nuclear norm and the l 2 , p norm. Moreover, it is more compact than a single norm during the approximation process of rank-minimization problems. Therefore, the truncated l 2 , p norm preserves the complete features of large singular values in a truncated manner and applies l 2 , p contraction only to the singular values of min ( n 1 , n 2 ) r after achieving the truncation. During the contraction, smaller singular values incur larger threshold penalties, ensuring the retention of a sufficient amount of internal strong correlation information to the maximum extent. Thus, the l 2 , p norm is defined as follows [22]:
Given a matrix X m × n , for any p ,
X p 2 = 1 2 ( i = 1 min ( m , n ) | σ i | p 2 2 )
where . p 2 refers to the l 2 , p norm of the matrix. Assuming that the singular value decomposition of matrix X is U Σ V T , then one can write Σ = D i a g ( σ i ( X ) ) , where σ i represents the singular value of the matrix, i.e., the nuclear norm of the matrix. Moreover, the singular values are ranked in descending order as σ 1 σ 2 σ min { m , n } 0 , and p indicates the contraction parameter.
Then, the l 2 , p norm is expanded to the tensor mode. For a given X n 1 × n 2 × n 3 , l 2 , p can be expressed as follows [23]:
X p 2 = 1 n 3 k = 1 n 3 X ¯ ( i ) p 2 = 1 n 3 k = 1 n 3 i = 1 min { n 1 , n 2 } | σ ¯ i | p 2 2
According to the tensor singular value decomposition principle, a Fourier transform is performed for X along the third dimension, n 3 . In more detail, X ¯ ( i ) refers to the forward slice after transformation in matrix form, while | σ ¯ i | indicates the singular value obtained by singular value decomposition for the forward slice X ¯ ( i ) , e.g., the tensor nuclear norm.
The truncated nuclear norm of the tensor is defined as follows [24]:
Given a third-order tensor X n 1 × n 2 × n 3 and a positive integer t = min ( n 1 , n 2 ) , the truncated nuclear norm represents the sum of the minimum singular values of the tensor:
X r , = 1 n 3 X ¯ r , = 1 n 3 j = 1 n 3 i = r + 1 t σ i ( X ¯ ( j ) ) = 1 n 3 j = 1 n 3 i = 1 t σ i ( X ¯ ( j ) ) 1 n 3 j = 1 n 3 i = 1 r σ i ( X ¯ ( j ) ) = i = 1 t σ i ( X ) i = 1 r σ i ( X )
where . r , refers to the truncated nuclear norm of the tensor, σ i ( X ) refers to the i-th singular value of X , and the following constraint is to be respected: r < min ( n 1 , n 2 ) .
Combining both tensor norm definitions, the truncated l 2 , p norm (P2TNN) of the tensor is defined as follows:
X p , r 2 = 1 n 3 j = 1 n 3 i = r + 1 t | σ ¯ i | p 2 2
where . p , r 2 refers to the truncated l 2 , p norm, and in the tensor mode, for any given X ,   Y n 1 × n 2 × n 3 , μ > 0 , p , and r < min ( n 1 , n 2 ) .
In the context of a given parameter τ , for a tensor X n 1 × n 2 × n 3 , the tensor singular value decomposition (t-SVD) can be expressed as X = U S V . The singular value threshold for the tensor X is consequently defined.
D τ ( A ) = U S τ V
The contraction threshold operator is defined for the truncated l 2 , p as D p , r μ ( . ) . Then, the following equation is obtained:
D p , r μ ( Y ) = arg min X n 1 × n 2 × n 3 X p , r 2 + 1 2 μ X Y F 2
According to the t-SVD theorem, D p , r μ ( Y ) represents the solution to the X optimization problem, μ denotes the threshold parameter for singular value thresholding decomposition, and p indicates the contraction parameter. When < p < 1 , the larger the truncated l 2 , p contraction threshold, the smaller the received penalty. Moreover, . F refers to the Frobenius norm. Regarding D p , r μ ( Y ) , we have D p , r μ ( Y ) = U S r ( Σ ) V T ; therefore, one can write
( D p , r μ ( Σ ) ) i j k = D p , r μ ( Σ i j k ) = D i a g ( ( σ 1 , , σ r , [ σ r + 1 μ | σ r + 1 | p 2 2 ] + , , [ σ min { m , n } μ | σ min { m , n } | p 2 2 ] + ) )
For any value a , the addition function ( . ) + is defined as a + = max { a , 0 } .

3.2. LRTC-P2TN Model

The NWLRTC model [25] was further optimized using the l 2 , p norm to develop the LRTC-P2TN model:
min X 0 , 0 , Z 0 , W 0 : α 1 p , r 2 + α 2 Z p , r 2 + α 3 W p , r 2 s . t . : P Ω ( X i ) = P Ω ( M i ) X 1 = B , X 2 = Z , X 3 = W
where n 1 × n 2 × n 3 , Z n 2 × n 3 × n 1 , and W n 3 × n 1 × n 2 refer to the spatiotemporal traffic data input in three different directions, whereas the truncated nuclear norm of the tensor is introduced to avoid dependencies on the input direction. Moreover, α 1 , α 2 , and α 3 represent the weights of each direction.
Equation (8) denotes a convex optimization problem that incorporates equality constraints. To convert this problem into one without such constraints, the augmented Lagrangian function is employed. The definition of the augmented Lagrangian function, which takes into account only the X 1 = ,   X 2 = Z ,   X 3 = W constraint conditions, is formulated as follows:
( X , , Z , W , Y ) = α 1 p , r 2 + α 2 Z p , r 2 + α 3 W p , r 2 + < Y 1 , X 1 > + ρ 2 X 1 F 2 + < Y 2 , X 2 Z > + ρ 2 X 2 Z F 2 + < Y 3 , X 3 W > + ρ 2 X 3 W F 2
where Y 1 n 1 × n 2 × n 3 , Y 2 n 2 × n 3 × n 1 , Y 3 n 3 × n 1 × n 2 are the defined Lagrange multipliers and ρ is the penalty parameter. The Frobenius norm is an added regularization term that can enhance the convergence of the function.
The Frobenius norm of tensor X n 1 × n 2 × n 3 is defined as follows:
X F 2 = 1 2 ( k = 1 n 3 i = 1 n 2 j = 1 n 1 | x i , j | 2 ) 1 2
Equation (9) is addressed using the alternating direction method of multipliers (ADMM), an algorithm that simplifies multiparameter optimization challenges by reducing them to problems concerning a single optimization variable. By leveraging the ADMM approach, the objective function of the primary problem can be divided into a set of manageable sub-problems, which may then be simultaneously solved. The resolution of these sub-problems subsequently converges through coordination, yielding the comprehensive solution to the initial problem.
When solving the model via the ADMM algorithm, it updates one variable at a time, maintaining the others as constants. Specifically, for Equation (9), it sequentially employs gradient descent to update the variables , Z , W , X , and Y in an alternating fashion.
Step 1: Updating l + 1 . Keeping X i l , Y i l , Z l , and W l fixed, let = 0 ; the update is as follows:
l + 1 : = arg min ( X i l , , Z l , W l , Y 1 l ) = α 1 p , r 2 + < Y 1 l , X 1 l > + ρ 2 X 1 F 2 = α 1 p , r 2 + ρ 2 < X 1 l + 1 ρ Y 1 l > F 2 = D α 1 ρ , r , p ( X 1 l + 1 ρ Y 1 l )
According to the definition of tensor singular value thresholding, it can be inferred that D α i ρ , r , * ( X i l + 1 ρ Y i l ) is the threshold applied to truncate the tensor ( X i l + 1 ρ Y i l ) after performing singular value decomposition, with the threshold τ = α i ρ .
Step 2: Updating Z l + 1 . Keeping X i l , Y i l , l + 1 , and W l fixed, let Z = 0 ; the update is as follows:
Z l + 1 = arg min Z ( X l , l + 1 , Z , W l , Y 2 l ) = α 2 Z p , r 2 + < Y 2 l , X 2 l Z > + ρ 2 X 2 l Z F 2 = α 2 Z p , r 2 + ρ 2 Z < X 2 l + 1 ρ Y 2 l > F 2 = D α 2 ρ , r , p ( X 2 l + 1 ρ Y 2 l )
Step 3: Updating W l + 1 . Keeping X i l , Y i l , l + 1 , and Z l + 1 fixed, let W = 0 ; the update is as follows:
W l + 1 = arg min W ( X 3 l , l + 1 , Z l + 1 , W , Y 3 l ) = α 3 W p , r 2 + < Y 3 l , X 3 l W > + ρ 2 X 3 l W F 2 = α 3 W p , r 2 + ρ 2 W < X 3 l + 1 ρ Y 3 l > F 2 = D α 3 ρ , p , r ( X 3 l + 1 ρ Y l )
Step 4: Updating X i l + 1 . Keeping W l + 1 , Y i l , l + 1 , and Z l + 1 fixed, let X = 0 ; the update is as follows:
X i l + 1 : = arg min X ( X i , l + 1 , Z l + 1 , W l + 1 , Y i l ) = < Y 1 l , X 1 l + 1 > + ρ 2 X 1 l + 1 F 2 + < Y , X 2 Z l + 1 > + ρ 2 X 2 Z l + 1 F 2 + < Y 3 l , X 3 W l + 1 > + ρ 2 X 3 W l + 1 F 2 = i = 1 3 ( ( l + 1 + Z l + 1 + W l + 1 ) 1 ρ Y i l )
Moreover, the remaining constraint P Ω ( X i ) = P Ω ( M i ) is added. Thus, X i l + 1 can be expressed as follows:
X i l + 1 = i = 1 3 ( ( l + 1 + Z l + 1 + W l + 1 ) 1 ρ Y i l ) s . t . : P Ω ( X i ) = P Ω ( M i )
Consequently, the constraint is removed, and elements at the positions of the observation points in the completion tensor are replaced by elements at the positions of the original observation points:
X i l + 1 = i = 1 3 ( ( l + 1 + Z l + 1 + W l + 1 ) 1 ρ Y i l ) Ω ¯ + P Ω ( M )
Step 4: Updating Y i l + 1 . Keeping W l + 1 , X i l + 1 , l + 1 , and Z l + 1 fixed, the value Y i l + 1 is updated based on the following process:
Y i l + 1 : = Y i l + ρ a 1 ( l + 1 X 1 l + 1 ) + a 2 ( Z l + 1 X 2 l + 1 ) + a 3 ( W l + 1 X 3 l + 1 )
The number of iterations is set and updated for each iteration according to the above steps. Moreover, the optimal solution for the model is computed. The steps of the LRTC- P2TN algorithm are defined as follows (Algorithm 1):
Algorithm 1 LRTC-P2TN
Input: X ≥ 0, A, P Ω ( M i ) = P Ω ( X i ) , Y i = 0 , ρ , k , λ
Output: the recovered tensor X
Calculate W using Formula (1)
for  i = 1  to  k
  Update l + 1 using Formula (11)
  Update Z l + 1 using Formula (12)
  Update W l + 1 using Formula (13)
  Update X i l + 1 using Formula (16)
  Update Y i l + 1 using Formula (17)
    If  X l + 1 X l F 2 P Ω ( X ) F 2 < then break
end for
During the construction and solution process of the entire mathematical model, the influence of noise was not considered. In more detail, the LRTC-P2TNN algorithm only considered the reconstruction accuracy of the given data.

3.3. Multi-Source Data-Fusion Tensor Completion Framework

The core of the multi-source data-fusion completion framework represents the LRTC-P2TN module. As depicted in Figure 1, the correlated data, gathered from different sources, were represented as tensors and were set with different deletion rates. After completion with the LRTC-P2TN module, the final completion data were acquired. The truncated l 2 , p norm integrated into the model played a crucial role in ensuring the convergence of the algorithm during the completion process.

4. Results and Discussion

4.1. Data Preparation

The experimental data included two datasets: one set of Shenzhen Tong data, referred to as SZ, and one set of Beijing Transportation Smart Card data, known as BJ.
The Shenzhen Tong dataset contained Shenzhen Tong Smart Card data for 22 October 2013, covering data gathered from 117 bus and subway stations. The methodology included both single-source and multi-source data completion techniques. Single-source data adopted a three-dimensional array formulated as day × time interval × subway station, whereas multi-source data extended to a four-dimensional array featuring passenger flow type × day × time interval × station. Analysis of bus flows at subway stations leveraged user ID sequencing to unravel transfer dynamics, with subway flows exclusively accounting for entry volumes, tallied at five-minute intervals. Shenzhen’s data footprint for a single day manifested as a three-dimensional array (1 × 288 × 117) and a four-dimensional array (2 × 1 × 288 × 117) for single-source and multi-source contexts, respectively. By examining RMSE and MAPE metrics, we evaluated the augmentation of subway data completion accuracy via the incorporation of bus data.
Conversely, the BJ dataset encompassed entries from 239 stations, spanning 1 and 10 May 2019. It mirrored the SZ data’s analytical approach, yielding a four-dimensional structure of day × passenger flow type × time interval × station, quantified in five-minute intervals. The resulting dataset, encapsulating a ten-day span, consisted of a sizably expansive array at 2 × 10 × 288 × 239.
The time-series datasets that we analyzed and constructed can be downloaded from the website https://github.com/yongmeizhao/smartcard-data/blob/main/data.rar, accessed on 2 December 2023.

4.2. Single-Source Data Completion

During this experiment, we used Beijing subway data for 7 May 2019 and Shenzhen smart card data for 22 October 2013. Figure 2 and Figure 3 illustrate the visualization results of both datasets, where the horizontal axis represents the station and the vertical axis is the passenger flow at each time interval. Based on these figures, it can be observed that the Beijing subway data for 7 May showed a more obvious periodicity, with clearer peaks in the morning and evening.
Moreover, the single-source data experiments were considered for subway flow data. The experimental data used three-dimensional data consisting of station × time interval × subway flow (data can be downloaded from the address). The truncation value was set to 4.5, and ρ represents the iteration step size. The initial iteration step size ρ was set to 1.15, and the maximum value ρ max of ρ was equal to 1 × 10−5. In each iteration, ρ was updated through the equation ρ = min { 1.05 ρ , ρ max } , and the number of iterations was set to 150. For convenience of recording and accuracy, the MAPE values in the table below were multiplied by ten. Moreover, Table 2 illustrates the completion results under deletion rates of 10%, 30%, 50%, and 70%, using LRTC-P2TN.

4.3. Multi-Source Data Completion

In this study, we used the Beijing subway and bus data for 7 May 2019 and Shenzhen smart card data for 22 October 2013. The experimental data were constructed as a four-dimensional tensor dataset of subway station name, time, subway station passenger flow, and bus station passenger flow. For missing data processing, deletion rates were only set for the subway flow data, while the bus flow data were complete. The experimental parameters were the same as those specified for the single-source data completion.
According to Table 2 and Table 3, the completion performance with the fusion of multi-source data of buses led to better behavior compared to the single-source subway data completion.
Furthermore, Figure 4 and Figure 5 display the data completion effects of single- and multi-source data on Route No. 7 at a deletion rate of 0.5. As indicated, the fusion of bus transfer data enhanced the algorithm’s capability to capture the overall trend of the data.

4.4. Comparison of the LRTC-P2TN Model with Other Models

The proposed LRTC-P2TN model was compared to the LATC [26], BGCP [27], NWLRTC [25], and BTTF [28] models. In more detail, LATC is an autoregressive low-rank tensor completion algorithm, whereas BGCP is a Bayesian tensor decomposition model that learns latent factor matrices (i.e., low-rank structures) using the Markov chain Monte Carlo method. Moreover, the Bayesian temporal tensor factorization (BTTF) method merges low-rank matrix/tensor decomposition with vector autoregressive (VAR) processes within a probabilistic graphical model framework to facilitate problem-solving. Finally, NWLRTC is a non-negative multidirectional weighted low-rank tensor completion method.
On the BJ dataset, each baseline algorithm was executed. The deletion rate was set to 20%, 30%, 40%, 50%, 60%, 70%, and 80%; ρ represents the iteration step size, which was initially set to 1 × 10−5 and updated by ρ = min { 1.05 ρ , ρ max } , while is the convergence accuracy, which was initially set to 1. The experimental results showed that, under random deficiency, setting ρ = 1 and = 1 could achieve higher convergence speed and completion accuracy. Moreover, Figure 6 shows the RMSE values for the different algorithms at various deletion rates under random deficiency. A comparison of the computed results indicates that, compared to the LATC, BGCP, NWLRTC, and BTTF models, the LRTC-P2TN model achieved better imputation accuracy at deletion rates between 20% and 60%. Furthermore, at deletion rates of 70% and 80%, its MAPE values surpassed those of the other algorithms, although the RMSE values yielded higher values (as depicted in Table 4). Additionally, the LRTC-P2TN model achieved higher completion accuracy than LATC, BGCP, NWLRTC, and BTTF at deletion rates of 20–60%; when the deletion rates were 70% and 80%, our model achieved higher MAPE values than the other algorithms, but its RMSE values were relatively higher (see Table 4).
Table 4 displays the RMSE and MAPE values for a range of algorithms using the BJ data with an input orientation of 1 × 288 × 239. A minimal variation can be observed in the RMSE and MAPE metrics for the LATC, BGCP, NWLRTC, and BTTF algorithms when the input orientation shifted to 239 × 288 × 1. The incorporation of directional weighting factors in the LRTC-P2TN and NWLRTC algorithms resulted in negligible changes to their RMSE and MAPE values.

5. Conclusions

Multi-source data can assist in the completion of single-source data, especially when the single-source data have a lot of missing values where correlated data could effectively contribute to data completion. In this study, a four-dimensional tensor dataset of subway and bus transfer was constructed, and experiments were conducted to demonstrate the fusion of multi-source bus transfer data that could achieve higher completion accuracy than single-source subway data alone. Since the tensor is a natural representation of high-dimensional data, tensor-based multi-source data fusion has the advantages of simple data representation and high completion accuracy. Low-rank tensor completion uses the truncated l 2 , p norm as an alternative approach for rank minimization, integrating the advantages of the truncated nuclear norm and the l 2 , p norm; therefore, it achieves more compact results than a single norm during the approximation process of rank minimization.

Funding

This work was supported by the National Natural Science Foundation of China (No. 6200238).

Data Availability Statement

(1) The numerical data analyzed during the present study are available at: https://github.com/yongmeizhao/smartcard-data/blob/main/data.rar (accessed on 2 December 2023). (2) The datasets generated during this study are included in this published article.

Conflicts of Interest

The author declares no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Liang, J.; Härdle, W.K.; Tian, M. Imputed quantile tensor regression for near-sited spatial-temporal data. Comput. Stat. Data Anal. 2023, 182, 107713. [Google Scholar] [CrossRef]
  2. Chen, G.; Liu, Z.; Yu, G.; Liang, J. A new view of multisensor data fusion: Research on generalized fusion. Math. Probl. Eng. 2021, 2021, 5471242. [Google Scholar] [CrossRef]
  3. Mitchell, H.B. Data Fusion: Concepts and Ideas; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  4. Ma, X.; Ye, Q.; Yan, H. L2P-norm distance twin support vector machine. IEEE Access 2017, 5, 23473–23483. [Google Scholar] [CrossRef]
  5. Muravyov, S.V.; Khudonogova, L.I.; Dai Ho, M. Analysis of heteroscedastic measurement data by the self-refining method of interval fusion with preference aggregation–IF&PA. Measurement 2021, 183, 109851. [Google Scholar] [CrossRef]
  6. Roy, D.; Li, Y.; Jian, T.; Ioannidis, S. Multi-modality sensing and data fusion for multi-vehicle detection. IEEE Trans. Multimedia 2023, 25, 2280–2295. [Google Scholar] [CrossRef]
  7. Senel, N.; Kefferpütz, K.; Doycheva, K.; Elger, G. Multi-Sensor Data Fusion for Real-Time Multi-Object Tracking. Processes 2023, 11, 501. [Google Scholar] [CrossRef]
  8. Chen, R.; Ning, J.; Lei, Y.; Cheng, N. Mixed Traffic Flow State Detection: A Connected Vehicles Assisted Roadside Radar and Video Data Fusion Scheme. IEEE Open J. Intell. Transp. Syst. 2023, 4, 360–371. [Google Scholar] [CrossRef]
  9. Lin, L.; Li, J.; Chen, F.; Ye, J.; Huai, J. Road traffic speed prediction: A probabilistic model fusing multi-source data. IEEE Trans. Knowl. Data Eng. 2017, 30, 1310–1323. [Google Scholar] [CrossRef]
  10. Zißner, P.; Rettore, P.H.; Santos, B.P.; Loevenich, J.F.; Lopes, R.R.F. DataFITS: A Heterogeneous Data Fusion Framework for Traffic and Incident Prediction. IEEE Trans. Intell. Transp. Syst. 2023, 24, 11466–11478. [Google Scholar] [CrossRef]
  11. Wang, P.; Hao, W.; Jin, Y. Fine-grained traffic flow prediction of various vehicle types via fusion of multisource data and deep learning approaches. IEEE Trans. Intell. Transp. Syst. 2020, 22, 6921–6930. [Google Scholar] [CrossRef]
  12. Ji, B.; Chen, Z.; Mumtaz, S.; Liu, J.; Zhang, Y.; Zhu, J.; Li, C. SWIPT enabled intelligent transportation systems with advanced sensing fusion. IEEE Sens. J. 2020, 21, 15643–15650. [Google Scholar] [CrossRef]
  13. Satish, V.; Kumar, P.A.R. Variance based data fusion for k-means++. In Proceedings of the 2017 2nd International Conference for Convergence in Technology (I2CT), Mumbai, India, 7–9 April 2017. [Google Scholar]
  14. Zhang, W.; Qi, Y.; Zhou, Z.; Biancardo, S.A.; Wang, Y.H. Method of Speed Data Fusion Based on Bayesian Combination Algorithm and Markov Model. IET Intell. Transp. syst. 2018, 12, 1312–1321. [Google Scholar] [CrossRef]
  15. Sofuoglu, S.E.; Aviyente, S. Graph Regularized Tensor Train Decomposition. In Proceedings of the ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3912–3916. [Google Scholar]
  16. Zhao, L. Research on Multimodal Data Fusion Algorithms; Dalian University of Technology: Dalian, China, 2018. [Google Scholar]
  17. Pan, B.; Li, C.; Che, H.; Leung, M.F.; Yu, K. Low-Rank Tensor Regularized Graph Fuzzy Learning for Multi-View Data Processing. IEEE Trans. Consum. Electron. 2023, 1. [Google Scholar] [CrossRef]
  18. Shen, G.; Han, X.; Zhou, J.; Ruan, Z.; Pan, Q. Research on Intelligent Analysis and depth Fusion of Multi-source Traffic data. IEEE Access 2018, 6, 59329–59335. [Google Scholar] [CrossRef]
  19. Xing, J.; Liu, R.; Anish, K.; Liu, Z. A Customized Data Fusion Tensor Approach for Interval-Wise Missing Network Volume Imputation. IEEE Trans. Intell. Transp. Syst. 2023, 24, 12107–12122. [Google Scholar] [CrossRef]
  20. Respati, S.; Chung, E.; Zheng, Z.; Bhaskar, A. ABAFT: An adaptive weight-based fusion technique for travel time estimation using multi-source data with different confidence and spatial coverage. J. Intell. Transp. Syst. 2023, 1–14. [Google Scholar] [CrossRef]
  21. Liu, Y.X. Research on Urban Public Transit System Passenger Travel Pattern Mining and Its Application Based on Multi-Source Data Fusion; South China University of Technology: Guangzhou, China, 2018. [Google Scholar]
  22. Li, Y.F.; Shang, K.; Huang, Z.H. A singular value p-shrinkage thresholding algorithm for low rank matrix recovery. Comput. Optim. Appl. 2019, 73, 453–476. [Google Scholar] [CrossRef]
  23. Liu, C.; Shan, H.; Chen, C. Tensor p-shrinkage nuclear norm for low-rank tensor completion. Neurocomputing 2020, 387, 255–267. [Google Scholar] [CrossRef]
  24. Zhang, D.; Hu, Y.; Ye, J.; Li, X. Matrix completion by Truncated Nuclear Norm Regularization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2192–2199. [Google Scholar]
  25. Zhao, Y.M.; Tuo, M.; Zhang, H.; Zhang, H.; Wu, J.; Gao, F. Nonnegative low-rank tensor completion method for spatiotemporal traffic data. Multimed. Tools Appl. 2023, 1–16. [Google Scholar] [CrossRef]
  26. Chen, X.; Cheng, Z.; Saunier, N.; Sun, L. Laplacian convolutional representation for traffic time series imputation. arXiv 2022, arXiv:2212.01529. [Google Scholar]
  27. Chen, X.; He, Z.; Sun, L. A Bayesian tensor decomposition approach for spatiotemporal traffic data imputation. Transp. Res. Pt. C-Emerg. Technol. 2019, 98, 73–84. [Google Scholar] [CrossRef]
  28. Chen, X.; Sun, L. Bayesian Temporal Factorization for Multidimensional Time Series Prediction. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4659–4673. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Multi-source data-fusion tensor completion framework.
Figure 1. Multi-source data-fusion tensor completion framework.
Mathematics 12 00223 g001
Figure 2. Passenger flow data of the Shenzhen subway on 22 October 2013.
Figure 2. Passenger flow data of the Shenzhen subway on 22 October 2013.
Mathematics 12 00223 g002
Figure 3. Passenger flow data of the Beijing subway on 7 May 2019.
Figure 3. Passenger flow data of the Beijing subway on 7 May 2019.
Mathematics 12 00223 g003
Figure 4. Results of single-source data completion.
Figure 4. Results of single-source data completion.
Mathematics 12 00223 g004
Figure 5. Results of multi-source data completion.
Figure 5. Results of multi-source data completion.
Mathematics 12 00223 g005
Figure 6. RMSE of different models under different missing rates.
Figure 6. RMSE of different models under different missing rates.
Mathematics 12 00223 g006
Table 1. Structure of multi-source bus data [18].
Table 1. Structure of multi-source bus data [18].
Data SourceKey Fields
Passenger card swiping dataIC card no., card swiping time, license plate no., and route no.
Vehicle station announcement dataLicense plate no., route no., station name, station inbound time, and station outbound time
Vehicle coordinates and route tablesRoute name, direction, station list, station coordinates, and route trajectory
Vehicle dispatch recordsVehicle dispatch record, departure schedule, license plate no., departure time, and stops
Table 2. RMSE/MAPE of single-source data completion under random deficiency.
Table 2. RMSE/MAPE of single-source data completion under random deficiency.
10%30%50%70%
SZ30.52/37.0126.13/40.1632.81/62.0850.37/83.48
BJ3.25/5.753.25/5.823.31/5.903.93/6.05
Table 3. RMSE/MAPE of multi-source data completion under random deficiency.
Table 3. RMSE/MAPE of multi-source data completion under random deficiency.
10%30%50%70%
SZ22.74/32.2320.39/36.8325.12/47.7426.77/48.17
BJ3.15/5.763.20/5.733.28/5.873.86/6.02
Table 4. RMSE/MAPE of different models under random deficiency.
Table 4. RMSE/MAPE of different models under random deficiency.
Missing RateLATCBGCPBTTFNWLRTCLRTC-P2TN
20%3.47/6.503.15/5.533.22/5.553.17/5.733.15/5.76
30%3.48/6.693.31/5.653.35/5.553.34/5.843.18/5.63
40%3.56/6.683.28/5.553.49/5.673.25/5.873.17/5.78
50%3.64/6.923.33/5.723.36/5.743.41/5.953.19/5.83
60%3.64/7.033.41/5.823.48/5.963.47/6.103.26/5.85
70%3.77/7.323.51/6.133.64/6.173.86/6.053.78/5.99
80%3.79/7.313.93/6.934.12/6.955.29/5.965.06/5.79
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, Y. A Novel Truncated Normal Tensor Completion Method for Multi-Source Fusion Data. Mathematics 2024, 12, 223. https://doi.org/10.3390/math12020223

AMA Style

Zhao Y. A Novel Truncated Normal Tensor Completion Method for Multi-Source Fusion Data. Mathematics. 2024; 12(2):223. https://doi.org/10.3390/math12020223

Chicago/Turabian Style

Zhao, Yongmei. 2024. "A Novel Truncated Normal Tensor Completion Method for Multi-Source Fusion Data" Mathematics 12, no. 2: 223. https://doi.org/10.3390/math12020223

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop