Abstract
In order to avoid the loss of optimality of the optimal weighting factor in some cases and to further reduce the estimation error of an unbiased estimator, a multi-sensor adaptive weighted data fusion algorithm based on biased estimation is proposed. First, it is proven that an unbiased estimator can further optimize estimation error, and the reasons for the loss of optimality of the optimal weighting factor are analyzed. Second, the method of constructing a biased estimation value by using an unbiased estimation value and calculating the optimal weighting factor by using estimation error is proposed. Finally, the performance of least squares estimation data fusion, batch estimation data fusion, and biased estimation data fusion is compared through simulation tests, and test results show that biased estimation data fusion has a greater advantage in accuracy, stability, and noise resistance.
1. Introduction
Data fusion can utilize the information of multiple different data sources to complement each other, so the application of data fusion algorithms can suppress negative environmental factors and get more accurate results [1,2]. Data fusion algorithms in combination with various algorithms can achieve better results [3,4], so they have been widely used in various fields [5,6]. However, data fusion can only be applied to linear systems where all noises are independent of each other; if it is applied to a nonlinear system, it needs to be combined with the extended Kalman filter or the unscented Kalman filter [7]. If noises are correlated, they need to be decoupled by linear changes [8].
It is necessary to first process the original data, which includes removing abnormal data, filtering system deviations, and lossless compression of data. Abnormal data is removed using the Grubbs criterion. Filtering bias uses bias estimation algorithms [9,10], and in distributed fusion, if data need to be restored in the data center, lossless compression is required [11].
Before conducting multi-sensor data fusion, each sensor can also use various algorithms to estimate the true value (called the estimation value), which can improve the fusion accuracy because the estimation value has higher accuracy than the measurement value.
When the measured physical quantity does not change over time, if there is sufficient understanding of the system, linear minimum mean square error estimation is the best method. This algorithm can find a linear function that minimizes mean square error through a single measurement value [12,13,14]. Otherwise, it is best to use least squares estimation [15,16] and batch estimation [17]. When the measured physical quantity changes over time, it can only be suitable to use Kalman filtering [18,19].
Adaptive weighted data fusion is a widely used multi-sensor data fusion algorithm with a unique optimal weighting factor that enables the fusion estimation value to have a lower estimation variance than the measurement variance of all sensors, but it requires prior knowledge of the measurement variance to run [20]. However, its combination with a partially unbiased estimation algorithm (hereinafter referred to as unbiased estimation data fusion) not only further reduces the estimation variance but also removes the need for prior knowledge.
Unbiased estimation data fusion has two shortcomings. The first one is that unbiased estimation does not mean that the estimation results must be reliable. According to the Gauss-Markov theorem, the unbiased estimation variance has a lower bound, and when the lower bound itself is very large, even if the unbiased estimation variance is minimized, the difference between the estimation results and the true value (hereinafter referred to as estimation error) may not be within the acceptable range [21]. The other one is that the optimal weighting factor is not necessarily optimal; the optimal weighting factor is calculated according to the measurement variance; a smaller measurement variance does not necessarily mean a smaller measurement error, so in some cases, the optimal weighting factor is not optimal. Therefore, this paper proposes a biased estimation data fusion algorithm that also does not require priori knowledge, which uses unbiased estimation values to construct biased estimation values with lower estimation errors and calculates the optimal weighting factor according to the estimation error to ensure its optimality.
This paper is organized as follows: Section 2 reviews some existing unbiased estimation data fusion algorithms, Section 3 describes the shortcomings of the existing algorithms and the feasibility of the improved methods, Section 4 describes the specific implementation of the algorithms in this paper, Section 5 compares the performance of the different algorithms, and Section 6 summarizes the entire content and provides prospects for the future.
2. Unbiased Estimation Data Fusion
When N sensors are measuring an invariant physical quantity and all the sensor nodes in a network measure the observed physical quantity at the same time, the measurement equation is:
where is the unknown quantity to be measured, is the true value of , and is the discrete time indicator, is the measurement value of the ith sensor at the moment , is the measurement matrix of the ith sensor (the are often unit matrices), is the measurement noise, and are independent of each other, and be called measurement variance. Moreover, according to the additivity of the normal distribution, .
In this system, only is a known quantity; how one can use to calculate the true value of ? Since the measurement equations are designed for multiple sensors and the measurement variances may vary, adaptive weighted data fusion is often used to solve such problems, but adaptive weighted data fusion needs prior knowledge of the measurement variance of all the sensors, and in practice, the measurement covariance can be obtained by experiments.
2.1. Unbiased Estimation
Unbiased estimation refers to: if the mathematical expectation of an estimator is equal to the true value of the estimated parameter (this property is called unbiasedness), then this estimator is called the unbiased estimation of the estimated parameter. The mathematical expectation of the estimator constructed by unbiased estimation is equal to the true value of the estimated parameter, so using unbiased estimators instead of measurement values will not cause an increase in error. Unbiased estimation of sample variance can be used to replace the measurement variance of sensors, allowing adaptive weighted data fusion to operate without knowing the measurement variance of sensors. There are two commonly used unbiased estimation algorithms, as follows:
Least squares estimation:
Assuming that a particular sensor collects n data over a period of time, , the estimation value, and , estimation variance, are given as follows:
Batch estimation:
Assuming that a particular sensor collects n data over a period of time, these n data are randomly divided into two groups, where the data in the ith groups are: , . , the sample mean, and , sample variance, are given as follows:
According to the theory of batch estimation in statistics, the optimal estimation value and the optimal estimation variance of these 2 sets of data are given as follows:
2.2. Adaptive Weighted Data Fusion
Adaptive weighted data fusion principle: assume that there are sensors working on the same unknown quantity (the true value is ) is measured, and the measurement value of the ith sensor is . The measurement variance of the ith sensor is , and the measurement data of the ith sensor follows , then the unbiased estimator is given as follows:
Since is an unbiased estimator, so that
In order to find that minimizes (call this as the optimal weighting factor), the auxiliary function that is constructed according to the Lagrange multiplier method is given as follows:
According to multivariate extremum theory, it is known that the solution of the following partial differential equation makes the estimation variance minimized, as follows:
The simplified formula is given as follows:
Thus, the formula for are given as follows:
So, and are given as follows:
When adaptive weighted data fusion is combined with other unbiased estimation algorithms, it is possible to replace and in Equation (15) with an unbiased estimation value and an unbiased estimation variance. The Figure 1 shows the flowchart for batch estimation data fusion, and the flowchart for other unbiased estimation data fusion is more or less the same.
Figure 1.
Flowchart of batch estimation data fusion.
3. Rationale
Statistical inference is a statistical method of inferring the whole through samples, and there are two kinds of errors in statistical inference: systematic error and random error. Unbiased estimation has no systematic error because of its unbiased nature, but there is still random error, so its estimation results are not necessarily reliable.
The optimal weighting factor is calculated based on the measurement variance alone, but a smaller measurement variance does not mean a smaller measurement error. Although the measurement variance of the data measured by different sensors varies, the mathematical expectations are all equal, so, in practice, it is easy to have a smaller measurement variance but a larger measurement error.
3.1. Biased Estimators
Theorem 1.
The error of the unbiased estimator can be further optimized.
Proof of Theorem 1.
Assuming that the true value of the measured unknown quantity is , the sensor measurement data follows . is an unbiased estimator constructed using a linear combination of the sensor measurement data, and letting the estimation variance of be , then , defining the estimation error as follows:
is easy to obtain, and follows . Assuming that the estimation error is tolerable when , the probability that the estimation error is tolerable is:
If , we obtain: , which means that for any , there is about a 31.7% chance that is larger than the estimation standard deviation , so that even if is very small, may be able to be reduced further. □
Theorem 2.
A biased estimation value with a smaller error can be constructed from an unbiased estimation value.
Proof of Theorem 2.
Assuming unbiased estimator follows , the biased estimator corresponding to is defined as and , where is called offset, which is a nonzero variable. The following results can be obtained:
Because is not equal to the true value , B is a biased estimator.
Assuming that is a specific estimate of , its estimation error is and . is the offset corresponding to and , and the biased estimator corresponding to is and . The estimation error of is and , so:
So, when and , . In other words, when the sign of is opposite to and , the estimation error of biased estimation is smaller than that of unbiased estimation. □
3.2. Optimal Weighting Factors
Theorem 3.
In some cases, the optimal weighting factor is not optimal.
Proof of Theorem 3.
Suppose there are two sensors and that simultaneously measure the unknown quantity (the true value is ). Their measurement variances are and , and they satisfy . Their measurement data are and . For adaptive weighted data fusion, , the weights of , will always be greater than , the weights of , since < .
Define , the measurement error (unlike , only considers the magnitude of the error, not the sign of the error), as follows:
So , . If , then , which implies that the estimation error of the fusion estimation value computed from optimal weighting factors is not minimal. So, when < but , the optimal weighting factor is not optimal, and the probability is given as follows:
It can be shown that , which indicates that the weights of any two sensors may not be optimal. The above proof uses measured values and measured variances, but replacing them with estimated values and estimated variances is equally valid. □
Theorem 4.
It is optimal to compute the optimal weighting factor using the measurement error.
Proof of Theorem 4.
Suppose there are N sensors measuring the unknown quantity (the true value is ), their measurement values are , and their measurement errors are ; moreover, there are . The optimal weighting factor calculated from the measurement error is , because , then , so the corresponding fusion estimation value is as follows:
In this paper, only one case is proved, which has two weights different from , and the proof can be easily generalized to other cases.
Counterfactual: suppose that when the optimal weighting factor is and , the fusion estimation value is more accurate than , where is given as follows:
The common term of and is , and , where . So and become this form:
Because , so and set . Substituting for and , then we get the following equation:
Because , so the larger and are, the less accurate and are. Because and , so = and = . Set is given as follows:
So and . The derivative of is given as follows:
So, the larger is, the smaller . Because and , so , , and , is more accurate than . Contradicting the hypothesis, the optimal weighting factor calculated from the estimation error is better. The above proof uses measured values and measured variances, but replacing them with estimated values and estimated variances is equally valid. □
4. Algorithm Implementation
The algorithm in Section 4.1 only uses measurement data from a single sensor, so it can be implemented in a distributed manner. The algorithms in Section 4.2 and Section 4.3 include data fusion algorithms, so they can only be implemented using centralized methods.
4.1. Unbiased Estimation
Suppose that N sensors simultaneously measure the unknown quantity (the true value is µ), and all the sensor nodes in a network measure the observed physical quantity at the same time. For any of the sensors, if it collects 10 data over a period of time, in order from smallest to largest: , then construct the unbiased estimators as follows:
Assuming that the sensor follows , theoretically, and . Because ~ are data sorted in order of size, these estimators have lower estimated variances. Through experiments, it was found that the estimated variances of are approximately . Therefore, is independent of each other and has an estimated variance smaller than the measured variance.
, the mean of estimation values, and , the estimation variance, are given as follows:
An adaptive weighted data fusion algorithm is used to obtain , the unbiased estimation value, and , the unbiased estimation variance. And the specific formula is as follows:
Because is the least squares estimator, theoretically, the accuracy of the fused estimate will be higher than that of the least squares estimate.
If the number of measurement data collected by a sensor within a certain period is m and m is not equal to 10, the unbiased estimator can be constructed according to the following rules:
- Each group consists of at least two measurement values, and each group does not use duplicate measurement values. The unbiased estimator corresponding to each group is the mean of the measurement values within the group.
- When m is even (m = 2k), for any measurement value , it is necessary to ensure that and are in the same group.
- When m is odd (m = 2k + 1), for any measurement value , it is necessary to ensure that and are in the same group. In addition, it is also necessary to ensure that , , and are in the same group.
- Add an additional group that includes all measured values, such as in Equation (30).
4.2. Biased Estimation
, the size of the offset , needs to be determined first. For the ith sensor, define its unbiased estimation value and unbiased estimation variance as and . The fusion estimation value and the fusion estimation variance of these unbiased estimation values are computed by adaptive weighted data fusion algorithm.
Define as the degree of deviation of the ith unbiased estimation value from the true value. And is given as (use replaces true value) follows:
Define as the coefficient of , where the size of , and the value of are determined by . When , , when , . Otherwise, . is used to control the size of the offset and avoid excessive offset.
is given as follows:
Then, the sign of needs to be determined. takes a negative sign when > and a positive sign otherwise. So is given as follows:
4.3. Data Fusion
The application of the adaptive weighted data fusion algorithm requires measuring variance, which is replaced by estimated variance in unbiased estimation data fusion. In biased estimation data fusion, the estimated variance is calculated based on the estimation error, and the design concept of the calculation formula comes from the Kalman gain. Define as the biased estimation error of , and is given as follows:
The true value is unknown; use instead of the true value. So is given as:
Then, through 100,000 experiments, calculate , mean of , and , mean of , for different g-value intervals. The and over the three intervals are shown in the Table 1.
Table 1.
Mean estimation error and mean g for different g intervals.
and can be represented as follows:
So, , which represents the biased estimation variance of , is given as follows:
Thus, , the optimal weighting factor, and , the fusion estimation value, are given as follows:
The flowchart of biased estimation data fusion is shown in Figure 2:
Figure 2.
Flowchart of biased estimation data fusion.
5. Results
The simulation test is divided into two parts as follows: 1. a normal distribution white noise test (all noise follows a normal distribution), and 2. A uniform distribution white noise test (environmental noise is uniformly distributed white noise). The test items for the two tests are shown in Table 2:
Table 2.
The test items for the two tests.
The testing mode is as follows: ten rounds of experiments are conducted for each combination of test items, and ten tests are conducted for each round. The average of the ten test results is used as the result of this round of experiments. Three sensors are used for each test, with 10 measurement data used for each sensor.
The test indicators are mean relative error () and mean square error (). is used to compare the average performance of the three algorithms, and is used to compare the stability of the three algorithms. Their calculation formulas are as follows:
where represents the number of test rounds and represents the estimated value of the jth test in the ith round. Relative improvement () is used to compare the performance differences between two algorithms, and the calculation formula for is as follows:
where and represent the same indicator value (e.g., ) for algorithms and algorithms . And, if the indicator has multiple values, calculate the mean of the indicator first, and then calculate . If is greater than 0, it means that algorithm has improved compared to algorithm .
5.1. Normal Distribution White Noise Test
The test results are as follows (Figure 3, Figure 4 and Figure 5 show the test results when the measurement noise is 0.25, and the remaining result graphs are Figure A1, Figure A2, Figure A3, Figure A4, Figure A5 and Figure A6 in Appendix A):
Figure 3.
and when measurement variance = 0.25 and normal distribution noise variance = 0.2.
Figure 4.
and when measurement variance = 0.25 and normal distribution noise variance = 0.5.
Figure 5.
and when measurement variance = 0.25 and normal distribution noise variance = 1.
From Table 3, it can be seen that the of biased estimation data fusion (BED) decreased by an average of 9.25% and 10.26% compared to least squares estimation data fusion (LSD) and batch estimation data fusion (BTD). Therefore, BED has higher fusion accuracy.
Table 3.
of all algorithms when noise follows normal distribution.
From Table 4, it can be seen that the of BED decreased by an average of 16.24% and 20.28% compared to LSD and BTD. Therefore, the bed has higher stability.
Table 4.
of all algorithms when noise follows normal distribution.
And, as the noise variance increases, the average increases in and of LSD are 0.297% and 0.036, the average increases in and of BTD are 0.306% and 0.038, and the average increases in the and of BED are 0.269% and 0.03. So, and of BED decreased by an average of 11.59% and 18.85% compared to LSD and BTD, respectively. Therefore, BED has higher noise resistance.
5.2. Uniform Distribution White Noise Test
The test results are as follows (Figure 6, Figure 7 and Figure 8 show the test results when the measurement noise is 0.25, and the remaining result graphs are shown in Figure A7, Figure A8, Figure A9, Figure A10, Figure A11 and Figure A12):
Figure 6.
and when measurement variance = 0.25 and uniformly distributed noise variance = 0.2.
Figure 7.
and when measurement variance = 0.25 and uniformly distributed noise variance = 0.5.
Figure 8.
and when measurement variance = 0.25 and uniformly distributed noise variance = 1.
From Table 5, it can be seen that the of BED decreased by an average of 8.56% and 6.58% compared to LSD and BTD. Therefore, BED has higher fusion accuracy, but the advantage of is reduced by 22.39% compared to normal distribution white noise testing.
Table 5.
of all algorithms when noise follows uniform distribution.
From Table 6, it can be seen that the of BED decreased by an average of 17.42% and 21.01% compared to LSD and BTD. Therefore, BED has higher fusion accuracy, but the advantage of is increased by 5.23% compared to normal distribution white noise testing.
Table 6.
of all algorithms when noise follows uniform distribution.
As the noise variance increases, the average increases in and of LSD are 0.305% and 0.03, the average increases in and of BTD are 0.298% and 0.031, and the average increases in and of BED are 0.278% and 0.024. So, and of BED decreased by an average of 8.21% and 21.29% compared to LSD and BTD. BED has higher noise resistance, and the comprehensive noise resistance is equivalent to that of normal distribution white noise testing.
In summary, even if the noise does not conform to a normal distribution, as long as the mean noise is 0, the algorithm can perform well.
6. Discussion
In this paper, an adaptive weighting data fusion algorithm based on biased estimation is proposed to address the shortcomings of unbiased estimation data fusion. The algorithm in this paper first further reduces the estimation error by biased estimation and then uses the estimation error instead of the measurement variance to compute the optimal weighting factor, which avoids the fact that the optimal weighting factor is not optimal in some cases. The final results of the simulation experiments show that this algorithm outperforms the original algorithm in terms of fusion accuracy, stability, and noise resistance, which are better than other algorithms.
The essence of this algorithm is to sacrifice unbiasedness for higher estimation accuracy, so the overall accuracy of biased estimation is significantly higher than that of unbiased estimation. However, due to the loss of unbiasedness, the performance of the algorithm is not as good as expected when using adaptive weighted data fusion, so there is still much room for improvement.
Author Contributions
Writing—original draft preparation, M.Q.; writing—review and editing, B.L. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Major Science and Technology Projects in Yunnan Province (202202AE090032).
Data Availability Statement
Data are contained within the article.
Conflicts of Interest
The authors declare no conflict of interest.
Appendix A
Figure A1.
and when measurement variance = 1 and normal distribution noise variance = 0.2.
Figure A2.
and when measurement variance = 1 and normal distribution noise variance = 0.5.
Figure A3.
and when measurement variance = 1 and normal distribution noise variance = 1.
Figure A4.
and when measurement variance = 2 and normal distribution noise variance = 0.2.
Figure A5.
and when measurement variance = 2 and normal distribution noise variance = 0.5.
Figure A6.
and when measurement variance = 2 and normal distribution noise variance = 1.
Figure A7.
and when measurement variance = 1 and uniformly distributed noise variance = 0.2.
Figure A8.
and when measurement variance = 1 and uniformly distributed noise variance = 0.5.
Figure A9.
and when measurement variance = 1 and uniformly distributed noise variance = 1.
Figure A10.
and when measurement variance = 2 and uniformly distributed noise variance = 0.2.
Figure A11.
and when measurement variance = 2 and uniformly distributed noise variance = 0.5.
Figure A12.
and when measurement variance = 2 and uniformly distributed noise variance = 1.
References
- Kenyeres, M.; Kenyeres, J. Average Consensus over Mobile Wireless Sensor Networks: Weight Matrix Guaranteeing Convergence without Reconfiguration of Edge Weights. Sensors 2020, 13, 3677–3702. [Google Scholar]
- Sha, G.; Radzienski, M.; Soman, R.; Cao, M.; Ostachowicz, W.; Xu, W. Multiple damage detection in laminated composite beams by data fusion of Teager energy operator-wavelet transform mode shapes. Compos. Struct. 2020, 235, 111798. [Google Scholar] [CrossRef]
- Sun, S.-L.; Deng, Z.-L. Multi-sensor optimal information fusion Kalman filter. Automatic 2004, 40, 1017–1023. [Google Scholar] [CrossRef]
- Meng, T.; Jing, X.; Yan, Z. A Survey on Machine Learning for Data Fusion. Inf. Fusion 2019, 57, 115–129. [Google Scholar] [CrossRef]
- Xia, Y.; Leung, H. Performance analysis of statistical optimal data fusion algorithms. Inf. Sci. 2014, 277, 808–824. [Google Scholar] [CrossRef]
- Kim, H.; Suh, D. Hybrid Particle Swarm Optimization for Multi-Sensor Data Fusion. Sensors 2018, 9, 2792. [Google Scholar] [CrossRef]
- Gao, B.; Hu, G.; Gao, S. Multi-Sensor Optimal Data Fusion Based on the Adaptive Fading Unscented Kalman Filter. Sensors 2018, 18, 488. [Google Scholar] [CrossRef]
- Yan, L.; Xia, Y.; Fu, M. Optimal fusion estimation for stochastic systems with cross-correlated sensor noises. Sci. China Inf. Sci 2017, 60, 120205. [Google Scholar] [CrossRef]
- Zhang, T.; Li, H.; Yang, L. Multi-Radar Bias Estimation without a Priori Association. IEEE Access 2018, 6, 44616–44625. [Google Scholar] [CrossRef]
- Tian, W.; Huang, G.; Peng, H. Sensor Bias Estimation Based on Ridge Least Trimmed Squares. IEEE Trans. Aerosp. Electron. Syst. 2019, 56, 1645–1651. [Google Scholar] [CrossRef]
- Duan, Z.; Li, X.R. Lossless Linear Transformation of Sensor Data for Distributed Estimation Fusion. IEEE Trans. Signal Process. 2011, 59, 362–372. [Google Scholar] [CrossRef]
- Yaseen, M.; Canbilen, A.E.; Ikki, S. Channel Estimation in Visible Light Communication Systems: The Effect of Input Signal-Dependent Noise. IEEE Trans. Veh. Technol. 2023, 72, 14330–14340. [Google Scholar] [CrossRef]
- Li, Q.; El-Hajjar, M.; Hemadeh, I.; Shojaeifard, A.; Hanzo, L. Low-Overhead Channel Estimation for RIS-Aided Multi-Cell Networks in the Presence of Phase Quantization Errors. IEEE Trans. Veh. Technol. 2023, 73, 6626–6641. [Google Scholar] [CrossRef]
- Li, Q.; El-Hajjar, M.; Hemadeh, I.; Jagyasi, D.; Shojaeifard, A.; Hanzo, L. Performance analysis of active RIS-aided systems in the face of imperfect CSI and phase shift noise. IEEE Trans. Veh. Technol. 2023, 72, 8140–8145. [Google Scholar] [CrossRef]
- Fortunati, S.; Farina, A.; Gini, F. Least Squares Estimation and Cramér-Rao Type Lower Bounds for Relative Sensor Registration Process. IEEE Trans. Signal Process. 2011, 59, 1075–1087. [Google Scholar] [CrossRef]
- Zheng, Z.-W.; Zhu, Y.-S. New least squares registration algorithm for data fusion. IEEE Trans. Aerosp. Electron. Syst. 2004, 40, 1410–1416. [Google Scholar] [CrossRef]
- Wang, H.D.; Wang, D.Y. An improved adaptive weighted fusion algorithm for batch estimation of multi-wireless sensor data. J. Sens. Technol. 2015, 28, 1239–1243. [Google Scholar]
- Shao, S.; Zhang, K. An Improved Multisensor Self-Adaptive Weighted Fusion Algorithm Based on Discrete Kalman Filtering. Complexity 2020, 2020, 9673764. [Google Scholar] [CrossRef]
- Gan, Q.; Harris, C.J. Comparison of two measurement fusion methods for Kalman-filter-based multisensor data fusion. IEEE Trans. Aerosp. Electron. Syst. 2001, 37, 273–279. [Google Scholar] [CrossRef]
- Chen, Y.; Si, X.; Li, Z. Research on Kalman-filter based multisensor data fusion. J. Syst. Eng. Electron. 2007, 18, 497–502. [Google Scholar]
- Yue, Y.; Zuo, X.; Luo, X. A biased estimation method for multi-sensor data fusion to improve measurement reliability. J. Autom. 2014, 40, 1843–1852. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).