Next Article in Journal
Blockchain-Powered Gaming: Bridging Entertainment with Serious Game Objectives
Previous Article in Journal
A Robust Timing Synchronization Algorithm Based on PSSS for LTE-V2X
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparative Study on Recent Automatic Data Fusion Methods †

Instituto de Telecomunicaciones y Aplicaciones Multimedia, Universitat Politècnica de València, 46022 Valencia, Spain
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in 17th International Work-Conference on Artificial Neural Networks (IWANN2023), Ponta Delgada, Portugal, 19–21 June 2023.
Computers 2024, 13(1), 13; https://doi.org/10.3390/computers13010013
Submission received: 5 December 2023 / Revised: 26 December 2023 / Accepted: 28 December 2023 / Published: 30 December 2023

Abstract

:
Automatic data fusion is an important field of machine learning that has been increasingly studied. The objective is to improve the classification performance from several individual classifiers in terms of accuracy and stability of the results. This paper presents a comparative study on recent data fusion methods. The fusion step can be applied at early and/or late stages of the classification procedure. Early fusion consists of combining features from different sources or domains to form the observation vector before the training of the individual classifiers. On the contrary, late fusion consists of combining the results from the individual classifiers after the testing stage. Late fusion has two setups, combination of the posterior probabilities (scores), which is called soft fusion, and combination of the decisions, which is called hard fusion. A theoretical analysis of the conditions for applying the three kinds of fusion (early, late, and late hard) is introduced. Thus, we propose a comparative analysis with different schemes of fusion, including weaknesses and strengths of the state-of-the-art methods studied from the following perspectives: sensors, features, scores, and decisions.

1. Introduction

Currently, information is growing exponentially in complexity, volume, variety, and veracity. We can extract and derive valuable information from data to learn about the nature of things [1,2,3]. The recent development of machine learning (ML) methods, in conjunction with the increasing capacity of sensor devices, has posed a critical problem to efficiently solve data classification. This problem comprises two questions: how to efficiently combine information from several sources or modalities of data, i.e., multimodal measurements, and how to combine the results from several classifiers. This latter question assumes that there are relationships between the random variables of the score distributions provided by each of the classifiers to be combined that allow the results to be improved.
In general, processes that require parameter estimation from multiple sources can benefit from data fusion in a context of multiple classifiers. Data fusion aims to combine the information received from the real world to make the results more heterogeneous and informative than the original ones [4]. In this way, the objectives are to increase the reliability of the classification and the quality of the extracted information. Data fusion aims to improve classification accuracy by combining the predictions of multiple models or classifiers to obtain a more robust final classification. Thus, the individual classifier biases and generalization for data not seen in classifier training can be solved. Data fusion aims to effectively handle noisy data, adapt to different changing scenarios and conditions, and improve the stability of problem solving [5,6].
The exponential expansion of data in recent years due to new technologies has also brought with it an increase in the interest of researchers in data fusion, driven by the need to extract relevant information from increasingly complex datasets [7,8,9]. Let us show a descriptive example of the objectives of decision fusion. Figure 1 shows a toy example of the decision boundaries formed by different methods of classification: linear discriminant analysis (LDA); quadratic discriminant analysis (QDA); and a Bayesian classifier implementing non-parametric probability estimation. In addition, the boundaries obtained by the fusion of the three single classifiers are also shown. The problem outlined in Figure 1 is two-class classification, i.e., to distinguish between the samples of classes labeled 0 and 1. The geometry of the data is 3D, considering three features of the multidimensional vector observation.
We can see that the data of the two classes are squeezed together, and it is difficult to separate them. The separation boundaries of the three classifiers are different: that provided by LDA consists of a hyperplane; QDA’s is a paraboloid; and the boundary provided by the Bayesian classifier describes a complex surface. The efficiency of those surfaces to separate the data of the two classes in terms of missed detections (100-probability of detection (PD)) and probability of false alarm (PFA) is shown. Setting the value of the miss detection at 4.92%, it is clearly shown that the best PFA is obtained by the Bayesian classifier (53.13%) consistently with the classifier boundary decision surface that best fits the data. However, the fusion of the three classifiers yields a better suitable separation surface, which is recorded in a much lower PFA (38.02%).
In order to perform the comparative study, we propose the following four schemes to analyze the state of the art of data fusion methods: (i) early fusion from sensors; (ii) early fusion from features; (iii) late fusion from scores; and (iv) late fusion from decisions. The comparison of the methods studied considers aspects such as complexity, optimization, accuracy, and volume of data. Moreover, a theoretical analysis of the conditions for applying the different kinds of fusion is presented.

Data Fusion Concepts

Data fusion techniques attempt to combine multiple sources of information to achieve accuracy and precision in decision-making that would not be possible to achieve with the use of a single source of information in isolation. The fusion of multiple sources of information, in addition to adding a certain redundancy that provides reliability and robustness, can provide complementary information with which to increase performance and accuracy in the decision-making process. For simplicity, we will discuss the two-class classification (detection) problem. Figure 2 shows a diagram of the fusion process from a statistical standpoint. It shows how various sources of information (with different probability density functions (PDFs) for the two hypotheses— H 1 : detection and H 0 : non-detection) are merged into a single one. The separation between hypotheses is improved, i.e., there is less data confusion in the fusion result.
We can find areas or applications where it is common to work with various sources of information and, thus, a combination of information is necessary. For example, distributed sensing systems [10] use a multi-sensor system composed of a set of identical sensors placed with a certain spatial distribution. Multimodal systems [11], where sources of information of various kinds from different sensors and/or sources of unsensed information (such as web resources, databases, metadata, etc.) are used, are common in biometric systems or multimedia analysis problems. There are also detection techniques, such as expert combination [12], where an attempt is made to combine a series of simple detectors to achieve results that would be very difficult to achieve with a single detector and/or reduce the complexity and computational resources required to implement it.
The random variables that characterize the data from each information source can follow different probability distributions. It may be because they are data from different types of sensors that capture different phenomena produced by the same event, or from sensors of the same type, but whose position relative to the source that produces the event translates into different distributions of their observations. It is very common to use data that come from the extraction of multiple features related to various physical aspects or that are of a different nature from a single flow of information provided by a single sensor. In other cases, the aim is to combine different algorithms or processing techniques with very different output information.
The use of multiple sources and redundant information can provide complementary information that can be exploited by combining all available sources. Thus, the discriminatory information on the occurrence or not of the event to be detected can be increased, and both overall detection performance, robustness, and reliability can be improved. The sources of information generally share a common origin, in which the event to be detected originates, so it is common to find the existence of statistical dependence between them. Statistical dependency between sources can introduce complementary information that we can benefit from to improve detection performance.
Data fusion is a booming area of research. There are several quite different real-world problems where data fusion has been recently applied, including the following: detecting breast tumors in tomosynthesis images [13]; video summarization [14]; multi-scenario violence detection [15]; energy-efficient grid-based routing protocols [16]; image sparse representation [17]; accurate skin lesion classification [18]; multimodal inference of mental workload for cognitive human machine systems [19]; fingerprint and online signature for multimodal biometrics [20]; and image fusion using per-pixel saliency visualization [21].
Different levels of fusion can be defined depending on the stage of the detection process in which the integration of the information is carried out. A detection problem can generally be divided into four stages (Figure 3). The first of these is the sensing stage, where one or more sensors are responsible for obtaining a series of measurements of the environment where the event occurs. These raw data ( y ) are processed to extract certain characteristics (features) of the event ( x ), from which detectors or binary classification algorithms will be able to yield data sample scores ( z ), usually related to the probability that the event to be detected has occurred. After thresholding scores, a set of binary decisions ( u ) is obtained on the occurrence or not of the event.
Thus, the sensor fusion level relates to when the different sample flows provided by sensors are combined ( y y f u s ); the feature fusion level is where the raw data features extracted are fused ( x x f u s ); the score fusion level concerns obtaining fused scores from the ones obtained by the classifiers ( z z f u s ). Those scores are thresholded to obtain decisions that are fused at the decision level ( u u f u s ). The levels of sensor and feature fusion can be grouped into what is known as early or pre-detection fusion. The levels of score and decision fusion are often referred to as late fusion or detector fusion [22,23,24]. The different levels of fusion are not mutually exclusive, so systems can be designed that combine fusion at different levels based on information provided by different sources throughout all stages.
Depending on the type of data to be combined, we can classify the fusion into two categories, soft fusion and hard fusion. (i) Soft fusion refers to the combination of continuous data, which can be modeled by continuous random variables, usually characterized by PDFs under each of the hypotheses. Within the detection process, we can generally find soft information in the data provided by the sensors ( y ), the features extracted from them ( x ), and in the scores provided by some types of detectors ( z ). (ii) Hard fusion combines discrete data, which are modeled by discrete random variables under each of the hypotheses. Hard information fusion is commonly associated only with the detector fusion stage, where binary random variables are used.

2. Early Fusion/Late Fusion Comparison

2.1. Early Fusion

Let us assume for simplicity the scenario of two classes ( k = 1 , 2 ), and two sets of features x 1 ,   x 2 , x 1 ,   x 2 , from two different modalities that we will assume conditionally independent, i.e., p x 1 , x 2 / k = p x 1 / k p x 2 / k . Suppose we perform early fusion, choosing the class that maximizes the a posteriori probability given the two sets of features x 1 ,   x 2 i.e., the decision rule will be:
P k = 1 / x 1 , x 2 > k = 1 < k = 2 P k = 2 / x 1 , x 2           P k = 1 / x 1 , x 2 > k = 1 < k = 2 0.5 ,
where we have taken into account that P k = 1 / x 1 , x 2 + P k = 2 / x 1 , x 2 = 1 . Note that rule (1) implies minimization of the probability of error if we assume that the costs of being wrong in the decision are symmetric and normalized to 1. That is, considering that we had exact knowledge of P k = 1 / x 1 , x 2 , no other fusion rule allows us to reduce the probability of error more than (1).

2.2. Late Soft Fusion

Suppose now that we perform late soft fusion. For this purpose, we generate an a posteriori probability (score) separately for each modality: s 1 = P k = 1 / x 1 and s 2 = P k = 1 / x 2 . Note that when it is the case that P k = 1 / x 1 + P k = 2 / x 1 = 1 and P k = 1 / x 2 + P k = 2 / x 2 = 1 , it is enough to consider the scores s 1 and s 2 defined. We then generate a new score s fusing s 1 and s 2 through a certain fusion function s = f s 1 ,   s 2             0 s 1 and the new decision rule will be:
s > k = 1 < k = 2 0.5 ,
Note that s = f s 1 ,   s 2 = f P k = 1 / x 1 ,   P k = 1 / x 2 is ultimately a function of multivariate random variables x 1 ,   x 2 , which will generally be different from P k = 1 / x 1 , x 2 in (1) and therefore rule (2) can never achieve a probability of error lower than rule (1), assuming knowledge of P k = 1 / x 1 , x 2 . For instance, let us analyze the case of the fusion method based on α -integration [25,26,27,28]. First, we apply Bayes’ rule:
s 1 = P k = 1 / x 1 = p x 1 / k = 1 P 1 p x 1 ;         s 2 = P k = 1 / x 2 = p x 2 / k = 1 P 1 p x 2 ,
We will now consider the α -integration of both scores to obtain s :
s = w 1 s 1 1 α 2 + w 2 s 2 1 α 2 2 1 α = w 1 p x 1 / k = 1 P 1 p x 1 1 α 2 + w 2 p x 2 / k = 1 P 1 p x 2 1 α 2 2 1 α = = w 1 p x 1 / k = 1 p x 2 P 1 1 α 2 + w 2 p x 2 / k = 1 p x 1 P 1 1 α 2 p x 1 p x 2 1 α 2 2 1 α = = w 1 p x 1 / k = 1 p x 2 w 1 2 1 α 1 α 2 + w 2 p x 2 / k = 1 p x 1 w 2 2 1 α 1 α 2 2 1 α p x 1 p x 2 P 1 ,
On the other hand, the score generated in rule (1) can be written as:
P k = 1 / x 1 , x 2 = p x 1 , x 2 / k = 1 P 1 p x 1 , x 2 = p x 1 / k = 1 p x 2 / k = 1 p x 1 , x 2 P 1 = = p x 1 / k = 1 p x 2 / k = 1 p x 1 p x 2 / x 1 P 1 ,
where we have taken into account the assumed conditional independence between the characteristics of the two different modalities x 1 ,   x 2 . Comparing (4) and (5), we can conclude an interpretation of the α -integration as an attempt to approximate the optimal score in (5). On the one hand, in relation to the denominator, when operating the two modalities through separate channels before merging, it is not taken into account that x 1 and x 2 are (unconditionally) dependent. On the other hand, we can see that the parameters α ,   w 1   and   w 2 , which are estimated by minimizing a certain cost function (e.g., error probability), should tend to be adjusted so that p x 2 w 1 2 1 α p x 2 / k = 1 and p x 1 w 2 2 1 α p x 1 / k = 1 , in order to obtain the minimum probability of error of rule (1). In any case, it is clear that, assuming perfect knowledge of late soft fusion can never provide a lower probability of error than early fusion. However, we must also consider that the use of particular classifiers is conditioned to the requirements of the size of the training datasets, and thus conditioning the fusion of the classifiers. The estimation of the sample size for training to obtain an estimated error is a complex task; see for instance the proxy learning curve for the Bayes classifier [29].

2.3. Late Hard Fusion

In this case, first, the binary decisions of each modality are generated, and from them a decision fusion rule is established to generate the final decision. Let us denominate d 1 ,   d 2 the binary variables corresponding, respectively, to the binary decisions of each modality, i.e., we can write:
d 1 = u s 1 0.5 ;         d 2 = u s 2 0.5 ,
where u x = 1       x > 0 0       x < 0 is the step function. As we can see, the binary variables are equal to 1 if we choose class 1 and equal to 0 if we choose class 2. The final decision will be the result of applying a fusion rule to d 1 ,   d 2 . In the case of two modalities, we have only two possible reasonable rules:
d = g 1 d 1 ,   d 2 = 1         si   d 1 = 1 ,     d 2 = 1 0       all   other   cases                       d = g 2 d 1 ,   d 2 = 0         si   d 1 = 0 ,     d 2 = 0 1         all   other   cases   ,
It seems natural that late hard fusion, in exchange for its simplicity, leads to worse performance than late soft fusion, since binarization prior to fusion entails the loss of information. Somewhat heuristically, we can reason as follows. Suppose that the late soft fusion function f s 1 ,   s 2 is optimized to minimize a certain cost function of the binary classifier, such as the probability of error. If we take into account that d 1 ,   d 2 are a function of the scores s 1 ,   s 2 , we can think that ultimately the functions g 1 d 1 ,   d 2 and g 2 d 1 ,   d 2 in (7) are functions of s 1 ,   s 2 , which will, in general, be different from the optimal function used in late soft fusion and which will, therefore, not minimize the chosen cost function. To be more specific, let us consider again the integration. It is easy to check that:
α                       s min s 1 ,   s 2           α                       s max s 1 ,   s 2 ,
But the following fusion rules are equivalents:
min s 1 ,   s 2 > k = 1 < k = 2 0.5                       g 1 d 1 ,   d 2           max s 1 ,   s 2 > k = 1 < k = 2 0.5                       g 2 d 1 ,   d 2 ,
That is, we can interpret late hard fusion as a late soft fusion for a particular choice of integration parameters of α -integration, which will not, in general, be those that minimize the cost function and will therefore lead to worse (or, at best, equal) performances than those achievable with late soft fusion.
In short, we conclude that, if the a posteriori probability or score P k = 1 / x 1 , x 2 is known precisely, late fusion will not be able to outperform early fusion. On the other hand, the hard fusion will not be able to outperform late soft fusion either, as we have justified above.
The exact knowledge of P k = 1 / x 1 , x 2 is only strictly possible if we have an infinite training set and we assume statistical consistency, i.e., P ^ k = 1 / x 1 , x 2 N P k = 1 / x 1 , x 2 , where N is the size of the training set of each class (for simplicity we assume the same for both classes). From a practical point of view, accurate estimation implies having sufficiently large training sizes for each class. Unfortunately, it is not easy to determine in general what the adequate minimum size is. This is outside the scope of this paper. However, there are recent works that provide certain criteria in this regard [30].

3. Early Fusion from Sensor and Features

In Section 3 and Section 4 we will study the state-of-the-art data fusion methods and briefly describe them comparatively according to different points in the data processing where fusion can take place. The study includes the analysis of strengths and weaknesses of the methods and related works. Figure 4 shows the proposed outline to explain the different fusion approaches.

3.1. Early Fusion from Sensor

The process of combining disparate sensor data is known as sensor fusion. Figure 5 shows a schematic illustrating early sensor fusion. There are several ways of acquiring data from the sensors:
  • ✓ Individual sensor, multiple samples.
This first case is the simplest as it involves a single sensor from which we obtain multiple samples.
  • ✓ Multisensors.
The essence of this data acquisition is to combine information from multiple sensors into a single sensor. In this way, greater accuracy can be achieved than could be achieved using a single sensor [31].
  • ✓ Multimodal.
Multimodal acquisition integrates information from multiple sources to compensate for each of them. An example is the simultaneous acquisition of electroencephalogram (EEG) and functional magnetic resonance imaging (fMRI) data. On the one hand, we have a high temporal resolution, as in the case of EEG, and on the other hand, we have fMRI data with a high spatial resolution [32].
Figure 5. Early sensor-level fusion scheme.
Figure 5. Early sensor-level fusion scheme.
Computers 13 00013 g005
At the sensor level, there are several methods of data fusion. Among them, we can find the Kalman Filter, Bayesian Inference, Fuzzy Logic, Artificial Neural Networks (ANNs) and Dempster–Shafer (DS). In Table 1, we present a brief description of these methods, and finally, we offer a comparative table where we will find the weaknesses and strengths of each of these methods and related works.
  • ✓ Kalman Filter.
The Kalman filter is an ideal recursive statistical data processing algorithm that continuously computes an estimate of a continuous-valued state from periodic observations [33]. It uses an explicit statistical model of how the vector of interest x t changes over time and an explicit statistical model of how the vector of observations z t made are related [1]. The gains used in a Kalman filter are chosen to ensure that the resulting estimate minimizes the mean square error with certain assumptions about the observation models and processes used [34].
  • ✓ Bayesian Inference.
Bayesian inference is a statistical data fusion algorithm based on Bayes’ theorem of a posteriori probability for estimating an n-dimensional state vector Z , given the observation or measurement X [35].
According to Bayes’ theorem, given a set of measurements X k = x 1 ; ; x k in time k, the Bayesian interference is formulated as:
p z k | X k = p x k z k p z k | X k 1 p X k | X k 1 ,
where p x k z k is the probability function based on a given sensor measurement model, p z k | X k 1 is the a priori probability distribution, and p X k | X k 1 is the probability density function considering all hypotheses.
If we assume that the statistical independence between the set of measurements X k = x 1 ; ; x k , then we can combine them to infer the state of the observed system [36]. This method requires knowledge of the a priori probability distribution of the states.
  • ✓ Fuzzy Logic.
Fuzzy Logic is a logic that uses neither wholly true nor false expressions. It is applied to concepts that can acquire any value of veracity within a set of values that oscillate between two extremes: absolute truth and total falsehood [37].
The main benefits of using fuzzy logic techniques are the simplicity of the approach, its ability to handle ambiguous information, and its capacity to include heuristic knowledge about the phenomenon under consideration [2,38]. The use of fuzzy logic for sensor fusion has demonstrated a high degree of accuracy and precision, but the complicated computations required are an obstacle [39].
  • ✓ Artificial Neural Networks (ANNs).
ANNs are mathematical models of non-linear computational elements operating in parallel and linked together in a topology distinguished by several ponderable connections. Comparing ANN analysis with conventional linear or non-linear analysis, it has proven to be a more powerful and flexible method [40].
One of the features of ANNs is their adaptive learning capability. Developing a priori models or specifying probability distribution functions is unnecessary, because neural networks can learn differentially through examples and training [41].
A typical ANN structure comprises three main layers: the input layer, one or more hidden caps, and the output layer. One or more neuron-like nodes distinguish each. The number of input and output variables determines the number of neurons in the input and output layers. Depending on the problem’s difficulty, the number of hidden layers and the number of neurons associated with each hidden layer varies [42].
  • ✓ Dempster–Shafer (DS).
Dempster–Shafer theory offers an alternative to the traditional probabilistic approach for the mathematical representation of uncertainty. It has been widely applied in various applications, such as target tracking, surveillance, robotic navigation, and signal and image processing [43].
As is well known, DS represents the uncertainty or imprecision in a hypothesis that characterizes all possible system states. A probability mass assignment (PMA) is applied to such a hypothesis, which results in a decision when combined. Therefore, creating a function for the mass assignment and combining it is essential for accurate prediction. By applying a combinatorial rule to the sources of evidence, DS achieves the goal of data fusion [9].
Table 1. Comparison between early sensor-level fusion methods and related works.
Table 1. Comparison between early sensor-level fusion methods and related works.
Sensor-Level
Fusion Methods
StrengthsWeaknessesWorks
Kalman FilterThe Kalman filter can provide highly efficient and accurate results in contexts where the system conditions are well understood and the models are correct.
Estimates are performed recursively. This property makes it computationally efficient and suitable for real-time applications.
It is not a method designed for optimization; therefore, it cannot converge to local or global minima. It is intended for state estimation and prediction in dynamic systems.
It requires a broad knowledge of probabilities, especially the subject of Gaussian conditionality in random variables.
[44,45,46,47,48]
Bayesian InferenceIt is a recursive technique and can compute probabilities and posterior probabilities for multiple hypotheses.
If the conditions are well understood and the models are correct, it offers a convenient setup for various models, such as hierarchical models and missing data problems.
The probability distribution of the states must be known a priori. It often comes with a high computational cost, especially in models with many parameters.
As the size of the data increases, handling the distributions becomes more difficult.
[35,49,50,51,52]
Fuzzy LogicAccurate results in non-linear and challenging-to-model processes. It is based on logical sets and reasoning that are easy to understand and, therefore, to use.
Provides a simple mechanism for reasoning with vague, ambiguous, or imprecise information.
Extensive validation and verification of fuzzy algorithms are necessary.
Accurately defining fuzzy sets or membership functions requires time and effort. In addition, increasing the dimension of the data makes it more challenging to model the problem.
Fuzzy control systems depend on human experience and knowledge.
[2,39,53,54,55]
Artificial Neural NetworksIt is self-learning and can execute tasks that a linear program cannot and is able to process unorganized data. Its structure is adaptive in nature.
When an element of the neural network slows down, it can continue without problems, thanks to its parallel characteristics and is efficient at handling data noise, separating only the necessary information.
Requires prior training to operate, a large amount of data to achieve adequate efficiency and a lot of processing time for large neural networks.
Requires specific hardware equipment to operate due to their computational complexity.
If not handled properly, neural networks may be overfitted to the training data and not generalize well to new data.
It can converge to local minima instead of global minima, although there are solutions to this problem, such as weight initialization and regularization in terms of L1 and L2.
[40,56,57,58,59,60]
Dempster–ShaferSuch a theory can provide accurate results if the evidence is accurate and reliable.
Although there is no hard and fast rule regarding data, a limited amount of data can help manage uncertainty and consistently combine information.
Generally, it presents a high computational complexity, although this may vary depending on the amount of data, hypotheses, and uncertainty of the problem to be treated.[9,61,62]

3.2. Early Fusion from Features

These methods or techniques aim to reduce the feature vector’s dimension while preserving as much information as possible. Figure 6 shows a schematic of the feature-level fusion methods.
Among the feature-level fusion methods, we can find Principal Component Analysis (PCA), Singular Value Decomposition (SVD), Multidimensional Scaling (MDS), and Deep Learning. Below, in Table 2, we will briefly describe them, detailing their strengths and weaknesses and the related works that have been most cited.
  • ✓ Principal Component Analysis (PCA).
PCA is a mathematical procedure transforming several correlated variables into uncorrelated variables called principal components [63]. This technique reduces the data set (in this case, the features extracted from multiple sensors) while maintaining the statistical information and at the same time minimizing the loss of information [6]. Principal components are calculated as linear combinations of the original data or variables [64].
We refer to each main component as Y j and a set of variables X 1 , X 2 , , X p . The first principal component is the normalized linear combination of these variables with the highest variance.
Y 1 = ϕ 11 X 1 + ϕ 21 X 2 + + ϕ p 1 X p ,
The normalized linear combination implies that:
j = 1 p ϕ j 1 2 = 1 ,
The terms ϕ 11 , , ϕ 1 p define the component and are called loading. They can be interpreted as each variable’s weight in each component.
  • ✓ Singular Value Decomposition (SVD).
SVD is a powerful mathematical technique to decompose a matrix into constituent parts. It is a generalization of the autovalues decomposition of a matrix and can be applied to any rectangular matrix, not just square matrices. It helps to reduce data sets containing a large number of values and is useful for generalizing meaningful solutions with fewer values [65,66,67].
A matrix X of m × n with m and n dimensions of the row and column vectors can be related to the diagonal matrix ∑ de m × n , which satisfies the following equation:
= U T X V ,
where U is an orthogonal matrix of m × n and V is an orthogonal matrix of n × n . When rewritten, it is satisfied that:
X = U V T ,
  • ✓ Multidimensional Scaling (MDS).
MDS is a dimensionality reduction technique that converts multidimensional data into a lower dimensional space while maintaining intrinsic information [68]. This set of statistical techniques takes as input similarity estimates between a set of data [69]. This method is a much more flexible alternative to other multivariate analysis methods, since it only requires a matrix containing the similarity or dissimilarity between the input data.
In general, MDS takes a proximity matrix as input M n x n , where n is the number of stimuli and each element δ i j of Δ represents the proximity between stimulus i and stimulus j.
Δ = δ 11 δ 1 n δ n 1 δ n n ,
From this proximity matrix, MDS provides us with an output matrix M n x m , where n is the number of stimuli as before and m is the number of dimensions. Each value of x i j represents the coordinate of stimulus i in dimension j.
X = x 11 x 1 m x n 1 x n m ,
From the matrix X we can calculate the distance between any two stimuli by applying the general Minkowski formula:
d i j = t = 1 m x i t x j t p p ,
where p can be a value between 1 and infinity. From these distances, we can obtain a matrix of distances D M n x n .
D = d 11 d 1 n d n 1 d n n ,
The solution provided by the MDS should be such that there is a maximum correspondence between the initial proximity matrix ∆ and the matrix of distances obtained D .
  • ✓ Deep Learning.
We can say that deep learning is a type of machine learning that aims to make an artificial intelligence (AI) capable of learning by itself and performing tasks similar to those performed by humans, such as image recognition, speech recognition, and making predictions, over time and with large amounts of data and processing with algorithms.
Deep learning algorithms are applied to ANNs structured into input, hidden, and output layers. Nowadays, feature fusion using deep learning is very fashionable, especially in the medical field [70,71] and in social networks [72,73].
Table 2. Comparison between early feature-level fusion methods and related works.
Table 2. Comparison between early feature-level fusion methods and related works.
Feature-Level
Fusion Methods
StrengthsWeaknessesWorks
Principal Component AnalysisReduces the complexity of the data and identifies the most important features.
This method captures the directions of maximum variability in the data. This means that the most informative features are retained as valuable in data fusion where relevant information is sought to be preserved.
After the transformation, the variance of the data is preserved.
It is necessary to choose the correct number of principal components needed for the data set to avoid some loss of information.
Although the method works quickly for large data sets, it requires high computational complexity and memory requirements.
[6,64,74,75,76]
Singular Value DecompositionThis mathematical technique is precious for reducing the dimensionality of the data, capturing the most relevant and distinctive feature information, and eliminating redundancies and noise.SVD can be computationally expensive for large data sets, mainly when applied to high-dimensional arrays.
It only makes use of a single data set, and by default, the resulting dimension reduction cannot incorporate any additional information that may be relevant.
The accuracy of the data may decrease if the data patterns are not linear.
[65,66,67,77,78,79]
Multidimensional ScalingThe solutions are relatively accurate.
It can be used to fuse different types of data into a shared space, which is useful when the features of the sensors are different.
The method provides a visual representation of the data in a two- or three-dimensional space, which can help to understand patterns and relationships.
It does not allow quantifying the level of quality of the result. Since it is based on the relationship between dimensions or factors, evaluating this relationship in numbers is tough.
Can be computationally expensive for large data sets and may require iterative optimization.
As the data are projected into a lower dimensional space, there may be a loss of information, which could affect the quality of the fused data.
[68,69,80,81,82,83]
Deep LearningIt assists in trend and pattern detection and does not need human assistance, i.e., it makes its own decisions.
It can handle many multidimensional data and constantly improves the algorithm to achieve more accurate results. It can fuse data from multiple sources, such as images, text, and signals, into a single architecture, leveraging information from different modalities.
Requires a large amount of data for training, which is time-consuming and computationally complex. Therefore, more powerful computers are needed for it to work.
Limited data availability can affect performance, as large amounts of data are needed for effective training.
Can be trapped in local minima instead of reaching the best possible solution (global minimum).
[70,71,72,73,84]

4. Late Fusion from Scores and Decisions

4.1. Late Fusion from Scores (Late Soft Fusion)

Information from different detectors or classifiers is often referred to as scoring [85]. Score-level fusion is usually preferred because it offers the best compensation in terms of information content and ease of fusion [86,87]. Figure 7 shows a schematic of the workflow for these methods.
According to the literature, Sum Rules, Likelihood Ratio (LR), Fusion by classifiers, Alpha Integration, and Behavior-Knowledge Space (BKS) are widely used score fusion methods. These techniques are described below, and Table 3 with the strengths and weaknesses of these methods and related work is presented.
  • ✓ Sum and Weighted Sum Rules.
The Sum Rule is a simple fusion technique that operates directly on the raw data of the match scores [88]. It is given by the following equation:
f s = s 1 + s 2 + s n ,
where f s is the fused score and s i represent the scores results. The weighted sum is similar, except that a weighting is assigned w i to each score depending on performance. It can be calculated as:
f s = w 1 s 1 + w 1 s 2 + w n s n ,
  • ✓ Likelihood Ratio (LR).
LR is a fusion approach based on the Neyman–Pearson theorem. This method does not require parametric adjustments and achieves a maximum valid acceptance rate (GAR) [89]. This approach is density-based and can achieve optimal performance at any desired false acceptance rate (FAR) point, provided that score density is accurately calculated.
The LR method reduces the feature values into a single metric (score) representing the similarity between the comparison items [90]. The genuine and impostor match score vectors are modeled using Gaussian mixture models. The likelihood ratio is calculated as the ratio of the genuine probability density over the impostor probability density.
  • ✓ Fusion by classifiers.
Score fusion using classifiers is a technique in which classification models are used to combine scores. It aims to improve the accuracy and reliability of final decisions by taking advantage of the diversity of information provided by different classifiers.
The choice of classifier depends on the nature of the input scores and the application in question. It can be a simple classifier, such as logistic regression, or a more complex one, such as random forest (RF), support vector machine (SVM), and quadratic discriminant analysis (QDA). In addition, the right fit between performance and generalization must be achieved.
  • ✓ Alpha Integration.
Alpha integration is a group of integrators containing many combinations, as may be the particular case of the alpha parameter. This approach was first proposed by Amari [91] to integrate multiple stochastic models by minimizing their alpha divergence. It has also been used to perform optimal integration of scores in binary classification problems [10].
This method was extended to integrate multiclass classifiers by considering the scores of each class in isolation in a technique called separated score integration (SSI) [26]. A new alpha integration method for late fusion of multiple classifiers that considers the combined effect of all classes in the multiclass problem called vector score integration (VSI) is proposed in [27].
  • ✓ Behavior-Knowledge Space (BKS).
The BKS method is a trainable combination scheme at the abstract level, requiring neither measurements nor ordered sets of candidate classes. It attempts to estimate a posteriori probability by calculating each class’s frequency for each possible set of classifier decisions based on a given training set [92].
In BKS, each classifier can assign a sample to one of the M possible classes. Each unit of a BKS represents a particular intersection of the decisions of an individual classifier. Thus, all possible combinations of the decisions of the individual classifiers are represented [8].
Table 3. Comparison between late score-level fusion methods and related works.
Table 3. Comparison between late score-level fusion methods and related works.
Score-Level Fusion MethodsStrengthsWeaknessesWorks
Sum RuleIt does not require training samples. No sample distribution modeling is required.
The addition process is fast and computationally efficient.
This method works well for significant data inputs.
Requires estimation of normalized parameter and weights vector, and its accuracy is rarely consistent.
It requires that match scores be of the same nature.
It assumes comparable scales and strengths for input match scores
[88,93,94,95,96]
Likelihood RatioIt has the potential to converge to maxima when maximizing the likelihood. It is able to handle discrete values in the score distribution.
It does not involve the normalization of the score vector but the transformation of its respective likelihood ratio.
If the densities of the scores are accurate, an optimum level is reached at any desired value of false acceptance rate (FAR).
Requires detailed modeling of score distributions.
It is complex to implement due to the estimation of densities and is computationally complicated.
It is very time-consuming as it involves a large amount of training samples.
Requires a high knowledge of statistical techniques.
[89,90,97,98,99,100]
Fusion by classifierIt increases the overall accuracy of predictions by combining the strengths of different algorithms and reducing their weaknesses.
Reduces the bias inherent in any algorithm and achieves greater flexibility by adapting to different data patterns.
Data fusion usually requires more data to train and validate the classifiers.
It can increase computational complexity, especially if the data sets are large.
Convergence to local and global minima is related to the type of classifier to be used.
[101,102,103,104,105]
Alpha Integration It integrates many classic fusion operators and classifiers, optimizing fusion parameters and achieving more results that are accurate.Optimizing the parameters is done by the gradient method, which may not converge to the global optimum. This method would inherit the weaknesses of the optimization method used.[10,26,27,91,106]
Behavior-Knowledge SpaceIt does not depend on a prior hypothesis, such as statistical independence between classifier outputs.
It allows the creation of a knowledge model that can organize and represent the knowledge extracted from the classifiers’ prior behavior.
A limitation of this model is that with increasing data size the memory requirements increase exponentially.[107,108,109,110]

4.2. Late Fusion from Decisions (Late Hard Fusion)

Decision-level fusion aims to combine the decisions made by different classifiers to reach a common consensus and obtain a more accurate decision. Figure 8 shows where late fusion is performed at the decision level.
Majority Voting, Bagging, Boosting and Copula fusion are the most commonly used decision-level fusion methods. Table 4 below presents a description of these techniques, as well as the work related to these methods and a comparison taking into account their strengths and weaknesses.
  • ✓ Majority Voting.
This type of method is based on a weighted process that combines the decisions provided by the classifiers. The simplest and most intuitive approach is based on voting by the predominant class, i.e., assigning a sample based on the most frequent class. In case of a tie, the sample is not classified. Protective approaches derive from applying specific limits, meaning that a sample is assigned only if the frequency of assignments to a class is greater than a selected threshold [111,112].
  • ✓ Bagging.
It was originally introduced in 1996 [113] and expanded by [114]. It is based on sampling in its simplest form. The idea here is to sample subsets of the training set, generating independent bootstrap random replicates. The classifier is then constructed on each of these bootstrap samples, and finally, the constituent classifiers are aggregated by a simple majority vote [115].
The strength of bagging lies in unstable classifiers. These can be sensitive to minor alterations in the data set. Training the same classifier with slightly different data can give substantially different classifiers. Bootstrap sampling provides small random perturbations of the data set [116].
  • ✓ Boosting.
Boosting has been proposed and perfected in the works of [117] which led to the most successful implementation called AdaBoost. Unlike the bagging method, which is based on random changes of the bootstrap sampling, the boosting method relies on classifiers built on weighted versions of the training set, which depend on previous classification results [115].
Generally, when there is little noise, boosting offers a reasonably accurate classifier; however, for situations with substantial noise, it is better to use other techniques [118,119] as previously shown.
  • ✓ Copula fusion.
The theory of copulas relates multivariate distribution functions to their univariate marginal distribution functions and is explained by Sklar’s theorem [120].
Considering an m-dimensional distribution function with marginal distribution functions F 1 ,   ,   F m , a copula function C exists for all continuous random variables x 1 ,   , x m .
F x 1 ,   , x m = C F 1 x 1 ,   , F m x m
The function C is nothing more than a joint probability distribution of uniformly distributed random variables in the interval [0, 1].
Table 4. Comparison between late decision-level fusion methods and related works.
Table 4. Comparison between late decision-level fusion methods and related works.
Decision-Level Fusion MethodsStrengthsWeaknessesWorks
Majority VotingSince this method is based on the linear combination of multiple detection algorithms, errors or misclassifications of one model do not affect the result. The excellent performance of the others can compensate for the poor performance of one classifier. It allows the results to be more robust and prone to overfitting. It does not take into account the accuracy of the individual predictions of each classifier. If one classifier is more accurate, it will not have more influence than a less accurate one. Therefore, the result may be erratic.
It should also be noted that the computational complexity could be high.
[111,121,122,123]
BaggingReduces variance and, in many cases, improves the accuracy of some predictors, especially if individual classifiers are prone to bias.
Increases stability and eliminates the problem of overfitting for large amounts of data.
Introduces a loss of model interpretability; may experience biases when proper procedure is ignored.
This method involves training and maintaining several models, which can significantly increase computational requirements compared to a single model.
[113,114,115,119]
BoostingReduces variance and bias. Can generate a combined model that minimizes errors by avoiding the drawbacks of individual models.
Weights those classifiers with better performance on the training data. Therefore, the accuracy of the model generally tends to improve.
It does not help solve the overfitting problem; on the contrary, it may increase it for large data sets.
The computational complexity of the Boosting method can be considerable, especially in terms of training time and storage space. The number of iterations and the complexity of the base classifier are vital factors affecting complexity.
[115,117,118,119,124]
Copula fusionIt helps improve the fusion model’s accuracy and generalization, especially when there are uncertain classifier outputs.
If the choice and settings of copulas are correct, it converges to parameters that adequately represent the dependence between classifier outputs.
Depending on the copulas’ complexity, the method may require complex computational calculations.
Having sufficiently large data sets to train and evaluate the machine learning models is advisable.
[125,126,127]

5. Conclusions

A comparative study on recent automatic data fusion methods has been presented. The fusion approaches were divided in four categories for ease of comparison: early fusion from sensors; early fusion from features; late fusion from scores (late soft fusion); and late fusion from decisions (late hard fusion). A theoretical analysis of the conditions for applying early or late fusion was included. It was demonstrated that if the available data samples are sufficient, i.e., if the a posteriori probability or score is known precisely, late fusion will not be able to outperform early fusion. Early fusion can lead to the curse of dimensionality, which can be alleviated by using dimension reduction methods such as principal component analysis. On the other hand, late fusion is the practical option in real scenarios, and it was demonstrated that late hard fusion will not be able to outperform late soft fusion. Nevertheless, determining the best option in a specific application context requires a strong experimental component. The specific comparison of weaknesses and strengthens of the state-of-the-art methods shows that, generally, the computational complexity is usually higher for early fusion, as it involves integrating information in the early stages of the data flow. Optimization may be easier to achieve in late fusion, as it focuses on scores and decisions, and accuracy, which may be more robust for specific contexts despite the loss of information. In summary, the choice of method to apply requires a thorough understanding of the context of the problem.
As a general conclusion, it is evident that fusion is a current topic in machine learning with significant room for continued research. Specifically, it is possible to suggest some lines of work. Firstly, a good part of the experimental work and most of the theoretical contributions only consider the two-class problem, the interest of an extension to an arbitrary number of classes being obvious. Along these lines, it should be mentioned that there is a great need for theoretical analyses that allow a better understanding of the very high number of experimental works existing in this area. For example, it would be most convenient to have an analysis of the different kinds of fusion when a posteriori probabilities are unknown and have to be estimated, as well as including in the theoretical analysis other factors such as computational complexity for real-time implementation. Another possible line of research could be called “fusion of fusions”, that is, we could combine different levels of fusion, e.g., the output from an early feature fuser could be fused with the output of a soft late fuser. This could lead to the definition of complex fusion schemes. In a sense this is related to fusion in the context of deep learning, where the system itself can learn the best combination at different levels of the elements to be fused.

Author Contributions

The three researchers’ contributions to this article are made equally in the following points: comparison between early fusion and late fusion, review of fusion methods at different levels, and literature search of works related to data fusion. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Innovation of the Government of Spain MCIN/AEI/10.13039/501100011033 under Grants PRE2018-085092 and TEC2017-84743-P.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fung, M.L.; Chen, M.Z.Q.; Chen, Y.H. Sensor fusion: A review of methods and applications. In Proceedings of the 29th Chinese Control and Decision Conference, CCDC 2017, Chongqing, China, 28–30 May 2017; pp. 3853–3860. [Google Scholar]
  2. Usa, H.; Escamilla-Ambrosio, P.J.; Escamilla, J. Hybrid Kalman Filter-Fuzzy Logic Adaptive Multisensor Data Fusion Architectures. In Proceedings of the 42nd IEEE International Conference on Decision and Control, Maui, HI, USA, 9–12 December 2003. [Google Scholar]
  3. Vergara, L.; Soriano, A.; Safont, G.; Salazar, A. On the fusion of non-independent detectors. Digit. Signal Process. 2016, 50, 24–33. [Google Scholar] [CrossRef]
  4. Hang, G.; Min, Y. Data fusion in distributed multi-sensor system. Geo-Spat. Inf. Sci. 2012, 7, 214–217. [Google Scholar] [CrossRef]
  5. Bloch, I. Information combination operators for data fusion: A comparative review with classification. IEEE Trans. Syst. 1996, 26, 52–67. [Google Scholar] [CrossRef]
  6. Jollife, I.T.; Cadima, J. Principal component analysis: A review and recent developments. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 2016, 374, 2065. [Google Scholar] [CrossRef] [PubMed]
  7. Pereira, L.M.; Salazar, A.; Vergara, L. On Comparing Early and Late Fusion Methods. In Proceedings of the 17th International Work-Conference on Artificial Neural Networks, Ponta Delgada, Portugal, 19–21 June 2023; pp. 365–378. [Google Scholar]
  8. Ruta, D.; Gabrys, B. An overview of classifier fusion methods. Comput. Inf. Syst. 2000, 7, 1–10. [Google Scholar]
  9. Nesa, N.; Banerjee, I. IoT-Based Sensor Data Fusion for Occupancy Sensing Using Dempster-Shafer Evidence Theory for Smart Buildings. IEEE Internet Things J. 2017, 4, 1563. [Google Scholar] [CrossRef]
  10. Drakopoulos, E.; Lee, C.C. Optimum fusion of correlated local decisions. In Proceedings of the 27th IEEE Conference on Decision and Control, Austin, TX, USA, 7–9 December 1988; Volume 3, pp. 2489–2494. [Google Scholar]
  11. Atrey, P.; Hossain, M.; El Saddik, A.; Kankanhalli, M. Multimodal fusion for multimedia analysis: A survey. Multimed. Syst. 2010, 16, 345–379. [Google Scholar] [CrossRef]
  12. Yuksel, S.E.; Wilson, J.N.; Gader, P.D. Twenty Years of Mixture of Experts. IEEE Trans. Neural Netw. Learn. Syst. 2012, 23, 1177–1193. [Google Scholar] [CrossRef]
  13. Hassan, L.; Saleh, A.; Singh, V.K.; Puig, D.; Abdel-Nasser, M. Detecting Breast Tumors in Tomosynthesis Images Utilizing Deep Learning-Based Dynamic Ensemble Approach. Computers 2023, 12, 220. [Google Scholar] [CrossRef]
  14. Psallidas, T.; Spyrou, E. Video Summarization Based on Feature Fusion and Data Augmentation. Computers 2023, 12, 186. [Google Scholar] [CrossRef]
  15. Jebur, S.A.; Hussein, K.A.; Hoomod, H.K.; Alzubaidi, L. Novel Deep Feature Fusion Framework for Multi-Scenario Violence Detection. Computers 2023, 12, 175. [Google Scholar] [CrossRef]
  16. Tan, N.D.; Nguyen, D.N.; Hoang, H.N.; Le, T.T. EEGT: Energy Efficient Grid-Based Routing Protocol in Wireless Sensor Networks for IoT Applications. Computers 2023, 12, 103. [Google Scholar] [CrossRef]
  17. Qi, G.; Hu, G.; Mazur, N.; Liang, H.; Haner, M. A Novel Multi-Modality Image Simultaneous Denoising and Fusion Method Based on Sparse Representation. Computers 2021, 10, 129. [Google Scholar] [CrossRef]
  18. Adjobo, E.C.; Mahama, A.T.S.; Gouton, P.; Tossa, J. Towards Accurate Skin Lesion Classification across All Skin Categories Using a PCNN Fusion-Based Data Augmentation Approach. Computers 2022, 11, 44. [Google Scholar] [CrossRef]
  19. Planke, L.J.; Gardi, A.; Sabatini, R.; Kistan, T.; Ezer, N. Online Multimodal Inference of Mental Workload for Cognitive Human Machine Systems. Computers 2021, 10, 81. [Google Scholar] [CrossRef]
  20. Leghar, M.; Memon, S.; Dhomeja, L.D.; Jalbani, A.H.; Chandio, A.A. Deep Feature Fusion of Fingerprint and Online Signature for Multimodal Biometrics. Computers 2021, 10, 21. [Google Scholar] [CrossRef]
  21. Kumar, N.; Gumhold, S. FuseVis: Interpreting Neural Networks for Image Fusion Using Per-Pixel Saliency Visualization. Computers 2020, 9, 98. [Google Scholar] [CrossRef]
  22. Hall, D.L.; Llinas, J. Handbook of Multisensor Data Fusion; CRC Press: Boca Raton, FL, USA, 2001. [Google Scholar]
  23. Adams, W.H.; Iyengar, G.; Lin, C.-Y.; Naphade, M.R.; Neti, C.; Nock, H.J.; Smith, J.R. Semantic Indexing of Multimedia Content Using Visual, Audio, and Text Cues. EURASIP J. Adv. Signal Process. 2003, 2003, 170–185. [Google Scholar] [CrossRef]
  24. Sridharan, H.; Sundaram, H.; Rikakis, T. Computational models for experiences in the arts, and multimedia. In Proceedings of the 2003 ACM SIGMM Workshop on Experiential Telepresence, New York, NY, USA, 7 November 2003; pp. 31–44. [Google Scholar]
  25. Soriano, A.; Vergara, L.; Ahmed, B.; Salazar, A. Fusion of scores in a detection context based on Alpha integration. Neural Comput. 2015, 27, 1983–2010. [Google Scholar] [CrossRef]
  26. Safont, G.; Salazar, A.; Vergara, L. Multiclass Alpha Integration of Scores from Multiple Classifiers. Neural Comput. 2019, 31, 806–825. [Google Scholar] [CrossRef]
  27. Safont, G.; Salazar, A.; Vergara, L. Vector score alpha integration for classifier late fusion. Pattern Recognit. Lett. 2020, 136, 48–55. [Google Scholar] [CrossRef]
  28. Salazar, A.; Safont, G.; Vergara, L.; Vidal, E. Graph Regularization Methods in Soft Detector Fusion. IEEE Access, 2023; in press. [Google Scholar] [CrossRef]
  29. Salazar, A.; Vergara, L.; Vidal, E. A proxy learning curve for the Bayes classifier. Pattern Recognit. 2023, 136, 109240. [Google Scholar] [CrossRef]
  30. Pereira, L.; Salazar, A.; Vergara, L. A comparative analysis of early and late fusion for the multimodal two-class problem. IEEE Access 2023, 11, 84283–84300. [Google Scholar] [CrossRef]
  31. Hall, D.; Llinas, J. An introduction to multisensor data fusion. Proc. IEEE 1997, 85, 6–23. [Google Scholar] [CrossRef]
  32. Pereira, L.M.; Salazar, A.; Vergara, L. Simultaneous analysis of fMRI and EEG biosignals: A multimodal fusion approach. In Proceedings of the 2021 International Conference on Computational Science and Computational Intelligence (CSCI), Las Vegas, NV, USA, 15–17 December 2021; pp. 1673–1677. [Google Scholar]
  33. Sasiadek, J.Z. Sensor fusion. Annu. Rev. Control 2002, 26, 203–228. [Google Scholar] [CrossRef]
  34. Durrant-Whyte, H.; Henderson, T.C. Multisensor data fusion. Springer Handb. Robot. 2016, 35, 867–892. [Google Scholar]
  35. Abdulhafiz, W.A.; Khamis, A. Bayesian approach to multisensor data fusion with Pre- and Post-Filtering. In Proceedings of the 2013 10th IEEE International Conference on Networking, Sensing and Control, ICNSC 2013, Evry, France, 10–12 April 2013; pp. 373–378. [Google Scholar]
  36. Siaterlis, C.; Maglaris, B. Towards multisensor data fusion for DoS detection. Proc. ACM Symp. Appl. Comput. 2004, 1, 439–446. [Google Scholar]
  37. Bello, E. Lógica Difusa o Fuzzy Logic: Qué es y cómo funciona + Ejemplos. Think. Innov. Available online: https://www.iebschool.com/blog/fuzzy-logic-que-es-big-data/ (accessed on 4 December 2023).
  38. Zadeh, L.A. Outline of a New Approach to the Analysis of Complex Systems and Decision Processes. IEEE Trans. Syst. Man Cybern. 1973, SMC-3, 28–44. [Google Scholar] [CrossRef]
  39. Amin, M.; Akhoundi, A.; Valavi, E. Multi-Sensor Fuzzy Data Fusion Using Sensors with Different Characteristics. arXiv 2010, arXiv:1010.6096. [Google Scholar]
  40. Jiang, D.; Yang, X.; Clinton, N.; Wang, N. An artificial neural network model for estimating crop yields using remotely sensed information. Int. J. Remote Sens. 2004, 25, 9. [Google Scholar] [CrossRef]
  41. Matich, D.J. Redes Neuronales: Conceptos Básicos y Aplicaciones. 2001. Available online: https://www.frro.utn.edu.ar/repositorio/catedras/quimica/5_anio/orientadora1/monograias/matich-redesneuronales.pdf (accessed on 4 December 2023).
  42. Luwei, K.C.; Yunusa-Kaltungo, A.; Sha’aban, Y.A. Integrated Fault Detection Framework for Classifying Rotating Machine Faults Using Frequency Domain Data Fusion and Artificial Neural Networks. Machines 2018, 6, 59. [Google Scholar] [CrossRef]
  43. Elmore, P.A.; Petry, F.E.; Yager, R.R. Dempster–Shafer Approach to Temporal Uncertainty. IEEE Trans. Emerg. Top. Comput. Intell. 2017, 1, 316–325. [Google Scholar]
  44. Kalman, R.E. A New Approach to Linear Filtering and Prediction Problems. J. Basic Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef]
  45. Kalman, R.E.; Bucy, R.S. New Results in Linear Filtering and Prediction Theory. J. Basic Eng. 1961, 83, 95–108. [Google Scholar] [CrossRef]
  46. Moon, S.; Park, Y.; Ko, D.W.; Suh, I.H. Multiple Kinect Sensor Fusion for Human Skeleton Tracking Using Kalman Filtering. Int. J. Adv. Robot. Syst. 2016, 13, 65. [Google Scholar] [CrossRef]
  47. Yazdkhasti, S.; Sasiadek, J.Z. Multi Sensor Fusion Based on Adaptive Kalman Filtering. In Advances in Aerospace Guidance, Navigation and Control; Springer: Cham, Switzerland, 2018; pp. 317–333. [Google Scholar]
  48. Zhu, M.; Sui, T.; Wang, R.; Zhu, M.; Sui, T.; Wang, R. Distributed Kalman filtering over sensor networks with fading measurements and random link failures. J. Frankl. Inst. 2023, 360, 2457–2475. [Google Scholar] [CrossRef]
  49. Dempster, A.P. A Generalization of Bayesian Inference. J. R. Stat. Soc. Ser. B 1968, 30, 205–232. [Google Scholar] [CrossRef]
  50. Coninx, A.; Bessiere, P.; Mazer, E.; Droulez, J.; Laurent, R.; Aslam, M.A.; Lobo, J. Bayesian sensor fusion with fast and low power stochastic circuits. In Proceedings of the 2016 IEEE International Conference on Rebooting Computing (ICRC), San Diego, CA, USA, 17–19 October 2016. [Google Scholar]
  51. Coué, C.; Fraichard, T.; Bessière, P.; Mazer, E. Multi-sensor data fusion using Bayesian programming: An automotive application. IEEE Int. Conf. Intell. Robot. Syst. 2002, 1, 141–146. [Google Scholar]
  52. Ban, Y.; Alameda-Pineda, X.; Girin, L.; Horaud, R. Variational Bayesian Inference for Audio-Visual Tracking of Multiple Speakers. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 5. [Google Scholar] [CrossRef]
  53. Stover, J.A.; Hall, D.L.; Gibson, R.E. A fuzzy-logic architecture for autonomous multisensor data fusion. IEEE Trans. Ind. Electron. 1996, 43, 403–410. [Google Scholar] [CrossRef]
  54. Zhu, J.; Cao, H.; Shen, J.; Liu, H. Data fusion for magnetic sensor based on fuzzy logic theory. In Proceedings of the 2011 Fourth International Conference on Intelligent Computation Technology and Automation, ICICTA 2011, Shenzhen, China, 28–29 March 2011; Volume 1, pp. 87–92. [Google Scholar]
  55. Ren, X.; Li, C.; Ma, X.; Chen, F.; Wang, H.; Sharma, A.; Gaba, G.S.; Masud, M. Design of Multi-Information Fusion Based Intelligent Electrical Fire Detection System for Green Buildings. Sustainability 2021, 13, 3405. [Google Scholar] [CrossRef]
  56. McCulloch, W.S.; Pitts, W. A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 1943, 5, 115–133. [Google Scholar] [CrossRef]
  57. Quan, Y.; Zhou, M.C.; Luo, Z. On-line robust identification of tool-wear via multi-sensor neural-network fusion. Eng. Appl. Artif. Intell. 1998, 11, 717–722. [Google Scholar] [CrossRef]
  58. Lee, J.; Steele, C.M.; Chau, T. Swallow segmentation with artificial neural networks and multi-sensor fusion. Med. Eng. Phys. 2009, 31, 1049–1055. [Google Scholar] [CrossRef] [PubMed]
  59. Kańtoch, E. Human activity recognition for physical rehabilitation using wearable sensors fusion and artificial neural networks. Comput. Cardiol. 2010, 44, 1–4. [Google Scholar]
  60. Li, H.-C.; Hu, W.-S.; Li, W.; Li, J.; Du, Q.; Plaza, A. A 3 CLNN: Spatial, Spectral and Multiscale Attention ConvLSTM Neural Network for Multisource Remote Sensing Data Classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 747–761. [Google Scholar] [CrossRef]
  61. Jiaming, L.; Suhuai, L.; Jesse, S.J. Sensor Data Fusion for Accurate Cloud Presence Prediction Using Dempster-Shafer Evidence Theory. Sensors 2010, 10, 9384–9396. [Google Scholar]
  62. Yu, X.; Zhang, F.; Zhou, L.; Liu, Q. Novel Data Fusion Algorithm Based on Event-Driven and Dempster–Shafer Evidence Theory. Wirel. Pers. Commun. 2018, 100, 1377–1391. [Google Scholar] [CrossRef]
  63. Sahu, D.; Parsai, M.P. Different Image Fusion Techniques–A Critical Review. Int. J. Mod. Eng. Res. 2012, 2, 4298–4301. [Google Scholar]
  64. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  65. Kleibergen, F.; Paap, R. Generalized reduced rank tests using the singular value decomposition. J. Econ. 2006, 133, 97–126. [Google Scholar] [CrossRef]
  66. Klema, V.; Laub, A. The singular value decomposition: Its computation and some applications. IEEE Trans. Autom. Control. 1980, 25, 164–176. [Google Scholar] [CrossRef]
  67. Maruyama, K.; Sheng, Y.; Watanabe, H.; Fukuzawa, K.; Tanaka, S. Application of singular value decomposition to the inter-fragment interaction energy analysis for ligand screening. Comput. Theor. Chem. 2018, 1132, 23–34. [Google Scholar] [CrossRef]
  68. Saeed, N.; Nam, H.; Haq, M.I.U.; Bhatti, D.M.S. A Survey on Multidimensional Scaling. ACM Comput. Surv. 2018, 51, 3. [Google Scholar] [CrossRef]
  69. Goldstone, R.L.; Medin, D.L.; Gentner, D. Relational similarity and the nonindependence of features in similarity judgments. Cogn. Psychol. 1991, 23, 222–262. [Google Scholar] [CrossRef] [PubMed]
  70. He, Q.-H.; Feng, J.-J.; Lv, F.-J.; Jiang, Q.; Xiao, M.-Z. Deep learning and radiomic feature-based blending ensemble classifier for malignancy risk prediction in cystic renal lesions. Insights Imaging 2023, 14, 6. [Google Scholar] [CrossRef] [PubMed]
  71. Wang, K.; Xu, C.; Li, G.; Zhang, Y.; Zheng, Y.; Sun, C. Combining convolutional neural networks and self-attention for fundus diseases identification. Sci. Rep. 2023, 13, 76. [Google Scholar] [CrossRef] [PubMed]
  72. Koshy, R.; Elango, S. Multimodal tweet classification in disaster response systems using transformer-based bidirectional attention model. Neural Comput. Appl. 2022, 35, 1607–1627. [Google Scholar] [CrossRef]
  73. Jing, J.; Wu, H.; Sun, J.; Fang, X.; Zhang, H. Multimodal fake news detection via progressive fusion networks. Inf. Process. Manag. 2023, 60, 103120. [Google Scholar] [CrossRef]
  74. Hotelling, H. Analysis of a complex of statistical variables into principal components. J. Educ. Psychol. 1933, 24, 417–441. [Google Scholar] [CrossRef]
  75. Hasan, M.M.; Islam, N.; Rahman, M.M. Gastrointestinal polyp detection through a fusion of contourlet transform and Neural features. J. King Saud Univ. Comput. Inf. Sci. 2022, 34, 526–533. [Google Scholar] [CrossRef]
  76. Huang, Z.K.; Li, P.W.; Hou, L.Y. Segmentation of textures using PCA fusion based Gray-Level Co-Occurrence Matrix features. In Proceedings of the 2009 International Conference on Test and Measurement, Hong Kong, China, 5–6 December 2009; Volume 1, pp. 103–105. [Google Scholar]
  77. Nasir, H.; Stanković, V.; Marshall, S. Singular value decomposition based fusion for super-resolution image reconstruction. Signal Process. Image Commun. 2012, 27, 180–191. [Google Scholar] [CrossRef]
  78. Zhao, X.; Ye, B. Singular value decomposition packet and its application to extraction of weak fault feature. Mech. Syst. Signal Process. 2016, 70–71, 73–86. [Google Scholar] [CrossRef]
  79. Zhu, H.; He, Z.; Wei, J.; Wang, J.; Zhou, H. Bearing Fault Feature Extraction and Fault Diagnosis Method Based on Feature Fusion. Sensors 2021, 21, 2524. [Google Scholar] [CrossRef] [PubMed]
  80. Ye, X.; Gao, W.; Yan, Y.; Osadciw, L.A. Multiple Tests for Wind Turbine Fault Detection and Score Fusion Using Two-Level Multidimensional Scaling (MDS); SPIE: Washington, DC, USA, 2010; Volume 7704, pp. 70–77. [Google Scholar]
  81. Tian, G.Y.; Taylor, D. Colour image retrieval using virtual reality. In Proceedings of the 2000 IEEE Conference on Information Visualization, London, UK, 19–21 July 2000; Volume 2000, pp. 221–225. [Google Scholar]
  82. Choo, J.; Bohn, S.; Nakamura, G.C.; White, A.M.; Park, H. Heterogeneous Data Fusion via Space Alignment Using Nonmetric Multidimensional Scaling; SIAM: Philadelphia, PA, USA, 2012; pp. 177–188. [Google Scholar]
  83. Entezami, A.; Sarmadi, H.; Salar, M.; De Michele, C.; Arslan, A.N. A novel data-driven method for structural health monitoring under ambient vibration and high-dimensional features by robust multidimensional scaling. Struct. Health Monit. 2021, 20, 2758–2777. [Google Scholar] [CrossRef]
  84. Alam, N.-A.; Ahsan, M.; Based, A.; Haider, J.; Kowalski, M. COVID-19 Detection from Chest X-ray Images Using Feature Fusion and Deep Learning. Sensors 2021, 21, 1480. [Google Scholar] [CrossRef] [PubMed]
  85. Salazar, A.; Safont, G.; Soriano, A.; Vergara, L. Automatic Credit Card Fraud Detection based on Non-linear Signal Processing. In Proceedings of the 2012 IEEE International Carnahan Conference on Security Technology (ICCST), Newton, MA, USA, 15–18 October 2012; pp. 207–212. [Google Scholar]
  86. Salazar, A.; Safont, G.; Vergara, L. Surrogate Techniques for Testing Fraud Detection Algorithms in Credit Card Operations. In Proceedings of the 2014 International Carnahan Conference on Security Technology (ICCST), Rome, Italy, 13–16 October 2014; pp. 124–129. [Google Scholar]
  87. Vergara, L.; Salazar, A.; Belda, J.; Safont, G.; Moral, S.; Iglesias, S. Signal Processing on Graphs for Improving Automatic Credit Card Fraud Detection. In Proceedings of the 2017 International Carnahan Conference on Security Technology (ICCST), Madrid, Spain, 23–26 October 2017; pp. 1–6. [Google Scholar]
  88. Ezichi, S.I.; Ezika, I.J.F.; Nkpume, C.; Iloanusi, O.N. Biometric Security: A Review of the Sum Rule and the Likelihood Ratio Fusion Algorithms for Multibiometric Systems. In Proceedings of the 2020 LGT-ECE-UNN International Conference: Technological Innovation for Holistic Sustainable Development, Nsukka, Nigeria, 21–22 September 2020. [Google Scholar]
  89. Tran, L.B.; Le, T.H. Multimodal personal verification using likelihood ratio for the match score fusion. Comput. Intell. Neurosci. 2017, 2017, 9345969. [Google Scholar] [CrossRef]
  90. Ishihara, S.; Carne, M. Likelihood ratio estimation for authorship text evidence: An empirical comparison of score- and feature-based methods. Forensic Sci. Int. 2022, 334, 111268. [Google Scholar] [CrossRef]
  91. Amari, S. Integration of Stochastic Models by Minimizing α-Divergence Shun-ichi Amari. Neural Comput. 2007, 19, 10. [Google Scholar] [CrossRef]
  92. Tulyakov, S.; Jaeger, S.; Govindaraju, V.; Doermann, D. Review of classifier combination methods. Stud. Comput. Intell. 2008, 90, 361–386. [Google Scholar]
  93. Hube, J.P. Neyman-Pearson Biometric Score Fusion as an Extension of the Sum Rule; SPIE: Washington, DC, USA, 2007; Volume 6539, pp. 200–208. [Google Scholar]
  94. Hammouche, R.; Attia, A.; Akhrouf, S. Score level fusion of major and minor finger knuckle patterns based symmetric sum-based rules for person authentication. Evol. Syst. 2022, 13, 469–483. [Google Scholar] [CrossRef]
  95. Kittler, J.; Hatef, M.; Duin, R.P.W.; Matas, J. On combining classifiers. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 226–239. [Google Scholar] [CrossRef]
  96. Eskandari, M.; Toygar, Ö. Fusion of face and iris biometrics using local and global feature extraction methods. Signal Image Video Process. 2014, 8, 995–1006. [Google Scholar] [CrossRef]
  97. Salazar, A.; Safont, G.; Rodriguez, A.; Vergara, L. Combination of Multiple Detectors for Credit Card Fraud Detection. In Proceedings of the 2016 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Limassol, Cyprus, 12–14 December 2016; pp. 138–143. [Google Scholar]
  98. Nanni, L.; Lumini, A.; Brahnam, S. Likelihood ratio based features for a trained biometric score fusion. Expert Syst. Appl. 2011, 38, 58–63. [Google Scholar] [CrossRef]
  99. Zafar, R.; Dass, S.C.; Malik, A.S.; Kamel, N.; Rehman, M.J.U.; Ahmad, R.F.; Abdullah, J.M.; Reza, F. Prediction of Human Brain Activity Using Likelihood Ratio Based Score Fusion. IEEE Access 2017, 5, 13010–13019. [Google Scholar] [CrossRef]
  100. Sharma, V.; Bains, M.; Verma, R.; Verma, N.; Kumar, R. Novel use of logistic regression and likelihood ratios for the estimation of gender of the writer from a database of handwriting features. Aust. J. Forensic Sci. 2021, 55, 89–106. [Google Scholar] [CrossRef]
  101. Fierrez-Aguilar, J.; Ortega-Garcia, J.; Garcia-Romero, D.; Gonzalez-Rodriguez, J. A comparative evaluation of fusion strategies for multimodal biometric verification. Lect. Notes Comput. Sci. 2003, 2688, 830–837. [Google Scholar]
  102. Arigbabu, O.A.; Ahmad, S.M.S.; Adnan, W.A.W.; Yussof, S. Integration of multiple soft biometrics for human identification. Pattern Recognit. Lett. 2015, 68, 278–287. [Google Scholar] [CrossRef]
  103. Salazar, A.; Prieto, J.R.; Vidal, E.; Safont, G.; Vergara, L. Fusion of Visual and Textual Features for Table Header Detection in Handwritten Text Images. In Proceedings of the 2021 International Conference on Computational Science and Computational Intelligence (CSCI), Las Vegas, NV, USA, 15–17 December 2021. [Google Scholar]
  104. Ma, Y.; Cukic, B.; Singh, H. A Classification Approach to Multi-biometric Score Fusion. In Proceedings of the 5th International Conference Audio- and Video-Based Biometric Person Authentication, Hilton Rye Town, NY, USA, 20–22 July 2005; pp. 484–493. [Google Scholar]
  105. Fu, Q.; Ding, X.Q.; Li, T.Z.; Liu, C.S. An Effective and Practical Classifier Fusion Strategy for Improving Handwritten Character Recognition. In Proceedings of theInternational Conference on Document Analysis and Recognition, Curitiba, Brazil, 23–26 September 2007. [Google Scholar]
  106. Salazar, A.; Safont, G.; Vergara, L.; Vidal, E. Pattern recognition techniques for provenance classification of archaeological ceramics using ultrasounds. Pattern Recognit. Lett. 2020, 135, 441–450. [Google Scholar] [CrossRef]
  107. Huang, Y.S.; Suen, C.Y. The behavior-knowledge space method for combination of multiple classifiers. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 15–17 June 1993. [Google Scholar]
  108. Ferreira, A.; Felipussi, S.C.; Alfaro, C.; Fonseca, P.; Vargas-Muñoz, J.E.; dos Santos, A.; Rocha, A. Behavior Knowledge Space-Based Fusion for Copy-Move Forgery Detection. IEEE Trans. Image Process. 2016, 25, 4729–4742. [Google Scholar] [CrossRef]
  109. Zhang, X.; Dong, G.; Ramamohanarao, K. Building behaviour knowledge space to make classification decision. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Hong Kong, China, 16–18 April 2001; Volume 2035, pp. 488–494. [Google Scholar]
  110. Dainotti, A.; Pescapé, A.; Sansone, C.; Quintavalle, A. Using a behaviour knowledge space approach for detecting unknown IP traffic flows. Int. Work. Mult. Classif. Syst. 2011, 6713, 360–369. [Google Scholar]
  111. Ballabio, D.; Todeschini, R.; Consonni, V. Recent Advances in High-Level Fusion Methods to Classify Multiple Analytical Chemical Data. Data Handl. Sci. Technol. 2019, 31, 129–155. [Google Scholar]
  112. Singh, L.; Janghel, R.R.; Sahu, S.P. A hybrid feature fusion strategy for early fusion and majority voting for late fusion towards melanocytic skin lesion detection. Int. J. Imaging Syst. Technol. 2022, 32, 1231–1250. [Google Scholar] [CrossRef]
  113. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
  114. Breiman, L. Arcing classifier (with discussion and a rejoinder by the author). Ann. Stat. 1998, 26, 801–849. [Google Scholar] [CrossRef]
  115. Ćwiklińska-Jurkowska, M. Boosting, bagging and fixed fusion methods performance for aiding diagnosis. Biocybern. Biomed. Eng. 2012, 32, 17–31. [Google Scholar]
  116. Kuncheva, L.I.; Skurichina, M.; Duin, R.P.W. An experimental study on diversity for bagging and boosting with linear classifiers. Inf. Fusion 2002, 3, 245–258. [Google Scholar] [CrossRef]
  117. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. Lect. Notes Comput. Sci. 1995, 904, 23–37. [Google Scholar] [CrossRef]
  118. Ferreira, A.J.; Figueiredo, M.A.T. Boosting Algorithms: A Review of Methods, Theory, and Applications. In Ensemble Machine Learning; Springer: New York, NY, USA, 2012; pp. 35–85. [Google Scholar]
  119. Dietterich, T.G. Experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting, and randomization. Mach. Learn. 2000, 40, 139–157. [Google Scholar] [CrossRef]
  120. Nelsen, R. An Introduction to Copulas; Springer: New York, NY, USA, 2006. [Google Scholar]
  121. Ruta, D.; Gabrys, B. Classifier selection for majority voting. Inf. Fusion 2005, 6, 63–81. [Google Scholar] [CrossRef]
  122. Jimenez, L.O.; Morales-Morell, A. Classification of hyperdimensional data based on feature and decision fusion approaches using projection pursuit, majority voting, and neural networks. IEEE Trans. Geosci. Remote Sens. 1999, 37, 1360–1366. [Google Scholar] [CrossRef]
  123. Özçift, A.; Bozuyla, M. Majority Vote Decision Fusion System to Assist Automated Identification of Vertebral Column Pathologies. Celal Bayar Univ. J. Sci. 2023, 19, 53–65. [Google Scholar]
  124. Louk, M.H.L.; Tama, B.A. Dual-IDS: A bagging-based gradient boosting decision tree model for network anomaly intrusion detection system. Expert Syst. Appl. 2023, 213, 119030. [Google Scholar] [CrossRef]
  125. Sundaresan, A.; Varshney, P.K.; Rao, N.S.V.; Sundaresan, A.; Varshney, P.K. Copula-Based Fusion of Correlated Decisions. IEEE Trans. Aerosp. Electron. Syst. 2011, 47, 454–471. [Google Scholar] [CrossRef]
  126. Zhang, S.; Theagarajan, L.N.; Choi, S.; Varshney, P.K. Fusion of Correlated Decisions Using Regular Vine Copulas. IEEE Trans. Signal Process. 2019, 67, 2066–2079. [Google Scholar] [CrossRef]
  127. Ni, L.; Zhang, D.; Wang, Z.; Liang, J.; Zhuang, J.; Wan, Q. Fast copula-based fusion of correlated decisions for distributed radar detection. Signal Process. 2022, 201, 108676. [Google Scholar] [CrossRef]
Figure 1. Decision boundaries established by different classifiers and the fusion of the classifiers.
Figure 1. Decision boundaries established by different classifiers and the fusion of the classifiers.
Computers 13 00013 g001
Figure 2. Improving detection capabilities by fusing information.
Figure 2. Improving detection capabilities by fusing information.
Computers 13 00013 g002
Figure 3. Outline of a data fusion pipeline in a detection system.
Figure 3. Outline of a data fusion pipeline in a detection system.
Computers 13 00013 g003
Figure 4. Diagram of fusion levels.
Figure 4. Diagram of fusion levels.
Computers 13 00013 g004
Figure 6. Early feature-level fusion scheme.
Figure 6. Early feature-level fusion scheme.
Computers 13 00013 g006
Figure 7. Late score-level fusion scheme.
Figure 7. Late score-level fusion scheme.
Computers 13 00013 g007
Figure 8. Late decision-level fusion scheme.
Figure 8. Late decision-level fusion scheme.
Computers 13 00013 g008
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pereira, L.M.; Salazar, A.; Vergara, L. A Comparative Study on Recent Automatic Data Fusion Methods. Computers 2024, 13, 13. https://doi.org/10.3390/computers13010013

AMA Style

Pereira LM, Salazar A, Vergara L. A Comparative Study on Recent Automatic Data Fusion Methods. Computers. 2024; 13(1):13. https://doi.org/10.3390/computers13010013

Chicago/Turabian Style

Pereira, Luis Manuel, Addisson Salazar, and Luis Vergara. 2024. "A Comparative Study on Recent Automatic Data Fusion Methods" Computers 13, no. 1: 13. https://doi.org/10.3390/computers13010013

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop