Abstract
This paper focuses on data fusion, which is fundamental to one of the most important modules in any autonomous system: perception. Over the past decade, there has been a surge in the usage of smart/autonomous mobility systems. Such systems can be used in various areas of life like safe mobility for the disabled, senior citizens, and so on and are dependent on accurate sensor information in order to function optimally. This information may be from a single sensor or a suite of sensors with the same or different modalities. We review various types of sensors, their data, and the need for fusion of the data with each other to output the best data for the task at hand, which in this case is autonomous navigation. In order to obtain such accurate data, we need to have optimal technology to read the sensor data, process the data, eliminate or at least reduce the noise and then use the data for the required tasks. We present a survey of the current data processing techniques that implement data fusion using different sensors like LiDAR that use light scan technology, stereo/depth cameras, Red Green Blue monocular (RGB) and Time-of-flight (TOF) cameras that use optical technology and review the efficiency of using fused data from multiple sensors rather than a single sensor in autonomous navigation tasks like mapping, obstacle detection, and avoidance or localization. This survey will provide sensor information to researchers who intend to accomplish the task of motion control of a robot and detail the use of LiDAR and cameras to accomplish robot navigation.
Keywords:
datafusion; data fusion; multimodal; fusion; information fusion; survey; review; RGB; SLAM; localization; obstacle detection; obstacle avoidance; navigation; deep learning; neural networks; mapping; LiDAR; optical; vision; stereo vision; autonomous systems; data integration; data alignment; robot; mobile robot 1. Introduction
Autonomous systems can play a vital role in assisting humans in a variety of problem areas. This could potentially be in a wide range of applications like driver-less cars, humanoid robots, assistive systems, domestic systems, military systems, and manipulator systems, to name a few. Presently, the world is at a bleeding edge of technologies that can enable this even in our daily lives. Assistive robotics is a crucial area of autonomous systems that helps persons who require medical, mobility, domestic, physical, and mental assistance. This research area is gaining popularity in applications like autonomous wheelchair systems [1,2], autonomous walkers [3], lawn movers [4,5], vacuum cleaners [6], intelligent canes [7], and surveillance systems in places like assisted living [8,9,10,11].
Data are one of the most important components to optimally start, continue, or complete any task. Often, these data are obtained from the environment that the autonomous system functions in; examples of such data could be the system’s position and location coordinates in the environment, the static objects, speed/velocity/acceleration of the system or its peers or any moving object in its vicinity, vehicle heading, air pressure, and so on. Since this is obtained directly from the operational environment, the information is up-to-date and can be accessed through either built-in or connected sensing equipment/devices. This survey is focused on the vehicle navigation of an autonomous vehicle. We review the past and present research using Light Imaging Detection and Ranging (LiDAR) and Imaging systems like a camera, which are laser and vision-based sensors, respectively. The autonomous systems use sensor data for tasks like object detection, obstacle avoidance, mapping, localization, etc. As we will see in the upcoming sections, these two sensors can complement each other and hence are being used extensively for detection in autonomous systems. The LiDAR market alone is expected to reach USD Billion by the year 2032, as given in a recent survey by the Yole group, documented by “First Sensors” group [12].
In a typical autonomous system, a perception module inputs the optimal information into the control module. Refer Figure 1. Crowley et al. [13] define perception.
Figure 1.
High level Perception Architecture.
The process of maintaining an internal description of the external environment.
This paper is organized as follows:
This section, Section 1 introduces autonomous systems and how data fusion is used. Section 2 introduces data fusion, techniques, need and compares single vs. multi sensor fusion. Section 3 discusses some of hardware that could be used for autonomous navigation. Section 4 reviews autonomous vehicle navigation. It considers mapping, localization, and obstacle avoidance. Section 5 details how data fusion is used in autonomous navigation. Section 6 gives the conclusions after reviewing the present research.
2. Data Fusion
Data fusion entails combining information to accomplish something. This ’something’ is usually to sense the state of some aspect of the universe [14]. The applications of this ’state sensing’ are versatile, to say the least. Some high level areas are: neurology, biology, sociology, engineering, physics, and so on [15,16,17,18,19,20,21]. Due to the very versatile nature of the application of data fusion, throughout this manuscript, we will limit our review to the usage of data fusion using LiDAR data and camera data for autonomous navigation. More information about data fusion will be provided in the upcoming sections.
2.1. Sensors and Their Input to Perception
A sensor is an electronic device that measures physical aspects of an environment and outputs machine (a digital computer) readable data. They provide a direct perception of the environment they are implemented in. Typically, a suite of sensors is used since it is the inherent property of an individual sensor, in order to provide a single aspect of an environment. This not only enables the completeness of the data, but also improves the accuracy of measuring the environment.
The Merriam-Webster dictionary defines a sensor [22] as
The Collins dictionary defines a sensor as [23]:A device that responds to a physical stimulus (such as heat, light, sound, pressure, magnetism, or a particular motion) and transmits a resulting impulse (as for measurement or operating a control).
Many applications require multiple sensors to be present to achieve a task. This gives rise to the technique of data fusion, wherein the user will need to provide guidelines and rules for the best usage of the data that is given by the sensors. Several researchers have given their definition of data fusion. JDL’s definition of data fusion is quoted by Hall et al. [24] as:A sensor is an instrument which reacts to certain physical conditions or impressions such as heat or light, and which is used to provide information.
Stating that the JDL definition is too restrictive, Hall et al. [21,25,26] re-define data fusion as:A process dealing with the association, correlation, and combination of data and information from single and multiple sources to achieve refined position and identity estimates, and complete and timely assessments of situations and threats, and their significance. The process is characterized by continuous refinements of its estimates and assessments, and the evaluation of the need for additional sources, or modification of the process itself, to achieve improved results.
In addition to the sensors like LiDAR and Camera that are the focus in this survey, any sensor like sonar, stereo vision, monocular vision, radar, LiDAR, etc. can be used in data fusion. Data fusion at this high level will enable tracking moving objects as well, as given in the research conducted by Garcia et al. [27].Data fusion is the process of combining data or information to estimate or predict entity states.Data fusion involves combining data—in the broadest sense—to estimate or predict the state of some aspect of the universe.
The initial step is raw data capture using the sensors. The data is then filtered and an appropriate fusion technology implemented this is fed into localization and mapping techniques like SLAM; the same data can be used to identify static or moving objects in the environment and this data can be used to classify the objects, wherein classification information is used to finalize information in creating a model of the environment which in turn can be fed into the control algorithm [27]. The classification information could potentially give details of pedestrians, furniture, vehicles, buildings, etc. Such a classification is useful in both pre-mapped i.e., known environments and unknown environments since it increases the potential of the system to explore its environment and navigate.
- Raw Data sensing: LiDAR is the primary sensor due to its accuracy of detection and also the higher resolution of data and it is effective in providing the shape of the objects in the environment that may contain hazardous obstacles to the vehicle. A stereo vision sensor can provide depth information in addition to the LiDAR. The benefit of using this combination is the accuracy, speed, and resolution of the LiDAR and the quality and richness of data from the stereo vision camera. Together, these two sensors provide an accurate, rich, and fast data set for the object detection layer [18,28,29].In a recent study in 2019, Rövid et al. went a step further to utilize the raw data and fuse it to realize the benefits early on in the cycle [30]. They fused camera image data with LiDAR pointclouds closest to the raw level of data extraction and its abstraction.
- Object Detection: Object Detection is the method of locating an object of interest in the sensor output. LiDAR data scan objects differently in their environment than a camera. Hence, the methodology to detect objects in the data from these sensors would be different as well. The research community has used this technique to detect objects in aerial, ground, and underwater environments [30,31,32,33,34].
- Object Classification: The Objects are detected and then they are classified into several types so that they can be grouped into small, medium, and large objects, or hazard levels of nonhazardous or hazardous, such that the right navigation can be handled for the appropriate object. Chavez-Garcia et al. [35] fuse multiple sensors including camera and LiDAR to classify and track moving objects.
- Data Fusion: After the classification, the data are fused to finalize information as input to the control layer. The data fusion layer output will provide location information of the objects in the map of the environment, so that the autonomous vehicle can, for instance, avoid the obstacle or stop if the object is a destination or wait for a state to be reached for further action if the object is deemed a marker or milestone. The control segment will take the necessary action, depending on the behavior as sensed by the sensor suite [18,28,29,35,36,37].
2.2. Multiple Sensors vs. Single Sensor
It is a known fact that most of the autonomous systems require multiple sensors to function optimally. However, why should we use multiple sensors? Individual usage of any sensor could impact the system where they are used, due to the limitations in each of those sensors. Hence, to get acceptable results, one may utilize a suite of different sensors and utilize the benefits of each of them. The diversity offered by the suite of sensors contributes positively to the sensed data perception [38,39]. Another reason could be the system failure risk due to the failure of that single sensor [21,27,40] and hence one should introduce a level of redundancy. For instance, while executing the obstacle avoidance module, if the camera is the only installed sensor and it fails, it could be catastrophic. However, if it has an additional camera or LiDAR, it can navigate itself to a safe place after successfully avoiding the obstacle, if such logic is built-in for that failure. Roggen et al., Luo et al., and Foo et al. [41,42,43] performed a study on high-level decision data fusion and concluded that using multiple sensors with data fusion is better than individual sensors without data fusion. In addition to the above, several researchers [27,39,44,45,46] discovered that every sensor used provides a different type, sometimes unique type of information in the selected environment, which includes the tracked object, avoided object, the autonomous vehicle itself, the world it is being used, and so on and so forth, and the information is provided with differing accuracy and differing details.
There are some disadvantages while using multiple sensors and one of them is that they have additional levels of complexity; however, using an optimal technique for fusing the data can mitigate this challenge efficiently. When data are optimally combined, information from different views of the environment gives an accurate model of the environment the system is being used in.
The second was highlighted by Brooks et al. [47] who state:
A man with one clock knows what time it is. A man with two clocks is never sure!
That is, there may be the presence of a level of uncertainty in the functioning, accuracy, and appropriateness of the sensed raw data. Due to these challenges, the system must be able to diagnose accurately when a failure occurs and ensure that the failed component(s) are identified for apt mitigation. At a high level, we can term two types of sensor fusion: Homogeneous data fusion and Heterogeneous data fusion. As the name states, homogeneous data fusion comprises sensor data of the same types of sensors; there may or may not be the same make or model—for example, a stereo vision camera only, GPS data only, or LiDAR data only, etc. On the other hand, heterogeneous data fusion will have varied sensor data. There could be a suite of sensors like GPS, LiDAR, stereo vision camera or GPS and LiDAR or IMU and GPS, etc. In addition, it must be able to tolerate small differences between the same-sensor readings and be able to merge their small discrepancies into a single sensor reading that is reliable. This is done through data fusion, which we will address later. As an example, let us consider humans; redundancy is built into us, which is we have five different senses and among these senses, we have two eyes and two ears, and an entire body of skin that can sense. We use these senses subconsciously, i.e., without specifically instructing our brains to use them appropriately. This should be implemented purposefully, specifically and carefully into an autonomous system. The above-mentioned researchers [40,47] state that the information obtained by the intelligent system using a single sensor will tend to be incomplete and sometimes inaccurate, due to its inherent limitations and uncertainty.
Consider a graphical representation of a simple perception system as given in Figure 2. The system takes in as input, the sensor data of the perception sensors like LiDAR, sonar, camera, etc., and motion sensors like the odometric, navigational sensors, etc. The output comprises location, distance of the objects in the vicinity, and the current state of the robot to name a few. Although these outputs seem similar, details clearly state that they vary in many ways; for example, a vehicular motion sensor will not provide information about obstacles in front of the robot; a camera cannot provide details about the robot’s location like latitude and longitude, etc. (unless a GPS is built into the camera); and therefore a single sensor will not be able to provide all the information that is necessary to optimally perform the complete suite of tasks. Hence, we have the need to use multiple sensors that may be redundant but are complementary and can provide the information to the perception module in the intelligent system. Therefore, the perception module uses information from sensors like LiDAR, camera, sonar, etc. We will detail these sensors and the above-mentioned tasks in the following sections. Combining information from several sensors is a challenging problem [39,48,49].
Figure 2.
Concepts of perception.
Rao et al. [29] provide metrics comparing the difference(s) between single sensor and multi-sensors. They state that, if the distribution function depicting measurement errors of one sensor is precisely known, an optimal fusion process can be developed, and this fusion process performs similar to if not better than a single sensor. Users can be reassured that the fused data is better than that of a single sensor. Since the sensing layer is better now, the control application can be standardized independently.
2.3. Need for Sensor Data Fusion
Some of the limitations of single sensor unit systems are as follows:
- Deprivation: If a sensor stops functioning, the system where it was incorporated in will have a loss of perception.
- Uncertainty: Inaccuracies arise when features are missing, due to ambiguities or when all required aspects cannot be measured
- Imprecision: The sensor measurements will not be precise and will not be accurate.
- Limited temporal coverage: There are initialization/setup time to reach a sensor’s maximum performance and transmit a measurement, hence limiting the frequency of the maximum measurements.
- Limited spatial coverage: Normally, an individual sensor will cover only a limited region of the entire environment—for example, a reading from an ambient thermometer on a drone provides an estimation of the temperature near the thermometer and may fail to correctly render the average temperature in the entire environment.
The problems stated above can be mitigated by using a suite of sensors, either homogeneous or heterogeneous [38,44,46,50,51] in addition to mitigating the issues of the above data fusion. Some of the advantages of using multiple sensors or a sensor suite are as follows:
- Extended Spatial Coverage: Multiple sensors can measure across a wider range of space and sense where a single sensor cannot
- Extended Temporal Coverage: Time-based coverage increases while using multiple sensors
- Improved resolution: A union of multiple independent measurements of the same property, the resolution is better, i.e., more than that of single sensor measurement.
- Reduced Uncertainty: As a whole, when we consider the entire sensor suite, the uncertainty decreases, since the combined information reduces the set of unambiguous interpretations of the sensed value.
- Increased robustness against interference: An increase in the dimensionality of the sensor space (measuring using a LiDAR and stereo vision cameras), the system becomes less vulnerable against interference.
- Increased robustness: The redundancy that is provided due to the multiple sensors provides more robustness, even when there is a partial failure due to one of the sensors being down.
- Increased reliability: Due to the increased robustness, the system becomes more reliable.
- Increased confidence: When the same domain or property is measured by multiple sensors, one sensor can confirm the accuracy of other sensors; this can be attributed to re-verification and hence the confidence is better.
- Reduced complexity: The output of multiple sensor fusion is better; it has lesser uncertainty, is less noisy, and complete.
2.4. Levels of Data Fusion Application
Data fusion can be applied at various levels of data gathering or data grouping and are dependent on the abstraction levels of data. We will see in the upcoming sections the abstraction levels of data fusions. The abstraction levels of data fusion are:
- Decision or High-level data fusion. At the highest level, the system decides the major tasks and takes decisions based on the fusion of information, which is input from the system features [41,43].
- Feature or mid-level data fusion. At the feature level, feature maps containing lines, corners, edges, textures, and lines are integrated and decisions made for tasks like obstacle detection, object recognition, etc. [52,53,54].
- Raw-data or low-level data fusion. At this most basic or lowest level, better or improved data are obtained by integrating raw data directly from multiple sensors; such data can be used in tasks. This new combined raw data will contain more information than the individual sensor data. We have summarized the most common data fusion techniques and the benefits of using that technique as well [55].
The versatility involved in the implementation of data fusion can be realized by the above levels of application.
2.5. Data Fusion Techniques
Nature provides us sensing as one of its most important methods for survival in the animal or plant kingdom. In the animal kingdom, this can be seen as a seamless integration of data from various sources, some overlapping and some non-overlapping to output information which is reliable and feature-rich that can be used in fulfilling goals. In nature, this capability is most essential for survival, to hunt for food or to escape from being hunted. As an example in wildlife, consider bears and compare their sensory capabilities; they have a sharp color close-up vision but do not have a good long distant vision [56]. However, their hearing is excellent because they have the capability to hear in all directions. Their sense of smell is extremely good. They use their paws very dexterously to manipulate wide-ranging objects, from picking little blueberries to lifting huge rocks. Often, bears touch objects with their lips, noses, and tongue to feel them. Hence, we can surmise that their sense of touch is very good. Surely they combine signals from the five body senses i.e., sound, sight, smell, taste, and touch) with information of the environment they are in, and create and maintain a dynamic model of the world. At the time of need, for instance, when a predator is around, it prepares itself and takes decisions regarding the current and future actions [56]. Over the years, scientists and engineers have applied concepts of such fusion into technical areas and have developed new disciplines and technologies that span over several fields. They have developed systems with multiple sensors and devised mechanisms and techniques to augment the data from all the sensors and get the ’best’ data as output from this set of sensors, also known as a ’suite of sensors’. In short, this augmentation or integration of data from multiple sensors can simply be termed as multi-sensor data fusion.
Kanade et al. in the early 1980s used aerial sensor data to obtain passive sensor fusion of stereo vision imagery. Crowley et al. performed fundamental research in the area of data fusion, perception, and world model development that is vital for robot navigation [57,58,59]. They realized that data fusion needs to be applied incrementally in their perception problem [59]. They developed similar techniques [58] that used Kanade’s incremental approach to build a world model for robot navigation. They generalized fusion work and documented that, using cyclical processes, one can achieve good perception. Brooks developed a visual ad-hoc technique [60] that was used in robot perception.
Bayesian estimation theory was recommended by Smith et al. for robotic vision [61]. Whyte documented in his research thesis the derivation techniques for optimizing and integrating sensor information, that may be considered as extensions of estimation theory [62]. It was also implemented in a recent study about system noise [63]. Faugeras et al. performed stereo vision calibration using an adaptation of estimation theory as well [64].
The community witnessed a growth in the development of techniques that performed the minimization of a required energy function which provided quantitative measurements and constraints and calculates how much the measurements and constraints are violated [65,66]. Further research was performed by Koch et al. [67,68], Blake [69], and so on, in the areas of implementing neural networks to implement regularization algorithms for the data fusion. Reinforcement learning networks were implemented to implement multisensor data fusion [70].
Symbolic reasoning techniques using artificial intelligence and machine learning contributed to rule-based inference which was studied in OPS5 [71,72], MYCIN [73], and BBI [74]. Any of these inference techniques coupled with constraint-based reasoning techniques.
Over the years, several techniques that have emerged as data fusion paradigms are Zadeh’s fuzzy logic [75], Duda’s symbolic uncertainty management [76], and Shafer’s combined evidence techniques that give a basis for inference under uncertainty [77]
Crowley et al. provide a set of numerical techniques that are represented by a primitive comprising a vector of property estimates and their respective precisions. They showed that Kalman filter prediction equations provide a means for prediction of the model’s state [57].
Waltz et al. [44] and Llinas and Hall [78] define the term multisensor data fusion as a technology concerned with combining data from multiple (and possibly diverse) sensors to make inferences about a physical environment, event, activity, or situation.
The International Society of Information Fusion defines information fusion as [79]
The definition of multi-sensor data fusion by Waltz and Llinas [44] and Hall [24] is given as:“Information Fusion encompasses theory, techniques, and tools conceived and employed for exploiting the synergy in the information acquired from multiple sources (sensor, databases, information gathered by human, etc.) such that the resulting decision or action is in some sense better (qualitatively or quantitatively, in terms of accuracy, robustness, etc.) than would be possible if any of these sources were used individually without such synergy exploitation.”
The definition, process, and one of the purposes of data fusion is elicited by Elmenreich et al. [80] as:The technology concerned with the combination of how to combine data from multiple (and possible diverse) sensors to make inferences about a physical event, activity, or situation
With respect to the output data types of the sensors, we can broadly categorize them into homogeneous sensor data and heterogeneous sensor data. Heterogeneous sensor data comprise of different types of sensing equipment, like imaging, laser, auditory, EEG, etc. For example, a monocular camera (RGB) will have pure image data, while a stereo vision camera (RGB-D) could have imaging data for both the cameras and a depth cloud for the depth information, an EEG could output signal details and LiDAR outputs’ location details of the object of interest with respect to the LiDAR. Systems with multi-sensor fusion are capable of providing many benefits when compared with single sensor systems. This is because all sensors suffer from some form of limitation, which could lead to the overall malfunction or limited functionality in the control system where it is incorporated.“Sensor Fusion is the combining of sensory data or data derived from sensory data such that the resulting information is in some sense better than would be possible when these sources were used individually”.
Garcia et al. in 2017 proposed a novel sensor data fusion methodology in which the augmented environment information is provided to the intelligent vehicles with LiDAR, camera, and GPS. They propose that their methodology leads to safer roads by data fusion techniques in single-lane carriage-ways where casualties are higher than in other road types. They rely on the speed and accuracy of the LiDAR for obstacle detection and camera-based identification techniques and advanced tracking and data association algorithms like Unscented Kalman Filter and Joint Probabilistic Data Association [81]. Jahromi et al. proposed a real-time hybrid data fusion technique in 2019 [82]. Extended Kalman Filter (EKF) based nonlinear state estimation and encoder–decoder based Fully Convolutional Neural Network (FCNN) are used on a suite of camera, LiDAR, and radar sensors. Data fusion is a vast area with numerous techniques; we provide advantages and disadvantages of data grouping/association, state estimations, and distributed systems [29,83,84,85]. The following subsections highlight some of the algorithms used in data fusion.
2.5.1. K-Means
K-Means is a popular algorithm that has been widely employed;
Some prominent advantages are:
- Simpler to implement compared to other techniques
- Good generalization to clusters of various shapes and sizes, such as elliptical clusters, circular, etc.
- Simpler and easy adaption to new examples
- Convergence is guaranteed.
- Scales to large data sets
- Centroid position can be warm-started
Some prominent disadvantages:
- Optimal solution for the cluster centers are not always found by the algorithm;
- The algorithm assumes that the covariance of the dataset is irrelevant or that it has been normalized already.
- The system must have knowledge of the number of clusters a priori.
- Assumption is made that this number is optimum.
2.5.2. Probabilistic Data Association (PDA)
PDA was proposed by Bar-Shalom and Tse, and it is also known by “modified filter of all neighbors” [86]. The functionality is to assign an association probability to each hypothesis from the correct measurement of a destination/target and then process it.
Prominent advantages are:
- Tracking target excellence: Excellent for tracking targets that do not make sudden changes in their navigationPDA is mainly good for tracking targets that do not make abrupt changes in their movement pattern. The prominent disadvantages are [86,87]:
- Track loss: PDA might display poor performance when the targets are close to each other because it ignores the interference with other targets and hence there is a possibility that it could wrongly classify the closest tracks.
- Suboptimal Bayesian approximation: PDA gives suboptimal Bayesian approximation when the source of information is uncertain; this might be seen when a LiDAR scans a pole.
- One target: PDA gives incorrect results in the presence of multiple targets since the false alarm model does not work well. The Poisson distribution typically models the number of false, with an assumption of uniform distribution
- Track management: Problems of tracking algorithms must be provided for track initialization and track deletion since PDA needs this a priori.
2.5.3. Joint Probabilistic Data Association
The prominent advantages are as follows [87,88,89]:
- Robust: JPDA is robust compared to PDA and MHT.
- Multiple object tracking: The algorithm can be used to track multiple agents (however, with a caveat)
- Representation of multimodal data: Can represent multimodal state densities, which represent the increase in robustness of the underlying state estimation process
The prominent disadvantages of JPDA are as follows [87,88,89]:
- Computationally expensive: JPDA is a computationally expensive algorithm when employed in multiple target environments since the number of hypotheses’ increments exponentially with the number of targets.
- Exclusive mechanism: It requires an exclusive mechanism for track initialization.
2.5.4. Distributed Multiple Hypothesis Test
The main advantages of MHT-D are [90]:
- Very useful in distributed and de-centralized systems
- Outperforms JPDA for the lower densities of false positives
- Efficient at tracking multiple targets in cluttered environments
- Functions also as an estimation and tracking technique
The main disadvantage of the MHT-D is as follows [90]:
- Exponential computational costs that are in the order of , where X is the number of variables to be estimated and n is the number of possible associations
Another type of fusion technique is by state estimation.
2.5.5. State Estimation
Also known as tracking techniques, they assist with calculating the moving target’s state, when measurements are given [87]. These measurements are obtained using the sensors. This is a fairly common technique in data fusion mainly for two reasons: (1) measurements are usually obtained from multiple sensors; and there could be noise in the measurements. Some examples are Kalman Filters, Extended Kalman Filters, Particle Filters, etc. [91]. We discuss state estimation techniques in Section 3.5.
2.5.6. Covariance Consistency Methods
These methods were proposed initially by Uhlmann et al. [84,87]. This is a distributed technique that maintains covariance estimations and means in a distributed system. They comprise of estimation-fusion techniques.
Some prominent advantages are:
- Efficient in distributed systems; i.e., multimodal multi-sensors as well
- Fault-tolerant for covariance means and estimates
Some prominent disadvantages are:
- If the Kalman filter is used for estimation, the exact cross-covariance information must be determined. This could pose a big challenge.
- Suboptimal results are realized if the iterative application of the technique is used to process a sequence of estimates for a batch application for simultaneous fusion of the estimates.
2.5.7. Decision Fusion Techniques
These techniques can be used when successful target detection occurs [87,92,93]. They enable high-level inference for such events.
Some prominent advantages are:
- Enables the user to arrive at a single decision from a set of multiple classifiers or decision-makers
- Provides compensatory advantage for other sensors when one sensor is deficient, in a multi-sensor system
- Enables a user to adjust the decision rules to arrive at the optimum.
Some prominent disadvantages are:
- Establishing a priori probabilities is difficult
- When a substantial number of events that depend on the multiple hypotheses occur, this will be very complex and a hypothesis must be mutually exclusive
- Decision uncertainty is difficult to finalize
2.5.8. Distributed Data Fusion
As the name suggests, this is a distributed fusion system and is often used in multi-agent systems, multisensor systems, and multimodal systems [84,94,95].
Some prominent advantages are:
- Enables usage across dynamic and distributed systems
- Communication costs can be low since systems can communicate with each other after onboard processing at the individual agents/nodes
Some prominent disadvantages are:
- Spatial and temporal information alignment
- Out-of-sequence measurements
- Data correlation challenges
- Systems may need robust communication systems to share information.
2.6. Classifications of Data Fusion Techniques
Classification of data fusion is fuzzy and fluid, in that it is quite tedious and complex to follow and adhere to strict processes and methodologies. Many criteria can be used for the classification of data fusion. Castanedo discussed [87] the techniques and algorithms for state estimation, data association and finally a higher-level decision fusion. Foo performed a study of high-level data fusion in tactical systems, biomedical systems, information science and security, disaster management, fault detection, and diagnosis [43]. Dasarathy et al. [96] discuss data fusion methods and several techniques. Luo et al. [38] discuss abstraction levels and Steinberg et al. via JDL [97] perform basic research in data fusion. The subsections below provide a brief introduction on how we can classify data fusion. Some of these techniques are given in Table 1.
Table 1.
Data fusion techniques and their classifications.
2.6.1. Data Type of Sensor Input and Output Values
Several types of classification emerged out of Dasarathy’s input–output data fusion [96]. They can be summarized as follows: Data-in-Data-out (DAI-DAO): Raw data are input and raw data are extracted out. Data-in-Feature-out (DAI-FEO): Raw data are sourced, but the system provides features extracted out of the data as output. Feature-in: Feature-out (FEI-FEO): Features from previous steps of fusion or other processes are fed into the fusion system and better features or higher-level features are output. New and improved features are output as part of this type of fusion. This is also called Feature-fusion [96]. Feature-in: Decision-out (FEI-DEO): The features fed into the input system as the source are processed to provide decisions for tasks and goals as output. This is where simple or high-level features are accepted as input, and processed and decisions are extracted for the system to follow. Most of the present-day fusion is of this type of classification technique. Decision-in-Decision-out (DEI-DEO): Simple and lower-level decisions are accepted by the system and higher-level better decisions are processed out. This is a type of fusion is also called Decision-fusion [96].
2.6.2. Abstraction Levels
In a typical perception system, one comes across the following abstraction of data: pixel, signal, symbols, feature-characteristics [38].
Pixel level classification: is performed on image input from sensors like monocular, stereo vision, or depth cameras, IR cameras, etc. to a system; image processing that is used to improve tasks that look for and extract objects; object features use this technique.
Signal level classification: is performed on data involving signals from sensors like LiDAR, sonar, audio, etc. The signal data are directly operated on and output rendered.
Symbol level classification: is a technique that employs methods to represent information as symbols. This is similar to the decision-fusion technique of Dasarathy [96] and called decision level.
Characteristic level classification: extracts features from signals or images while processing the data and is called feature level.
2.6.3. JDL Levels
Data fusion models divided into five processing layers, interconnected by a data bus to a relationship database [97,98]
- Layer 0: Processes source data comprised of pixel and signal. Information is extracted, processed, reduced, and output to higher layers.
- Layer 1: Data output from layer 0 is processed here and refined. Typical processes are alignment in the spatial-temporal information, correlation, clustering, association and grouping techniques, false-positive removal and reduction, state estimation, image feature data combination, and state estimations. Classification and identification: state and orientation are the typical outputs. It also performs input data transformation to obtain consistent and robust data-structures.
- Layer 2: Based on other output of layer 1 or the object refinement layer, analysis of the situation is performed. Based on the data input and the present and past decisions, the situation assessment is performed. A set of high-level inferences is the outcome of this layer. Identification of events and activities are performed.
- Layer 3: The output of layer 2 i.e., the significant activities and current events are assessed for impact on the system. Prediction of an outcome and threat analysis is performed at this layer.
- Layer 4: Overall processes from layer 0 through layer 3 are optimized and improved. Resource control and management, task scheduling, and prioritizing are performed to make improvements.
2.6.4. Data Source Relationships
This type of classification uses concepts of data redundancy, data complementing, and data combination [87]. Video data overlaps can be called redundant data sources and can be optimized. This is the area of data source classification wherein the same destination or target is identified by multiple data sources. Complementary data sources provide different inputs that can be combined to form a complete target or scene or object—for example, a complete scene if formed using different cameras and the scene can be put together from individual pieces. Combining data sources in a cooperative environment gives a result that is more complex than the input source information.
2.6.5. System Architecture
This type of classification is based on the architecture of the data fusion system. The architecture could be hierarchical, distributed or decentralized, centralized, etc. [85,87,96]. This prompts us to think that the researchers classified these systems based on how many agents/nodes are available, how the sensors are spread across these agents/nodes. In a decentralized architecture, all the agents take part in the data fusion task. Each system processes its own and its neighbor’s data. The advantages are processing faster since each system could be processing smaller chunks of data. The cons of this process are the high communication costs since several systems need to communicate with each other and the cost is , at each step of communication, and n is the number of nodes. The process is costliest if each node has to communicate with every one of its peers. Contrary to this, in a centralized architecture, a powerful single system will perform the data fusion. Suboptimal systems could end up being resource hogs that take up a lot of resources in the form of bandwidth since raw data are transferred from the sensors to the central processing system. When a higher number of sensors are used, this type of architecture will pose huge resource issues. Moreover, the central unit would need to be very powerful to process and perform data fusion, which could mean an expensive system.
- Distributed or decentralized systems: State estimation and data processing are performed locally and then communicated to the other systems. Single node to groups of systems form the range of processing in this architecture. The fusion node processes the result only after the individual data processing at the local level is completed [94,99,100].
- Hierarchical systems: A system architecture, wherein the higher-level nodes control the lower-level nodes and a mechanism of hierarchical control of data fusion is set up, is the hierarchical data fusion system. In this type of architecture, a combination of distributed decentralized nodes could be employed to achieve data fusion. Back in the second half of the 1990s, Bowman et al. proposed a hierarchical data fusion system [101] which was reviewed by Hall et al. [21]. Taropa et al. in 2006 proposed a hierarchical data fusion model [102] in which they use real-time objects in a highly flexible framework and provide these features through an API. Dieterle et al. proposed a data fusion system for object tracking [103]. In the publication, they combine sensor information using a hierarchical data fusion approach and show that this approach drastically improves robustness in object detection with respect to sensor failures and occlusions.
3. Sensor Hardware
We will now briefly introduce some of the hardware that could be used for data fusion in vehicular navigation.
3.1. LiDAR
Light Detection and Ranging (LiDAR) is a technology that is used in several autonomous tasks and functions as follows: an area is illuminated by a light source. The light is scattered by the objects in that scene and is detected by a photo-detector. The LiDAR can provide the distance to the object by measuring the time it takes for the light to travel to the object and back.
NOAA states:
LIDAR, which stands for Light Detection and Ranging, is a remote sensing method that uses light in the form of a pulsed laser to measure ranges (variable distances) to the Earth. These light pulses—combined with other data recorded by the airborne system—generate precise, three-dimensional information about the shape of the Earth and its surface characteristics [104].
3.1.1. Data Generation in a LiDAR
Different types of data are generated by a LiDAR. Some are highlighted below.
- Number of Returns: The Light pulses from a LiDAR can penetrate a canopy in a forest. This also means that LiDAR can hit the bare Earth or short vegetation.
- Digital Elevation Models: Digital Elevation Models (DEM) are earth (topographic) models of the Earth’s surface. A DEM can be built by using only ground returns. This is different from Digital Terrain Models (DTM), wherein contours are incorporated.
- Digital Surface Models: A Digital Surface Model (DSM) incorporates elevations from man-made and natural surfaces. For example, the addition of elevation from buildings, tree canopies, vehicular traffic, powerlines, vineyards, and other features.
- Canopy Height Model: Canopy Height Models (CHM) provides the true height of topographic features on the ground. This is also called a Normalized Digital Surface Model (nDSM).
- Light Intensity: Reflectivity or Light intensity varies with the composition of the object reflecting the LiDAR’s return. Light intensity is defined as the reflective percentages.
3.1.2. Classifying the LiDAR
LiDAR can be generally classified based on the data returned, technology used, area of usage [105].
- Data Returned by the LiDAR: LiDAR types based on storing the data returned from the object [106]:
- Discrete LiDAR: While scanning, the data returned are in the form of 1st, 2nd, and 3rd returns, due to the light hitting multiple surfaces. Finally, a large-final pulse is returned. This can be seen when a LiDAR hits a forest canopy [107]. When the LiDAR stores the return data individually/discretely, it takes each peak and separates each return.
- Continuous/Full waveform LiDAR: When the entire waveform is saved as one unit, its a continuous or full form LiDAR [108]. A lot of LiDARs use this form of recording.
- Lidar types based on technology: The following technology types can be considered as well while classifying LiDARs [105,109]
- Mechanical-scanners: Macro-scanners, Risley prisms, Micro-motion.
- Non-Mechanical-scanners: MEMS, Optical phased arrays, electro-optical, liquid crystal.
- Flash-LiDAR-non-scanners
- Structured light-non-scanners
- Multicamera-stereo-non-scanners
- Based on area of usage: Two types of LiDAR broadly used are: topographic and bathymetric [104]. Topographic LiDARs are typically used in land mapping, and they use near-infrared laser and bathymetric LiDARs use green light technology for water-penetration to measure river bed elevations and seafloor.In Topographic LiDAR, the two main types are 2D (single scan) and 3D (multiple scan). Some brands of topographic LiDAR are Velodyne [110], another model from Velodyne, the HDL-64E provides a 3D laser scan i.e., 360° horizontal and 26.9° vertical field of view (FOV), while 2D LiDARs like the TiM571 LiDAR scanning range finder from SICK provide a 2D 220° FOV this is very similar to RPLidar [111] from Slamtech, Ouster [112] from Ouster laser scanners, Eclipse mapping systems [113]. The Bathymetric LiDARs use the green spectrum technology and are predominantly used for water surface and underwater mapping tasks. A small listing and background of Bathymetric LiDARs are given by Quadros et al. from Quadros [114]. However, bathymetric LiDARs are out of the scope of this survey due to its nature of use.
3.1.3. Advantages and Disadvantages in Using LiDAR
LiDARs are very useful in detecting objects and developing an environment model [93,104,114]. It does have both usage advantages and disadvantages. Advantages include Safety in usage, fast scans of the environment, high accuracy, and some can capture data even at 2500 m and have better resolution compared to other scan systems like Radar.
Disadvantages include: Many products are still very expensive, data are not as rich as an RGB camera with a good resolution, a single data point may not be accurate and high volume data points will need to be used, their scans and eventual point clouds are too big and consume a lot of space, and 2D LiDARs are useful mainly as line scanners and hence are sparingly used.
3.2. Camera
The types of camera are Conventional color cameras like USB/web camera; RGB, RGB-mono, and RGB cameras with depth information; RGB-Depth (RGB-D), 360° camera, and Time-of-Flight (TOF) camera.
3.2.1. RGB Family of Camera
An RGB camera is typically a camera equipped with a standard CMOS sensor through which the colored images of the world are acquired. The acquisition of static photos is usually expressed in megapixels [115].
Advantages and disadvantages of RGB cameras are as follows:
Advantages include availability of several inexpensive cameras, and they do not need any specialized drivers, simplicity in usage, etc.
The disadvantages include that the presence of good lighting is essential, some of the high-end cameras that have great resolution are very expensive, and there are RGB-D cameras that cannot efficiently capture surfaces that are reflective, absorptive, and transparent such as glass and plastic.
3.2.2. 360° Camera
A 360° camera captures dual images or video files from dual lenses with 180° field of view and either performs an on-camera automatic stitch of the images/video or lets the user perform off-board stitching of the images, to give a full 360° view of the world [28,116,117,118].
Some advantages and disadvantages are as follows:
Advantages include new technology possibilities in usage and improvements being higher, and hardware or software may be used to get 360 images, etc.
Disadvantages include diminished quality, few cameras are expensive, long rendering time, storage may be needed more in high-resolution cameras, etc.
3.2.3. Time-of-Flight (TOF)
The TOF gives depth information based on IR and camera technology. It works by emitting an infrared light signal and measures how long the signal takes to return and calculates the depth based on extracted data. This information can be used with several navigation-related modules like mapping and obstacle avoidance [119,120,121].
Some advantages and disadvantages are highlighted in [122] as follows:
Advantages include high speed, efficient usage of computation since TOF uses a one look approach compared to the multiple scans of laser scanners, long working distance, depth information up to 5 m given in real-time, wide application range (feature-filled or featureless, depth information given by camera in the presence or absence of ambient light).
Disadvantages include low resolution, relatively high power consumption due to which high heat may be generated, affected by object’s reflective, color and complexity of the environment, may need additional management of subjects’ background lighting, multiple path reflections, usage of multiple TOF at the same time may have interference with each other, supported application scenarios are less, and development and support groups are low in number.
In some autonomous vehicles, radar is used in addition to camera [123,124] (however, the study of radar is out of the scope of this paper)
3.3. Implementation of Data Fusion with the Given Hardware
We review an input–output type of the fusion as described by Dasarathy et al. [96]. They propose a classification strategy based on input–output of entities like data, architecture, features, and decisions. The fusion of raw data in the first layer, a fusion of features in the second, and finally the decision layer fusion. In the case of the LiDAR and camera data fusion, two distinct steps effectively integrate/fuse the data [28,117,125]:
- Geometric alignment of the sensor data
- Resolution match between the sensor data
Let us review these two steps in greater detail.
3.3.1. Geometric Alignment of the Sensor Data
The first and foremost step in the data fusion methodology is the alignment of the sensor data. In this step, the logic finds LiDAR data points for each of the pixel data points from the optical image. This ensures the geometric alignment of the two sensors [28].
3.3.2. Resolution Match between the Sensor Data
Once the data is geometrically aligned, there must be a match in the resolution between the sensor data of the two sensors. The optical camera has the highest resolution of 1920 × 1080 at 30 fps, followed by the depth camera output that has a resolution of 1280 × 720 pixels at 90 fps and finally the LiDAR data have the lowest resolution. This step performs an extrinsic calibration of the data. Madden et al. performed a sensor alignment [126] of a LiDAR and 3D depth camera using a probabilistic approach. De Silva et al. [28] performed a resolution match by finding a distance value for the image pixels for which there is no distance value. They solve this as a missing value prediction problem, which is based on regression. They formulate the missing data values using the relationship between the measured data point values by using a multi-modal technique called Gaussian Process Regression (GPR), developed by Lahat et al. [39]. The resolution matching of two different sensors can be performed through extrinsic sensor calibration. Considering the depth information of a liDAR and the stereo vision camera, 3D depth boards can be developed out of simple 2D images. Figure 3 shows the depth calibration board. The dimensions of this board are: length 58 width 18 height 41.5.
Figure 3.
Depth calibration [127].
For a stereo vision or depth camera like the Intel Realsense d435, there is a need to perform a depth scale calibration. Figure 4 shows the phone calibration tool [127]. Another addition to the calibration toolkit is the speck pattern board. These pattern boards in (not to scale) Figure 5 give us better results since there is a higher spatial frequency content with limited or no laser speckle. It has been documented that a passive target or LED-based projector gives about 25–30% better depth accuracy than a laser-based projector [127]. After using adequate turning mechanisms, the depth accuracy can be improved even more. The projector can be a drawback in some cases, and it may help to turn off the projection from the camera and light up the subject using clean white light [128]. It is also observed that the RealSense cameras have better performance in open bright sunlight since there is better visibility of the natural textures. It should be noted that, in the case of the depth cameras, the stereo vision has a limitation due to the quality differences between the left and right images.
Figure 4.
Realsense phone calibration tool [127].
Figure 5.
Realsense iPhone speck pattern for calibration [128].
There are several calibration techniques for the LiDAR and camera, wherein Mirzaei et al. [129] have provided techniques for intrinsic calibration of a LiDAR and extrinsic calibration based on camera readings.
Dong et al. [130] have provided a technique for extrinsic calibration of a 2D LiDAR and camera. Li et al. [131] also have developed a technique for 2D LiDAR and camera calibration—however for an indoor environment. Kaess et al. [132] developed a novel technique to calibrate a 3D LiDAR and camera.
3.4. Challenges with Sensor Data Fusion
Several challenges have been observed while implementing multisensor data fusion. Some of them could be data related to like: complexity in data, conflicting and/or contradicting data, or they can be technical such as resolution differences between the sensors, the difference in alignment between the sensors [28], etc. We review two of the fundamental challenges surrounding sensor data fusion, which are the resolution differences in the heterogeneous sensors and understanding and utilizing the heterogeneous sensor data streams [28] while accounting for many uncertainties in the sensor data sources [39]. We focus on reviewing the utilization of the fused information in the autonomous navigation, which is challenging since many autonomous systems work in complex environments, be it at home or work, which is to assist persons with severe motor disabilities to handle their navigational requirements and hence pose significant challenges for decision-making due to the safety, efficiency, and accuracy requirements. For reliable operation, decisions on the system need to be made by considering the entire set of multi-modal sensor data they acquire, keeping in mind a complete solution. In addition to this, the decisions need to be made considering the uncertainties associated with both the data acquisition methods and the implemented pre-processing algorithms. Our focus in this review is to survey the data fusion techniques that consider the uncertainty in the fusion algorithm.
Some researchers used mathematical and/or statistical techniques for data fusion. Others used techniques comprised of reinforcement learning in implementing multisensor data fusion [70], where they encountered conflicting data. In this study, they fitted smart mobile systems with sensors that enabled the systems to be sensitive to the environment(s) they were active in. The challenge they try to solve is mapping the multiple streams of raw sensory data Smart agents to their tasks. In their environment, the tasks were different and conflicting, which complicated the problem. This resulted in their system learning to translate the multiple inputs to the appropriate tasks or sequence of system actions.
Brooks et al. [47] achieve sensor data robustness, reliability, and resolve issues like mechanical failures, noise, transient errors using multiple sensors, whose data is fused. They recommend fusing readings from multiple heterogeneous sensors. This made their overall system less sensitive to failures from one technology. Crowel et al. developed mathematical tools to counter uncertainties with fusion and perception [133]. Other implementations include adaptive learning techniques [134], wherein the authors use D-CNN techniques in a multisensor environment for fault diagnostics in planetary gearboxes.
The other challenges are dependent on the sensor itself, i.e., the hardware, or the physics that are used by the hardware. Structural errors in the hardware are an example. These errors are the difference(s) between a sensor’s expected value and measured value, whenever the sensor is used for data collection. Repeated differences can be calculated using a technique called sensor calibration. Before using any sensor, it needs to be calibrated. This will ensure a consistent measurement, i.e., where all the sensors can be fused uniformly.
Broadly, one can differentiate calibration into extrinsic and intrinsic. Extrinsic calibration entails finding external parameters that are used in the sensors—for example, parameter differences between a LiDAR’s alignment/orientation and a camera’s alignment/orientation [130,135]. In another case, it may be the LiDAR’s orientation and location in its working environment or world. In contrast, intrinsic calibration entails finding the differences within the same sensor. For example, relationship(s) between the camera coordinates and its pixel coordinates. Usually, the manufacturer performs intrinsic calibration and communicates the details to the end-user in the user guide/manual.
Researchers have found that extrinsic calibration can be challenging when the number of agents is high as in cases of swarms of robots [129,130,132]. For example, senior living where the swarms of autonomous wheelchairs work together to share information about location, situation awareness, etc.; this could be attributed to the variations that exist between sensors due to manufacturing differences, types of sensors, and autonomous system types. In such an example, the calibration duration will be large if there is a large number of autonomous systems; in fact, it could be exponential and hence exorbitant and unacceptable. Reducing both the time required for the process and the complexity is essential.
3.5. Sensor Data Noise
Every sensor has an amount of noise that is inherent to its properties. There have been many attempts at reducing or removing the noise—for instance, in object detection [136] wherein the authors provide a method and technique to remove noise in LiDAR intensity images. They use a type of diffusion filtering called anisotropic filtering to retain the scanned object space details and characteristics. The second research is where the background noise is removed [137], wherein the authors develop a methodology to identify background noise under the clear atmospheric condition and derive equations to calculate the noise levels. Topics other than object detection are speech recognition [138,139]. In this section, we discuss filtering noise using the Kalman Filter. Kalman filter is over five decades old and is one of the most sought after filtering techniques. We will discuss two flavors of Kalman filter, namely: Extended Kalman Filter and Unscented Kalman Filter.
In addition to the sensing information, every sensor is bound to have a level of noise and, while using these sensors, one will soon realize that at least a small amount of noise is bound to exist in addition to measurement and estimation of uncertainties. When such errors or uncertainties occur, it is required to use techniques that mitigate their effects on the system. This now becomes a complex problem of estimating the state(s) of the system after the system becomes observable. Mathematical algorithms that accomplish this are the filtering techniques. Filtering techniques are applicable in several domains like economics, science, and engineering. Localization systems can make use of these techniques as there is an innate level of sensor measurement noise and uncertainty with their pose estimation. Filtering techniques have been used in many localization systems and two of the most popular filtering algorithms are Kalman filters and particle filters.
3.5.1. Kalman Filters
Kalman filters (KF) were introduced by Rudolf Kalman in 1960 [140]. It is also known as Linear Quadratic Estimation (LQE) in the field of controls and autonomous systems. KF is versatile and has been extensively used in the areas of autonomous systems, signal processing, system navigation, defense, aerospace, etc., and it is an iterative algorithm that uses Bayesian inference to estimate the probabilistic distribution of the uncertain/unknown variables. They use a series of measurements that have noise from measurements and process(es). This is because unknown variables can be estimated better with multiple measurements than with a single measurement. The algorithm is optimized to run in real-time and needs only the previous system state and the current input measurement. The KF starts with the system model and the known control inputs to that system, and multiple sequential measurements (measurements from sensors) and forms an estimate of the system’s varying quantities (provided in the state matrices). Incidentally, it is found to be better than the estimate obtained using a single measurement. Kalman Filter can also be broadly categorized as a common sensor fusion and data fusion algorithm.
A Dynamic System Model can be represented as follows:
where:
- : Current estimate,
- : Estimate of the signal in Previous state,
- : Control signal,
- : Measured value from the sensors,
- : Process noise in the previous iteration,
- : Measurement noise in the present iteration.
Equation (1) denotes the current estimate of a state variable , which is comprised of the previous system state , the control signal , and the process noise in the previous iteration .
Equation (2) calculates the current measurement value , which is a linear combination of the unknown variable and the measurement noise and this is usually a Gaussian. A, B, and H are matrices that provide the weights of the corresponding component of the equation. These values can be provided a priori and are system dependent. A Gaussian distribution with a zero mean contributes two noise values, namely and ; these have covariance matrices named and , respectively, and they are estimated a priori, although they initially provide a coarse estimate; over the set of iterations, the algorithm does converge to the accurate estimators.
There are two steps that dominate the process and they are: the time update and the measurement update; in turn, each step has a set of equations that must be solved to calculate the present state. The following is the algorithm:
- Predict state
- Measurement Update—Calculate the Kalman gain (weights): Kalman gain—The main and unknown value in this equation
- Update state
- Update state covariance
- Loop (now k becomes ), which is the next and subsequent iterations.where:: Prior error covariance Matrix,: Current Covariance Matrix, updated during each iteration,: Covariance Matrix,: Measurement Noise Covariance Matrix.
This filter’s output is the result of the state update and state-covariance update equations. These provide the combined estimate from the prediction model and measurements from sensors. The mean value of the distribution for each state variable is provided by state matrix and the variances by the covariance matrix. A set of measurements are taken in the present state. The system initializes many matrices. The state variables can be set based on the initial measurements from the sensors. The covariance of the state can be initialized using the identity matrix or the covariance matrix . Initially, the covariance matrix is not stable but will stabilize as time progresses and the system runs. Measurement noise covariance matrix is calculated using calibrations performed earlier. The measurement sensors will be developed to measure a large number of readings of the ground truth state, from which the variances can be calculated. The variance of the measurements provides the value of in .
Using literal interpretation(s) from state transition, equations can be used to place the much-needed bounds on dynamic noise. This is because it will be harder to calculate the dynamic noise covariance . For instance, 3 sigma in in can be calculated by interpreting the target acceleration as a constant velocity model with dynamic noise.
The relative ratio of the measurement noise to the dynamic noise is an important factor. This helps calculate the gains. In the Kalman Filter, it is known to keep one of the noise covariance matrices constant while adjusting the other continuously until the desired performance is achieved. The family of Kalman Filters is to be used in systems that can be run continuously for better accuracy or performance and cannot be used for quick/few iterations since it takes several iterations just to stabilize while using Kalman Filters.
The Kalman filter can become very inefficient and the convergence to the required values can take several steps; to reduce this, i.e., for the system to convergence in fewer steps, the system must be modeled more elegantly and precise estimation of the noise must be achieved.
3.5.2. Extended Kalman Filter
The world functions mostly in a nonlinear manner. Hence, if the techniques used to measure, estimate, predict, analyze, etc. are nonlinear, it is practical, convenient, or accurate. This applies to Kalman Filter as well. The nonlinear filtering problem heuristic is the Extended Kalman Filter (EKF). This technique is naturally the most sought after filtering and estimation for nonlinear systems.
The EKF is based on linearizing dynamics and output functions at an existing estimate(s). In an EKF, the state distribution is usually approximated by a Gaussian Random Variable (GRV), which is then analytically propagated through a first-order linearization of the given nonlinear system under consideration [141,142,143,144]. For example, it functions by propagating an approximation of the conditional expectation and covariance [141,142,144,145,146,147].
3.5.3. Unscented Kalman Filters
Unscented Kalman Filters (UKF) belong to the class of filters called Linear Regression Kalman Filters. These filters are also called Sigma-Point Kalman Filters [148,149]. This type of filter linearizes a nonlinear function of a random variable using a linear regression algorithm between n points drawn from the previous distribution of the given random variable. This is also called statistical linearization.
We have seen that the EKF propagates the state distribution through the first order linearization; this may corrupt the posterior mean and covariance. The flaws of EKF have been highlighted by Wan et al. [150]. The UKF is robust to this issue since its derivative-free and uses a deterministic sampling [151]. This logic chooses a set of points called sigma points to represent the state distribution. UKF has an additional step in the selection of sigma points. Broadly, the following are the steps involved:
- Select sigma points
- Model forecasting
- Data assimilation
When data in the input system is symmetric, a deterministic sampling of the data points can approximate the probability density in which the underlying distribution is Gaussian. The nonlinear transformation of the points is an estimation of the posterior distribution. Julier and Uhlmann [148,149,151] state that Unscented transformation is
Founded on the intuition that it is easier to approximate a probability distribution than it is to approximate an arbitrary nonlinear function or transformation
3.5.4. Distributed Kalman Filter
Over the past decade, a new technique of filtering that can be used in distributed and dynamic systems has been proposed by Olfati-Saber [91,152]. Techniques of consensus are used to fuse and filter the sensor data and apply covariance information to sensor networks with varying observation matrices. They prove that this provides a collective observer for the processes in the environment that the model uses. They propose a continuous-time distributed Kalman Filter (DKF) that performs a local mean of the sensor data but reaches a consensus with other agents/nodes in the selected network. The above authors also proposed a micro Kalman filter technique wherein an embedded low pass and bandpass consensus filter was used. The consensus filters performed a fusion of the sensor data and co-variance data measured at each agent/node.
Broadly, there are two types of the DKF from the above author:
- Consensus on Estimates
- Local Kalman filtering
- Continuous-time Distributed Kalman filtering
- Iterative Kalman-Consensus filtering
- Consensus on sensor data fusion
Carli et al. proposed a distributed Kalman Filter based on consensus strategies [99], wherein they estimate the state of a dynamic system from distributed noisy measurements. Every agent/node constructs a local estimate based on its individual measurements and also estimates from its neighbors (connected agents). They perform this over a two-step process: the first one being a Kalman based measurement update and the second one being an estimate fusion that uses a consensus matrix. They document that optimizing the consensus matrix for fast convergence.
Spanos et al. proposed a DKF techniques in their research [153] in 2005. The performance of an approximate DKF is analyzed in this research. This technique admits systematic analysis of quantities of several networks like connection density, bandwidth, and topology. The contribution is a frequency domain characterization of the steady-state performance of the applicable DKF. They demonstrate a simple error transfer function with a bound while incorporating the connection density, network topology, and communication bandwidth that performs better using their approach.
Mahmoud et al. performed a review of the DKF during 2013 [100], wherein they compared a centralized Kalman Filter with a distributed Kalman Filter and bring out DKF’s advantages, its techniques, challenges involved, and applications.
Julier et al. wrote a handbook highlighting decentralized data fusion (DDF) with co-variance intersection. This follows a distributed framework in the area of control and estimation. The DDF provides increased robustness and scalability as compared to centralized versions. They state that the time required to implement new computational and sensing components is reduced using DDF.
Recent studies have been performed including optimization of several factors. Some include DKF with finite-time max consensus, DKF over networks with random link failures, etc. These studies suggest that the techniques of DKF are vital in the field of autonomous systems to optimize the system, reduce noise and optimal estimation, etc.
3.5.5. Particle Filters
Particle filters were first introduced in 1993 [154], and have continuously become a very popular class of numerical methods for optimizing the solution of nonlinear non-Gaussian scenarios [31,155,156]. While Kalman filters are linear quadratic estimators(LQE), particle filters, like any member of the family of Bayes filters such as Kalman filters and Hidden Markov Model(HMMs), estimate the posterior distribution of the state of the dynamical system conditioned on the data:
where is a sequence of target probability densities with increasing dimension, in which every distribution is defined through the space .
We need to know only: . , which is a normalizing constant is given by:
Note that may be unknown. The particle filter provides an approximation of and an estimate of at time 1. Then, an approximation of is also an estimate of at time 2.
Considering the simplest implementation wherein , we find that it yields and
Broadly, there are three steps involved in implementing a particle filter [157,158]. They are:
- Importance sampling:
- Sample the present trajectories and update
- Normalize the weights
- Selection:
- Samples that have high importance weights are multiplied
- Samples that have low importance weights are suppressed
- Markov Chain Monte Carlo Transition:
- Apply Markov transition kernel with an invariant distribution that is given by and obtain
In comparison with standard approximation methods, such as the popular Extended Kalman Filter, the principal advantage of particle methods is that they do not rely on any local linearization techniques or any crude functional approximation [158,159]. They can be used in areas like large systems, where Kalman Filters tend to fail [160]. This technique, however, has its drawbacks, which are expensive computational processes and complexity. Back in 1993, this was an issue, but, nowadays, we can make use of CPU, GPU, and similar high power computing to reduce the computational effort. One of the main deficiencies in a particle filter is that: Particle filters are insensitive to costs that might arise from the approximate nature of the particle representation. The other is that, in uninformative sensor readings, samples tend to congregate and a process that times how long it takes for the samples to congregate is essential.
3.6. Research Patents in Data Fusion
Some of the patents in this research area of data fusion have been as follows:
- Publication number: US9128185B2, Publication date: 08 September 2015, Inventors: Shuqing Zeng at GM Global Technology Operations LLC, Title: Methods and apparatus of fusing radar/camera object data and LiDAR scan points
- Publication number: US20100157280A1, Publication date: 24 June 2010; Inventors: Kresimir Kusevic, Paul Mrstik, Len Glennie
- Publication number: WO 2016/100814 A1, Publication date: 23 June 2016, Inventors: Michael J. GIERING, Kishore Reddy, Vivek Venugopalan, Title: Multi-Modal sensor data fusion for perception systems
- Publication number: EP 3396408 A1 20181031 (EN), Publication date: January 2013, Title: LiDAR and camera data fusion for automated vehicle
6. Conclusions
As part of this survey, we have briefly introduced sensor data fusion and autonomous navigation. We have reviewed the most popular data fusion techniques that can be used in navigation tasks for intelligent mobility systems. This survey is by no means exhaustive, due to the nature of the research area. However, it provides adequate information to the audience by reviewing the laser and optical sensors like LiDAR and camera, respectively. A brief look into the task of autonomous navigation, while explaining its sub-tasks namely mapping, localization, and obstacle avoidance is accomplished. The multi-disciplinary nature of data fusion was researched, and it was found that multiple sensors are better than one when used for autonomous vehicle tasks like robot navigation. The acute need for a robust data fusion process, methodology, and logic are described, and a discussion of the concepts of robot perception is provided, in addition to presenting some of the previous works that have performed seminal research in this area.
We have observed from research publications how data fusion can drive the future of autonomous systems and extend algorithms into areas of commercial autonomous systems, in addition to military systems. Estimation and filtering techniques such as Kalman filters, particle filters, and similar techniques are briefly discussed and also the need for their usage is provided.
A comparison of the different types of data fusion and their pros and cons are provided as well. Some inexpensive but robust sensors like the Intel Realsense D435 and RPLiDAR were researched, and their performance and capabilities are documented and references to top performers (although expensive sensors) sensors like Velodyne and eclipse are given. As a first look into sensor fusion, calibration techniques suggested by some leading manufacturers are provided. Multimodal sensor architectures are discussed in Section 1 and Section 5. A summary of the application of data fusion for the four sub tasks of navigation is given in tabular form in Table 3, in Section 5. In conclusion, we state again that using a good perception system with an appropriate data fusion system is vital for the optimal functioning of an autonomous system and its task of navigation.
References and Note
Author Contributions
Conceptualization, P.K. and M.J.; methodology, P.K.; software, P.K.; validation, P.K.; formal analysis, P.K.; investigation, P.K.; resources, P.K. and P.B.; data curation, P.K.; writing–original draft preparation, P.K.; writing–review and editing, P.K.; visualization, P.K.; supervision, M.J..; project administration, P.K. and P.B.; funding acquisition, M.J., P.B. and P.K. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported by Grant No. FA8750-15-2-0116 from Air Force Research Laboratory and OSD through a contract with North Carolina Agricultural and Technical State University.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Simpson, R.C. Smart wheelchairs: A literature review. J. Rehabil. Res. Dev. 2005, 42, 423. [Google Scholar] [CrossRef] [PubMed]
- Fehr, L.; Langbein, W.E.; Skaar, S.B. Adequacy of power wheelchair control interfaces for persons with severe disabilities: A clinical survey. J. Rehabil. Res. Dev. 2000, 37, 353–360. [Google Scholar] [PubMed]
- Martins, M.M.; Santos, C.P.; Frizera-Neto, A.; Ceres, R. Assistive mobility devices focusing on smart walkers: Classification and review. Robot. Auton. Syst. 2012, 60, 548–562. [Google Scholar] [CrossRef]
- Noonan, T.H.; Fisher, J.; Bryant, B. Autonomous Lawn Mower. U.S. Patent 5,204,814, 20 April 1993. [Google Scholar]
- Bernini, F. Autonomous Lawn Mower with Recharge Base. U.S. Patent 7,668,631, 23 February 2010. [Google Scholar]
- Ulrich, I.; Mondada, F.; Nicoud, J. Autonomous Vacuum Cleaner. Robot. Auton. Syst. 1997, 19. [Google Scholar] [CrossRef]
- Mutiara, G.; Hapsari, G.; Rijalul, R. Smart guide extension for blind cane. In Proceedings of the 4th International Conference on Information and Communication Technology, Bandung, Indonesia, 25–27 May 2016; pp. 1–6. [Google Scholar] [CrossRef]
- Bharucha, A.J.; Anand, V.; Forlizzi, J.; Dew, M.A.; Reynolds, C.F., III; Stevens, S.; Wactlar, H. Intelligent assistive technology applications to dementia care: current capabilities, limitations, and future challenges. Am. J. Geriatr. Psychiatry 2009, 17, 88–104. [Google Scholar] [CrossRef]
- Cahill, S.; Macijauskiene, J.; Nygård, A.M.; Faulkner, J.P.; Hagen, I. Technology in dementia care. Technol. Disabil. 2007, 19, 55–60. [Google Scholar] [CrossRef]
- Furness, B.W.; Beach, M.J.; Roberts, J.M. Giardiasis surveillance–United States, 1992–1997. MMWR CDC Surveill. Summ. 2000, 49, 1–13. [Google Scholar]
- Topo, P. Technology studies to meet the needs of people with dementia and their caregivers: A literature review. J. Appl. Gerontol. 2009, 28, 5–37. [Google Scholar] [CrossRef]
- First Sensors. Impact of LiDAR by 2032, 1. Available online: https://www.first-sensor.com/cms/upload/investor_relations/publications/First_Sensors_LiDAR_and_Camera_Strategy.pdf (accessed on 1 August 2019).
- Crowley, J.L.; Demazeau, Y. Principles and techniques for sensor data fusion. Signal Process. 1993, 32, 5–27. [Google Scholar] [CrossRef]
- Steinberg, A.N.; Bowman, C.L. Revisions to the JDL data fusion model. In Handbook of Multisensor Data Fusion; CRC Press: Boca Raton, FL, USA, 2017; pp. 65–88. [Google Scholar]
- McLaughlin, D. An integrated approach to hydrologic data assimilation: interpolation, smoothing, and filtering. Adv. Water Resour. 2002, 25, 1275–1286. [Google Scholar] [CrossRef]
- Van Mechelen, I.; Smilde, A.K. A generic linked-mode decomposition model for data fusion. Chemom. Intell. Lab. Syst. 2010, 104, 83–94. [Google Scholar] [CrossRef]
- McGurk, H.; MacDonald, J. Hearing lips and seeing voices. Nature 1976, 264, 746–748. [Google Scholar] [CrossRef] [PubMed]
- Caputo, M.; Denker, K.; Dums, B.; Umlauf, G.; Konstanz, H. 3D Hand Gesture Recognition Based on Sensor Fusion of Commodity Hardware. In Mensch & Computer; Oldenbourg Verlag: München, Germany, 2012. [Google Scholar]
- Lanckriet, G.R.; De Bie, T.; Cristianini, N.; Jordan, M.I.; Noble, W.S. A statistical framework for genomic data fusion. Bioinformatics 2004, 20, 2626–2635. [Google Scholar] [CrossRef] [PubMed]
- Aerts, S.; Lambrechts, D.; Maity, S.; Van Loo, P.; Coessens, B.; De Smet, F.; Tranchevent, L.C.; De Moor, B.; Marynen, P.; Hassan, B.; et al. Gene prioritization through genomic data fusion. Nat. Biotechnol. 2006, 24, 537–544. [Google Scholar] [CrossRef] [PubMed]
- Hall, D.L.; Llinas, J. An introduction to multisensor data fusion. Proc. IEEE 1997, 85, 6–23. [Google Scholar] [CrossRef]
- Webster Sensor Definition. Merriam-Webster Definition of a Sensor. Available online: https://www.merriam-webster.com/dictionary/sensor (accessed on 9 November 2019).
- Collins Dictionary Definition. Collins Dictionary Definition of a Sensor. Available online: https://www.collinsdictionary.com/dictionary/english/sensor (accessed on 9 November 2019).
- Hall, D.L.; McMullen, S.A. Mathematical Techniques in Multisensor Data Fusion; Artech House: Norwood, MA, USA, 2004. [Google Scholar]
- Hall, D.L.; Linn, R.J. A taxonomy of algorithms for multisensor data fusion. In Proceedings of the 1990 Joint Service Data Fusion Symposium, Gold Coast, Australia, 27–31 August 1990. [Google Scholar]
- Liggins, M.E.; Hall, D.L.; Llinas, J. Handbook of Multisensor Data Fusion: Theory and Practice; The Electrical Engineering and Applied Signal Processing Series; CRC Press: Boca Raton, FL, USA, 2009. [Google Scholar]
- Chavez-Garcia, R.O. Multiple Sensor Fusion for Detection, Classification and Tracking of Moving Objects in Driving Environments. Ph.D. Thesis, Université de Grenoble, Grenoble, France, 2014. [Google Scholar]
- De Silva, V.; Roche, J.; Kondoz, A. Fusion of LiDAR and camera sensor data for environment sensing in driverless vehicles. arXiv 2018, arXiv:1710.06230v2. [Google Scholar]
- Rao, N.S. A Fusion Method that Performs Better than Best Sensor; Technical Report; Oak Ridge National Lab.: Oak Ridge, TN, USA, 1998.
- Rövid, A.; Remeli, V. Towards Raw Sensor Fusion in 3D Object Detection. In Proceedings of the 2019 IEEE 17th World Symposium on Applied Machine Intelligence and Informatics (SAMI), Herl’any, Slovakia, 24–26 January 2019; pp. 293–298. [Google Scholar] [CrossRef]
- Thrun, S. Particle Filters in Robotics. In Proceedings of the 17th Annual Conference on Uncertainty in AI (UAI), Edmonton, AB, Canada, 1–4 August 2002. [Google Scholar]
- Wu, B.; Nevatia, R. Detection of multiple, partially occluded humans in a single image by bayesian combination of edgelet part detectors. In Proceedings of the Tenth IEEE International Conference on Computer Vision, ICCV 2005, Beijing, China, 17–20 October 2005; Volume 1, pp. 90–97. [Google Scholar]
- Borenstein, J.; Koren, Y. Obstacle avoidance with ultrasonic sensors. IEEE J. Robot. Autom. 1988, 4, 213–218. [Google Scholar] [CrossRef]
- Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D.; Ramanan, D. Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 1627–1645. [Google Scholar] [CrossRef]
- Chavez-Garcia, R.O.; Aycard, O. Multiple Sensor Fusion and Classification for Moving Object Detection and Tracking. IEEE Trans. Intell. Transp. Syst. 2016, 17, 525–534. [Google Scholar] [CrossRef]
- Qi, C.R.; Liu, W.; Wu, C.; Su, H.; Guibas, L.J. Frustum pointnets for 3d object detection from rgb-d data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 918–927. [Google Scholar]
- Baltzakis, H.; Argyros, A.; Trahanias, P. Fusion of laser and visual data for robot motion planning and collision avoidance. Mach. Vis. Appl. 2003, 15, 92–100. [Google Scholar] [CrossRef]
- Luo, R.C.; Yih, C.C.; Su, K.L. Multisensor fusion and integration: Approaches, applications, and future research directions. IEEE Sens. J. 2002, 2, 107–119. [Google Scholar] [CrossRef]
- Lahat, D.; Adali, T.; Jutten, C. Multimodal Data Fusion: An Overview of Methods, Challenges, and Prospects. Proc. IEEE 2015, 103, 1449–1477. [Google Scholar] [CrossRef]
- Shafer, S.; Stentz, A.; Thorpe, C. An architecture for sensor fusion in a mobile robot. In Proceedings of the 1986 IEEE International Conference on Robotics and Automation, San Francisco, CA, USA, 7–10 April 1986; Volume 3, pp. 2002–2011. [Google Scholar]
- Roggen, D.; Tröster, G.; Bulling, A. Signal processing technologies for activity-aware smart textiles. In Multidisciplinary Know-How for Smart-Textiles Developers; Elsevier: Amsterdam, The Netherlands, 2013; pp. 329–365. [Google Scholar]
- Foo, P.H.; Ng, G.W. High-level information fusion: An overview. J. Adv. Inf. Fusion 2013, 8, 33–72. [Google Scholar]
- Luo, R.C.; Su, K.L. A review of high-level multisensor fusion: Approaches and applications. In Proceedings of the 1999 IEEE/SICE/RSJ International Conference on Multisensor Fusion and Integration for Intelligent Systems, MFI’99 (Cat. No. 99TH8480), Taipei, Taiwan, 18 August 1999; pp. 25–31. [Google Scholar] [CrossRef]
- Waltz, E.; Llinas, J. Multisensor Data Fusion; Artech House: Boston, MA, USA, 1990; Volume 685. [Google Scholar]
- Hackett, J.K.; Shah, M. Multi-sensor fusion: A perspective. In Proceedings of the 1990 IEEE International Conference on Robotics and Automation, Cincinnati, OH, USA, 13–18 May 1990; pp. 1324–1330. [Google Scholar]
- Grossmann, P. Multisensor data fusion. GEC J. Technol. 1998, 15, 27–37. [Google Scholar]
- Brooks, R.R.; Rao, N.S.; Iyengar, S.S. Resolution of Contradictory Sensor Data. Intell. Autom. Soft Comput. 1997, 3, 287–299. [Google Scholar] [CrossRef]
- Vu, T.D. Vehicle Perception: Localization, Mapping with dEtection, Classification and Tracking of Moving Objects. Ph.D. Thesis, Institut National Polytechnique de Grenoble-INPG, Grenoble, France, 2009. [Google Scholar]
- Vu, T.D.; Aycard, O. Laser-based detection and tracking moving objects using data-driven markov chain monte carlo. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA’09), Kobe, Japan, 12–17 May 2009; pp. 3800–3806. [Google Scholar]
- Bosse, E.; Roy, J.; Grenier, D. Data fusion concepts applied to a suite of dissimilar sensors. In Proceedings of the 1996 Canadian Conference on Electrical and Computer Engineering, Calgary, AL, Canada, 26–29 May 1996; Volume 2, pp. 692–695. [Google Scholar]
- Jeon, D.; Choi, H. Multi-sensor fusion for vehicle localization in real environment. In Proceedings of the 2015 15th International Conference on Control, Automation and Systems (ICCAS), Busan, Korea, 13–16 October 2015; pp. 411–415. [Google Scholar]
- Chibelushi, C.C.; Mason, J.S.; Deravi, F. Feature-Level Data Fusion for Bimodal Person Recognition. Feature Level Datafusion. 1997. Available online: https://www.imvc.co.il/Portals/117/Shmoolik_Mangan.pdf (accessed on 1 August 2019).
- Ross, A.A.; Govindarajan, R. Feature level fusion of hand and face biometrics. In Proceedings of the Biometric Technology for Human Identification II, International Society for Optics and Photonics, Boston; Artech house: Boston, MA, USA, 2005; Volume 5779, pp. 196–204. [Google Scholar]
- Ross, A. Fusion, Feature-Level. In Encyclopedia of Biometrics; Springer: Boston, MA, USA, 2009; pp. 597–602. [Google Scholar] [CrossRef]
- Nehmadi, Y.; Mangan, S.; Shahar, B.-E.; Cohen, A.; Cohen, R.; Goldentouch, L.; Ur, S. Redundancy Schemes with Low-Level Sensor Fusion for Autonomous Vehicles; Google Patents publisher. U.S. Patent 10,445,928, 15 October 2019. [Google Scholar]
- bear.com. North American Bear. Senses and Abilities-North American Bear Center. Available online: https://bear.org/senses-and-abilities/ (accessed on 1 August 2019).
- Crowley, J.L. A Computational Paradigm for Three Dimensional Scene Analysis; Technical Report CMU-RI-TR-84-11; Carnegie Mellon University: Pittsburgh, PA, USA, 1984. [Google Scholar]
- Crowley, J. Navigation for an intelligent mobile robot. IEEE J. Robot. Autom. 1985, 1, 31–41. [Google Scholar] [CrossRef]
- Herman, M.; Kanade, T. Incremental reconstruction of 3D scenes from multiple, complex images. Artif. Intell. 1986, 30, 289–341. [Google Scholar] [CrossRef]
- Brooks, R. Visual map making for a mobile robot. In Proceedings of the 1985 IEEE International Conference on Robotics and Automation, St. Louis, MO, USA, 25–28 March 1985; Volume 2, pp. 824–829. [Google Scholar]
- Smith, R. On the estimation and representation of spatial uncertainty. Int. J. Robot. Res. 1987, 5, 113–119. [Google Scholar] [CrossRef]
- Durrant-Whyte, H.F. Consistent integration and propagation of disparate sensor observations. Int. J. Robot. Res. 1987, 6, 3–24. [Google Scholar] [CrossRef]
- Maheswari, R.U.; Umamaheswari, R. Wind Turbine Drivetrain Expert Fault Detection System: Multivariate Empirical Mode Decomposition based Multi-sensor Fusion with Bayesian Learning Classification. Intell. Autom. Soft Comput. 2019, 10, 296–311. [Google Scholar] [CrossRef]
- Faugeras, O.; Ayache, N.; Faverjon, B. Building visual maps by combining noisy stereo measurements. In Proceedings of the 1986 IEEE International Conference on Robotics and Automation, San Francisco, CA, USA, 7–10 April 1986; Volume 3, pp. 1433–1438. [Google Scholar]
- Hopfield, J.J. Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. USA 1982, 79, 2554–2558. [Google Scholar] [CrossRef] [PubMed]
- Li, S.Z. Invariant surface segmentation through energy minimization with discontinuities. Int. J. Comput. Vis. 1990, 5, 161–194. [Google Scholar] [CrossRef]
- Koch, C.; Marroquin, J.; Yuille, A. Analog “neuronal” networks in early vision. Proc. Natl. Acad. Sci. USA 1986, 83, 4263–4267. [Google Scholar] [CrossRef] [PubMed]
- Poggio, T.; Koch, C. III-Posed problems early vision: From computational theory to analogue networks. Proc. R. Soc. London. Ser. B Biol. Sci. 1985, 226, 303–323. [Google Scholar]
- Blake, A.; Zisserman, A. Visual Reconstruction; MIT Press: Cambridge, MA, USA, 1987. [Google Scholar]
- Ou, S.; Fagg, A.H.; Shenoy, P.; Chen, L. Application of reinforcement learning in multisensor fusion problems with conflicting control objectives. Intell. Autom. Soft Comput. 2009, 15, 223–235. [Google Scholar] [CrossRef]
- Brownston, L.; Farrell, R.; Kant, E.; Martin, N. Programming Expert Systems in OPS5; Addison-Wesley: Boston, MA, USA, 1985. [Google Scholar]
- Forgy, C.L. Rete: A fast algorithm for the many pattern/many object pattern match problem. In Readings in Artificial Intelligence and Databases; Elsevier: Amsterdam, The Netherlands, 1989; pp. 547–559. [Google Scholar]
- Shortliffe, E.H.; Buchanan, B.G. A model of inexact reasoning in medicine. Addison-Wesley Reading, MA. In Rule-Based Expert Systems; John Wiley & Sons: New York, NY, USA, 1984; pp. 233–262. [Google Scholar]
- Hayes-Roth, B. A blackboard architecture for control. Artif. Intell. 1985, 26, 251–321. [Google Scholar] [CrossRef]
- Zadeh, L.A.; Hayes, J.; Michie, D.; Mikulich, L. Machine intelligence. In A Theory of Approximate Reasoning; IEEE: Berkeley, CA, USA, 20 August 1979; Volume 9, pp. 1004–1010. [Google Scholar]
- Duda, R.O.; Hart, P.E.; Nilsson, N.J. Subjective Bayesian methods for rule-based inference systems. In Readings in artificial intelligence; Elsevier: Amsterdam, The Netherlands, 1981; pp. 192–199. [Google Scholar]
- Shafer, G. A mathematical theory of evidence; Princeton University Press: Princeton, NJ, USA, 1976; Volume 42. [Google Scholar]
- Hall, D.L.; McMullen, S.A. Mathematical Techniques in Multisensor Data Fusion; Artech House Inc.: Norwood, MA, USA, 1992; Volume 57. [Google Scholar]
- inforfusion. Information Fusion Definition. Available online: http://www.inforfusion.org/mission.htm (accessed on 1 August 2019).
- Elmenreich, W. An Introduction to Sensor Fusion; Vienna University of Technology: Vienna, Austria, 2002; Volume 502. [Google Scholar]
- Garcia, F.; Martin, D.; de la Escalera, A.; Armingol, J.M. Sensor Fusion Methodology for Vehicle Detection. IEEE Intell. Transp. Syst. Mag. 2017, 9, 123–133. [Google Scholar] [CrossRef]
- Shahian Jahromi, B.; Tulabandhula, T.; Cetin, S. Real-Time Hybrid Multi-Sensor Fusion Framework for Perception in Autonomous Vehicles. Sensors 2019, 19, 4357. [Google Scholar] [CrossRef]
- Uhlmann, J. Simultaneous Map Building and Localization for Real Time Applications. Transfer Thesis, Univ. Oxford, Oxford, UK, 1994. [Google Scholar]
- Uhlmann, J.K. Covariance consistency methods for fault-tolerant distributed data fusion. Inf. Fusion 2003, 4, 201–215. [Google Scholar] [CrossRef]
- Castanedo, F.; Garcia, J.; Patricio, M.A.; Molina, J.M. Analysis of distributed fusion alternatives in coordinated vision agents. In Proceedings of the 2008 11th International Conference on Information Fusion, Cologne, Germany, 30 June–3 July 2008; pp. 1–6. [Google Scholar]
- Bar-Shalom, Y.; Willett, P.K.; Tian, X. Tracking and Data Fusion; YBS Publishing: Storrs, CT, USA, 2011; Volume 11. [Google Scholar]
- Castanedo, F. A review of data fusion techniques. Sci. World J. 2013, 2013. [Google Scholar] [CrossRef]
- Fortmann, T.; Bar-Shalom, Y.; Scheffe, M. Sonar tracking of multiple targets using joint probabilistic data association. IEEE J. Ocean. Eng. 1983, 8, 173–184. [Google Scholar] [CrossRef]
- He, S.; Shin, H.S.; Tsourdos, A. Distributed joint probabilistic data association filter with hybrid fusion strategy. IEEE Trans. Instrum. Meas. 2019, 69, 286–300. [Google Scholar] [CrossRef]
- Goeman, J.J.; Meijer, R.J.; Krebs, T.J.P.; Solari, A. Simultaneous control of all false discovery proportions in large-scale multiple hypothesis testing. Biometrika 2019, 106, 841–856. [Google Scholar] [CrossRef]
- Olfati-Saber, R. Distributed Kalman filtering for sensor networks. In Proceedings of the 2007 46th IEEE Conference on Decision and Control, New Orleans, LA, USA, 12–14 December 2007; pp. 5492–5498. [Google Scholar]
- Zhang, Y.; Huang, Q.; Zhao, K. Hybridizing association rules with adaptive weighted decision fusion for personal credit assessment. Syst. Sci. Control Eng. 2019, 7, 135–142. [Google Scholar] [CrossRef]
- Caltagirone, L.; Bellone, M.; Svensson, L.; Wahde, M. LIDAR–camera fusion for road detection using fully convolutional neural networks. Robot. Auton. Syst. 2019, 111, 125–131. [Google Scholar] [CrossRef]
- Chen, L.; Cetin, M.; Willsky, A.S. Distributed data association for multi-target tracking in sensor networks. In Proceedings of the IEEE Conference on Decision and Control, Plaza de España Seville, Spain, 12–15 December 2005. [Google Scholar]
- Dwivedi, R.; Dey, S. A novel hybrid score level and decision level fusion scheme for cancelable multi-biometric verification. Appl. Intell. 2019, 49, 1016–1035. [Google Scholar] [CrossRef]
- Dasarathy, B.V. Sensor fusion potential exploitation-innovative architectures and illustrative applications. Proc. IEEE 1997, 85, 24–38. [Google Scholar] [CrossRef]
- Steinberg, A.N.; Bowman, C.L. Revisions to the JDL data fusion model. In Handbook of Multisensor Data Fusion; CRC Press: Boca Raton, FL, USA, 2008; pp. 65–88. [Google Scholar]
- White, F.E. Data Fusion Lexicon; Technical Report; Joint Directors of Labs: Washington, DC, USA, 1991. [Google Scholar]
- Carli, R.; Chiuso, A.; Schenato, L.; Zampieri, S. Distributed Kalman filtering based on consensus strategies. IEEE J. Sel. Areas Commun. 2008, 26, 622–633. [Google Scholar] [CrossRef]
- Mahmoud, M.S.; Khalid, H.M. Distributed Kalman filtering: A bibliographic review. IET Control Theory Appl. 2013, 7, 483–501. [Google Scholar] [CrossRef]
- Bowman, C. Data Fusion and Neural Networks, 1643 Hemlock Way Broomfield, CO. Personal communication, regarding Revisions to the JDL Data Fusion Model. 1995. [Google Scholar] [CrossRef]
- Taropa, E.; Srini, V.P.; Lee, W.J.; Han, T.D. Data fusion applied on autonomous ground vehicles. In Proceedings of the 2006 8th International Conference Advanced Communication Technology, Phoenix Park, Korea, 20–22 February 2006; Volume 1, p. 6. [Google Scholar]
- Dieterle, T.; Particke, F.; Patino-Studencki, L.; Thielecke, J. Sensor data fusion of LIDAR with stereo RGB-D camera for object tracking. In Proceedings of the 2017 IEEE SENSORS, Glasgow, UK, 29 October–1 November 2017; pp. 1–3. [Google Scholar]
- NOAA. What Is LiDAR? Available online: https://oceanservice.noaa.gov/facts/lidar.html (accessed on 19 March 2020).
- Yole Developpement, W. Impact of LiDAR by 2032, 1. The Automotive LiDAR Market. Available online: http://www.woodsidecap.com/wp-content/uploads/2018/04/Yole_WCP-LiDAR-Report_April-2018-FINAL.pdf (accessed on 23 March 2020).
- Kim, W.; Tanaka, M.; Okutomi, M.; Sasaki, Y. Automatic labeled LiDAR data generation based on precise human model. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 43–49. [Google Scholar]
- Miltiadou, M.; Michael, G.; Campbell, N.D.; Warren, M.; Clewley, D.; Hadjimitsis, D.G. Open source software DASOS: Efficient accumulation, analysis, and visualisation of full-waveform lidar. In Proceedings of the Seventh International Conference on Remote Sensing and Geoinformation of the Environment (RSCy2019), International Society for Optics and Photonics, Paphos, Cyprus, 18–21 March 2019; Volume 11174, p. 111741. [Google Scholar]
- Hu, P.; Huang, H.; Chen, Y.; Qi, J.; Li, W.; Jiang, C.; Wu, H.; Tian, W.; Hyyppä, J. Analyzing the Angle Effect of Leaf Reflectance Measured by Indoor Hyperspectral Light Detection and Ranging (LiDAR). Remote Sens. 2020, 12, 919. [Google Scholar] [CrossRef]
- Warren, M.E. Automotive LIDAR technology. In Proceedings of the 2019 Symposium on VLSI Circuits, Kyoto, Japan, 9–14 June 2019; pp. C254–C255. [Google Scholar]
- Velodyne. Velodyne Puck Lidar. Available online: https://velodynelidar.com/products/puck/ (accessed on 2 February 2020).
- A1, R.L. RP Lidar A1 Details. Available online: http://www.ksat.com/news/alarming-40-percent-increase-in-pedestrian-deaths-in-2016-in-san-antonio (accessed on 1 August 2019).
- Ouster. Ouster Lidar. Available online: https://ouster.com/lidar-product-details/ (accessed on 2 February 2020).
- Eclipse. Eclipse Mapping Systems. Available online: https://geo-matching.com/airborne-laser-scanning/eclipse-autonomous-mapping-system (accessed on 2 February 2020).
- Quadros, N. Unlocking the characteristics of bathymetric lidar sensors. LiDAR Mag. 2013, 3. Available online: http://lidarmag.com/wp-content/uploads/PDF/LiDARMagazine_Quadros-BathymetricLiDARSensors_Vol3No6.pdf (accessed on 2 February 2020).
- igi global. RGB Camera Details. Available online: https://www.igi-global.com/dictionary/mobile-applications-for-automatic-object-recognition/60647 (accessed on 2 February 2020).
- Sigel, K.; DeAngelis, D.; Ciholas, M. Camera with Object Recognition/data Output. U.S. Patent 6,545,705, 8 April 2003. [Google Scholar]
- De Silva, V.; Roche, J.; Kondoz, A. Robust fusion of LiDAR and wide-angle camera data for autonomous mobile robots. Sensors 2018, 18, 2730. [Google Scholar] [CrossRef] [PubMed]
- Guy, T. Benefits and Advantages of 360° Cameras. Available online: https://www.threesixtycameras.com/pros-cons-every-360-camera/ (accessed on 10 January 2020).
- Myllylä, R.; Marszalec, J.; Kostamovaara, J.; Mäntyniemi, A.; Ulbrich, G.J. Imaging distance measurements using TOF lidar. J. Opt. 1998, 29, 188–193. [Google Scholar] [CrossRef]
- Nair, R.; Lenzen, F.; Meister, S.; Schäfer, H.; Garbe, C.; Kondermann, D. High accuracy TOF and stereo sensor fusion at interactive rates. In Proceedings of the ECCV: European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Volume 7584, pp. 1–11. [Google Scholar] [CrossRef]
- Hewitt, R.A.; Marshall, J.A. Towards intensity-augmented SLAM with LiDAR and ToF sensors. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 1956–1961. [Google Scholar] [CrossRef]
- Turnkey, A. Benefits and Advantages of TOF Industrial Cameras. Available online: http://www.adept.net.au/news/newsletter/201111-nov/article_tof_Mesa.shtml (accessed on 10 January 2020).
- Hinkel, R.; Knieriemen, T. Environment perception with a laser radar in a fast moving robot. In Robot Control 1988 (Syroco’88); Elsevier: Amsterdam, The Netherlands, 1989; pp. 271–277. [Google Scholar]
- fierceelectronics.com. Sensor Types Drive Autonomous Vehicles. Available online: https://www.fierceelectronics.com/components/three-sensor-types-drive-autonomous-vehicles (accessed on 2 October 2019).
- John Campbell, D. Robust and Optimal Methods for Geometric Sensor Data Alignment. Ph.D. Thesis, The Australian National University, Canberra, Australia, 2018. [Google Scholar]
- Maddern, W.; Newman, P. Real-time probabilistic fusion of sparse 3d lidar and dense stereo. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Korea, 9–14 October 2016; pp. 2181–2188. [Google Scholar]
- Realsense, I. Intel Realsense D435 Details. Realsense D435. Available online: https://click.intel.com/intelr-realsensetm-depth-camera-d435.html (accessed on 1 August 2019).
- Realsense, I. Tuning Depth Cameras for Best Performance. Realsense D435 Tuning. Available online: https://dev.intelrealsense.com/docs/tuning-depth-cameras-for-best-performance (accessed on 23 March 2020).
- Mirzaei, F.M.; Kottas, D.G.; Roumeliotis, S.I. 3D LIDAR–camera intrinsic and extrinsic calibration: Identifiability and analytical least-squares-based initialization. Int. J. Robot. Res. 2012, 31, 452–467. [Google Scholar] [CrossRef]
- Dong, W.; Isler, V. A novel method for the extrinsic calibration of a 2d laser rangefinder and a camera. IEEE Sensors J. 2018, 18, 4200–4211. [Google Scholar] [CrossRef]
- Li, J.; He, X.; Li, J. 2D LiDAR and camera fusion in 3D modeling of indoor environment. In Proceedings of the 2015 National Aerospace and Electronics Conference (NAECON), Dayton, OH, USA, 15–19 June 2015; pp. 379–383. [Google Scholar]
- Zhou, L.; Li, Z.; Kaess, M. Automatic Extrinsic Calibration of a Camera and a 3D LiDAR using Line and Plane Correspondences. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS, Madrid, Spain, 1–5 October 2018. [Google Scholar]
- Crowley, J.; Ramparany, F. Mathematical tools for manipulating uncertainty in perception. In Proceedings of the AAAI Workshop on Spatial Reasoning and Multi-Sensor Fusion, St. Charles, IL, USA, 5–7 October 1987. [Google Scholar]
- Jing, L.; Wang, T.; Zhao, M.; Wang, P. An adaptive multi-sensor data fusion method based on deep convolutional neural networks for fault diagnosis of planetary gearbox. Sensors 2017, 17, 414. [Google Scholar] [CrossRef]
- Guindel, C.; Beltrán, J.; Martín, D.; García, F. Automatic extrinsic calibration for lidar-stereo vehicle sensor setups. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017; pp. 1–6. [Google Scholar]
- Nobrega, R.; Quintanilha, J.; O’Hara, C. A noise-removal approach for lidar intensity images using anisotropic diffusion filtering to preserve object shape characteristics. In Proceedings of the ASPRS Annual Conference 2007: Identifying Geospatial Solutions, American Society for Photogrammetry and Remote Sensing, Tampa, FL, USA, 7–11 May 2007; Volume 2, pp. 471–481. [Google Scholar]
- Cao, N.; Zhu, C.; Kai, Y.; Yan, P. A method of background noise reduction in lidar data. Appl. Phys. B 2013, 113. [Google Scholar] [CrossRef]
- Hänsler, E.; Schmidt, G. Topics in Acoustic Echo and Noise Control: Selected Methods for the Cancellation of Acoustical Echoes, the Reduction of Background Noise, and Speech Processing; Springer Science & Business Media: Boston, MA, USA, 2006. [Google Scholar]
- Gannot, S.; Burshtein, D.; Weinstein, E. Iterative and sequential Kalman filter-based speech enhancement algorithms. IEEE Trans. Speech Audio Process. 1998, 6, 373–385. [Google Scholar] [CrossRef]
- Kalman, R.E. Contributions to thetheory of optimal control. Bol. Soc. Mat. Mex. 1960, 5, 102–119. [Google Scholar]
- Gelb, A. Applied Optimal Estimation; MIT Press: Cambridge, MA, USA, 1974. [Google Scholar]
- Julier, S.J.; Uhlmann, J.K. A General Method for Approximating Nonlinear Transformations of Probability Distributions; Technical Report; Robotics Research Group, Department of Engineering Science: Oxford, UK, 1996. [Google Scholar]
- Nørgaard, M.; Poulsen, N.K.; Ravn, O. Advances in Derivative-Free State Estimation for Nonlinear Systems; Informatics and Mathematical Modelling, Technical University of Denmark, DTU: Lyngby, Denmark, 2000. [Google Scholar]
- NøRgaard, M.; Poulsen, N.K.; Ravn, O. New developments in state estimation for nonlinear systems. Automatica 2000, 36, 1627–1638. [Google Scholar] [CrossRef]
- Lefebvre, T.; Bruyninckx, H.; De Schutter, J. Kalman filters for nonlinear systems: A comparison of performance. Int. J. Control 2004, 77, 639–653. [Google Scholar] [CrossRef]
- Julier, S.; Uhlmann, J.; Durrant-Whyte, H.F. A new method for the nonlinear transformation of means and covariances in filters and estimators. IEEE Trans. Autom. Control 2000, 45, 477–482. [Google Scholar] [CrossRef]
- Sorenson, H.W. Kalman Filtering: Theory and Application; IEEE: New York, NY, USA, 1985. [Google Scholar]
- Julier, S.J.; Uhlmann, J.K. New extension of the Kalman filter to nonlinear systems. In Proceedings of the International Society for Optics and Photonics, Signal Processing, Sensor Fusion, and Target Recognition VI, Orlando, FL, USA, 28 July 1997; Volume 3068, pp. 182–193. [Google Scholar]
- Julier, S.J.; Uhlmann, J.K. Unscented filtering and nonlinear estimation. Proc. IEEE 2004, 92, 401–422. [Google Scholar] [CrossRef]
- Wan, E.A.; Van Der Merwe, R. The unscented Kalman filter for nonlinear estimation. In Proceedings of the IEEE 2000 Adaptive Systems for Signal Processing, Communications, and Control Symposium (Cat. No. 00EX373), Lake Louise, AB, Canada, 4 October 2000; pp. 153–158. [Google Scholar]
- Julier, S.J. The spherical simplex unscented transformation. In Proceedings of the 2003 American Control Conference, Denver, CO, USA, 4–6 June 2003; Volume 3, pp. 2430–2434. [Google Scholar]
- Olfati-Saber, R. Distributed Kalman filter with embedded consensus filters. In Proceedings of the 44th IEEE Conference on Decision and Control, Seville, Spain, 15 December 2005; pp. 8179–8184. [Google Scholar]
- Spanos, D.P.; Olfati-Saber, R.; Murray, R.M. Approximate distributed Kalman filtering in sensor networks with quantifiable performance. In Proceedings of the IPSN 2005, Fourth International Symposium on Information Processing in Sensor Networks, Boise, ID, USA, 15 April 2005; pp. 133–139. [Google Scholar]
- Gordon, N.J.; Salmond, D.J.; Smith, A.F. Novel approach to nonlinear/non-Gaussian Bayesian state estimation. Proc. IEE F-Radar Signal Process. 1993, 140, 107–113. [Google Scholar] [CrossRef]
- Thrun, S. Particle filters in robotics. In Proceedings of the Eighteenth Conference on Uncertainty In Artificial Intelligence; Morgan Kaufmann Publishers Inc.: Burlington, MA, USA, 2002; pp. 511–518. [Google Scholar]
- Doucet, A. Sequential Monte Carlo Methods in Practice. Technometrics 2003, 45, 106. [Google Scholar] [CrossRef]
- Bugallo, M.F.; Xu, S.; Djurić, P.M. Performance comparison of EKF and particle filtering methods for maneuvering targets. Digit. Signal Process. 2007, 17, 774–786. [Google Scholar] [CrossRef]
- Van Der Merwe, R.; Doucet, A.; De Freitas, N.; Wan, E.A. The unscented particle filter. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, UK, 3–8 December 2001; pp. 584–590. [Google Scholar]
- Carpenter, J.; Clifford, P.; Fearnhead, P. Improved particle filter for nonlinear problems. IEE Proc.-Radar Sonar Navig. 1999, 146, 2–7. [Google Scholar] [CrossRef]
- Hsiao, K.; Miller, J.; de Plinval-Salgues, H. Particle filters and their applications. Cogn. Robot. 2005, 4. [Google Scholar]
- Waxman, A.; Moigne, J.; Srinivasan, B. Visual navigation of roadways. In Proceedings of the 1985 IEEE International Conference on Robotics and Automation, Louis, MO, USA, 25–28 March 1985; Volume 2, pp. 862–867. [Google Scholar]
- Delahoche, L.; Pégard, C.; Marhic, B.; Vasseur, P. A navigation system based on an ominidirectional vision sensor. In Proceedings of the 1997 IEEE/RSJ International Conference on Intelligent Robot and Systems, Innovative Robotics for Real-World Applications, IROS’97, Grenoble, France, 11 September1997; Volume 2, pp. 718–724. [Google Scholar]
- Zingaretti, P.; Carbonaro, A. Route following based on adaptive visual landmark matching. Robot. Auton. Syst. 1998, 25, 177–184. [Google Scholar] [CrossRef]
- Research, B. Global Vision and Navigation for Autonomous Vehicle.
- Thrun, S. Robotic Mapping: A Survey; CMU-CS-02–111; Morgan Kaufmann Publishers: Burlington, MA, USA, 2002. [Google Scholar]
- Thorpe, C.; Hebert, M.H.; Kanade, T.; Shafer, S.A. Vision and navigation for the Carnegie-Mellon Navlab. IEEE Trans. Pattern Anal. Mach. Intell. 1988, 10, 362–373. [Google Scholar] [CrossRef]
- Zimmer, U.R. Robust world-modelling and navigation in a real world. Neurocomputing 1996, 13, 247–260. [Google Scholar] [CrossRef][Green Version]
- Research, K. Autonomous Navigation Market: Investigation and Growth Forecasted until the End of 2025. Marketwath.com Press Release. Available online: https://www.marketwatch.com/press-release/autonomous-navigation-market-investigation-and-growth-forecasted-until-the-end-of-2025-2019-11-13 (accessed on 12 November 2019).
- Brooks, R. A robust layered control system for a mobile robot. IEEE J. Robot. Autom. 1986, 2, 14–23. [Google Scholar] [CrossRef]
- Danescu, R.G. Obstacle detection using dynamic Particle-Based occupancy grids. In Proceedings of the 2011 International Conference on Digital Image Computing: Techniques and Applications, Noosa, QLD, Australia, 6–8 December 2011; pp. 585–590. [Google Scholar]
- Leibe, B.; Seemann, E.; Schiele, B. Pedestrian detection in crowded scenes. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 878–885. [Google Scholar]
- Lwowski, J.; Kolar, P.; Benavidez, P.; Rad, P.; Prevost, J.J.; Jamshidi, M. Pedestrian detection system for smart communities using deep Convolutional Neural Networks. In Proceedings of the 2017 12th System of Systems Engineering Conference (SoSE), Waikoloa, HI, USA, 18–21 June 2017; pp. 1–6. [Google Scholar]
- Kortenkamp, D.; Weymouth, T. Topological mapping for mobile robots using a combination of sonar and vision sensing. Proc. AAAI 1994, 94, 979–984. [Google Scholar]
- Engelson, S.P.; McDermott, D.V. Error correction in mobile robot map learning. In Proceedings of the IEEE International Conference on Robotics and Automation, Nice, France, 12–14 May 1992; pp. 2555–2560. [Google Scholar]
- Kuipers, B.; Byun, Y.T. A robot exploration and mapping strategy based on a semantic hierarchy of spatial representations. Robot. Auton. Syst. 1991, 8, 47–63. [Google Scholar] [CrossRef]
- Thrun, S.; Bücken, A. Integrating grid-based and topological maps for mobile robot navigation. In Proceedings of the National Conference on Artificial Intelligence, Oregon, Portland, 4–8 August 1996; pp. 944–951. [Google Scholar]
- Thrun, S.; Buecken, A.; Burgard, W.; Fox, D.; Wolfram, A.B.; Fox, B.D.; Fröhlinghaus, T.; Hennig, D.; Hofmann, T.; Krell, M.; et al. Map Learning and High-Speed Navigation in RHINO; MIT/AAAI Press: Cambridge, MA, USA, 1996. [Google Scholar]
- Moravec, H.P. Sensor fusion in certainty grids for mobile robots. AI Mag. 1988, 9, 61. [Google Scholar]
- Elfes, A. Occupancy Grids: A Probabilistic Framework for Robot Perception and Navigation. Ph.D. Thesis, Carnegie-Mellon University, Pittsburgh, PA, USA, 1989. [Google Scholar]
- Borenstein, J.; Koren, Y. The vector field histogram-fast obstacle avoidance for mobile robots. IEEE Trans. Robot. Autom. 1991, 7, 278–288. [Google Scholar] [CrossRef]
- Ramsdale, J.D.; Balme, M.R.; Conway, S.J.; Gallagher, C.; van Gasselt, S.A.; Hauber, E.; Orgel, C.; Séjourné, A.; Skinner, J.A.; Costard, F.; et al. Grid-based mapping: A method for rapidly determining the spatial distributions of small features over very large areas. Planet. Space Sci. 2017, 140, 49–61. [Google Scholar] [CrossRef]
- Jiang, Z.; Zhu, J.; Jin, C.; Xu, S.; Zhou, Y.; Pang, S. Simultaneously merging multi-robot grid maps at different resolutions. In Multimedia Tools and Applications; Springer Science & Business Media: Berlin, Germany, 2019; pp. 1–20. [Google Scholar]
- Burgard, W.; Fox, D.; Hennig, D.; Schmidt, T. Estimating the absolute position of a mobile robot using position probability grids. In Proceedings of the National Conference on Artificial Intelligence, Portland, Oregon, 4–8 August 1996; pp. 896–901. [Google Scholar]
- Gutmann, J.S.; Schlegel, C. Amos: Comparison of scan matching approaches for self-localization in indoor environments. In Proceedings of the First Euromicro Workshop on Advanced Mobile Robots (EUROBOT’96), Kaiserslautern, Germany, 9–11 October 1996; pp. 61–67. [Google Scholar]
- Zhang, Z.; Deriche, R.; Faugeras, O.; Luong, Q.T. A robust technique for matching two uncalibrated images through the recovery of the unknown epipolar geometry. Artif. Intell. 1995, 78, 87–119. [Google Scholar] [CrossRef]
- Lu, F.; Milios, E. Robot pose estimation in unknown environments by matching 2d range scans. J. Intell. Robot. Syst. 1997, 18, 249–275. [Google Scholar] [CrossRef]
- Buschka, P. An Investigation of Hybrid Maps for Mobile Robots. Ph.D. Thesis, Örebro universitetsbibliotek, Örebro, Sweden, 2005. [Google Scholar]
- Fernández-Madrigal, J.A. Simultaneous Localization and Mapping for Mobile Robots: Introduction and Methods: Introduction and Methods; IGI Global: Philadelphia, PA, USA, 2012. [Google Scholar]
- Thrun, S. Robotic mapping: A survey. Explor. Artif. Intell. New Millenn. 2002, 1, 1–35. [Google Scholar]
- Leonard, J.J.; Durrant-Whyte, H.F.; Cox, I.J. Dynamic map building for an autonomous mobile robot. Int. J. Robot. Res. 1992, 11, 286–298. [Google Scholar] [CrossRef]
- Dissanayake, M.W.M.G.; Newman, P.; Clark, S.; Durrant-Whyte, H.F.; Csorba, M. A solution to the simultaneous localization and map building (SLAM) problem. IEEE Trans. Robot. Autom. 2001, 17, 229–241. [Google Scholar] [CrossRef]
- Mirowski, P.; Grimes, M.; Malinowski, M.; Hermann, K.M.; Anderson, K.; Teplyashin, D.; Simonyan, K.; Zisserman, A.; Hadsell, R.; et al. Learning to navigate in cities without a map. In Proceedings of the Advances in Neural Information Processing Systems; Montreal Convention Centre, Montreal, QC, Canada, 3–8 December 2018; pp. 2419–2430. [Google Scholar]
- Pritsker, A. Introduction to Stimulation and Slam II, 3rd ed.; U.S. Department of Energy Office of Scientific and Technical Information: Oak Ridge, TN, USA, 1986. [Google Scholar]
- Davison, A.J.; Reid, I.D.; Molton, N.D.; Stasse, O. MonoSLAM: Real-time single camera SLAM. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1052–1067. [Google Scholar] [CrossRef] [PubMed]
- Sturm, J.; Engelhard, N.; Endres, F.; Burgard, W.; Cremers, D. A benchmark for the evaluation of RGB-D SLAM systems. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura, Portugal, 7–12 October 2012; pp. 573–580. [Google Scholar]
- Wikipedia.com. List of Slam Methods. Available online: https://en.wikipedia.org/wiki/List_of_SLAM_Methods (accessed on 1 August 2019).
- Aguilar, W.G.; Morales, S.; Ruiz, H.; Abad, V. RRT* GL based optimal path planning for real-time navigation of UAVs. In International Work-Conference on Artificial Neural Networks; Springer: Boston, MA, USA, 2017; pp. 585–595. [Google Scholar]
- Huang, S.; Dissanayake, G. Robot Localization: An Introduction. In Wiley Encyclopedia of Electrical and Electronics Engineering; John Wiley & Sons: New York, NY, USA, 1999; pp. 1–10. [Google Scholar]
- Huang, S.; Dissanayake, G. Convergence and consistency analysis for extended Kalman filter based SLAM. IEEE Trans. Robot. 2007, 23, 1036–1049. [Google Scholar] [CrossRef]
- Liu, H.; Darabi, H.; Banerjee, P.; Liu, J. Survey of wireless indoor positioning techniques and systems. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2007, 37, 1067–1080. [Google Scholar] [CrossRef]
- Leonard, J.J.; Durrant-Whyte, H.F. Mobile robot localization by tracking geometric beacons. IEEE Trans. Robot. Autom. 1991, 7, 376–382. [Google Scholar] [CrossRef]
- Betke, M.; Gurvits, L. Mobile robot localization using landmarks. IEEE Trans. Robot. Autom. 1997, 13, 251–263. [Google Scholar] [CrossRef]
- Thrun, S.; Fox, D.; Burgard, W.; Dellaert, F. Robust Monte Carlo localization for mobile robots. Artif. Intell. 2001, 128, 99–141. [Google Scholar] [CrossRef]
- Kwon, S.; Yang, K.; Park, S. An effective kalman filter localization method for mobile robots. In Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 9–15 October 2006; pp. 1524–1529. [Google Scholar]
- Ojeda, L.; Borenstein, J. Personal dead-reckoning system for GPS-denied environments. In Proceedings of the IEEE International Workshop on Safety, Security and Rescue Robotics, SSRR 2007, Rome, Italy, 27–29 September 2007; pp. 1–6. [Google Scholar]
- Levi, R.W.; Judd, T. Dead Reckoning Navigational System Using Accelerometer to Measure Foot Impacts. U.S. Patent 5,583,776, 1996. [Google Scholar]
- Elnahrawy, E.; Li, X.; Martin, R.P. The limits of localization using signal strength: A comparative study. In Proceedings of the 2004 First Annual IEEE Communications Society Conference on Sensor and Ad Hoc Communications and Networks, Santa Clara, CA, USA, 4–7 October 2004; pp. 406–414. [Google Scholar]
- Neves, A.; Fonseca, H.C.; Ralha, C.G. Location agent: A study using different wireless protocols for indoor localization. Int. J. Wirel. Commun. Mob. Comput. 2013, 1, 1–6. [Google Scholar] [CrossRef]
- Whitehouse, K.; Karlof, C.; Culler, D. A practical evaluation of radio signal strength for ranging-based localization. ACM SIGMOBILE Mob. Comput. Commun. Rev. 2007, 11, 41–52. [Google Scholar] [CrossRef]
- He, S.; Chan, S.H.G. Wi-Fi fingerprint-based indoor positioning: Recent advances and comparisons. IEEE Commun. Surv. Tutor. 2016, 18, 466–490. [Google Scholar] [CrossRef]
- Wang, Y.; Ye, Q.; Cheng, J.; Wang, L. RSSI-based bluetooth indoor localization. In Proceedings of the 2015 11th International Conference on Mobile Ad-hoc and Sensor Networks (MSN), Shenzhen, China, 16–18 December 2015; pp. 165–171. [Google Scholar]
- Howell, E.; NAV Star. Navstar: GPS Satellite Network. Available online: https://www.space.com/19794-navstar.html (accessed on 1 August 2019).
- Robotics, A. Experience the New Mobius. Available online: https://www.asirobots.com/platforms/mobius/ (accessed on 1 August 2019).
- Choi, B.S.; Lee, J.J. Sensor network based localization algorithm using fusion sensor-agent for indoor service robot. IEEE Trans. Consum. Electron. 2010, 56, 1457–1465. [Google Scholar] [CrossRef]
- Ramer, C.; Sessner, J.; Scholz, M.; Zhang, X.; Franke, J. Fusing low-cost sensor data for localization and mapping of automated guided vehicle fleets in indoor applications. In Proceedings of the 2015 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI), San Diego, CA, USA, 14–16 September 2015; pp. 65–70. [Google Scholar]
- Fontanelli, D.; Ricciato, L.; Soatto, S. A fast ransac-based registration algorithm for accurate localization in unknown environments using lidar measurements. In Proceedings of the IEEE International Conference on Automation Science and Engineering, CASE 2007, Scottsdale, AZ, USA, 22–25 September 2007; pp. 597–602. [Google Scholar]
- Wan, K.; Ma, L.; Tan, X. An improvement algorithm on RANSAC for image-based indoor localization. In Proceedings of the 2016 International Conference on Wireless Communications and Mobile Computing Conference (IWCMC), An improvement algorithm on RANSAC for image-based indoor localization, Paphos, Cyprus, 5–9 September 2016; pp. 842–845. [Google Scholar]
- Biswas, J.; Veloso, M. Depth camera based indoor mobile robot localization and navigation. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation (ICRA), Saint Paul, MN, USA, 14–18 May 2012; pp. 1697–1702. [Google Scholar]
- Vive, W.H. HTC Vive Details. Available online: https://en.wikipedia.org/wiki/HTC_Vive (accessed on 1 August 2019).
- Buniyamin, N.; Ngah, W.W.; Sariff, N.; Mohamad, Z. A simple local path planning algorithm for autonomous mobile robots. Int. J. Syst. Appl. Eng. Dev. 2011, 5, 151–159. [Google Scholar]
- Popović, M.; Vidal-Calleja, T.; Hitz, G.; Sa, I.; Siegwart, R.; Nieto, J. Multiresolution mapping and informative path planning for UAV-based terrain monitoring. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 1382–1388. [Google Scholar]
- Laghmara, H.; Boudali, M.; Laurain, T.; Ledy, J.; Orjuela, R.; Lauffenburger, J.; Basset, M. Obstacle Avoidance, Path Planning and Control for Autonomous Vehicles. In Proceedings of the 2019 IEEE Intelligent Vehicles Symposium (IV), Paris, France, 9–12 June 2019; pp. 529–534. [Google Scholar]
- Rashid, A.T.; Ali, A.A.; Frasca, M.; Fortuna, L. Path planning with obstacle avoidance based on visibility binary tree algorithm. Robot. Auton. Syst. 2013, 61, 1440–1449. [Google Scholar] [CrossRef]
- Wagner, G.; Choset, H. M*: A complete multirobot path planning algorithm with performance bounds. In Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011; pp. 3260–3267. [Google Scholar]
- Urdiales, C.; Bandera, A.; Arrebola, F.; Sandoval, F. Multi-level path planning algorithm for autonomous robots. Electron. Lett. 1998, 34, 223–224. [Google Scholar] [CrossRef]
- Mac, T.T.; Copot, C.; Tran, D.T.; De Keyser, R. Heuristic approaches in robot path planning: A survey. Robot. Auton. Syst. 2016, 86, 13–28. [Google Scholar] [CrossRef]
- Vokhmintsev, A.; Timchenko, M.; Melnikov, A.; Kozko, A.; Makovetskii, A. Robot path planning algorithm based on symbolic tags in dynamic environment. In Proceedings of the Applications of Digital Image Processing XL, International Society for Optics and Photonics, San Diego, CA, USA, 7–10 August 2017; Volume 10396, p. 103962E. [Google Scholar]
- Marin-Plaza, P.; Hussein, A.; Martin, D.; de la Escalera, A. Global and local path planning study in a ROS-based research platform for autonomous vehicles. J. Adv. Transp. 2018. [Google Scholar] [CrossRef]
- Bhattacharya, P.; Gavrilova, M.L. Voronoi diagram in optimal path planning. In Proceedings of the 4th International Symposium on Voronoi Diagrams in Science and Engineering (ISVD 2007), Glamorgan, UK, 9–11 July 2007; pp. 38–47. [Google Scholar]
- Canny, J. A new algebraic method for robot motion planning and real geometry. In Proceedings of the 28th Annual Symposium on Foundations of Computer Science (sfcs 1987), Los Angeles, CA, USA, 12–14 October 1987; pp. 39–48. [Google Scholar]
- Skiena, S. Dijkstra’s algorithm. In Implementing Discrete Mathematics: Combinatorics and Graph Theory with Mathematica; Addison-Wesley: Reading, MA, USA, 1990; pp. 225–227. [Google Scholar]
- Dechter, R.; Pearl, J. Generalized best-first search strategies and the optimality of A. J. ACM 1985, 32, 505–536. [Google Scholar] [CrossRef]
- Yang, S.X.; Luo, C. A neural network approach to complete coverage path planning. IEEE Trans. Syst. Man Cybern. Part B Cybernet. 2004, 34, 718–724. [Google Scholar] [CrossRef]
- Piazzi, A.; Bianco, C.L.; Bertozzi, M.; Fascioli, A.; Broggi, A. Quintic G/sup 2/-splines for the iterative steering of vision-based autonomous vehicles. IEEE Trans. Intell. Transp. Syst. 2002, 3, 27–36. [Google Scholar] [CrossRef]
- Rastelli, J.P.; Lattarulo, R.; Nashashibi, F. Dynamic trajectory generation using continuous-curvature algorithms for door to door assistance vehicles. In Proceedings of the 2014 IEEE Intelligent Vehicles Symposium, Dearborn, MI, USA, 8–11 June 2014; pp. 510–515. [Google Scholar]
- Reeds, J.; Shepp, L. Optimal paths for a car that goes both forwards and backwards. Pac. J. Math. 1990, 145, 367–393. [Google Scholar] [CrossRef]
- Gim, S.; Adouane, L.; Lee, S.; Derutin, J.P. Clothoids composition method for smooth path generation of car-like vehicle navigation. J. Intell. Robot. Syst. 2017, 88, 129–146. [Google Scholar] [CrossRef]
- Kumar, P.B.; Sahu, C.; Parhi, D.R.; Pandey, K.K.; Chhotray, A. Static and dynamic path planning of humanoids using an advanced regression controller. Sci. Iran. 2019, 26, 375–393. [Google Scholar] [CrossRef]
- Tuba, E.; Strumberger, I.; Bacanin, N.; Tuba, M. Optimal Path Planning in Environments with Static Obstacles by Harmony Search Algorithm. In International Conference on Harmony Search Algorithm; Springer: Boston, MA, USA, 2019; pp. 186–193. [Google Scholar]
- Dutta, A.K.; Debnath, S.K.; Das, S.K. Path-Planning of Snake-Like Robot in Presence of Static Obstacles Using Critical-SnakeBug Algorithm. In Advances in Computer, Communication and Control; Springer: Boston, MA, USA, 2019; pp. 449–458. [Google Scholar]
- Gabardos, B.I.; Passot, J.B. Systems and Methods for Dynamic Route Planning in Autonomous Navigation. U.S. Patent App. 16/454,217, 2 January 2020. [Google Scholar]
- Connell, D.; La, H.M. Dynamic path planning and replanning for mobile robots using rrt. In Proceedings of the 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Banff, AB, Canada, 5 October 2017; pp. 1429–1434. [Google Scholar]
- Liu, Y.; Ma, J.; Zang, S.; Min, Y. Dynamic Path Planning of Mobile Robot Based on Improved Ant Colony Optimization Algorithm. In Proceedings of the 2019 8th International Conference on Networks, Communication and Computing, Luoyang, China, 13–15 December 2019; 2019; pp. 248–252. [Google Scholar]
- Wang, C.C.; Thorpe, C.; Thrun, S.; Hebert, M.; Durrant-Whyte, H. Simultaneous localization, mapping and moving object tracking. Int. J. Robot. Res. 2007, 26, 889–916. [Google Scholar] [CrossRef]
- Saunders, J.; Call, B.; Curtis, A.; Beard, R.; McLain, T. Static and dynamic obstacle avoidance in miniature air vehicles. In Infotech@ Aerospace; BYU ScholarsArchive; BYU: Provo, UT, USA, 2005; p. 6950. [Google Scholar]
- Chu, K.; Lee, M.; Sunwoo, M. Local path planning for off-road autonomous driving with avoidance of static obstacles. IEEE Trans. Intell. Transp. Syst. 2012, 13, 1599–1616. [Google Scholar] [CrossRef]
- Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2005, San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
- Fox, D.; Burgard, W.; Thrun, S. The dynamic window approach to collision avoidance. IEEE Robot. Autom. Mag. 1997, 4, 23–33. [Google Scholar] [CrossRef]
- Elfes, A. Using occupancy grids for mobile robot perception and navigation. Computer 1989, 22, 46–57. [Google Scholar] [CrossRef]
- Cho, J.H.; Pae, D.S.; Lim, M.T.; Kang, T.K. A Real-Time Obstacle Avoidance Method for Autonomous Vehicles Using an Obstacle-Dependent Gaussian Potential Field. J. Adv. Transp. 2018. [Google Scholar] [CrossRef]
- Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. arXiv 2017, arXiv:1612.08242. [Google Scholar]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 1137–1149. [Google Scholar] [CrossRef]
- Akhtar, M.R.; Qin, H.; Chen, G. Velodyne LiDAR and monocular camera data fusion for depth map and 3D reconstruction. In Proceedings of the Eleventh International Conference on Digital Image Processing (ICDIP 2019), International Society for Optics and Photonics, Guangzhou, China, 10–13 May 2019; Volume 11179, p. 111790. [Google Scholar]
- Jin, Z.; Shao, Y.; So, M.; Sable, C.; Shlayan, N.; Luchtenburg, D.M. A Multisensor Data Fusion Approach for Simultaneous Localization and Mapping. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, NZ, USA, 27–30 October 2019; pp. 1317–1322. [Google Scholar]
- Andresen, L.; Brandemuehl, A.; Hönger, A.; Kuan, B.; Vödisch, N.; Blum, H.; Reijgwart, V.; Bernreiter, L.; Schaupp, L.; Chung, J.J.; et al. Fast and Accurate Mapping for Autonomous Racing. arXiv 2020, arXiv:2003.05266. [Google Scholar]
- Zhang, X.; Rad, A.B.; Wong, Y.K. A robust regression model for simultaneous localization and mapping in autonomous mobile robot. J. Intell. Robot. Syst. 2008, 53, 183–202. [Google Scholar] [CrossRef]
- Zhang, X.; Rad, A.B.; Wong, Y.K. Sensor fusion of monocular cameras and laser rangefinders for line-based simultaneous localization and mapping (SLAM) tasks in autonomous mobile robots. Sensors 2012, 12, 429–452. [Google Scholar] [CrossRef] [PubMed]
- Wei, P.; Cagle, L.; Reza, T.; Ball, J.; Gafford, J. LiDAR and camera detection fusion in a real-time industrial multi-sensor collision avoidance system. Electronics 2018, 7, 84. [Google Scholar] [CrossRef]
- Wang, X. A Driverless Vehicle Vision Path Planning Algorithm for Sensor Fusion. In Proceedings of the 2019 IEEE 2nd International Conference on Automation, Electronics and Electrical Engineering (AUTEEE), Shenyang, China, 22–24 November 2019; pp. 214–218. [Google Scholar]
- Ali, M.A.; Mailah, M. Path planning and control of mobile robot in road environments using sensor fusion and active force control. IEEE Trans. Veh. Technol. 2019, 68, 2176–2195. [Google Scholar] [CrossRef]
- Gwon, J.; Kim, H.; Bae, H.; Lee, S. Path Planning of a Sweeping Robot Based on Path Estimation of a Curling Stone Using Sensor Fusion. Electronics 2020, 9, 457. [Google Scholar] [CrossRef]
- Xi, Y. Improved Intelligent Water Droplet Navigation Method for Mobile Robot Based on Multi-sensor Fusion. In Proceedings of the 2019 IEEE International Conference on Power, Intelligent Computing and Systems (ICPICS), Shenyang, China, 12 July 2019; pp. 417–420. [Google Scholar]
- Sabe, K.; Fukuchi, M.; Gutmann, J.S.; Ohashi, T.; Kawamoto, K.; Yoshigahara, T. Obstacle avoidance and path planning for humanoid robots using stereo vision. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA’04), New Orleans, LA, USA, 26 April–1 May 2004; Volume 1, pp. 592–597. [Google Scholar]
- Rasshofer, R.H.; Spies, M.; Spies, H. Influences of weather phenomena on automotive laser radar systems. Adv. Radio Sci. 2011, 9, 49–60. [Google Scholar] [CrossRef]
- Kytö, M.; Nuutinen, M.; Oittinen, P. Method for measuring stereo camera depth accuracy based on stereoscopic vision. In Proceedings of the Three-Dimensional Imaging, Interaction, and Measurement, International Society for Optics and Photonics, San Francisco, CA, USA, 24–27 January 2011; Volume 7864, p. 78640. [Google Scholar]
- Duong Pham, T.; Shrestha, R.; Virkutyte, J.; Sillanpää, M. Recent studies in environmental applications of ultrasound. Can. J. Civ. Eng. 2009, 36, 1849–1858. [Google Scholar] [CrossRef]
- Dan, B.K.; Kim, Y.S.; Jung, J.Y.; Ko, S.J.; et al. Robust people counting system based on sensor fusion. IEEE Trans. Consum. Electron. 2012, 58, 1013–1021. [Google Scholar] [CrossRef]
- Pacha, A. Sensor Fusion for Robust Outdoor Augmented Reality Tracking on Mobile Devices; GRIN Verlag: München, Germany, 2013. [Google Scholar]
- Breitenstein, M.D.; Reichlin, F.; Leibe, B.; Koller-Meier, E.; Van Gool, L. Online multiperson tracking-by-detection from a single, uncalibrated camera. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1820–1833. [Google Scholar] [CrossRef]
- Stein, G. Barrier and Guardrail Detection Using a Single Camera. U.S. Patent 9,280,711, 8 March 2016. [Google Scholar]
- Boreczky, J.S.; Rowe, L.A. Comparison of video shot boundary detection techniques. J. Electron. Imag. 1996, 5, 122–129. [Google Scholar] [CrossRef]
- Sheikh, Y.; Shah, M. Bayesian modeling of dynamic scenes for object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1778–1792. [Google Scholar] [CrossRef] [PubMed]
- John, V.; Long, Q.; Xu, Y.; Liu, Z.; Mita, S. Sensor Fusion and Registration of Lidar and Stereo Camera without Calibration Objects. IEICE TRANSACTIONS Fundam. Electron. Commun. Comput. Sci. 2017, 100, 499–509. [Google Scholar] [CrossRef]
- Huber, D.; Kanade, T. Integrating LIDAR into stereo for fast and improved disparity computation. In Proceedings of the 2011 International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission (3DIMPVT), Hangzhou, China, 16–19 May 2011; pp. 405–412. [Google Scholar]
- Banerjee, K.; Notz, D.; Windelen, J.; Gavarraju, S.; He, M. Online Camera LiDAR Fusion and Object Detection on Hybrid Data for Autonomous Driving. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium (IV), Changshu, Suzhou, China, 26–30 June 2018; pp. 1632–1638. [Google Scholar]
- Manghat, S.K.; El-Sharkawy, M. A Multi Sensor Real-time Tracking with LiDAR and Camera. In Proceedings of the 2020 10th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 6–8 January 2020; pp. 668–672. [Google Scholar]
- Asvadi, A.; Garrote, L.; Premebida, C.; Peixoto, P.; Nunes, U.J. Multimodal vehicle detection: Fusing 3D-LIDAR and color camera data. Pattern Recognit. Lett. 2018, 115, 20–29. [Google Scholar] [CrossRef]
- Dollar, P.; Appel, R.; Belongie, S.; Perona, P. Fast Feature Pyramids for Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1532–1545. [Google Scholar] [CrossRef]
- Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 1440–1448. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision; Springer: Boston, MA, USA, 2016; pp. 21–37. [Google Scholar]
- Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? the kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, Rhode Island, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
- Simony, M.; Milzy, S.; Amendey, K.; Gross, H.M. Complex-YOLO: An Euler-region-proposal for real-time 3D object detection on point clouds. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).