Next Article in Journal
The Design of a Glycerol Concentration Sensor Based on an LRSPP Hybrid Photonic Biosensor
Next Article in Special Issue
Correction of Error of Airborne Anemometers Caused by Self-Excited Air Turbulence
Previous Article in Journal
Innovative Spectrum Handoff Process Using a Machine Learning-Based Metaheuristic Algorithm
Previous Article in Special Issue
Advanced Sensing and Safety Control for Connected and Automated Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Autonomous Vehicle Dataset with Real Multi-Driver Scenes and Biometric Data

División de Sistemas e Ingeniería Electrónica (DSIE), Campus Muralla del Mar, s/n, Universidad Politécnica de Cartagena, 30202 Cartagena, Spain
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 2009; https://doi.org/10.3390/s23042009
Submission received: 13 December 2022 / Revised: 27 January 2023 / Accepted: 30 January 2023 / Published: 10 February 2023

Abstract

:
The development of autonomous vehicles is becoming increasingly popular and gathering real-world data is considered a valuable task. Many datasets have been published recently in the autonomous vehicle sector, with synthetic datasets gaining particular interest due to availability and cost. For a real implementation and correct evaluation of vehicles at higher levels of autonomy, it is also necessary to consider human interaction, which is precisely something that lacks in existing datasets. In this article the UPCT dataset is presented, a public dataset containing high quality, multimodal data obtained using state-of-the-art sensors and equipment installed onboard the UPCT’s CICar autonomous vehicle. The dataset includes data from a variety of perception sensors including 3D LiDAR, cameras, IMU, GPS, encoders, as well as driver biometric data and driver behaviour questionnaires. In addition to the dataset, the software developed for data synchronisation and processing has been made available. The quality of the dataset was validated using an end-to-end neural network model with multiple inputs to obtain the speed and steering wheel angle and it obtained very promising results.

1. Introduction

Rapid advances in artificial intelligence, electronics, information and communications technology (leading to miniaturisation and improved performance of computers, sensors and networks) has led to the development of new approaches to Autonomous Vehicle technologies [1]. This together with new consumption habits and environmental awareness, where technology is vital and allows us to be more efficient and sustainable, has led to a considerable increase in the amount of research carried out on autonomous vehicles, making it the latest trend in the automotive industry [2]. Evidently, there is plenty of motivation and enthusiasm for speeding up progress, especially with the recent success of Big Data, Machine Learning and Deep Neural Networks.
Given the growing popularity of the development of autonomous vehicles, the collection of real data is considered a valuable task, with it being necessary for this sector to provide high-quality, multimodal and real-world datasets which can be used for benchmarking, simulation development, algorithms testing and diverse computer vision training exercises, among others.
The vehicle used for the data collection is usually equipped with a variety of sensors, such as cameras, Light Detection and Ranging (LiDAR) sensors, RADAR, GPS and Inertial Measurement Units (IMU). The raw data obtained by these sensors is recorded on a disk while the vehicle is being driven manually. Subsequently, the recorded data can be used to train and test different algorithms for autonomous driving, e.g., vehicle/pedestrian detection and tracking, Simultaneous Localization and Mapping (SLAM) and motion estimation [3].
In this context, many datasets have been published, a summary of the most popular datasets and their features is presented in Table 1. These datasets vary greatly in terms of traffic conditions, sensor configuration, application focus, data format, size, tool support, as well as other aspects. The most sought-after datasets dedicated to autonomous vehicle systems (AVS) are the so-called multimodal datasets. These datasets have gained particular attention recently, as datasets containing data from an individual sensor are insufficient to provide a complete perception of the environment. Furthermore, the most exploited sensors in this field, such as cameras, LIDARs, radars, etc., offer complementary data and their collaboration can guarantee a better understanding of the surroundings [3].
Table 1 provides a comparison of the main existing datasets, at both an academic and professional level and consists of a brief survey of datasets relevant to the development of autonomous driving systems. We focus on the most comparable and recent datasets, which strongly emphasise multimodal sensor data. Although they are not recent, we also include the KITTI and Udacity datasets as we consider them to be two of the most significant early driving datasets. We present the datasets in chronological order.
Despite the large number of existing studies, most of these datasets do not provide raw data, but instead offer labelled data to support training and evaluation, in particular semantic segmentation techniques. Obtaining real labelled data in large quantities is far from trivial. To start with, it is arduous and expensive to deploy multiple vehicles to collect images and data in a wide range of environmental, weather and lighting conditions. Secondly, the task of manually classifying each image is extremely time-consuming. Lastly, the accuracy of manually produced labels may be inconsistent across the dataset. These reasons, along with the level of fidelity achieved by 3D graphics engines, have encouraged the creation of synthetic datasets of artificial data based on scenes recreated by simulators [5].
As stated in the work by [19], this method of offering already labelled and even segmented data often presents problems in data quality due to the methods or models used. Another disadvantage of those models trained using only synthetic datasets is that in real-world scenarios, these tend to perform poorly, suffering from domain shift [20,21].
On the other hand, for a real implementation and correct evolution of autonomous vehicles at levels 4–5, it is also necessary to consider human interaction. Whether to infer a pedestrian’s intent to cross the road, identify a driver’s intent to perform a certain manoeuvre or detect potentially reckless moves, autonomous vehicles must have a high-level understanding of human behaviour. In most existing datasets, it is precisely this human data factor which is lacking. As can be seen in Table 1, apart from our proposal, the UPCT dataset, existing datasets dedicated to autonomous vehicles do not include biometric data or driver behaviour data.
In this article, we present the UPCT dataset, a public dataset of high-quality, multimodal data, obtained using state-of-the-art sensors equipped by the CICar autonomous vehicle belonging to the UPCT. The CICar includes sensors such as cameras, LiDAR, IMU, GPS and encoders, as well as biometric data from the drivers and driver behaviour questionnaires. The UPCT dataset offers the data acquired during 20 manual driving tests carried out by different drivers on an urban circuit, which consists of a circular route in the Spanish town of Fuente Alamo. To facilitate the use of the dataset, three large subgroups of data have been differentiated: Perception, Positioning and Driver data (biometrics and Driver Behaviour Questionnaire) and both the pre-processed raw data and the code which facilitates its use have been made available for download.

2. Materials and Methods

2.1. Experimental Design

To obtain the data, we decided to carry out ad hoc driving tests with a group of 50 healthy subjects of different ages and sex from the Region of Murcia (Spain), following the distribution shown in Table 2. The subjects were in possession of a valid type B driving licence (for driving cars, vans and, in general, vehicles with a maximum authorised mass of 3500 kg) at the time of the test. After performing the tests, the results of some subjects were excluded due to technical problems during the performance of the test or during the recording of the results, leaving a total of n = 20 subjects (11 male/9 female) with valid raw data to make up the final dataset.

2.1.1. Driver Test Design

Before starting the experiment, in addition to the informed consent, each subject filled in two questionnaires: (1) the Biographic Questionnaire and (2) the Driver Behaviour Questionnaire.
  • The Biographic Questionnaire identifies key facts about the subject, such as gender, age and driving record.
  • The Driver Behaviour Questionnaire (DBQ) collects self-reported data from the drivers, as there are no objective records of driving behaviour and previous traffic violations. The original DBQ consists of 50 items and is used to score the following three underlying factors: errors, violations, and lapses.
For this experiment, we have chosen to use the Spanish Driver Behaviour Questionnaire (SDBQ) [22], a shorter version adapted to Spanish drivers consisting of 28 items adapted to the peculiarities of the Spanish population. The version used consists of four factors, composed as follows: 6 traffic law violation items, 6 violation/aggressive manifestation items, 8 error items, and 8 lapse items. Participants were asked to indicate, on a 5-point scale, how often they had been involved in the behaviours or situations mentioned in the questionnaire.

2.1.2. Driving Test Design

The driving test consists of one of the participating drivers, who has been equipped with a non-invasive smart band device, manually driving the UPCT-CICar vehicle (the equipment onboard and its characteristics will be explained in more detail in the following platform setup subsection) and following a previously established and identical route which is the same for all tests. Each driver had to complete one lap of the circuit, which included a parking exercise situated approximately halfway along the circuit.
The selected route is an urban circuit in the town of Fuente Álamo in the Region of Murcia, Spain, with the tests performed by multiple drivers manually driving the UPCT-CICar in real traffic situations (see Figure 1). This route provides a significant Point of Interest (POI) of typical urban driving situations: (a) intersections with priority and with “Give way”; (b) joining a roundabout, internal circulation and leaving the roundabout; (c) circulation in streets with “green wave” traffic lights; (d) traffic jams; (e) rapid incorporation to a high-density road through a side lane; and (f) pedestrian traffic on public roads.
In order to contemplate a variety of environmental and driving conditions, the tests were carried out at different times of the day (morning, afternoon or night). Figure 2 shows some images from the dataset, where different situations captured during the tests are shown.
After each driving test, the data acquired from the vehicle’s perception systems (LiDARs and cameras), positioning systems (IMU, GPS, rotation angle, acceleration, etc.) and biometric data from the driver are transferred to the central server.

2.1.3. Platform Setup

For this work, the UPCT autonomous vehicle (UPCT-CICar [23]), was driven by a human pilot in manual mode. CICar is a real-world prototype, based on a commercial electric vehicle, the Renault Twizy, which has undergone a series of modifications to provide it with the required functionality. The CICar has been equipped with multiple sensors, including LiDAR, cameras, IMU, GPS, encoders, etc., necessary for the vehicle to perform autonomous driving tasks. This platform setup integrates a perception system, a control system, and a processing system on board the vehicle.
Perception System: The purpose of a sensor system is to collect data from the surrounding environment of the AV and send that data to the control system. These sensors measure different physical quantities, which are typically selected to overlap each other, providing the redundant information needed to correctly merge and correlate the information. In our autonomous vehicle, two types of sensors are used to measure the environment: short-range sensors (up to 10 m) and long-range sensors. Installed short-range sensors include a Sick 2D laser ranging scanner and time-of-flight camera. The long-range sensors are a 3D LIDAR scanner and a camera in the visible spectrum. Table 3 and Figure 3 show the different devices involved in data acquisition during the tests, as well as the details of the variables involved in obtaining them.
Driver Biometric System: The drivers’ biometric signal collection system has been carried out using a non-invasive wearable device, bracelet type, called Empatica E4. The Empatica E4 is a wrist-worn top-quality sensor device considered a Class IIa Medical Device according to 93/42/EEC Directive. Empatica E4 device measures the acceleration data (ACC), as well as other physiological parameters, namely the Blood Volume Pulse (BVP), from which the Heart Rate Variability (HRV) and the Inter-Beat Interval (IBI) are derived as well, skin temperature (TEMP) and also changes in certain electrical properties of the skin such as the Electrodermal Activity (EDA). For the creation of our dataset, among the several measurements recorded by the Empatica E4, this signal was considered, since it provides information better suited for activity recognition. A summary of the technical specifications of the accelerometer sensor is detailed in Table 4.
Control System: The main control systems of the Renault Twizy have been automated in order to allow the vehicle to be autonomously controlled. The modified systems are the steering wheel, the brake pedal and the accelerator pedal (see mechanical modification in Figure 3). Despite the fact that all driving will be manual and not autonomous, the system will record the data with two controller drives through a CAN bus. The Compact Rio cRIO 9082 controls the accelerator, brake and steering wheel movements with the CAN-Open communication protocol, as well as I/O signals.
Processing System: Each sensor works with its own sample rate, and in most cases, this is different between devices. The achieve the synchronisation of the data and accurately reconstruct the temporal sequence, time stamps have been generated to synchronise the operating start and finish times. All of this is controlled and synchronised by the on-board processing system.

3. Results

As a result of the different executions of the experiment with the participating subjects, a raw data set has been obtained that has been curated and published in a repository. The data in the repository is organized under three major directories: (1) Driver, (2) Perception and (3) Position. The distribution of the data in the different directories is detailed below.

3.1. Driver Directory

This directory contains information regarding the drivers, from the questionnaires completed before the test and the biometric data obtained during the test. The directory contains 20 Biometric_XX.csv files (one per driver, where XX is the driver identifier number) and a DBQ.csv file with the data collected from the Biographic Questionnaire and the Driver Behaviour Questionnaire forms.
For the composition of the Biometric_XX.csv files, a normalised sampling frequency of 4 Hz has been used and in the case of sensors with lower frequencies, the table has been completed with NaN fields, with the HR column being the only one affected as the sample rate of this field is 1 Hz. The Biometric_XX.csv files have the following table format, where each column contains the following information:
  • (TIME): The first column corresponds to the time stamp expressed as a unix timestamp in UTC.
  • (TEMP): Data from the temperature sensor expressed as degrees in Celsius (°C).
  • (EDA) Measurement of electrodermal activity by capturing electrical conductance (inverse of resistance) across the skin. The data provided is raw data obtained directly from the sensor expressed in micro siemens (μS).
  • (BVP) The BVP is the blood-volume pulse and the raw output of the PPG sensor. The PPG/BVP is the input signal to algorithms that calculate Inter beat Interval Times (IBI) and Heart Rate (HR) as outputs.
  • (HR): This file contains the average heart rate values calculated at 10-s intervals. They are not derived from real-time readings but are processed after the data is loaded into a session.
  • (ACC_X, ACC_Y, ACC_Z) Data from the three-axis accelerometer sensor. The accelerometer is configured to measure acceleration in the range [−2 g, 2 g]. Therefore, the unit in this file is 1/64 g. Data from x, y and z axis are displayed in the sixth, seventh and eighth columns, respectively.
The DBQ.csv file is made up of a total of 45 columns, where the first column contains the subject identifier. The rest of the columns correspond to each of the items from the Biographic Questionnaire and the Driver Behaviour Questionnaire forms, where the last 25 columns are the questions from the DBQ form.

3.2. Perception Directory

This directory contains:
  • Twenty .bin type files, called perceptionXX.bin, where XX corresponds to the identifier number assigned to each driver at the time of the test.
  • Twenty images from the RGB-D camera (front view).
In the .bin file, the data from the 3D LiDAR sensor, 2D LiDAR sensors and TOF cameras, obtained by the CICar perception system is saved. The data from these sensors were recorded continuously during the driving test, with data packets being written by the different sensors one after the other and at the exact moment in which they arrived at the system, without contemplating an established order of sensor reading and recording.
Each data packet consists of a header made up of two 32-bit integers, which identify the source of the data followed by the data as it was received from the sensors. The header format is as follows:
uint32_t head [2];
  • head [0]: indicates the size in bytes of the packet received by the sensor.
  • head [1]: contains a sensor identifier which shows the source of the data packet received.
The sensor identifiers are the following:
//Packet identifiers in the data file
static const uint32_t LIDAR_PACKET_ID = 0 × 4C494452;  // 3D LiDAR
static const uint32_t GPS_PACKET_ID = 0 × 475053;   // GPS
static const uint32_t NMEA_STRING_ID = 0 × 4E4D4541;//
static const uint32_t M420_FRAME_ID = 0 × 4D343230;  // Camera ToF
static const uint32_t T551_FRONT_ID = 0 × 54354652;   // Front 2D LiDAR
static const uint32_t T551_BACK_ID = 0 × 5435424B;   // Rear 2D LiDAR
The following is a .bin file example:
(uint32_t) 1206 // Data packet size 1206 bytes
(uint32_t) 0 × 4C494452 // Data source: 3D LiDAR
(char [1206]) { ... } // 1206-byte vector with 3D LiDAR data
(uint32_t) 230,524 // Size of data packet 230,524 bytes
(uint32_t) 0 × 4D343230  // Data source: ToF camera
(char [230524]) { .. }   // 230524-byte vector with ToF data.
(uint32_t) 1206 // Size of data packet 1206 bytes
(uint32_t) 0 × 4C494452 //Data source: 3D LiDAR
(char [1206]) { .... }   //1206-byte vector with 3D LiDAR data.
(uint32_t) 1206 //Size of data packet 1206 bytes
(uint32_t) 0 × 4C494452  //Data source: 3D LiDAR
(char [1206]) { .... }   //1206-byte vector with 3D LiDAR data.
(uint32_t) 921  //Size of data packet 921 bytes
(uint32_t) 0 × 54354652  //Data source: Front 2D LiDAR
(char [921]) { ..... }   //921-byte vector with Front 2D LiDAR data.
To facilitate the use and processing of the data, a programme has been developed that allows the data from each sensor to be extracted separately and independently into an additional .bin file. In this case, by separating the data into different files, the data packet identifier is not necessary, but synchronisation with the system is lost. Therefore, to avoid loss of synchronisation between the data packet of each sensor, the time stamp of the exact moment of capture must be included. The .bin file format for each independent sensor is as follows:
uint32_t segundos    // Capture timestamp seconds
uint32_t microseg    // Capture timestamp microseconds
uint32_t numbytes    // Number of bytes in the data packet
char   datos[numbytes] // ‘raw’ data packet from the sensor
This structure is repeated continuously for each data packet until the end of the file.
Furthermore, the data has been pre-processed and the 3D LiDAR, 2D LIDAR and ToF camera data from each test carried out They have been merged into a single point cloud and extracted to a .csv file called POINTCLOUD_XX.csv, where XX is the identifier assigned to each driver at the start of the test.

3.3. Position Directory

This directory contains information regarding the position system, obtained during the driving tests. The directory contains 20 Position_XX.csv files, one for each driver where XX is the driver identifier number. Each of these systems collects information from the GPS, IMU and Encoder sensors.
The Position_XX.csv files are saved in the following table format, where each column contains the following information:
  • (TIME) This first column contains the timestamp of the session expressed as a unix UTC timestamp.
  • (LATITUDE) latitude values obtained by the GPS.
  • (LONGITUD) longitude values obtained by the GPS.
  • (ALTITUDE) altitude values obtained by the GPS.
  • (STERING_ANGLE): Steering wheel angle.
  • (SPEED): Speed/(m/s).
  • (DISTANCE_TRAVELLED).
  • (LIN_ACEL_X): acceleration obtained around the x-axis, obtained in g.
  • (LIN_ACEL_Y): acceleration obtained around the y-axis, obtained in g.
  • (LIN_ACEL_Z): acceleration obtained around the z-axis, obtained in g.
  • (ANG_VEL_X): angular velocity obtained around the x-axis, in degrees/second.
  • (ANG_VEL_Y): angular velocity obtained around the y-axis, in degrees/second.
  • (ANG_VEL_Z): angular velocity obtained around the x-axis, in degrees/second.
The acceleration or angular velocity values are given by four bytes. These bytes correspond to a real number according to the IEEE-754 standard. The IEEE-754 standard is the most widely used for the representation of floating-point numbers.

4. Technical Validation

4.1. Driver Test Validation

To validate the data regarding the drivers, the following actions were carried out:
  • A first validation is carried out by measuring the reliability of the data obtained from the DBQS tests carried out on the drivers. The reliability of the questionnaires was obtained with the entire sample, finding Cronbach’s alpha indices and the two Guttman halves. The values to interpret the reliability were: <0.50 unacceptable; 0.50 ≥ poor < 0.60; 0.60 ≥ questionable/doubtful < 0.70; 0.70 ≥ acceptable < 0.80; 0.90 ≥ good < 0.90; and ≥0.90 excellent. The Cronbach Alpha coefficient is 0.796, which shows that the DBQ data set in this experiment has credibility [24].
  • Missing data E4 data of seven participants (driver 1, driver 5, driver 17, driver 21, driver 30, driver 42, driver 45) were excluded due to a device malfunction during data collection. While physiological signals in the dataset are mostly error-free with most of the files complete above 95%, a portion of data is missing due to issues inherent to devices or a human error.
Raw data from the Empatica device was downloaded in csv format and analysed with the Kubios tool [25]. Kubios offers five artefact correction options based on very low to very high thresholds. No correction of the artefacts analysed by Kubios was necessary This is not surprising since the Empatica E4 already uses an algorithm that removes wrong IBIs or other wrong signals [26].

4.2. Driving Test Validation

Once the driving tests have been completed, a manual verification phase has been carried out on the data obtained (see Figure 4), where the data from those tests where reading or writing failures occurred, or failures in the test itself (routes, drivers, etc.) has been discarded.
  • Checking for abnormalities during the test. The time elapsed for the completion of each test has been checked, passing a filter, and discarding those tests in which the time has been either very short or too long. Data of five participants (driver 4, driver 17, driver 23, driver 29, driver 40) were excluded.
  • Checking for errors in reading the sensors or writing to the disk. For each of the tests, the correct sending of information by the sensors during the test is verified. Those tests where a total or partial failure has been detected have been discarded. To detect these failures, the following aspects were checked:
    a.
    All files exist on the disk. At the end of each test, the number of files generated has been checked. The absence of any of the files implies a failure to read or write the data occurred, therefore this test was discarded completely. Data of four participants (driver 1, driver 10, driver 31, driver 34) were excluded.
    b.
    Empty files. It has been verified that the files generated all contain data, discarding those tests where empty files have been detected. Data of two participants (driver 35, driver 36) were excluded.
    c.
    Exploratory data analysis. Considering the different types of data processed, different types of descriptive analytics have been chosen: (1) Analysis of data deviation. A standard deviation analysis has been applied to those data with discrete values (average speed, time travelled, etc.), discarding those data with a sharp deviation. Data of two participants (driver 11, driver 38) were excluded (2) Time series analysis: most of the data correspond to time series of data, with a certain variation of speed, for this reason, it has been decided to use the Dynamic Time Warping (DTW) technique.
  • Checking for driving route failures. For each of the tests carried out, the route taken by the driver during the test has been verified, to make sure the driver stuck to the route initially stipulated. The test where a small deviation from the track occurred has been discarded. To verify this, the following checks were made: (1) steering wheel rotation pattern during the test, given that for the same trajectory the steering wheel rotation pattern must be similar for all the tests. (2) GPS trajectory, the trajectory has been painted and the tests that do not comply with the marked route have been eliminated.
After this first screening process, a quality validation of the resulting data is performed to guarantee the quality of the data (see Figure 4). Our validation method comprised three steps: (1) Quality control of variables. (2) Quality control of support media. (3) Experimental validation.

4.2.1. Quality Control of Variables

An analysis of the internal structure of the set of circumstances of the DBQ form (content validity) has been performed. To be able to apply a factorial analysis correctly, those items with a declaration frequency of less than 5% were eliminated. Subsequently, and since the items on the form are dichotomous variables, the tetrachoric correlation coefficient was applied to obtain the correlation matrix between the 28 items.
The reliability of the questionnaires was obtained with the entire sample, finding Cronbach’s alpha indices and the two Guttman halves. The values to interpret the reliability were: <0.50 unacceptable; 0.50 ≥ poor < 0.60; 0.60 ≥ questionable/doubtful < 0.70; 0.70 ≥ acceptable < 0.80; 0.90 ≥ good < 0.90; and ≥0.90 excellent. The Cronbach Alpha coefficient is 0.796.
Secondly, outliers in the acquired data, those values notably different compared to the patterns present in the rest of the data, may be due to errors in reading and writing the data from the sensors. Certain deviations were detected in the data from the GPS, due to momentary loss of signal or where the position has been calculated with a fewer number of satellites. In those cases in which the losses are less than two consecutive time intervals, a prediction of the vehicle’s position is made. For cases where the loss is greater, the tests have been discarded. To apply this prediction a constant acceleration Kalman filter has been used.

4.2.2. Quality Control of Support Media

The clocks of all the sensors and devices were synchronised at the start of each experimental test session. All devices are controlled by the control unit on board the vehicle, which provides a perfect temporal and spatial synchronisation of the data obtained by the sensors (see Figure 5).
It has been verified that the data obtained by the encoder is synchronised with the rest of the sensor data. This was achieved by checking the distance indicated by the encoder coincides with the distance calculated between two consecutive GPS timestamps (GNSS). This was done using the Havershine expression shown in Equation (1), where d is the distance in metres between two points on the Earth’s surface; r is the Earth’s radius (6378 km); φ1 and φ2 are the latitudes in radians; Ψ1 and Ψ2 are the longitudes in radians of two consecutive timestamps.
d = 2 r sin 1 ( sin 2 ( φ 2 φ 1 2 ) + cos φ 1 cos φ 2 sin 2 ( Ψ 2 Ψ 1 2 ) )
The indirect data provided by the encoder has also been verified, for example, that the speed matches the direct data measurements provided by the GPS.

4.2.3. Experimental Validation

Finally, the most conclusive validation was performed: the usability analysis of the data contained in the final dataset. The work by Navarro et al. [20] presents the implementation of six end-to-end deep learning models trained using the UPCT dataset. The different end-to-end models were tested using different data sources from the vehicle, including RGB images, linear accelerations and angular velocities. We trained two models using only RGB image data, two using both the image data and IMU data as input to the models, and the last two used sequences of images as an input.
The best results were obtained using a mixed data input type end-to-end deep neural network model which used the front images obtained by the vehicle camera and angular speeds from the IMU to predict the speed and steering wheel angle, obtaining a mean error of 1.06%. An exhaustive optimization process of the convolutional blocks has demonstrated that it is possible to design lightweight end-to-end architectures with a high performance more suitable for the final implementation in autonomous driving.

5. Conclusions

In this work, we have presented the UPCT dataset, a real-world public driving dataset with 20 sets of driving data from 20 drivers which performed a driving test on an urban circuit in real traffic situations. The dataset contains different types of data which we have divided into three categories: (1) Driver, (2) Perception and (3) Position.
The dataset has been validated and tested with six end-to-end deep neural network models, using the RGB image data and IMU data, obtaining very promising results considering the size of the dataset. The detailed results are published in the work by Navarro et al. [20]. We plan to continue this research by making use of the depth images and comparing the results to those obtained when using just RGB images, as well as performing data augmentation to increase the sample sizes.
The main novelty of this dataset is the collection of biometric driver data which allows the behaviour of autonomous driving models to be compared to human drivers. In future research, we plan to use biometric driver data to perform driver behaviour studies. An interesting approach would be to relate the stress levels of the driver to certain driving situations, such as entering a roundabout, entering a main road or parking, for example. As each of the 20 drivers completed the same circuit, it would also be possible to compare the different driving styles and relate these tendencies to certain age groups or to a particular sex. In addition, in the driving behaviour questionnaire, each driver was asked about their driving style, and with the driving test, this can be compared to their real-life performance to determine if drivers correctly perceive their attitudes whilst driving.

Author Contributions

Conceptualization, F.R., P.J.N., L.M. and E.S.; methodology, F.R., P.J.N., L.M. and E.S.; software, P.J.N. and F.R.; validation, F.R. and L.M.; formal analysis, F.R., P.J.N., L.M. and E.S.; investigation, F.R., P.J.N., L.M. and E.S.; resources, F.R. and P.J.N.; data curation, P.J.N., L.M. and F.R.; writing—original draft preparation, F.R., P.J.N. and L.M.; writing—review and editing, F.R., P.J.N., L.M. and E.S.; All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by AEI METROPOLIS (ref. PLEC2021-007609) Spanish Government projects.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data set presented here, which is available within Figshare and released under a CC-BY 4.0 license. https://figshare.com/s/4b9a25a958c3ec578362 (accessed on 1 December 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Olaverri-Monreal, C. Autonomous vehicles and smart mobility related technologies. Infocommun. J. 2016, 8, 17–24. [Google Scholar]
  2. Alawadhi, M.; Almazrouie, J.; Kamil, M.; Khalil, K.A. A systematic literature review of the factors influencing the adoption of autonomous driving. Int. J. Syst. Assur. Eng. Manag. 2020, 11, 1065–1082. [Google Scholar] [CrossRef]
  3. Rosique, F.; Navarro, P.J.; Fernández, C.; Padilla, A. A Systematic Review of Perception System and Simulators for Autonomous Vehicles Research. Sensors 2019, 19, 648. [Google Scholar] [CrossRef] [PubMed]
  4. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The KITTI dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
  5. Yang, Z.; Zhang, Y.; Yu, J.; Cai, J.; Luo, J. End-to-end Multi-Modal Multi-Task Vehicle Control for Self-Driving Cars with Visual Perceptions. In Proceedings of the Proceedings—International Conference on Pattern Recognition, Beijing, China, 20–24 August 2018; Institute of Electrical and Electronics Engineers Inc.: New York, NY, USA, 2018; Volume 2018-Augus, pp. 2289–2294. [Google Scholar]
  6. Kesten, R.; Usman, M.; Houston, J.; Pandya, T.; Nadhamuni, K.; Ferreira, A.; Yuan, M.; Low, B.; Jain, A.; Ondruska, P.; et al. Lyft Level 5 AV Dataset. 2019. Available online: https://www.woven-planet.global/en/woven-alpha/future-automated-driving-applications (accessed on 1 December 2022).
  7. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. Nuscenes: A multimodal dataset for autonomous driving. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Juan, PR, USA, 17–19 June 1997; IEEE Computer Society: Washington, DC, USA, 2020; pp. 11618–11628. [Google Scholar]
  8. Scale, A.I. Hesai Pandaset Open Dataset 2019. Available online: https://scale.com/open-av-datasets/pandaset (accessed on 1 December 2022).
  9. Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Gou, J.; Zhou, Y.; Chai, Y.; Caine, B.; et al. Scalability in Perception for Autonomous Driving: Waymo Open Dataset. 2019. Available online: https://waymo.com/open/ (accessed on 1 December 2022).
  10. Hurl, B.; Czarnecki, K.; Waslander, S. Precise Synthetic Image and LiDAR (PreSIL) Dataset for Autonomous Vehicle Perception. In Proceedings of the 2019 IEEE Intelligent Vehicles Symposium (IV), Paris, France, 9–12 June 2019; pp. 2522–2529. [Google Scholar]
  11. Wang, D.; Wen, J.; Wang, Y.; Huang, X.; Pei, F. End-to-End Self-Driving Using Deep Neural Networks with Multi-auxiliary Tasks. Automot. Innov. 2019, 2, 127–136. [Google Scholar] [CrossRef]
  12. Geyer, J.; Kassahun, Y.; Mahmudi, M.; Ricou, X.; Durgesh, R.; Chung, A.S.; Hauswald, L.; Pham, V.H.; Mühlegg, M.; Dorn, S.; et al. A2D2: Audi Autonomous Driving Dataset. arXiv 2020, arXiv:2004.06320. [Google Scholar]
  13. Alberti, E.; Tavera, A.; Masone, C.; Caputo, B. IDDA: A Large-Scale Multi-Domain Dataset for Autonomous Driving. IEEE Robot. Autom. Lett. 2020, 5, 5526–5533. [Google Scholar] [CrossRef]
  14. Huang, X.; Wang, P.; Cheng, X.; Zhou, D.; Geng, Q.; Yang, R. The ApolloScape Open Dataset for Autonomous Driving and Its Application. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2702–2719. [Google Scholar] [CrossRef] [PubMed]
  15. Dataset Overview—Cityscapes Dataset. Available online: https://www.cityscapes-dataset.com/dataset-overview/ (accessed on 1 December 2022).
  16. Mimouna, A.; Alouani, I.; Ben Khalifa, A.; El Hillali, Y.; Taleb-Ahmed, A.; Menhaj, A.; Ouahabi, A.; Ben Amara, N.E. OLIMP: A Heterogeneous Multimodal Dataset for Advanced Environment Perception. Electronics 2020, 9, 560. [Google Scholar] [CrossRef]
  17. Déziel, J.; Merriaux, P.; Tremblay, F.; Lessard, D.; Plourde, D.; Stanguennec, J.; Goulet, P.; Olivier, P. PixSet: An Opportunity for 3D Computer Vision to Go Beyond Point Clouds With a Full-Waveform LiDAR Dataset. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 September 2021; pp. 2987–2993. [Google Scholar]
  18. Mao, J.; Niu, M.; Jiang, C.; Liang, H.; Chen, J.; Liang, X.; Li, Y.; Ye, C.; Zhang, W.; Li, Z.; et al. One Million Scenes for Autonomous Driving: ONCE Dataset. 2021. Available online: https://arxiv.org/abs/2106.11037 (accessed on 1 December 2022).
  19. Koopman, P.; Wagner, M. Challenges in Autonomous Vehicle Testing and Validation. SAE Int. J. Trans. Saf. 2016, 4, 15–24. [Google Scholar] [CrossRef]
  20. Navarro, P.J.; Miller, L.; Rosique, F.; Fernández-Isla, C.; Gila-Navarro, A. End-to-End Deep Neural Network Architectures for Speed and Steering Wheel Angle Prediction in Autonomous Driving. Electronics 2021, 10, 1266. [Google Scholar] [CrossRef]
  21. Cai, J.; Deng, W.; Guang, H.; Wang, Y.; Li, J.; Ding, J. A Survey on Data-Driven Scenario Generation for Automated Vehicle Testing. Machines 2022, 10, 1101. [Google Scholar] [CrossRef]
  22. de Cózar, E.L.; Sanmartín, J.; Molina, J.G.; Aragay, J.M.; Perona, A. Behaviour Questionnaire (SDBQ). p. 15. Available online: https://www.uv.es/metras/docs/2005_metodologia_SDBQ.pdf (accessed on 1 December 2022).
  23. Borraz, R.; Navarro, P.J.; Fernández, C.; Alcover, P.M. Cloud Incubator Car: A Reliable Platform for Autonomous Driving. Appl. Sci. 2018, 8, 303. [Google Scholar] [CrossRef]
  24. Deng, Z.; Chu, D.; Wu, C.; He, Y.; Cui, J. Curve safe speed model considering driving style based on driver behaviour questionnaire. Transp. Res. Part F Traffic Psychol. Behav. 2019, 65, 536–547. [Google Scholar] [CrossRef]
  25. Kubios HRV—Heart Rate Variability. Available online: https://www.kubios.com/ (accessed on 7 December 2022).
  26. How Is IBI.csv Obtained? Available online: https://support.empatica.com/hc/en-us/articles/201912319-How-is-IBI-csv-obtained- (accessed on 7 December 2022).
Figure 1. Urban route selected for the driving tests.
Figure 1. Urban route selected for the driving tests.
Sensors 23 02009 g001
Figure 2. Images from dataset. (a) Pedestrian crossing; (b) saturation due to reflections on the road; (c) car braking; (d) complex shadows on the road.
Figure 2. Images from dataset. (a) Pedestrian crossing; (b) saturation due to reflections on the road; (c) car braking; (d) complex shadows on the road.
Sensors 23 02009 g002
Figure 3. Autonomous vehicle UPCT-CICar and its different sensors and devices.
Figure 3. Autonomous vehicle UPCT-CICar and its different sensors and devices.
Sensors 23 02009 g003
Figure 4. Verification process.
Figure 4. Verification process.
Sensors 23 02009 g004
Figure 5. Data synchronisation with timestamp. (a) accelerator; (b) break; (c) distance; (d) steering wheel angle.
Figure 5. Data synchronisation with timestamp. (a) accelerator; (b) break; (c) distance; (d) steering wheel angle.
Sensors 23 02009 g005
Table 1. Comparison of the main Datasets.
Table 1. Comparison of the main Datasets.
Ref./YearSamplesImage TypeLIDARRADARIMU/GPSControl
Actions
Raw DataDriver DataReal DataBiometrics DataDriver
Behaviour
UPCT78 KRGB, DepthYesNoYesSteering wheel, SpeedYesYesYesYesYes
KITTI [4]/201215 KRGBYesNoYes-YesNoYesNoNo
Udacity [5]/201634 KRGBYesNoYesSteering wheelYesNoNoNoNo
Lyft L5 [6]/2019323 KRGBYesNoYes-YesNoYesNoNo
nuScenes [7]/20191.4 MRGBYesYesYes-PartialNoYesNoNo
Pandaset [8]/201948 KRGBYesNoYes-PartialNoYesNoNo
Waymo [9]/20191 MRGBYesNoYes-YesNoYesNoNo
PreSIL [10]/201950 KRGBYesNoNo-NoNoNoNoNo
GAC [11]/20193.24 MRGBNoNoNoSteering wheel, SpeedN/ANoYesNoNo
A2D2 [12]/2020392 KRGBYesNoYesSteering angle, brake, acceleratorPartialNoYesNoNo
IDDA [13]/20201 MRGB, DepthNoNoNo-NoNoNoNoNo
Appollo Scape [14]/2020100 KRGBYesNoNo-NoNoYesNoNo
Cityscapes [15]/202025 KRGBNoNoYes-NoNoYesNoNo
OLIMP [16]/202047 KRGBNoYesNo-YesNoYesNoNo
PixSet [17]/202129 KRGBYesYesYes-NoNoYesNoNo
ONCE [18]/20211 MRGBYesNoNo-NoNoYesNoNo
Table 2. Demographic distribution of subjects by gender and age.
Table 2. Demographic distribution of subjects by gender and age.
Categoriesn Initial% Initialn Final% Final
GenderMale26521155
Female2448945
Age18–24612315
25–4422441050
45–641734525
>=65510210
Table 3. Sensor data in CICar.
Table 3. Sensor data in CICar.
DeviceVariableDetails
LiDAR 3DSceneLong-range sensors
3D High-Definition LIDAR (HDL64SE supplied by Velodyne)
Its 64 laser beams spin at 800 rpm and can detect objects up to 120 m away with an accuracy of 2 cm
1.3 Million Points per Second
Vertical FOV: 26.9°
2 × LiDAR 2DSceneShort-range sensors
Sick laser 2D TIM551
Operating range 0.05 m–10 m
Horizontal FOV 270°
Frequency 15 Hz
Angular resolution 1°
Range 10% of reflectance 8 m
2 × ToFSceneShort-range sensors
ToF Sentis3D-M420Kit cam
Range: Indoor: 7 m, Outdoor: 4 m
Horizontal FOV: 90°
RGB-DSceneShort-range sensors
Depth Camera D435 Intel RealSense
range 3 m
Up to 90 fps
Depth FOV: 87° × 58°
RGB FOV: 69° × 42°
IMULocalisation, longitudinal and transversal AccelerationNAV440CA-202 Inertial Measurement Unit (IMU)
3-axis accelerometer
Bandwidth: 25 Hz
Pitch and roll accuracy of <0.4°, Position Accuracy < 0.3 m
GPSLocalisationEMLID RTK GNSS Receiver
7 mm positioning precision
EncoderDistance
Biometric sensorsDriver Biometric signalsEmpatica E4
EDA Sensor (GSR Sensor), PPG Sensor, Infrared Thermopile
3-axis Accelerometer
Table 4. Biometric variables details.
Table 4. Biometric variables details.
VariableSampling FrequencySignal Range [Min, Max]Details
ACC32 Hz[−2 g, 2 g]Accelerometer 3 axes data (x, y, z).
EDA4 Hz[0.01 µS, 100 µS]Electrodermal activity by capturing electrical conductance (inverse of resistance) across the skin.
BVP64 Hzn/aBlood Volume Pulse.
IBI64 Hzn/aInter-beat interval (obtained from the BVP signal)
HR1 Hzn/aAverage Heart Rate (obtained from the BVP signal). Values are calculated at 10-s intervals.
TEMP4 Hz[−40 °C, 115 °C]Skin Temperature.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rosique, F.; Navarro, P.J.; Miller, L.; Salas, E. Autonomous Vehicle Dataset with Real Multi-Driver Scenes and Biometric Data. Sensors 2023, 23, 2009. https://doi.org/10.3390/s23042009

AMA Style

Rosique F, Navarro PJ, Miller L, Salas E. Autonomous Vehicle Dataset with Real Multi-Driver Scenes and Biometric Data. Sensors. 2023; 23(4):2009. https://doi.org/10.3390/s23042009

Chicago/Turabian Style

Rosique, Francisca, Pedro J. Navarro, Leanne Miller, and Eduardo Salas. 2023. "Autonomous Vehicle Dataset with Real Multi-Driver Scenes and Biometric Data" Sensors 23, no. 4: 2009. https://doi.org/10.3390/s23042009

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop