You are currently on the new version of our website. Access the old version .
SustainabilitySustainability
  • Article
  • Open Access

16 July 2021

A Crash Injury Model Involving Autonomous Vehicle: Investigating of Crash and Disengagement Reports

,
,
,
and
1
Research Centre for Integrated Transport Innovation (rCITI), School of Civil and Environmental Engineering, UNSW Sydney, Sydney, NSW 2052, Australia
2
School of Civil and Environmental Engineering, University of Technology Sydney (UTS), Sydney, NSW 2007, Australia
3
IAG of Risk in Smart Cities, UNSW Sydney, Sydney, NSW 2052, Australia
*
Author to whom correspondence should be addressed.

Abstract

Autonomous vehicles (AVs) are being extensively tested on public roads in several states in the USA, such as California, Florida, Nevada, and Texas. AV utilization is expected to increase into the future, given rapid advancement and development in sensing and navigation technologies. This will eventually lead to a decline in human driving. AVs are generally believed to mitigate crash frequency, although the repercussion of AVs on crash severity is ambiguous. For the data-driven and transparent deployment of AVs in California, the California Department of Motor Vehicles (CA DMV) commissioned AV manufacturers to draft and publish reports on disengagements and crashes. This study performed a comprehensive assessment of CA DMV data from 2014 to 2019 from a safety standpoint, and some trends were discerned. The results show that decrement in automated disengagements does not necessarily imply an improvement in AV technology. Contributing factors to the crash severity of an AV are not clearly defined. To further understand crash severity in AVs, the features and issues with data are identified and discussed using different machine learning techniques. The CA DMV accident report data were utilized to develop a variety of crash AV severity models focusing on the injury for all crash typologies. Performance metrics were discussed, and the bagging classifier model exhibited the best performance among different candidate models. Additionally, the study identified potential issues with the CA DMV data reporting protocol, which is imperative to share with the research community. Recommendations are provided to enhance the existing reports and append new domains.

1. Introduction

Autonomous vehicles (AVs) have the potential to revolutionize the transport industry by alleviating congestion, improving safety, and reducing accidents. Consequently, the transport industry will benefit from AV deployment, but the impacts on the auxiliary industries are debatable. As AV manufacturing companies move from research to prototype to production, the generation of data increases exponentially. This data lake (conditional to availability) could elucidate the ambiguity around crash risk.
The ability of AVs to mitigate crash risk caused by human error has raised the expectation of a downtrend in conventional critical safety events. However, the inclusion of AVs within the system can result in new safety challenges associated with poorly maintained road markings and light reflections affecting the vehicle sensors. AV communication faults, cybersecurity, and disengagements also influence crash prevalence [1]. Furthermore, it is essential to note that there is even less understanding about the influence these and other AV features have on crash severity.
The estimation of the economic and social burden caused by road traffic crashes involving AVs requires a thorough understanding of the factors affecting the crash frequency or, given that a crash has occurred, the attributes that can alleviate or aggravate the cost and injury in the crash. Although AV trials generate a vast amount of data, these data cannot be acquired because of data privacy: especially when personal information is involved. Furthermore, synthetic data created by traffic microsimulation models (e.g., [2,3]) and surrogate safety assessment models cannot represent the new or unprecedented crash sources present in AVs. As a result, these models are advantageous but not necessarily an exact representation of the AV scenario. For the safe and transparent deployment of AVs in California, the California Department of Motor Vehicles (CA DMV) commissioned AV manufacturers to draft and publish two reports: a disengagement report and a crash report. A disengagement report is a summary of events where the software (“brain”) of a car disengages automatically (automated) or the disengagement is initiated by the human driver (manual). A crash report is a detailed rundown of events that resulted in a collision and/or damage to property and injuries. These reports are filed by manufacturers conforming to guidelines given by the CA DMV. Whilst there are few implicit fields and fewer data points than typical crash datasets in the literature, it is the best accessible dataset because of open-source availability. Additionally, these data represent the real-world scenario since they comprehend the unprecedented crash causes, unlike microsimulation models. For these reasons, we used CA DMV data, which have been used in a few recent studies [4,5,6,7,8,9,10,11,12,13,14].
This paper provides a detailed review of critical issues associated with crash severity and develops a model for crash severity of AVs. Consequently, the model performance is presented. Furthermore, a detailed review of the limitations of reporting protocol is discussed. Additionally, a comprehensive analysis of the DMV crash data and disengagement data is performed, and expansion on previous studies’ conclusions is presented. This study will aid AV manufacturers and actuarial organizations to evaluate the risks and associated burden in an AV crash. Necessarily, this study also assists the policymakers in updating the present reporting system. The following sections provide a literature review, study area and data description, data analysis and crash model formulation, discussion, limitations, and extension.

2. Literature Review

Before detailing the data and modelling exercise, it is essential to acknowledge recent relevant literature concerning the focal topics of the research. The following Section 2.1 and Section 2.2 present a background to the development of crash severity models in the AV landscape and the application of CA DMV data by other researchers.

2.1. Existing Crash Severity Models and Their Limitations in AV Deployed Scenario

Since the number and severity of crashes affect overall road safety, numerous studies have investigated these measures. Existing crash severity models have been developed for different combinations of dependent (outcome) and independent (explanatory) variables. These studies can be clustered into three significant brackets: transport viewpoint, crash analysis viewpoint, and medical viewpoint [15].
The transport viewpoint studies concentrated on transport-related explanatory variables, e.g., traffic (speed, flow, and density), road, and weather characteristics [16,17,18,19]. Some of these characteristics are expected to alter with AV penetration, and existing severity models need alteration. For instance, AVs are expected to manoeuver efficiently in platoons with decreased headways resulting in different network throughput, compared to the present scenario [20,21]. In addition, AVs’ negligible or absent reaction time stipulates to revise roadway design manuals [4], e.g., implementation of taut turning radii because of negligible minimum sight distances of AVs [21]. Additionally, road maintenance (e.g., repainting markings) could reduce the disengagements [6], affecting crash frequency and severity. In CA DMV reports, Waymo and Mercedes specifically reported the weather as a disengagement cause, indicating the pivotal association of weather and AV performance.
The second category is the crash analysis [22,23,24] viewpoint, which focused on vehicle characteristics such as vehicle’s mass, size, age, and facilities [25] and impact characteristics such as impact speed, collision type, and angle of crash. Increase in manual/conventional vehicle age is found to elevate crash severity [26]. It is speculative that AV hardware’s age should elevate crash severity; however, the age of software (AV “brain”) should alleviate crash severity since AV brains have more training data and have been exposed to a greater variety of hazards. Additionally, an old study [27] evaluated the severity and frequency of rear-end crashes for automated and human driver highway systems. They concluded that automated systems are anticipated to collide at less than 30% of the velocity of the human driver system. In addition, AV technology is competent in preventing all other collision typologies except rear-end crashes [5]. These studies exhibit a change in crash dynamics in the AV deployed environment, which stipulates the necessity of AV-focused crash severity models.
The third category of researchers approached severity from a medical point of view. A severity level (dependent variable) was quantified using the Abbreviated Injury Scale and Injury Severity Score [26,28,29,30]. The exploratory variables encompassed are human behaviour, demographics, and safety facility usage. Existing medical studies fail to incorporate a few issues, for instance, the “driver in the loop” concern in AVs. Automation could make tasks more difficult by removing the easy parts of a task [31]. The out-of-loop performance problems are a salient issue that has been widely documented as a potential negative consequence of automation [32]. Drivers’ increasing trust in AV systems becomes disconcerting on the grounds of “driver in loop” concern [4], and thus new studies examining this concern are vital.

2.2. Previous Studies Employing CA DMV Data

CA DMV data are the best open-source data available for external validation. Our study attempted to use the crash data involving AVs to investigate the safety impact of AVs on the transport system. To date, CA DMV data have been predominantly examined for trends in disengagements, crashes, human trust and reaction time [4], crash frequencies, dynamics, and damage analysis using crash reports [5], triggers and contributory factors of disengagements [6], potential causal relationships between the crashes and disengagements [9], and reliability of AVs by assessing the failures across a wide range of AV manufacturers [8]. Table 1 provides a summary of past research utilizing CA DMV data, including significant findings and insights.
Table 1. Summary of past studies employing CA DMV data.
Building on the findings presented in Table 1, the novelty of our study was the investigation of the contributory and explanatory factors of AV damage crash severity and formulation of the AV crash damage severity model. Additionally, this study exhibited the dissonant implications of disengagement trends and proposed recommendations based on the rationales.

3. Study Area and Data Description

3.1. Study Area

CA DMV disengagement reports summarize failure events, categorized as autonomous disengagement or manual disengagement, albeit the distinction between these categories is implicit [6]. Our study used Google disengagement data because Google contributed 63.04% of total autonomous miles travelled (including 2019). Additionally, these data provide comprehensive details of disengagement events as compared with other manufacturers. On the other hand, the crash reports are a detailed description of the events that incurred property damage or injuries. From 2018, a new category of “vehicle damage description” has been appended in the crash reports. The vehicle damage description is perceived damage by the manufacturer’s authorized representative and hence not explicit. The data used in this study are the DMV disengagement and crash reports and Google traffic data. AV testing is permitted in San Francisco and San Jose cities in Silicon Valley, California, and the crash locations are mapped in Figure 1.
Figure 1. Crash locations superimposed on the street network: (a) San Francisco; (b) San Jose.

3.2. Data Description

A brief description of all the variables used in developing the AV damage severity model is presented in Table 2. Further, details of the data explicitly used during the modelling process are discussed in Section 4.3.1. A few important notes regarding the dataset are mentioned below.
Table 2. Descriptive statistics of variables.
  • The observed damage severity levels sustained by autonomous vehicles during the examination period have the following distribution: no damage (7.14%); minor damage (71.43%); moderate damage (20.41%); major damage (1.02%).
  • The kinetic energy of a crash plays a crucial role in gauging crash severity [33,34]. The relative velocity of vehicles involved in the crash is missing for 176 out of 259 incidents.
  • Vehicle type is the representation of vehicle size derived from the vehicle model stated in the reports. “Two wheels” type refers to vehicles with two wheels such as bikes, motorcycles, and scooters. Subcompact and compact cars are slightly smaller than mid-size cars with an inside volume between 2.4 m3 and 3.1 m3 by combining cargo and passenger volume. Mid-size vehicles are identified as possessing an interior volume of between 3.1 m3 and 3.4 m3. The last category is large vehicles, which represents trucks and buses in this study.

4. Data Analysis and Crash Model Formulation

CA DMV data from 14 October 2014 to 15 June 2020 were used for crash analysis and developing crash damage severity model. For disengagement analysis, we utilized CA DMV reports from 2014 to 2019 since disengagement reports were not available for 2020 at the time of this study. Testing was halted for a while due to the COVID-19 pandemic, and hence 2020 data were not considered for disengagement.

4.1. Crashes

The main concern regarding the adoption of AVs has been based on the grounds of safety. In this section, we perform an in-depth analysis, revisit some of the results, and expand on what was concluded in previous studies [4,5].
Figure 2 exhibits the distribution of crash typologies. Table 3 shows the distribution of crash typologies, driving mode, and party at fault. There were just two crash events where the AV was on autonomous mode and found to be at fault, highlighting the dearth of data related to at-fault AV crashes. In 88.42% (229/259) of all cases, the other party was found to be at fault. Within these cases, 52.90% (137/259) of the time the AV was found to be in autonomous mode. Of total crashes, 61.78% (160 out of 259) were rear-ended. Among these 160 rear-end cases, on 103 occasions, AV was on autonomous mode and not at fault. Rear-end crashes were the most prevalent (105 out of 160) of all the crashes when the AV was on autonomous mode. Authors posit that the dissonant and incompatible reaction time of AV and a human driver in a conventional car was the cause of several rear-end crashes. AVs are very efficient in avoiding collision with the leader (none or negligible reaction time involved). However, follower non-AV drivers cannot react efficiently and thus bump into the leader AV, which was postulated by [5]. Consequently, there is a need to revisit roadway design manuals and safety manuals, which still use the reaction time values determined empirically for manually operated vehicles [4]. Additionally, during location examination, 143 out of 259 rear-end crashes occurred at an intersection, and 83 of them were signalized. The cause of multiple rear-end crashes at signalized intersections could be attributed to the AVs’ homogenous translation of amber light time to green and the heterogeneity of this aspect by human drivers [35].
Figure 2. Crash typology distribution.
Table 3. Distribution of crash typologies, driving mode, and at-fault party.
Sideswipe crashes with AV on autonomous mode constitute half of the sideswipe crashes. Out of 55 sideswipe crashes, the AV was found at fault in only 8 cases. However, in 6 out of these 8 incidents, the AV was not on autonomous mode; thus, the incidents occurred while the driver was manually operating the vehicle. Rear-end collisions (62.55%) and sideswipes (20.46%) make up nearly 83.01% of recorded crashes. There are less than 5.79% (15/259) of crashes categorized as head-on. This is consistent with the previous study [5], which stated that AV technology can prevent all other crash typologies effectively, leaving rear-end collisions with AV as the most crucial failure scenario to be addressed. The Insurance Institute for Highway Safety (IIHS) notes that head-on crashes accounted for 56% of motor vehicle deaths, and 42% of deaths were caused by rear-end and side-impact accidents [36]. Thus, as AV technology can alleviate head-on collisions, it will have a positive effect on road fatality rates, a significant advantage for overall road safety.

4.2. Disengagements

Before presenting the disengagement analysis, it is essential to define some key terms. Dixit et al. (2016) stated that the AV disengagements could be initiated manually by the driver or autonomously. This distinction is crucial from a safety standpoint. Manual disengagements are cautionary in nature; for instance, if a driver feels uncomfortable in a specific situation and/or adopts a proactive approach to prevent a potential autonomous disengagement. On the other hand, automated disengagements represent a design limitation of the car and constitute a potential safety concern for the consumer and the general public. Google Inc. categorizes disengagements into two categories: Failure detection and safety operations. This study assumed that failure detection and safety operation indicate autonomous and manual disengagement, respectively. This ambiguous terminology is one of the limitations of this study; nevertheless, it is a sensible assumption for producing relevant and valuable disengagement trends.
Previous studies stated that the number of crashes observed is highly correlated with the autonomous miles travelled. The cumulative accident trend as a function of cumulative miles can reach a plateau. This plateau will signify that the AV technology training has been effective and is approaching a “crash-free” status with the more miles travelled [5]. Similar conclusions can be derived from the cumulative disengagements trend as a function of cumulative miles [6]. However, this theory can be contended as the road environment evolves with differing infrastructure, technology, and vehicle composition. More importantly, how humans interact with AVs is also changing; this is described in more detail in the following paragraph.
Figure 3 represents the plot of cumulative disengagements with cumulative miles. The blue-coloured plot represents the trend of manual disengagements, and the orange-coloured plot represents automated disengagements. The slope of the automated disengagement curve approaches zero (the curve plateaus) as the number of cumulative autonomous miles increases. Assuming that cumulative autonomous miles serve as a proxy for time/experience, we can infer that automated disengagement events are dropping and approaching zero with time/experience. Since automated disengagements indicate a design limitation and the system performance of the AV, we can infer that the system performance of AV is improving, and the AV “brain” can handle driving tasks, which were intractable before.
Figure 3. Disengagement trends (automated and manual).
Additionally, the non-decreasing slope of the manual disengagement curve indicates relative consistency in the degree of manual disengagement; in fact, the curve indicates that the rate of disengagement increases slightly as the cumulative autonomous miles increase. “Trust effect” can be defined as a phenomenon where the driver’s reliance on technology increases with increased miles driven, attributed to an increase in driver’s trust in the AV system [4]. Since manual disengagements initiated by the driver are cautionary in nature (manual disengagements can be potential automated disengagements that were avoided by quick and brisk action of safety operation driver), the trend implies no improvement in trust of drivers in AVs, and the “trust effect” may not be present in the light of new data. This implication is based on the conjecture that the set of potential automated disengagements and set of manual disengagements are mutually exclusive ( A   M = ϕ , where A is the set of potential automated disengagements and M is the set of manual disengagements). Alternatively, the above trends can also be explained on the grounds of drivers increasing acquaintance. The underlying assumption for this implication is that a set of manual disengagements is a proper subset of potential automated disengagements ( M A ). With more experience (cumulative autonomous miles) with the system, drivers can anticipate the potential intractable situations and decide to exercise caution and hence initiate manual disengagement. As a result, the number of automated disengagements decreases, despite no inordinate improvement in AV technology. Authors propose that in reality, the set of potential automated disengagements and set of manual disengagements are not mutually exclusive A   M ϕ . In addition, the set of manual disengagements is not a proper subset of potential automated disengagements, and in reality, the sets are situated somewhere between two extremes. These postulations can be revised in the light of new and improved data since manufacturers are implicitly and not clearly reporting events resulting in ambiguity. A recommendation from this study is that manufacturers should provide camera and sensors data at the time of the critical event for clarity. Furthermore, the research also supports the development of a driver survey that could be issued with current reporting requirements. Ultimately, it is essential for the CA DMV to include the cause of the disengagement, given that the crash occurred after a disengagement. These recommendations were also proposed in the previous study [6]; however, this was based on a different argument. The study categorized disengagements into macro (human factors, system failures, external conditions, and others) and micro categories. The study found that Tesla’s disengagement falls under the “others” category, indicating the necessity of providing a detailed explanation. Further, the study observed that Mercedes-Benz primarily reports human factors-related disengagement and recommends other manufacturers report contributory factors. Lastly, a reporting protocol similar to what is used in the aviation industry is recommended to enhance consistency.

4.3. Models

4.3.1. Methodology and Data Brief

The crash model developed in this study considered a binary dependent variable signifying the crash and 15 independent variables which are also referred to as features. The dependent variable has two categories: 0 stands for crashes with no injury and 1 for crashes with an injury. The authors investigated the data and discovered a few complications. Firstly, all the independent variables are categorical, and the sample size corresponding to each category is small. Additionally, there is a disparity in sample size in each class. The number of samples in classes 0 (no injury) and 1 (injury) are 224 and 35, respectively. This raises the problem of skewed distributions of a binary task known as an imbalance. The total sample size is 259, with 15 explanatory variables (features) after removing rows with empty entries. The independent variables are primarily associated with the AV, other party’s vehicle, traffic conditions, and roadway features stated in the CA DMV crash reports. The Shapiro–Wilk ranking [37] was used to rank features and is presented in Figure 4a. This ranking assisted in selecting the crucial features in the model. The variable names are presented in Appendix A.2.
Figure 4. (a) Shapiro ranking of 15 features, (b) feature importance of features using severity model.
To handle the imbalanced data, we used stratification while dividing the data into training and testing sets. The stratification preserves the percentage of samples for each class. Two models were developed: model excluding relative velocity as a feature and model including relative velocity since the sample size is different. Out of 200 instances, only 68 instances were reported with relative velocity. The class balance of respective models is presented in Figure 5a,b. Further, we provided class weights in the models, which adjust weights inversely proportional to class frequencies in the dataset. One of the elementary ways to address the class imbalance is to provide a weight for each class. This provides more emphasis on the minority classes such that the end result is a classifier that can learn equally from all classes [38]. Class weight uses the formulae presented in Equation (1).
n s a m p l e n c l a s s e s × n p . b i n c o u n t ( y )  
where n s a m p l e   is the number of samples (259), n c l a s s e s   is the number of classes, and n p . b i n c o u n t ( y )   counts the number of occurrences of element y .
Figure 5. Class balance: (a) overall data; (b) relative velocity feature included.
The model performance is gauged by accuracy, but in the case of class imbalance, accuracy might not be the best representation of the performance of the model. Considering a user preference bias towards the minority class examples, accuracy is not suitable. If the bias is neglected, the impact of the least represented but more important examples is reduced when compared to that of the majority class [39]. Therefore, balanced accuracy is reported, which performs better with imbalanced data. False positive (FP) or type 1 error means there was an injury, but the model reported no injury, and false negative (FN) or type 2 error indicates that there was no injury recorded, but the model predicted an injury. Precision  T P T P + F P   and recall  T P T P + F N are both important in model evaluation. Inevitably, higher values in FP can be dangerous and hence deploy more importance to the precision metric ( T P T P + F P ) rather than recall. Consequently, the authors proposed to use modified F1 score similar to previous studies [40], presented in Equation (2), for evaluating the performance.
F β = ( 1 + β 2 ) × p r e c i s i o n × r e c a l l   β 2 × p r e c i s i o n + r e c a l l
where F β   is the modified F1 score, and β   is a parameter; when 0 < β < 1, more emphasis is put on precision, and when β > 1, recall is given priority. Precision is defined as the percentage of results that is relevant, while recall is the percentage of total relevant results retrieved by the algorithm. This research uses a β of 0.5 to put more weighting on precision as it utilizes FP, and as aforementioned, FP is a crucial aspect in assessing model performance. Resampling is another technique widely used to handle imbalanced data. Khattak et al. (2020) resampled the data by oversampling the two minor classes and under-sampling the major category for balancing the dataset. However, they concluded that this dataset did not provide satisfactory results in comparison to the original data. The methodology applied for this study is presented in the flowchart shown in Figure 6. Appendix A.3 presents more information about the methodology.
Figure 6. Methodology.

4.3.2. Crash Severity Model

Crash severity models were developed using the DMV data and new extracted variables such as vehicle type. The most conservative model was selected as the final model. The confusion matrix and metric of the best-performing model based on modified F1 are presented in Table 4. The results for the lesser performing models used are provided in the appendix (refer to Appendix A.1). Models were developed using pre-built libraries available [41,42]. Feature importance was evaluated to rank them based on their in-model performance and is presented in Figure 4b. A bagging classifier was applied and used base estimators; in this case, the decision tree was used as a base estimator. Model-independent methods were used for computing feature importance. In the case of decision trees as base estimators, feature importance was computed by taking the average of tree’s feature importance among all trees in bagging estimators. Lastly, the precision–recall curve was used because there was a moderate to large class imbalance [43] and is presented in Appendix A.4.
Table 4. Severity model.
Vehicle 1 damage and Vehicle 2 damage rank high in feature importance, presented in Figure 4b, which is unequivocal. It is highly likely that if the vehicles are damaged to a higher degree, the passenger might get injured. The intersection geometry and intersection management (signalization) are essential features exhibited by rank in feature importance. Additionally, 143 out of 259 rear-end crashes occurred at an intersection, and 83 of them were signalized, indicating that signalized intersections are a hotspot for crashes. As mentioned previously, an explanation could be that the percentage of amber light time translated as green by AVs and human drivers is incongruous, resulting in crashes (idealistic behaviour of AVs.). These results highlight the importance of details such as the lantern state at the time of the crash, and signal phasing should be provided within the reporting. Road type is another factor that is ranked highly. It is clear that road infrastructure elements play a crucial role in the severity and require immediate attention from a transport engineering perspective. It is further recommended that these reports also have additional information such as camera images and Lidar cloud points to assess the dynamic location characteristics such as traffic lights, parking occupancy, advertising structure, etc. Rich AV datasets available [44,45,46,47] currently fail to provide data of critical events.
Figure 7a shows the relative speed values for each category of severity. The severity categories used are the same as in the original crash reports, which are different from the developed crash severity model. The graph indicates that higher levels of severity are related to higher relative velocity values. Previous studies also showed that relative velocity is an important explanatory variable for predicting crash severity [34] and can be useful for predicting injury. Relative velocity was used as one of the explanatory variables during model formulation, but it decreased the sample size (83 instances). Out of all the manufacturers, only a few reported the relative velocity. Additionally, collision type is ranked high in feature importance. The authors recommend the CA DMV includes mandatory fields such as relative velocity, time-to-collision, and post-encroachment time to get better insights into the severity of the crashes. The non-parametric distribution was estimated using a Gaussian kernel function calculated using [48] and is presented in Figure 7b. The figure illustrates that the relative velocities during a crash are generally below 15 miles per hour (lower values), indicating less severe accidents. Consequently, another model was developed, including relative velocity as one of the explanatory variables. The training and testing data class balance is provided in Figure 5b. Logistic regression was developed using the features based on feature ranking (Figure 4) and best parameters.
Figure 7. (a) Relative speed and accident severity, (b) probability density function (PDF) of relative velocities of vehicles during a crash.
The solver used for the model was LIBLINEAR, and the penalty was L2 [42]. The explanatory variables used were “Vehicle Type”, “Road Type”, “Intersection”, “Intersection Geometry”, “Parking Provision”, “Mode”, “Vehicle1 Status”, “Vehicle1 Damage”, “Vehicle2 Damage”, “Signalized”, and “Relative Speed”. The model performance was found to be weak with a balanced accuracy of 0.46 and confusion matrix [[11 1], [2 0]]. Further, incrementally trained logistic regression and logistic regression with cross-validation were also developed [42]. Nevertheless, these models also indicated poor model performance. As mentioned in Section 2, the severity can be gauged from three perspectives: transport viewpoint, vehicle damage viewpoint, and medical viewpoint. Relative speed is consistently considered a salient explanatory variable in all the three abovementioned assessment perspectives. Nevertheless, manufacturers are not reporting this critical variable due to optional and voluntary reporting of relative speed. Due to the lack of enough sample size, the model performed poorly when relative velocity was included.

5. Concluding Remarks and Future Work

Since automated disengagement events reduce and approach zero with time/experience, we can infer that the system performance of AV is improving, and the AV “brain” can handle driving tasks, which were intractable before. Additionally, the data reveal that there has been no reduction in manual disengagement events implying no improvement in the trust of humans with regards to AV technology. An alternate explanation for these trends is that with more experience (cumulative autonomous miles) drivers have with the system, they can anticipate the potential intractable situations better. In these scenarios, the driver decides to exercise caution and hence disengages the vehicle. With greater experience, the test drivers are actively shifting between autonomous and manual driving as they can better anticipate scenarios that AVs cannot negotiate.
Consequently, there is a decrease in automated disengagements, yet no inordinate improvement in AV technology. Additionally, increased disengagement frequency could mean that the manufacturer has broadened the conditions/scenarios in which the car is tested [6]. This explains the subsidence in plateauing in the graph of Figure 3. Using similar reasoning, it can be stated that any conclusions on technological improvement based on cumulative disengagements with cumulative miles plots are prone to indiscretion. As a result, the authors recommend averting from this approach to infer the findings of technological advancement and argue that the results in previous studies are premature. Furthermore, the authors also recommend the CA DMV includes mandatory spatial information in their reporting system.
The previous study by Xu et al. (2019) attempted to use the DMV data. The study investigated property damage only (PDO) and non-PDO crashes. They found that roadside parking, crash typology (rear-end or not), and the application of one-way roads were significant. The class imbalance of data was not considered by Xu et al. (2019), which is necessary. Additionally, the authors opine that a clear distinction between road parking provision and actual parked vehicle during the crash should be made. This was a limitation for Xu et al. (2019) and the current study. This reinforces the recommendation of appending image data at the crash locations (critical events). Additionally, further crash data must be collected to avoid any premature results due to limited data points. Consequently, the authors recommend extending this study with new data continuously collected and appended to the database.
Psychological attributes such as the risk attitude of the drivers could play a vital role in the control of AVs but are not available in the reports [49]. These would reveal other significant variables governing the damage severity. Therefore, developing a driver survey and issuing it along with the current reporting mechanism are also recommended. The crash severity field in the CA DMV reports fails to indicate the assessment outlook used to gauge severity level. The authors recommend using a specific assessment tool for potential vehicle crash damage, as stated in this study [50]. Such tools are utilitarian as their aim is to automatically make a decision for autonomous vehicles, which decide in a specific accident scenario, which type of crash is the least detrimental.
The ambiguous terminology of manual and automated disengagements because of implicit data is one of the limitations of this study. Regardless, rational assumptions are made for producing acceptable interpretations. Since the crash data are considered over a prolonged period, the time-varying explanatory variables may change significantly. Neglecting latent within-period variation may result in the loss of crucial explanatory variables. This loss of information using discrete-time intervals can institute error in model estimation because of unobserved heterogeneity [51]. Although this study provides valuable insights into AV crashes, the advancement of this technology may change the currently observed trends [14]. Additionally, using data collected over prolonged periods can also bring temporal instability to the crash severity models. The study can be improved in the future by considering temporal elements. Temporal elements can play a salient role in explaining accident trends [52]. They are typically overlooked which can lead to inaccurate and unreliable results [53]. In the future, the period for data collection can be long (over a decade). Different models can be developed for different periods instead of a single model due to temporal instability [54].
The authors identified that the available data lacked consistency. For example, crash severity levels required further processing; therefore, it is recommended that the DMV standardize data for future use.
Furthermore, improved, explicit, detailed, and increased quantities of data will inevitably produce credible and well-founded interpretations. For instance, if data from more locations (say different countries representing different driver behaviour) are available, the models could incorporate the explanatory variables such as human demographics and safety facilities (namely, medical viewpoint crash studies). This study captured the driver–AV interaction and the AV–conventional vehicle interaction. Due to limited data, it did not account for AV–AV interactions. For further investigation of the severity of AVs’ accidents, more research is required to explore AV–driver, AV–infrastructure, and driver–surrounding environment interactions, which are salient factors responsible for many crashes.
Lastly, the authors also propose that the testing location is a crucial factor that needs to be accounted for. The number of autonomous miles is a proxy for experience for a particular location. Due to the diverse nature of road infrastructure and driver behaviour in different parts of the world, it is not possible to generalize these empirical findings. It is clear from this research and that of Das et al. (2020) that there is a fundamental requirement for more advanced and robust collision narrative reporting to better assess collision likelihoods of autonomous vehicles.

Author Contributions

Conceptualization, A.S., S.C. and V.D.; methodology, A.S.; software, A.S. and V.V.; validation, A.S. and V.V.; formal analysis, A.S. and V.V.; investigation, A.S., V.V. and S.C.; resources, V.D.; data curation, A.S., V.V. and S.C.; writing—original draft preparation, A.S.; writing—review and editing, A.S., S.C., K.W. and V.D.; visualization, A.S. and V.V.; supervision, S.C., K.W. and V.D.; funding acquisition, A.S. and V.D. All authors have read and agreed to the published version of the manuscript.

Funding

We would like to acknowledge the Australian Research Council Grant LP160101021 for the resources and support that made this research possible.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The dataset used for this study can be requested from the CA DMV. Email AVarchive@dmv.ca.gov to request a digital copy of archived disengagement and collision reports. The dataset was last accessed by the authors on the 16 June 2020.

Acknowledgments

Authors want to thank Huang Chen and Jason Wang, graduate students at the UNSW Sydney, for their contribution to data collection.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Performance Comparison of Different Models

Table A1 presents the performance of different models. Bagging classifier and random forest have similar performances, and the final model is selected based on the best adjusted F1 score.
We attempted to choose the most conservative model. In this study context, false negative (type II error) is the event where there was an actual injury, but the model predicted no injury. The most conservative model should have the minimum number of false negatives. The true positives and false positives are the same in both the bagging classifier and random forest. Therefore, the final model selection was made based on false negatives, and the bagging classifier has fewer false negatives (4 < 5). Furthermore, it has a better adjusted F1 score as well.
Table A1. Performance of different models.
Table A1. Performance of different models.
ModelsBalanced AccuracyConfusion MatrixEvaluation Metric
TPFPFNTNFPRFDRFORPrecisionRecallF1Adjusted F1 Score
Decision Tree0.62323420.600.090.670.910.890.450.91
Decision Tree with Weights0.51305510.830.140.830.860.860.430.86
Bagging Classfier0.65341420.330.030.670.970.890.470.96
Balanced Bagging Classifier0.592411330.790.310.500.690.890.390.72
Random Forest0.57341510.500.030.830.970.870.460.95
Balanced Random Forest Classifier 0.382114510.930.400.830.600.810.340.63
Easy Ensemble Classifier0.731619060.760.540.000.461.000.310.51
RUSboost Classfier0.732213150.720.370.170.630.960.380.67
Extra Trees Classifier0.62323420.600.090.670.910.890.450.91
In the above table, TP = true positive; FP = false positive; FN = false negative; TN = true negative; FPR = false positive rate; FDR = false discovery rate; FOR = false omission rate.

Appendix A.2. Variables Used in the Modelling

Table A2 lists all the variables used in the modelling exercise. Details are presented in Table 2.
Table A2. Variables used in the modelling.
Table A2. Variables used in the modelling.
Vehicle TypeVehicles InvolvedCollision Type
Road TypeModeVehicle1 Damage
IntersectionVehicle1 StatusVehicle2 Damage
Intersection GeometryVehicle2 StatusSignalized
Parking ProvisionFaultRelative Speed

Appendix A.3. Methodology

This section provides additional details for Section 4.3.1. Skewed distributions of a binary task known as imbalance were found in the data. Stratification is used to handle the distribution of training and testing datasets and prevents sampling bias. Stratification preserves the percentage of samples for each class, which enables creating a test set with a population that best represents the original data. Sampling bias is an issue that occurs when testing a model. Model performance can be poor as test data are not representative of the whole population. Sampling bias is often observed because specific values of the variable are underrepresented or overrepresented with respect to the true distribution of the variable. Stratification, to prevent sample bias, is achieved by splitting the data into strata. The right number of instances are then sampled from each stratum to guarantee that the test set is representative of the original data. The next step is the training and development of the model. Most learning algorithms are incompetent with imbalanced data (biased class data). The training algorithm can be modified by providing different weights to the majority and minority classes. During the training, more weightage is given to the minority class in the cost function of the algorithm. Hence, it will provide a higher penalty to the minority class, and the algorithm can emphasize decreasing the errors for the minority class. Class weight uses the formulae presented in Equation (1). For our data, 259 is the number of samples ( n s a m p l e ) and there are two classes ( n c l a s s e s ) , and n p . b i n c o u n t ( y ) counts the number of occurrences of element y . Model selection is the next step after extracting the training and testing data.
For this study, the bagging classifier was found to be the best performing model for this data. Bagging [55] is a “bootstrap” [56] ensemble method that generates individuals for its ensemble by training each classifier on a random redistribution of the training set. It fits every base classifier on random subsets of the original dataset and then aggregates their individual predictions (either by voting or by averaging) to form a final prediction. In this study, we used a decision tree as our base estimator. Each classifier’s training set is generated by randomly drawing, with replacement, N examples, where N is the size of the original training set. Each classifier in the ensemble is generated with a different random sampling of the training set. Bagging classifier can reduce overfitting. It is deployed to decrease the variance of a base estimator (e.g., a decision tree) by introducing randomization into its construction procedure and then making an ensemble out of it. Details of the basic setup of the bagging classifier are presented in Table A3.
Table A3. Basic setup parameters of the bagging classifier method.
Table A3. Basic setup parameters of the bagging classifier method.
ParametersAssignment
Base EstimatorDecision Tree Classifier
n_estimators (the number of base estimators in th ensemble)20
random_state0
max_samples, max_features, bootstrap, bootstrap_features, oob_score, warm_start, n_jobs, verbose Default values as in [42]
Accuracy might not be the best representation of the performance model in biased class data due to user preference bias towards the minority class examples [57]. Balanced accuracy is a metric that one can use when evaluating the classifier’s performance, especially when the classes are imbalanced. Sensitivity, which is the true positive rate (also known as recall), and specificity, which is the true negative rate (also known as false positive rate), are used to define balanced accuracy. Balanced accuracy is simply the arithmetic mean of the two as presented in Equation (A1).
Balanced   accuracy = Sensitivity + Specificity 2
Along with balanced accuracy, a modified F1 score is also used to evaluate the model’s performance. FN indicates that there was no injury recorded, but the model predicted an injury, and FP means there was an injury, but the model reported no injury. Higher values in FP can be dangerous and hence deploy more importance to the precision instead of recall. Consequently, a modified F1 score, presented in Equation (2), is used to assess the model performance.

Appendix A.4. Precision–Recall Curve

Figure A1 presents the precision–recall curve for the bagging classifier.
Figure A1. Precision–recall curve for bagging classifier.

References

  1. Taeihagh, A.; Lim, H.S.M. Governing Autonomous Vehicles: Emerging Responses for Safety, Liability, Privacy, Cybersecurity, and Industry Risks. Transp. Rev. 2019, 39, 103–128. [Google Scholar] [CrossRef] [Green Version]
  2. Papadoulis, A.; Quddus, M.; Imprialou, M. Evaluating the Safety Impact of Connected and Autonomous Vehicles on Motorways. Accid. Anal. Prev. 2019, 124, 12–22. [Google Scholar] [CrossRef] [Green Version]
  3. Sinha, A.; Chand, S.; Wijayaratna, K.P.; Virdi, N.; Dixit, V. Comprehensive Safety Assessment in Mixed Fleets with Connected and Automated Vehicles: A Crash Severity and Rate Evaluation of Conventional Vehicles. Accid. Anal. Prev. 2020, 142, 105567. [Google Scholar] [CrossRef]
  4. Dixit, V.V.; Chand, S.; Nair, D.J. Autonomous Vehicles: Disengagements, Accidents and Reaction Times. PLoS ONE 2016, 11, e0168054. [Google Scholar] [CrossRef] [Green Version]
  5. Favarò, F.M.; Nader, N.; Eurich, S.O.; Tripp, M.; Varadaraju, N. Examining Accident Reports Involving Autonomous Vehicles in California. PLoS ONE 2017, 12, e0184952. [Google Scholar] [CrossRef]
  6. Favarò, F.; Eurich, S.; Nader, N. Autonomous Vehicles’ Disengagements: Trends, Triggers, and Regulatory Limitations. Accid. Anal. Prev. 2018, 110, 136–148. [Google Scholar] [CrossRef] [PubMed]
  7. Favarò, F.M.; Eurich, S.O.; Rizvi, S.S. “Human” Problems in Semi-Autonomous Vehicles: Understanding Drivers’ Reactions to Off-Nominal Scenarios. Int. J. Hum. Comput. Interact. 2019, 35, 956–971. [Google Scholar] [CrossRef]
  8. Banerjee, S.S.; Jha, S.; Cyriac, J.; Kalbarczyk, Z.T.; Iyer, R.K. Hands Off the Wheel in Autonomous Vehicles?: A Systems Perspective on over a Million Miles of Field Data. In Proceedings of the 48th Annual IEEE/IFIP International Conference on Dependable Systems and Networks (DSN), Luxembourg, 25–28 June 2018; pp. 586–597. [Google Scholar]
  9. Khattak, Z.H.; Fontaine, M.D.; Smith, B.L. Exploratory Investigation of Disengagements and Crashes in Autonomous Vehicles under Mixed Traffic: An Endogenous Switching Regime Framework. IEEE Trans. Intell. Transp. Syst. 2020, 1–11. [Google Scholar] [CrossRef]
  10. Boggs, A.M.; Wali, B.; Khattak, A.J. Exploratory Analysis of Automated Vehicle Crashes in California: A Text Analytics & Hierarchical Bayesian Heterogeneity-Based Approach. Accid. Anal. Prev. 2020, 135, 105354. [Google Scholar] [CrossRef] [PubMed]
  11. Wang, J.; Zhang, L.; Huang, Y.; Zhao, J. Safety of Autonomous Vehicles. J. Adv. Transp. 2020, 2020, e8867757. [Google Scholar] [CrossRef]
  12. Das, S.; Dutta, A.; Tsapakis, I. Automated Vehicle Collisions in California: Applying Bayesian Latent Class Model. IATSS Res. 2020, 44, 300–308. [Google Scholar] [CrossRef]
  13. Xu, C.; Ding, Z.; Wang, C.; Li, Z. Statistical Analysis of the Patterns and Characteristics of Connected and Autonomous Vehicle Involved Crashes. J. Saf. Res. 2019, 71, 41–47. [Google Scholar] [CrossRef]
  14. Khattak, Z.H.; Fontaine, M.D.; Smith, B.L. An Exploratory Investigation of Disengagements and Crashes in Autonomous Vehicles. In Proceedings of the Transportation Research Board 98th Annual Meeting, Washington, DC, USA, 13–17 January 2019. [Google Scholar]
  15. Sobhani, A.; Young, W.; Logan, D.; Bahrololoom, S. A Kinetic Energy Model of Two-Vehicle Crash Injury Severity. Accid. Anal. Prev. 2011, 43, 741–754. [Google Scholar] [CrossRef]
  16. Zeng, Q.; Hao, W.; Lee, J.; Chen, F. Investigating the Impacts of Real-Time Weather Conditions on Freeway Crash Severity: A Bayesian Spatial Analysis. Int. J. Environ. Res. Public Health 2020, 17, 2768. [Google Scholar] [CrossRef]
  17. Cheng, W.; Gill, G.S.; Sakrani, T.; Dasu, M.; Zhou, J. Predicting Motorcycle Crash Injury Severity Using Weather Data and Alternative Bayesian Multivariate Crash Frequency Models. Accid. Anal. Prev. 2017, 108, 172–180. [Google Scholar] [CrossRef] [PubMed]
  18. Holdridge, J.M.; Shankar, V.N.; Ulfarsson, G.F. The Crash Severity Impacts of Fixed Roadside Objects. J. Saf. Res. 2005, 36, 139–147. [Google Scholar] [CrossRef]
  19. Quddus, M.A.; Wang, C.; Ison, S.G. Road Traffic Congestion and Crash Severity: Econometric Analysis Using Ordered Response Models. J. Transp. Eng. 2009, 136, 424–435. [Google Scholar] [CrossRef]
  20. Olia, A.; Razavi, S.; Abdulhai, B.; Abdelgawad, H. Traffic Capacity Implications of Automated Vehicles Mixed with Regular Vehicles. J. Intell. Transp. Syst. 2018, 22, 244–262. [Google Scholar] [CrossRef]
  21. Virdi, N.; Grzybowska, H.; Waller, S.T.; Dixit, V. A Safety Assessment of Mixed Fleets with Connected and Autonomous Vehicles Using the Surrogate Safety Assessment Module. Accid. Anal. Prev. 2019, 131, 95–111. [Google Scholar] [CrossRef]
  22. Buzeman, D.G.; Viano, D.C.; Lövsund, P. Injury Probability and Risk in Frontal Crashes: Effects of Sorting Techniques on Priorities for Offset Testing. Accid. Anal. Prev. 1998, 30, 583–595. [Google Scholar] [CrossRef]
  23. Rana, T.A.; Sikder, S.; Pinjari, A.R. Copula-Based Method for Addressing Endogeneity in Models of Severity of Traffic Crash Injuries: Application to Two-Vehicle Crashes. Transp. Res. Rec. 2010, 2147, 75–87. [Google Scholar] [CrossRef]
  24. Farmer, C.M.; Braver, E.R.; Mitter, E.L. Two-Vehicle Side Impact Crashes: The Relationship of Vehicle and Crash Characteristics to Injury Severity. Accid. Anal. Prev. 1997, 29, 399–406. [Google Scholar] [CrossRef]
  25. Huelke, D.F.; Compton, C.P. The Effects of Seat Belts on Injury Severity of Front and Rear Seat Occupants in the Same Frontal Crash. Accid. Anal. Prev. 1995, 27, 835–838. [Google Scholar] [CrossRef]
  26. Bédard, M.; Guyatt, G.H.; Stones, M.J.; Hirdes, J.P. The Independent Contribution of Driver, Crash, and Vehicle Characteristics to Driver Fatalities. Accid. Anal. Prev. 2002, 34, 717–727. [Google Scholar] [CrossRef]
  27. Carbaugh, J.; Godbole, D.N.; Sengupta, R. Safety and Capacity Analysis of Automated and Manual Highway Systems. Transp. Res. Part C Emerg. Technol. 1998, 6, 69–99. [Google Scholar] [CrossRef] [Green Version]
  28. Langley, J.; Marshall, S.W. The Severity of Road Traffic Crashes Resulting in Hospitalisation in New Zealand. Accid. Anal. Prev. 1994, 26, 549–554. [Google Scholar] [CrossRef]
  29. Conroy, C.; Tominaga, G.T.; Erwin, S.; Pacyna, S.; Velky, T.; Kennedy, F.; Sise, M.; Coimbra, R. The Influence of Vehicle Damage on Injury Severity of Drivers in Head-on Motor Vehicle Crashes. Accid. Anal. Prev. 2008, 40, 1589–1594. [Google Scholar] [CrossRef]
  30. Lee, J.; Conroy, C.; Coimbra, R.; Tominaga, G.T.; Hoyt, D.B. Injury Patterns in Frontal Crashes: The Association between Knee–Thigh–Hip (KTH) and Serious Intra-Abdominal Injury. Accid. Anal. Prev. 2010, 42, 50–55. [Google Scholar] [CrossRef]
  31. Bainbridge, L. Ironies of Automation. In Analysis, Design and Evaluation of Man–Machine Systems; Johannsen, G., Rijnsdorp, J.E., Eds.; Pergamon: Oxford, UK, 1983; pp. 129–135. ISBN 978-0-08-029348-6. [Google Scholar]
  32. Wiener, B.E.L.; Curry, R.E. Flight-Deck Automation: Promises and Problems. Ergonomics 1980, 23, 995–1011. [Google Scholar] [CrossRef]
  33. Corben, B.; Senserrick, T.; Cameron, M.; Rechnitzer, G. Development of the Visionary Research Model: Application to the Car/Pedestrian Conflict; Monash University Accident Research Centre: Melbourne, Australia, 2004. [Google Scholar]
  34. Elvik, R. To What Extent Can Theory Account for the Findings of Road Safety Evaluation Studies? Accid. Anal. Prev. 2004, 36, 841–849. [Google Scholar] [CrossRef]
  35. Papaioannou, P. Driver Behaviour, Dilemma Zone and Safety Effects at Urban Signalised Intersections in Greece. Accid. Anal. Prev. 2007, 39, 147–158. [Google Scholar] [CrossRef]
  36. The Insurance Institute for Highway Safety (IIHS). Fatality Facts 2019: Passenger Vehicle Occupants. Available online: https://www.iihs.org/topics/fatality-statistics/detail/passenger-vehicle-occupants (accessed on 25 June 2021).
  37. Shapiro, S.S.; Wilk, M.B. An Analysis of Variance Test for Normality (Complete Samples). Biometrika 1965, 52, 591–611. [Google Scholar] [CrossRef]
  38. Jordan, J. Learning from Imbalanced Data. Available online: https://www.jeremyjordan.me/imbalanced-data/ (accessed on 25 June 2021).
  39. Branco, P.; Torgo, L.; Ribeiro, R.P. A Survey of Predictive Modeling on Imbalanced Domains. ACM Comput. Surv. 2016, 49, 1–50. [Google Scholar] [CrossRef]
  40. Li, Z.; Chen, W.; Pei, D. Robust and Unsupervised KPI Anomaly Detection Based on Conditional Variational Autoencoder. In Proceedings of the 2018 IEEE 37th International Performance Computing and Communications Conference (IPCCC), Orlando, FL, USA, 17–19 November 2018; pp. 1–9. [Google Scholar]
  41. Lemaître, G.; Nogueira, F.; Aridas, C.K. Imbalanced-Learn: A Python Toolbox to Tackle the Curse of Imbalanced Datasets in Machine Learning. J. Mach. Learn. Res. 2017, 18, 559–563. [Google Scholar]
  42. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  43. Saito, T.; Rehmsmeier, M. The Precision-Recall Plot Is More Informative than the ROC Plot When Evaluating Binary Classifiers on Imbalanced Datasets. PLoS ONE 2015, 10, e0118432. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. NuScenes: A Multimodal Dataset for Autonomous Driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11621–11631. [Google Scholar]
  45. Maddern, W.; Pascoe, G.; Linegar, C.; Newman, P. 1 Year, 1000 Km: The Oxford RobotCar Dataset. Int. J. Robot. Res. 2017, 36, 3–15. [Google Scholar] [CrossRef]
  46. Huang, X.; Cheng, X.; Geng, Q.; Cao, B.; Zhou, D.; Wang, P.; Lin, Y.; Yang, R. The ApolloScape Dataset for Autonomous Driving. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 954–960. [Google Scholar]
  47. Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Guo, J.; Zhou, Y.; Chai, Y.; Caine, B. Scalability in Perception for Autonomous Driving: Waymo Open Dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2446–2454. [Google Scholar]
  48. Terrell, G.R.; Scott, D.W. Variable Kernel Density Estimation. Ann. Stat. 1992, 20, 1236–1265. [Google Scholar] [CrossRef]
  49. Dixit, V.; Xiong, Z.; Jian, S.; Saxena, N. Risk of Automated Driving: Implications on Safety Acceptability and Productivity. Accid. Anal. Prev. 2019, 125, 257–266. [Google Scholar] [CrossRef]
  50. Wiseman, Y.; Grinberg, I. Circumspectly Crash of Autonomous Vehicles. In Proceedings of the 2016 IEEE International Conference on Electro Information Technology (EIT), Grand Forks, ND, USA, 19–21 May 2016; pp. 0387–0392. [Google Scholar]
  51. Washington, S.P.; Karlaftis, M.G.; Mannering, F. Statistical and Econometric Methods for Transportation Data Analysis; CRC Press: Boca Raton, FL, USA, 2010. [Google Scholar]
  52. Mannering, F. Temporal Instability and the Analysis of Highway Accident Data. Anal. Methods Accid. Res. 2018, 17, 1–13. [Google Scholar] [CrossRef]
  53. Al-Bdairi, N.S.S.; Behnood, A.; Hernandez, S. Temporal Stability of Driver Injury Severities in Animal-Vehicle Collisions: A Random Parameters with Heterogeneity in Means (and Variances) Approach. Anal. Methods Accid. Res. 2020, 26, 100120. [Google Scholar] [CrossRef]
  54. Theofilatos, A.; Antoniou, C.; Yannis, G. Exploring Injury Severity of Children and Adolescents Involved in Traffic Crashes in Greece. J. Traffic Transp. Eng. 2020. [Google Scholar] [CrossRef]
  55. Breiman, L. Bagging Predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  56. Efron, B.; Tibshirani, R.J. An Introduction to the Bootstrap; CRC Press: Boca Raton, FL, USA, 1994; ISBN 978-0-412-04231-7. [Google Scholar]
  57. Branco, P.; Torgo, L.; Ribeiro, R. A Survey of Predictive Modelling under Imbalanced Distributions. arXiv 2015, arXiv:1505.01658. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.