Next Article in Journal
CFD Analysis on Novel Vertical Axis Wind Turbine (VAWT)
Previous Article in Journal
Parameter Optimization Method for Centrifugal Feed Disc Discharging Based on Numerical Simulation and Response Surface
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensuring Driving and Road Safety of Autonomous Vehicles Using a Control Optimiser Interaction Framework Through Smart “Thing” Information Sensing and Actuation

1
Department of Civil and Environmental Engineering, College of Engineering, Majmaah University, Majmaah 11952, Saudi Arabia
2
Civil Engineering Department, College of Engineering, King Khalid University, Abha 61421, Saudi Arabia
3
Department of Civil Engineering, College of Engineering, Northern Border University, Arar 73222, Saudi Arabia
4
Civil Engineering Department, College of Engineering, Jouf University, Sakakah 72388, Saudi Arabia
5
Department of Electrical Engineering, College of Engineering, Jouf University, Sakakah 72388, Saudi Arabia
*
Author to whom correspondence should be addressed.
Machines 2024, 12(11), 798; https://doi.org/10.3390/machines12110798
Submission received: 11 October 2024 / Revised: 29 October 2024 / Accepted: 4 November 2024 / Published: 11 November 2024
(This article belongs to the Special Issue Safety and Security of AI in Autonomous Driving)

Abstract

:
Road safety through point-to-point interaction autonomous vehicles (AVs) assimilate different communication technologies for reliable and persistent information sharing. Vehicle interaction resilience and consistency require novel sharing knowledge for retaining driving and pedestrian safety. This article proposes a control optimiser interaction framework (COIF) for organising information transmission between the AV and interacting “Thing”. The framework relies on the neuro-batch learning algorithm to improve the consistency measure’s adaptability with the interacting “Things”. In the information-sharing process, the maximum extraction and utilisation are computed to track the AV with precise environmental knowledge. The interactions are batched with the type of traffic information obtained, such as population, accidents, objects, hindrances, etc. Throughout travel, the vehicle’s learning rate and the surrounding environment’s familiarity with it are classified. The learning neurons are connected to the information actuated and sensed by the AV to identify any unsafe vehicle activity in unknown or unidentified scenarios. Based on the risk and driving parameters, the safe and unsafe activity of the vehicles is categorised with a precise learning rate. Therefore, minor changes in vehicular decisions are monitored, and driving control is optimised accordingly to retain 7.93% of navigation assistance through a 9.76% high learning rate for different intervals.

1. Introduction

Autonomous vehicle control involves safe driving and navigation capability by incorporating advanced sensors, communication systems, and control mechanisms [1]. The vehicles are informed continuously in real-time by the environment, such as road conditions, traffic flow, object presence, speed, direction, and braking action, to decide on the appropriate control actions [2]. The primary components of autonomous vehicle control are stability, collision avoidance, and following traffic flow rules. Environmental sensing, communication with infrastructure, and object detection systems enable safe navigation in urban and highway environments [3,4]. The navigation systems within autonomous vehicles make necessary adaptations to changes through real-time route recalculations based on new data received [5]. The control systems should efficiently manage different driving scenarios, from congested traffic to unforeseeable conditions on the road. The systems must ensure responses are consistent and dependable to realise the safe operation of the autonomous vehicle under variant conditions [6].
The control of safe driving in an autonomous vehicle faces various problems and challenges in an unpredictable or complex environment. Sensors and controls are associated with the vehicle to interpret and respond appropriately to real-world data such as pedestrians, roadway obstacles, or rapid traffic flow changes [7,8]. Sensor data, weather conditions, and road scenarios include uncertainty that may cause control failures and make decisions hard for autonomous systems [9]. Smooth coordination is supposed to be present when integrating various data sources such as GPS, cameras, and lidar. The mismatch in interpretation from all data sources may create navigation errors [10]. Moreover, ensuring the reliability and safety of the decision-making process of a vehicle against incomplete or inaccurate data is another challenge [11]. System failures, hacking risks, and technical malfunctions presuppose other obstacles to keeping control in critical situations [12].
The solution of intelligent learning algorithms in enhancing autonomous vehicle navigation comes up with adaptive and dynamic solutions while interpreting complex driving environments [13]. Machine learning algorithms analyse vast data from multiple sensors to understand patterns for correct decision-making in runtime. These enable vehicles to learn from experience to continuously improve navigation and control performance using reinforcement learning methods [14,15]. The understanding of complex driving scenarios—for instance, pedestrian crossings or adverse weather conditions—becomes more profound with autonomous systems supported by neural networks and deep learning models [16,17]. The learning systems further enhance the detection and classification of objects in the vehicle’s environment for safer and more accurate navigation. The decision-making process will be refined by intelligent algorithms, wherein the vehicles can learn from their past driving experience and refine their strategy with time. Artificial intelligence applied at each stage can enable an autonomous vehicle to adapt to different driving conditions with very limited human intervention [15,18,19].
The COIF was created to make AVs safer on the road by simplifying their point-to-point connection with real-world objects and other “Things”. Unlike older AV frameworks that use preset environmental parameters, COIF uses a neuro-batch learning algorithm to adjust to real-time traffic and road conditions. This technology helps the AV make better decisions and controls by classifying vehicle actions as safe or risky. It achieves this despite constant data from population density, possible obstacles, and traffic incidents.
COIF assures that AVs can interact with fixed and moving objects in complicated contexts via adaptive learning and better knowledge-sharing mechanisms. This technology increases safety, driving control, and the AV’s capacity to anticipate and avoid problems by more precisely monitoring and reacting to environmental signals. The project uses information-sharing technology and adaptive learning algorithms to assure real-world autonomous driving safety and consistency.
The key highlights of the article are:
  • A brief discussion of the previous works related to AV safe driving and navigation support through different learning and optimisation methods
  • The introduction, discussion, and theoretical explanations of the proposed framework with illustrations and mathematical models
  • The discussion includes data-based analysis, functional representations, and partial outputs.
  • The metric-centric discussions through comparisons using related metrics and congruent methods with percentage improvement
The rest of the paper is followed by Section 2, which discusses the latest literature review of the proposed idea. Section 3 describes the proposed control optimiser interaction framework in detail. The results and discussion are given in Section 4 and Section 5, respectively, while the paper’s conclusion is drawn in Section 6.

2. Related Works

Kang et al. [20] developed a data-driven, control-policy-based system for automated driving safety analysis. It extracts the control policies based on historical driving data and simulates them to detect hazard scenarios. It achieves 89% coverage of hazard scenarios validated with Baidu Apollo driving data. The method exhibits great competence in the detection of hazards and safety analysis. Validations on extensive data sets prove the robustness of the system. Significant advances have been made in the system to enhance vehicle safety. Han et al. [21] introduced a framework of multi-agent reinforcement learning for connected autonomous vehicles. The framework enhances decision-making and processes by sharing information among agents and uses a truncated Q-function to protect itself. Simulations within CARLA environments demonstrate its ability to maintain appropriate distance apart and prevent unsafe actions. The approach enhances traffic safety and efficiency profile for autonomous vehicle networks.
Wang et al. [22] developed a methodology for lane change decision-making in autonomous vehicles. The fuzzy inference was implemented through the decision-making process as factors such as vehicle velocity, acceleration, and delay are all considered to control safe and comfortable lane changes. Simulation across several scenarios indicated the method’s effectiveness in increasing safety and comfort when changing lanes. The technique translates to a safe and comfortable lane change. Sun et al. [23] proposed an adaptive robust formation control of connected autonomous vehicle swarms. The method presents collision avoidance and time-varying uncertainties through a constraint-following scheme with diffeomorphism transformation. The method presents a compact formation of vehicle swarms with an effective management approach to the risk associated with the collision. Simulation runs showed its validity, which presented reliable performance in complex environments.
Li et al. [24] introduced a global sorting-local gaming framework for managing complex multivehicle interactions in the scope of autonomous driving scenarios. Through game theory, the global sorting-local gaming method has effectively quantified interaction disturbance and resolved multiplayer interaction. Simulations and human-in-the-loop experiments have indicated improvements in safety and traffic efficiency compared to the traditional models. Overall, traffic management can be improved because it deals with multivehicle scenarios. Jond et al. [25] proposed a differential game-theory-based optimal control method for autonomous vehicle convoys. The method transforms the dynamics of individual vehicles to a relative dynamics-based optimal control problem, which ensures an openness Nash equilibrium for distributed convoy control. Simulations illustrate the successful application of the method in stability and safety during highway driving. The method supports stable and safe operations in highway environments. Stability is enhanced in a convoy using this method.
Wang et al. [26] presented a hybrid approach combining reinforcement learning with model predictive control to apply to cars following manoeuvres of autonomous vehicles. An algorithm is used to dynamically adjust the parameters from model predictive control according to the calculated levels of risk, which applies reinforcement learning for the classification of risks in traffic scenes. It further points to improved safety and performance compared to pure model predictive control. It further provides adaptability through the real-time tuning of parameters. Nie et al. [27] proposed an autonomous highway driving strategy with a time-to-collision and reinforcement learning-based safety check system. Unsafe actions are only substituted by safer ones if the TTC is less than some certain value; therefore, safety is increased. Simulations conducted in ramp merging tasks demonstrate elevated arrival rates but decreased collisions; the approach has good prospects for dense traffic conditions. The approach enhances collision avoidance by its feasibility in making immediate action adjustments, and validation confirms improvements in safety and efficiency.
Deng et al. [28] proposed a deep reinforcement learning approach to decision-making in uncertain highway driving scenarios. The method densifies training data to focus the process on only the safety-critical events, accelerating the evaluation speed by as much as 105 times. It enhances the building of autonomous vehicles through enhanced efficiency in safety testing. The extensive testing and strategy demonstrate the method’s fast evaluation capabilities and greatly boost safety validation processes. Feng et al. [29] offered a dense reinforcement learning technique for the safety validation of self-driving vehicles. It adapts Markov decision processes to emphasise safety-critical events more, accelerating evaluation. Considering crucial safety events enhances the efficiency of testing as a whole. The results of validation validate its efficiency in accelerating evaluation tests. It is one of the efficient accelerative techniques for safety validation.
Lee et al. [30] offered a deep hybrid learning network for autonomous driving control in adverse weather conditions. The method utilised a variational autoencoder to learn environmental features and an inception-bidirectional, long, short-term memory network to predict the driving based on these features. Highly impressive performance improvements were reported, with a mean square error of 0.0464 and an inference time of 6 milliseconds. It is an enhancement technique of control while driving, which gains response when dealing with adverse weather conditions. He et al. [31] proposed a decision-making approach for safe autonomous driving with observation robust reinforcement learning. The adversarial agent simulates attacks to the observations and utilises a robust actor-critic model to optimise policies in bounded variations. Lane-change tasks showing resilience against adversarial attacks ensure safety and good performance in uncertain conditions. The approach ensures decision-making safety in complex driving environments by addressing perception uncertainties, and the method improves safety in complex driving environments.
Ben Elallid et al. [32] proposed a reinforcement learning-based control method for an autonomous vehicle under weather and daylight conditions. Methodology driver-specific behaviour is incorporated into a model predictive control scheme using a proportional integral differential feedback channel. Hardware-in-the-loop experiments have verified its suitability for personalised control and collision prevention. It addresses varied driving conditions and styles. Enhanced tracking control and safety were offered, contributing to better driver adaptation and comfort. Shi et al. [33] proposed an integrated deep reinforcement learning framework for intelligent autonomous driving. The approach integrates macrolevel routing with real-world, microlevel behaviours, lane changes, and car-following-the approach outperforms human drivers. Its performance significantly improves traffic efficiency and safety performance, especially in networks with high penetration rates of autonomous vehicles. The approach improves traffic efficiency and safety and supports advanced autonomous driving systems.
Li et al. [34] introduced a method for controlling an autonomous vehicle with an orientation oriented towards personalised driving behaviour. The method utilises Q-ABSAS optimisation within a framework of model predictive control, thus enabling the method to add driver-dependent characteristics. The method is flexible with various driving behaviours, so an improvement in tracking control, safety, and comfort occurs. The approach improves the driving experience through individual differences in adaptation and provides customised solutions in different driving conditions. Mihaly et al. [35] developed a supervised reinforcement learning scheme for applying trajectory-tracking control in an autonomous vehicle. The technology utilises reinforcement learning to improve stability and the precision of the applied control action on trajectory tracking. Simulation outputs depict a high degree of stability in the retention of vehicle stability throughout the trajectories. Different simulations have proven the performance of the approach to stable trajectory handling. The proposed method promotes the stabilisation of the handling trajectory.
Gao et al. [36] proposed an online safety verification approach for autonomous driving based on dynamic reachability analysis. It predicts the possible future moves of neighbouring vehicles and incorporates traffic rules to analyse trajectory overlaps with legal zones. Experimental results demonstrate the framework’s effectiveness in determining and mitigating safety threats. The approach ensures collision-free driving conditions, predicts the future movements of vehicles, and provides a sound solution to safety verification in autonomous vehicles. Wang et al. [37] designed a control learning design that learns to ensure the safety of traffic with both self-driving and human-driven vehicles. The technique blended a traditional method and a learning model to predict how human-driven vehicles act in mixed groups. The combined model reduced the root mean square error to 35.64% in speed prediction from the HV compared to when the first principles were used alone. GP-MPC improves safety by better dealing with uncertainties and performing well in challenging scenarios like emergency braking. A mathematical safety buffer for cars that collide is developed, this temporal separation is linearised, and a combination mixed-integer linear program is constructed [38]. Every possible scenario, including regular, emergency, and recovery scenarios, is detailed and tested via simulations. The tradeoff between fast response times (larger safety buffer) and low reaction times (more false positives) is shown via sensitivity analysis on different reaction times. The study’s main goal is to determine how much space should be left between cars to prevent a crash in the event of a vehicle failure.
In contrast to static or rule-based solutions, the suggested COIF can adapt in real time to changing settings because of the use of a neuro-batch learning algorithm. COIF’s exclusive emphasis on point-to-point interaction guarantees that the AV extracts data directly from its immediate environment, improving safety by providing accurate, situationally driven reactions. This contrasts with other approaches, such as ORRL, HDSE-DRL, and CRL-MPC, that frequently function in sterile settings and are not flexible enough for use in the actual world. As a result of its superior fit for complicated and unpredictable environments, COIF’s method guarantees constant AV performance and enhanced safety.

3. Proposed Control Optimiser Interaction Framework

3.1. Data Collection and Description

The data used in this article is a consolidated inheritance from [39,40]. The first source provides a “self-driving cars” dataset comprising images (of lanes) and acceleration parameters (brake, steering, speed, and throttle). This dataset observes 8036 entries for the above parameters under straight and left lane changing settings. The second dataset provides independent accelerometer, GPS, timestamp, distance covered, and magnetometer readings from in-vehicle and in-road sensors. Using the images from the first dataset, the navigation scenario is illustrated in Figure 1.
With an emphasis on improving navigational and driving safety via sophisticated environmental learning, the proposed study presents a new COIF for autonomous cars. In contrast to conventional methods, this architecture allows for consistent and uninterrupted communication between the AV and its environment’s “Things” using a neuro-batch learning algorithm. To maximise driving choices, this technique guarantees accurate, real-time information transmission. A highly adaptable and safety-oriented AV system is made possible by the framework’s batch optimisation methodology for risk assessment and a unique risk indicator to change navigation in response to risky situations quickly. There has been a giant leap forward in autonomous vehicle technology with the integration of real-time learning, adaptive response systems, and risk management. The first image marks the sensors and visual input in the AV driving scenario. The information acquired for processing towards navigation and safe driving is illustrated in Figure 1. The second image showcases the change of AV direction towards the lefthand side with the possibilities of θ (angle). The third image shows the “after lane change” position of the AV for which the θ variation [ 180 °   t o 45 ° , 60 ° , 90 °   t o   180 ° ] is illustrated. Based on this, 1695 × 4 entries, accelerometer, speed, GPS, etc. sensors are observed. The maximum distance accounted for is 1000 m, and the testing time is between 0 and 55 min from which the navigation data is extracted.
Autonomous vehicles employ a variety of sensors to alter their trajectory, as seen in the image. These include a camera, speed, acceleration, and global positioning system (GPS). In the initial lane view, with a steering angle of θ o r 180 ° , the sensors watch the automobile for its position and speed. This setup guarantees that the vehicle knows its intended path and orientation before changing lanes.
During the direction change phase, the steering angle is adjusted in a series of steps, beginning at 45 ° , 60 ° degrees and finishing at 90 ° degrees. The vehicle can adjust to environmental changes effectively since the camera sensor records this direction shift. In the final frame, After the direction change, the vehicle returns to a straight path at an angle between zero and ninety degrees after changing directions θ o r 180 ° . This series demonstrates the vehicle’s directional manoeuvring capabilities, including its ability to maintain a constant speed and orientation.

3.2. Framework Description

The COI framework aims to provide AV users with a safe driving and navigation experience. This framework considers vehicle, environmental, and computational intelligence to fulfil its purpose. This framework application for the AV scenario is depicted in Figure 2. In the above depiction, the “Things” refer to the access points, traffic lights, cloud devices, base stations, etc. that are interfaced using the wireless communication media to the AVs. These “Things” provide users with infotainment, navigation, location services, and messaging features. The AVs define their paths through a persistent interaction schedule and learn the environment to ensure safe driving. Environmental learning refers to the visual and communication sensed information the AV utilises for speed and direction control, movement, lane hanging, etc. The learning by the AV is facilitated by the actuating and communicating sensors/devices attached to the vehicle. The fundamental information requirement is the starting and ending location, lane, safe distance, and vehicle acceleration. The following/trailing vehicles communicate with the AV to broadcast their status (speed, direction, etc.), which further serves as the learning input for the vehicle (Figure 2). Communication between the autonomous car and the interacting “Things” allows it to extract and use local data, including location, lane, visual, etc., for reliable data sharing.
Road safety management is improved using consistent vehicle interaction that obtains the knowledge of novel sharing between the AV and interacting “Thing” to retain driving and vehicle safety. In the case of different locations or lanes or visuals in front, vehicle interaction and consistency change accordingly. Estimating the precise and accurate minimum safety distance between the vehicles and interacting “Things” is important. Hence, the minimum safety distance S a f e D is computed as
S a f e D = R V s 2 2 E r d O V s + L V s + c
where, R V s is the relative vehicle speed, E r d signifies the expected deceleration of the relative vehicle, O V s signifies the overtaking vehicle speed, L V s signifies the leading vehicle speed and c denotes the consistency of vehicle interaction. The point-to-point interaction is defined as a distance to the AV and interacting “Things” that exceeds the minimum safety distance, while the following extraction and utilisation process is characterised by a location, lane, and visual that is less than S a f e D .
To guarantee road safety, the shown equation determines the minimal safe distance (safe) that must be maintained between an autonomous vehicle (AV) and an interacting “Thing” (such as another AV or an object). This component considers the following: the relative vehicle’s speed R V s , the relative vehicle’s predicted deceleration E r d the relative vehicle’s speed when passing O V s , the relative vehicle’s speed when leading L V s and a stability factor ( c ) for the interaction between vehicles. Autonomous vehicles can maintain a safe following distance in all situations by figuring this out and considering things like acceleration and deceleration rates. An important safety buffer during AV encounters in many traffic conditions, this dynamic measure may respond to location, lane, or visual signal changes. The equation shows that the consistency factor in vehicle encounters is represented by c. It ensures that the AV’s interactions with other “Things” in its surroundings, such as other cars or barriers, are stable and dependable. This factor modifies the minimum safety distance to account for variations in interaction quality that occur in real-time. The model ensures that the AV has a flexible safety margin by integrating c, which makes it more resistant to changes on the road.
In this scenario, the information transmission refers to the AVs and interacting “Thing” undertaking a location, lane, and visual change trick.
I n t R V = m a x s ,     i n t e r a c t i n g m a x L V s , m a x s ,     e x t r a c t i n g m a x s ,     u t i l i z i n g
where, I n t R V represents the interaction of relevant vehicles, m a x s is the maximum speed, and L V s represents the leading vehicle speed.
The above equation describes the interaction, I n t R V that occurs during data transmission between AVs and an external “Thing” (such as other AVs or traffic-related factors). The three possible interactions—interacting, extracting, and utilising—depend on the top speed and leading vehicle speed. The AVs modify their pace and behaviour depending on their environment in each situation, which constitutes a unique kind of information exchange. When interacting, an AV takes its maximum speed into account; when extracting, it compares the speed of the leading vehicle L V s with its maximum speed and chooses the greater value. This method enables the AV to make real-time, behavioural adjustments that improve its interaction with its surroundings while minimising risk.

3.3. Information-Sharing Process

This proposed framework relies on consistent vehicle interaction to address decision issues suitable for adaptability using the interacting “Things”. This monitors the complete location, lane, visual extraction, and utilisation process. In this process, a typical consideration of extracted data, such as location, lane, and visual, is utilised to analyse the information on environmental knowledge. Based on the instance, the front vehicle (FV), interacting “Thing” (IT), and rear vehicle (RV) in the current location/lane/visual is required for appropriate information transmission, while the traffic, objects, the population in the current location/lane/visual is monitored on the roadside.
X i T = X i 0 + W i μ i T + i T Y i T = X i 0 + W i μ i T
In Equation (3), the variable X i T represents the initial extracted factors from a certain location, lane, and visual and μ i T is the environmental knowledge data observed from different time instances. Where, i T signifies the uncertainty in location, lane, and visual resulting in noise occurrence. The Y i T are the utilised factors and W i are the weights. Based on the above Equation (3), the uncertainty case is discussed as a lemma that impacts the vehicle’s behaviour.
Lemma 1. 
i T impacts c for which I n t R V is expected to be optimal.
Discussion 1. 
The first lemma of i T assessment over c and I n t R V impacts the μ i T of the AV at any distance or time. Therefore, the impact is described using the derivatives defined below in Table 1.
In this process, the extracted data addresses the unique characteristics of AVs and interacting “Things”, including environmental knowledge data. The environmental knowledge data are defined in Equation (4),
μ N L , N l , N V = X i 0 + i X i T + i T μ C L , C l , C V = X i 0 + i X i T
where, the variables N L and C L represent the nearby and current locations; N l and C l represent the nearby and current lanes;   N V and C V represent the nearby and current visual. If α and β are the position and interaction between the nearby vehicles and “Things”; γ signifies the type of AV and interacting “Things”. In this scenario, when the AVs interact with other “Things”, the output is 1 ; otherwise, it is 0 . In this process, the maximum extraction and utilisation are estimated. Using the estimation output, the autonomous vehicle with appropriate environmental knowledge is tracked to improve the consistency and adaptability measures through the neuro-batch learning algorithm. The interactions are batched with the environmental knowledge data, with the measures of 1 , 0 , and 1 corresponding to the batches and driving time. Each factor in environmental knowledge is categorised to avoid unsafe navigation responses.

3.4. Batch Optimisation Algorithm

The neuro-batch learning algorithm is based on a random utilisation scenario, which assumes that the random integer. q i is alone and follows the learning rate throughout its driving time. The learning, as mentioned in the above algorithm, implies that the probability of data extraction and utilisation corresponds to the precise tracking of AV.
P X i , Y i = e x p X i T N i e x p X N T
To enhance driving and road safety, the interactions are adaptively batched over time to determine the utilisation of location, lane, visual, and extraction probability from the current instance to the future. If N represents the type of traffic information observed from the roadside environment in different time intervals. Based on the objects, population, and traffic conditions, the environmental knowledge of the AV and its learning rate are categorised until its driving time through the learning process. This algorithm increases the environmental knowledge data for performing appropriate batching. The target location is chosen based on the actuated and sensed information from the vehicle’s environmental knowledge. This enhances driving and road safety by appropriately selecting the utility of each location and lane in the future.
B μ i = arg m a x i T T + 1 P X i , Y i . T d T
The batch optimisation model for risk assessment is illustrated in Figure 3.
The model of the batch optimisation is illustrated in Figure 3 for T · d T detection using X i inputs. The P X i , Y i is split into two combinations: X i = 1 , P i = 0 and X i = 1 , P i = 1 . In the other two conditions (i.e.,) X i = 0 the input is not fetched, and therefore, previous decisions are retained. Therefore, for the two conditions, the batch combinations are: c , W i and I n t R V , W i . Using these combinations, X i T , X i T 1 , X i T 2 , , X i 1 an X N 1 , X N 2 , , X N T integration, are verified to ensure μ i = m a x and i = 0 . This verification is pursued with W i to ensure T · d T outputs a unity value that holds P X i , Y i = 1 . Besides, μ N L , N l , N V is the T derivative for i X i T + i T such that Y i T = c + W i T = c W i = 0 . The second batch optimisation requires i T under i T X i T is validated to identify any W i = 0 in T . The μ i T is iterated until μ i T = m a x and i = 0 achieving high precision of P X i , Y i .
The neuro-batch learning algorithm is well suited to the proposed method, which enables the AV to adapt instantaneously to its environment. The system can effectively assess and classify traffic data, including vehicle speeds, obstacles, and pedestrian movements, using batch processing. It is safer to utilise the AV when things are continually changing because of its flexibility, allowing it to adapt to changing conditions precisely. In contrast to traditional methods, neuro-batch learning continuously improves the AV’s decision-making with new information, allowing it to manage the multiple interconnections required by the COIF.

3.5. Risk Assessment

Once the intention is to categorise, safe and unsafe activity of the vehicle has been identified using the connected learning neurons to the information for unidentified scenarios. The risk assessment technique is used to verify the risk occurrence in the vehicle’s decision over unavailable or inconsistent information to regularise its existing neuro batches for ensuring safe navigation responses. The framework for autonomous vehicles identifies the multiple risk factors from the different types of traffic information required in the current location, lane, and visual. However, this module considers only single-vehicle interaction and does not consider consistency measures adaptability. This article proposes a new risk indicator for easily cut navigation during unsafe driving.
By optimising risk assessment in batches and using the neuro-batch learning algorithm for dynamic communication between AVs and interacting “Things,” this technique aims to manage complicated, real-world situations better. Installing a new risk indication for early detection of hazardous driving is an extra precaution for preventative safety that allows swift navigational alterations. This technique aims to enhance road safety by making AVs more dependable and adaptive in unpredictable situations.
The risk indicators assessment is validated to categorise safe and unsafe driving. Subsequently, the learning neurons are connected to compute the likelihood of safe navigation at each level of risk. Ultimately, the identifying risk level with the interacting “Thing” is detected as the unsafe navigation risk level. The first assessment is to categorise the environmental knowledge of the vehicle δ throughout its driving time using neural learning. The neurons are connected to the actuated and sensed information, which can be employed to identify the unsafe activity where the extracted feature is close to zero, and interaction undergoes some variation.
δ = V f V l L f L l ,   V f > V l   0 ,     V f V l
In Equation (7), the variables L f and L l represent the longitudinal positions of the vehicles is observed in both leading and following of the certain AV, while V f and V l are the corresponding velocities of the vehicles. The second assessment is the computation of the safety distance index s d i of the vehicle. The environmental knowledge gained of the AV using the Y i ( T ) is discussed using Lemma 2. Table 2 explains the environmental knowledge gain Lemma 2.
Lemma 2. 
μ i ( T )  Gain using  Y i ( T )
Discussion 2. 
Table 2. Environmental knowledge gain Lemma 2.
Table 2. Environmental knowledge gain Lemma 2.
Batches
B μ i = T = 1 P 1,0 · T d T : B μ i = T = 1 P 1,1 · T d T
δ = 1 i = 0   ( or )     i 0
The batch derivatives are:
P 1,0 = X i T X n T , μ i = m a x
P 1,1 = X i T X n T W i × μ i
i = 1 Y i T X i T
μ i T = P 1,0 i δ = 0 P 1,1 W i i δ = 1
Therefore,
B μ i = μ i T · P 1,0 d T μ i T i · P 1,1 d T
Machines 12 00798 i004 δ = 1 i 0 & 1 ,   but     0 < i < 1
The batch derivatives are:
Y i T = X i o + W i μ i T
Subs   for     μ i T from LHS,
Y i T = X i o + W i · P 1,0 i X i o + W i · P 1,1 W i i
and thus,
μ i T · P 1,0 = X i o + W i · P 1,0
μ i T i · P 1,1 = X i o + W i P 1,1 W i i
The   Crain   is     μ i T   for     i and
μ i T   in   any     T · d T = 1
Machines 12 00798 i005
Machines 12 00798 i006
This validation ensures that AVs are maintained at the precise distance between interacting “Things”.
s d i = S a f e D L f L l
where, S a f e D signifies the minimum safety distance as defined in Equation (1). The third assessment is the computation of the total risk ω based on the traffic information obtained. The risk factor is related to the kinetic energy of the traffic unit, as shown in Equation (9).
ω = T M + 1 2 M V 2 + 2 N V l i m 2
In Equation (9), M indicates the mass value, V indicates the velocity, N is the type of traffic information required coefficient and V l i m 2 indicates the speed limit of the selected road lane. The risk indicator based on lane characteristics is completed.
D j i = 1 ε v j i x × ε v j i ε v j i
From the instance, the variable ε v j i represents the coordinate vector from j to i time interval and x means the descending gradient coefficient of the vehicle. The risk assessment is related to the unsafe navigation of the movement ϑ . The learning categorisation coefficients can be expressed as
D ϑ , j i = e x p R V × c o s θ
In Equation (11), R V means the relative velocity of i and j vectors and θ represents the angle between the AV and interacting “Things”. From the considered dataset values, the impact of batch optimisation over the metrics discussed in Equations (8)–(11) is studied in Figure 4.
The impact of S a f e D and P X i , Y i is induced by the batches in any T provided the ω lends are precisely identified. The derivatives in Lemma 2 are validated to maximise B μ i under T d T . In this assessment, the Y i T and it is corresponding W i adjustments are pursued to ensure i is suppressed. Through different 1 Y i T X i T < μ i T instances are extracted to ensure i is suppressed. In the batch optimisation process, the Y i T differentiations are pursued to ensure X i T = Y i T . If this case is true then T · d T generates a unity output. The chances of unity pose D j i and D v j i through ε v j i differences for V l i m 2 . Hence, ω from i and s d i > L f L l validations, W i modifications are pursued to ensure n o unsecured/information omission takes place. Besides, the physical parameter θ of the vehicle for L V s or O V s following or E r d are identified based on the rate of x any ω . Thus, Y i T = W i i is the ω deciding factor across P 1,0 and P 1,1 such that W i μ i is the balancing factor for B μ i . This case is inferred in the Lemmas 2’s μ i T · P 1,0 and μ i T i · P 1,1 derivatives for B μ i (Refer to Figure 4 Representations). The traffic regulations and lane marks are used to limit vehicle navigation, and the environmental knowledge of the vehicle will filter all risk factors on the roadside.
The coefficient of the filtering process can be expressed as
F i = N × cos p v × π 6 τ ,   p v > τ 2   1 ,     p v τ 2
In Equation (12), p v is the position vector, τ is the width of the lane and N is the type of traffic information coefficient from the centre line. The proposed framework precisely identifies the risk factors associated with vehicle velocity, mass, and consistency measures adaptability are taken to address the impact of unsafe driving navigation response. The final risk assessment index R i s k a can be expressed as
R i s k a = ω f r o n t l e a d i n g v e h i c l e + ω r e a r l e a d i n g v e h i c l e
The above three assessments are used to assess the accurate level of the risks as X = δ , s d i , ω , R i s k a and gives notification to the vehicle based on the three risk levels as expressed as Ψ = s a f e S , u n s a f e U , unsafe in the sense of providing dangerous or alert signals. Assume that the consistency of the vehicle interaction x and the risk level Ψ satisfies the following Equation:
p v X = x Ψ = ψ i e x p x G 2 2 σ + 1 r g 2 ,     i f   x > G   e x p x H 2 2 σ + 1 r g 2 ,     i f   x > H r g ,     o t h e r w i s e
In Equation (14), G and H are upper and bottom threshold conditions in decision-making. Where the variable r g denotes the regularisation factor, and σ is the uncertainty of observational information. The neurons’ connectivity can be estimated by incorporating the different risk weight assignments. In Lemma 3, the p v derivatives for R i s k a classification (safe and unsafe) are discussed. Table 3 describes the risk classification of Lemma 3.
Lemma 3. 
The role of p v in AV safety deciding is inevitable based on R i s k a
Discussion 3. 
Table 3. Risk classification of Lemma 3.
Table 3. Risk classification of Lemma 3.
P ( X = X ) α X G 2 2 σ
P ( X = X ) α X H 2 2 σ
X > G X > H
Consider     D v j i   as     e x p R V × C o s θ
for   the   lane   change     θ = 0 ° (no change)
D v j i = R V ; F i = N ;
Equating   with     p v
R V = x G 2 2 σ ,   r g = 0
and
2 σ R V + x G 2 = 0 is the safe
Driving criteria
For   the   lane   change     θ = 90 ° (maximum)
D v j i = 0 ; F i = 1
x G 2 2 σ = o x G 2 = 0 is the Safe driving criteria
Consider     F i = 1   for     Y i T = i
for   the   lane   change     θ = 0 ° (no change)
P X i = X i ψ = ψ i = p v 1 = p v
p v X = X ψ = ψ i α 1 r g 2 2 σ = x H 2
and
p v X X ψ = ψ i α r g = R i s k a
for   the   lane   change     θ = 90 ° (maximum)
D j i = 1 1 x × 1 r g 1 x = 1 r g 1 x
where     x = X H 2 2 σ
Machines 12 00798 i007 Machines 12 00798 i008
Machines 12 00798 i009Machines 12 00798 i010
Based on the instance, the probability of a particular risk level can be determined as
P X i = x i Ψ = ψ i = i = 1 N Ψ W i p v X = x Ψ = ψ i i = 1 N Ψ W j
When considering the unsafe activity of the vehicle, the risk factors originate from the actuated and sensed information of AV in the vehicle’s decision. Consequently, the probabilities of unsafe navigation responses should be estimated by joining all the respective risk levels by using a new function that is expressed as
P ψ i = S = 1 N = 0,1 P ψ N = S X i N = x M i , N P ψ i = U = 1 N = 0,1 P ψ N = U X i N = x M i , N P ψ i = S U = 1 P ψ N = S P ψ N = U
Equation (16) follows the information-sharing process described above that allows the environmental knowledge of the vehicle to be observed and the risk level to be identified through the batch optimisation algorithm and neural learning process. The risk of changing unsafe navigation responses is determined based on the risk level when the learning neurons are activated and connected to classify safe and unsafe driving. During an emergency, an alert signal indicates the vehicle is unsafe. The unsafe AV navigation function outputs above are analysed graphically using Figure 5.
The F i and R i s k a at G and H for the derivatives P ( ψ i = S U is represented in Figure 4. The p v variants for ( X > G ) and ( x > H ) are the key factors that decide the r g and ψ for A V navigation. The possibilities are ψ N = S , ψ N = U for the individual unity factors for providing precise driving controls. Therefore for F i the G and H are the inverse for ψ = X i ,   ω i ,   i ,   D based on the information acquired such that 1 P ψ N S (or) U maximises the chances of safe driving over R i s k a . Thus, the cases violating the above are unsafe navigations (H cases) in all of the above F i differentiated outputs. These points require ω and D v j i modifications for X G 2 and X H 2     r g .

3.6. Driving Control Design

In this training process, the AV uses the straight road to drive; if there is any abnormality, it takes a new lane using the environmental knowledge data. Therefore, the control optimiser interaction framework must extend the features to make appropriate decisions. The state space of the control optimiser interaction is given as
φ = x i , ω i , i , o f f s e t , l a s t s t e e r , D
In Equation (17), the vehicle’s position follows the safe navigation after regularisation. The angle between the vehicle and the interacting things is completed to follow the safe lane. In general, multiple factors interrupt the vehicle interaction consistency to ensure driving and pedestrian safety. This article presents a batch optimisation-based consistency measure adaptability prediction model using the neural learning paradigm. The neural learning process identifies the unsafe activity of the vehicle of different risk/abnormality factors towards safe navigation.

4. Results

In the discussion section, the learning rate, environment detection, navigation assistance, and information latency metrics are comparatively analysed between the proposed and existing methods. The existing methods considered are ORRL [31], HDSE-DRL [28], and CRL-MPC [26], which are introduced in the related works section. These metrics are comparatively analysed using distance (100 and 1000 m) and time (up to 55 min). This comparative discussion is used to verify the efficiency of the proposed method over the other methods through the percentage of improvement estimated and described under each metric discussed below. Therefore, in Figure 6, the comparative illustration of the learning rate is depicted.
The environmental data extraction and utilisation process improves the tracking of autonomous vehicles with appropriate environmental knowledge. Through batching of interactions, it is easy to identify the type of traffic information sharing between the AV and interacting “Thing” over different time domains. The precise observation and information extraction from the previous neuro batch and utilisation computation output are used for retaining driving and road safety. The unsafe activity of the AV is addressed by using COIF and batch optimisation algorithms based on the X i T and Y i T , the condition that satisfies successive observation of the safe navigation, preventing i T Y i T and Y i T i T X i T . Therefore, the following information-sharing sequence analysis is performed for a feasible solution and is not represented. Based on (3) and (4), the conditions satisfy a high learning rate in precise categorisation-based neurons connected. In this process, a high safe navigation response in AV and environmental knowledge of the vehicle are based on current observation, and it is comparatively less extraction and complexity. Hence, the distinct AV information for the navigation is reduced and satisfies a high learning rate due to changes in the sharing information series (Figure 6). This proposed framework improves the learning rate by 9.76% and 8.68%, respectively, for the above comparisons. Following the above illustrations, the environment detection ratio is presented in Figure 7.
Based on data extraction and utilisation, the proposed framework achieves high environment detection for safe and unsafe driving categorisation (Figure 7). The threat occurrence is mitigated based on arg m a x i T T + 1 P X i , Y i . T   d T with connected neurons. Information extraction and sharing are pursued to identify unsafe driving due to changes in location and lane, and the vehicle’s visual observations and environmental knowledge are categorised through neural learning. Maximum extraction and utilisation are useful for precise environmental detection to reduce the AV’s unsafe activity. The interactions are batched and addressed the unknown scenarios throughout the driving time based on both observations of V f > V l and V f V l and transmission of organising information through weight assignment. Contrarily, the computed unsafe environmental data observation is used to augment the unidentified scenarios for verifying driving and pedestrian safety, which relies on other factors in the point-to-point interaction. Therefore, the environmental detection is high, and batching also increases. The proposed COI framework improves environment detection by 9.65% and 10.97% for different times and distances. Figure 8 illustrates the navigation assistance rate for the distance and time variants.
The navigation assistance through neuron connectivity ensures consistency and adaptability with the interacting “Thing” for accurate data observation in the information-sharing process (Figure 8). This framework satisfies less learning categorisation by computing the maximum extraction and utilisation. In this sequential process, information actuated and sensed from the AV in different time intervals for preventing unsafe driving based on the condition X = δ , s d i , ω , R i s k a and Ψ = s a f e S , u n s a f e U is computed until driving time to retain road safety. The observed vehicle resilience and consistency depend on the position vector and their threshold conditions, wherein the regularisation and uncertainty in the environmental knowledge of the vehicle are identified through neural learning. The unsafe driving behaviour is identified using the observation of information between the AV and interacting “Thing” to compute retaining accuracy based on driving time. The unsafe navigation is addressed with the extracted data and retaining this condition with p v X = x Ψ = ψ i computation. Therefore, the maximum environmental knowledge data extraction is computed to maximise the interactions based on actuated and sensed information from the AV, which satisfies high navigation assistance. The maximum percentage of improvement is 7.93% and 8.1% for the time and distance.
The comparative illustrations of information latency for distance and time are in Figure 9.
The accurate information transmission between the AV and interacting “Thing” through the learning process classifies safe and unsafe navigation responses. The observed environmental information is extracted from the previous neuro batches to ensure highly safe navigation. The available data extraction and environment detection accuracy are considered to identify the risk factor level. If identifying the unsafe activity of the vehicle in any instance, the risk factors originate from the information actuated and sensed for accurate decision-making. Reliable and persistent information sharing is based on the environmental knowledge of the vehicle to be observed, and the risk level can be detected through the batch optimisation algorithm and neural learning process. This detection and batch optimisation accuracy is increased by streamlining its previous neuro batches for filtering safe navigation responses, preventing information latency. The extracted features are extended to the vehicle decision over unavailable information-based environmental data observation, for which the proposed framework satisfies less information latency (Figure 9). The proposed framework suppresses the latency by 10.22% and 10.58% for the time and distance from the above comparative analysis.

5. Discussion

The choices of ORRL, HDSE-DRL, and CRL-MPC as comparison algorithms for the COIF show their utility in increasing autonomous vehicle (AV) safety, flexibility, and decision-making under uncertain conditions.
(i)
COIF’s primary goal is the same as ORRL’s: to ensure AV decisions can withstand unpredictable environments. The overarching objective of ORRL and COIF is to enhance AV performance in novel environments. However, COIF enhances information-sharing consistency and improves control over safety responses by communicating with external “Things” in the environment in real time via neuro-batch learning.
(ii)
HDSE-DRL was selected for its strategic flexibility in uncertain highway driving, aligning with COIF’s emphasis on resilience. The versatility is further enhanced by COIF’s ability to classify and batch various types of data (such as traffic density, objects, and obstacles), which allows for more detailed environmental information and the real-time detection of harmful activities.
(iii)
CRL-MPC is included, which enhances vehicle navigation via predictive control and focuses on managing AV behaviours during car-following manoeuvres. Using CRL-MPC’s predicted performance as a yardstick, one may assess COIF’s real-time adaptability and decision consistency. While CRL-MPC uses reinforcement learning and model predictive control to anticipate how vehicles will behave and ensure safe following distances, COIF uses adaptive neuro-batch learning to enhance real-time responses to various road circumstances.
As a result, COIF can maximise safety by continually analysing hazards, keeping an eye on them, and correcting for small changes in vehicle choices. In contrast to these proven methods, COIF may adaptively enhance vehicle-pedestrian safety through consistent, robust, and classified interaction, and it takes a novel approach to improve AV control.
Indeed, the findings of the comparison study are encouraging, demonstrating that the suggested strategy outperforms current methods. Nevertheless, it would be much more beneficial to provide an exact estimate of the obtained accuracy since this would provide a clear and measurable way to evaluate the development. Such an absolute accuracy number might provide a precise measure of this method’s effect on autonomous vehicle operations’ performance, dependability, and safety. The method’s benefits would be better supported, and its efficacy might be more easily assessed with this information included if it is feasible.

6. Conclusions

This article proposes and discusses the control optimiser interaction framework for autonomous vehicle driving and navigation safety. The AV control optimisation with environmental adaptability through intense learning is designed to improve its performance. Based on the neuro-batch optimisation method, the communication between the roadside intelligent communication “Things” is used to improve the vehicles’ adaptability and environmental knowledge gain. The interaction extracts, computes, and utilises data through batches with different additive or depreciative information exchanged. This included the traffic information, pedestrian density, hindrances, etc., in the driving path of the vehicles. The batches are connected through neurons and are recurrently trained for any new information acquired throughout the interaction. Based on the risk and driving parameters, the safe and unsafe activity of the vehicles is categorised with a precise learning rate. Therefore, minor changes in vehicular decisions are monitored, and driving control is optimised accordingly to retain 7.93% of navigation assistance through a 9.76% high learning rate for different intervals.

Author Contributions

All authors planned the study and contributed to the idea and field of information. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to thank Deanship of Scientific Research at Majmaah University for supporting this work under Project Number No. R-2024-1392.

Data Availability Statement

The data will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Duan, J.; Wang, Z.; Jing, X. Digital Twin Test Method With LTE-V2X for Autonomous Vehicle Safety Test. IEEE Internet Things J. 2024, 11, 30161–30171. [Google Scholar] [CrossRef]
  2. Jiang, Z.; Pan, W.; Liu, J.; Dang, S.; Yang, Z.; Li, H.; Pan, Y. Efficient and unbiased safety test for autonomous driving systems. IEEE Trans. Intell. Veh. 2022, 8, 3336–3348. [Google Scholar] [CrossRef]
  3. Hosseinian, S.M.; Mirzahossein, H. Efficiency and safety of traffic networks under the effect of autonomous vehicles. Iran. J. Sci. Technol. Trans. Civ. Eng. 2024, 48, 1861–1885. [Google Scholar] [CrossRef]
  4. Noh, J.; Jo, Y.; Kim, J.; Min, K. Enhancing transportation safety with infrastructure cooperative autonomous driving system. Int. J. Automot. Technol. 2024, 25, 61–69. [Google Scholar] [CrossRef]
  5. Wang, H.; Song, L.; Wei, Z.; Peng, L.; Li, J.; Hashemi, E. Driving safety zone model oriented motion planning framework for autonomous truck platooning. Accid. Anal. Prev. 2023, 193, 107225. [Google Scholar] [CrossRef]
  6. Wei, S.; Shao, M. Existence of connected and autonomous vehicles in mixed traffic: Impacts on safety and environment. Traffic Inj. Prev. 2024, 25, 390–399. [Google Scholar] [CrossRef]
  7. Liu, W.; Hua, M.; Deng, Z.; Meng, Z.; Huang, Y.; Hu, C.; Song, S.; Gao, L.; Liu, C.; Shuai, B.; et al. A systematic survey of control techniques and applications in connected and automated vehicles. IEEE Internet Things J. 2023, 10, 21892–21916. [Google Scholar] [CrossRef]
  8. Tang, X.; Yan, Y.; Wang, B. Trajectory Tracking Control of Autonomous Vehicles Combining ACT-R Cognitive Framework and Preview Tracking Theory. IEEE Access 2023, 11, 137067–137082. [Google Scholar] [CrossRef]
  9. Sun, Z.; Wang, R.; Meng, X.; Yang, Y.; Wei, Z.; Ye, Q. A novel path tracking system for autonomous vehicle based on model predictive control. J. Mech. Sci. Technol. 2024, 38, 365–378. [Google Scholar] [CrossRef]
  10. Kim, J.S.; Quan, Y.S.; Chung, C.C. Koopman operator-based model identification and control for automated driving vehicle. Int. J. Control. Autom. Syst. 2023, 21, 2431–2443. [Google Scholar] [CrossRef]
  11. Chen, J.; Chen, F. Efficient vehicle lateral safety analysis based on Multi-Kriging metamodels: Autonomous trucks under different lateral control modes during being overtaken. Accid. Anal. Prev. 2024, 208, 107787. [Google Scholar] [CrossRef] [PubMed]
  12. He, S.; Liu, X.; Wang, J.; Ge, H.; Long, K.; Huang, H. Influence of conventional driving habits on takeover performance in joystick-controlled autonomous vehicles: A low-speed field experiment. Heliyon 2024, 10, e31975. [Google Scholar] [CrossRef] [PubMed]
  13. Ni, H.; Yu, G.; Chen, P.; Zhou, B.; Liao, Y.; Li, H. An Integrated Framework of Lateral and Longitudinal Behavior Decision-Making for Autonomous Driving Using Reinforcement Learning. IEEE Trans. Veh. Technol. 2024, 73, 9706–9720. [Google Scholar] [CrossRef]
  14. Liao, Y.; Yu, G.; Chen, P.; Zhou, B.; Li, H. Integration of Decision-Making and Motion Planning for Autonomous Driving Based on Double-Layer Reinforcement Learning Framework. IEEE Trans. Veh. Technol. 2023, 73, 3142–3158. [Google Scholar] [CrossRef]
  15. Peng, Z.; Xia, F.; Liu, L.; Wang, D.; Li, T.; Peng, M. Online deep learning control of an autonomous surface vehicle using learned dynamics. IEEE Trans. Intell. Veh. 2023, 9, 3283–3292. [Google Scholar] [CrossRef]
  16. Yu, J.; Arab, A.; Yi, J.; Pei, X.; Guo, X. Hierarchical framework integrating rapidly-exploring random tree with deep reinforcement learning for autonomous vehicle. Appl. Intell. 2023, 53, 16473–16486. [Google Scholar] [CrossRef]
  17. Nguyen, H.D.; Han, K. Safe reinforcement learning-based driving policy design for autonomous vehicles on highways. Int. J. Control. Autom. Syst. 2023, 21, 4098–4110. [Google Scholar] [CrossRef]
  18. Guangwen, T.; Mengshan, L.; Biyu, H.; Jihong, Z.; Lixin, G. Achieving accurate trajectory predicting and tracking for autonomous vehicles via reinforcement learning-assisted control approaches. Eng. Appl. Artif. Intell. 2024, 135, 108773. [Google Scholar] [CrossRef]
  19. Tóth, S.H.; Bárdos, Á.; Viharos, Z.J. Tabular Q-learning Based Reinforcement Learning Agent for Autonomous Vehicle Drift Initiation and Stabilization. IFAC-PapersOnLine 2023, 56, 4896–4903. [Google Scholar] [CrossRef]
  20. Kang, L.; Shen, H.; Li, Y.; Xu, S. A data-driven control-policy-based driving safety analysis system for autonomous vehicles. IEEE Internet Things J. 2023, 10, 14058–14070. [Google Scholar] [CrossRef]
  21. Han, S.; Zhou, S.; Wang, J.; Pepin, L.; Ding, C.; Fu, J.; Miao, F. A multi-agent reinforcement learning approach for safe and efficient behavior planning of connected autonomous vehicles. IEEE Trans. Intell. Transp. Syst. 2023, 25, 3654–3670. [Google Scholar] [CrossRef]
  22. Wang, X.; Hu, J.; Wei, C.; Li, L.; Li, Y.; Du, M. A Novel Lane-Change Decision-Making With Long-Time Trajectory Prediction for Autonomous Vehicle. IEEE Access 2023, 11, 137437–137449. [Google Scholar] [CrossRef]
  23. Sun, Q.; Wang, X.; Yang, G.; Chen, Y.H.; Ma, F. Adaptive robust formation control of connected and autonomous vehicle swarm system based on constraint following. IEEE Trans. Cybern. 2022, 53, 4189–4203. [Google Scholar] [CrossRef] [PubMed]
  24. Li, D.; Zhang, J.; Liu, G. Autonomous Driving Decision Algorithm for Complex Multi-Vehicle Interactions: An Efficient Approach Based on Global Sorting and Local Gaming. IEEE Trans. Intell. Transp. Syst. 2024, 25, 6927–6937. [Google Scholar] [CrossRef]
  25. Jond, H.B.; Platoš, J. Differential game-based optimal control of autonomous vehicle convoy. IEEE Trans. Intell. Transp. Syst. 2022, 24, 2903–2919. [Google Scholar] [CrossRef]
  26. Wang, L.; Yang, S.; Yuan, K.; Huang, Y.; Chen, H. A combined reinforcement learning and model predictive control for car-following maneuver of autonomous vehicles. Chin. J. Mech. Eng. 2023, 36, 80. [Google Scholar] [CrossRef]
  27. Nie, X.; Liang, Y.; Ohkura, K. Autonomous highway driving using reinforcement learning with safety check system based on time-to-collision. Artif. Life Robot. 2023, 28, 158–165. [Google Scholar] [CrossRef]
  28. Deng, H.; Zhao, Y.; Wang, Q.; Nguyen, A.T. Deep Reinforcement Learning Based Decision-Making Strategy of Autonomous Vehicle in Highway Uncertain Driving Environments. Automot. Innov. 2023, 6, 438–452. [Google Scholar] [CrossRef]
  29. Feng, S.; Sun, H.; Yan, X.; Zhu, H.; Zou, Z.; Shen, S.; Liu, H.X. Dense reinforcement learning for safety validation of autonomous vehicles. Nature 2023, 615, 620–627. [Google Scholar] [CrossRef]
  30. Lee, C.Y.; Khanum, A.; Sung, T.W. Robust autonomous driving control using deep hybrid-learning network under rainy/snown conditions. Multimed. Tools Appl. 2024, 1–15. [Google Scholar] [CrossRef]
  31. He, X.; Lv, C. Towards Safe Autonomous Driving: Decision Making with Observation-Robust Reinforcement Learning. Automot. Innov. 2023, 6, 509–520. [Google Scholar] [CrossRef]
  32. Ben Elallid, B.; Bagaa, M.; Benamar, N.; Mrani, N. A reinforcement learning based autonomous vehicle control in diverse daytime and weather scenarios. J. Intell. Transp. Syst. 2024, 1–14. [Google Scholar] [CrossRef]
  33. Shi, Y.; Liu, J.; Liu, C.; Gu, Z. DeepAD: An integrated decision-making framework for intelligent autonomous driving. Transp. Res. Part A Policy Pract. 2024, 183, 104069. [Google Scholar] [CrossRef]
  34. Li, H.; Wei, W.; Zheng, S.; Sun, C.; Lu, Y.; Zhou, T. Personalised driving behavior oriented autonomous vehicle control for typical traffic situations. J. Frankl. Inst. 2024, 361, 106924. [Google Scholar] [CrossRef]
  35. Mihály, A.; Do, T.T.; Gáspár, P. Supervised reinforcement learning based trajectory tracking control for autonomous vehicles. IFAC-PapersOnLine 2024, 58, 140–145. [Google Scholar] [CrossRef]
  36. Gao, F.; Luo, C.; Shi, F.; Chen, X.; Gao, Z.; Zhao, R. Online Safety Verification of Autonomous Driving Decision-Making Based on Dynamic Reachability Analysis. IEEE Access 2023, 11, 93293–93309. [Google Scholar] [CrossRef]
  37. Wang, J.; Jiang, Z.; Pant, Y.V. Improving safety in mixed traffic: A learning-based model predictive control for autonomous and human-driven vehicle platooning. Knowl. Based Syst. 2024, 293, 111673. [Google Scholar] [CrossRef]
  38. Kang, D.; Li, Z.; Levin, M.W. Evasion planning for autonomous intersection control based on an optimised conflict point control formulation. J. Transp. Saf. Secur. 2022, 14, 2074–2110. [Google Scholar]
  39. Available online: https://www.kaggle.com/datasets/roydatascience/training-car (accessed on 22 August 2024).
  40. Available online: https://www.kaggle.com/datasets/magnumresearchgroup/offroad-terrain-dataset-for-autonomous-vehicles (accessed on 22 August 2024).
Figure 1. AV navigation scenario from data collected.
Figure 1. AV navigation scenario from data collected.
Machines 12 00798 g001
Figure 2. Application of COI framework in AV scenarios.
Figure 2. Application of COI framework in AV scenarios.
Machines 12 00798 g002
Figure 3. Batch optimisation model for risk assessment.
Figure 3. Batch optimisation model for risk assessment.
Machines 12 00798 g003
Figure 4. Assessment of parameters from Equations (8)–(11).
Figure 4. Assessment of parameters from Equations (8)–(11).
Machines 12 00798 g004
Figure 5. Unsafe AV navigation function outputs analysis.
Figure 5. Unsafe AV navigation function outputs analysis.
Machines 12 00798 g005
Figure 6. Comparative illustrations for learning rate.
Figure 6. Comparative illustrations for learning rate.
Machines 12 00798 g006
Figure 7. Environment detection ratio comparisons.
Figure 7. Environment detection ratio comparisons.
Machines 12 00798 g007
Figure 8. Navigation assistance rate for distance and time.
Figure 8. Navigation assistance rate for distance and time.
Machines 12 00798 g008
Figure 9. Comparative illustrations of information latency.
Figure 9. Comparative illustrations of information latency.
Machines 12 00798 g009
Table 1. Impacts of vehicle’s behaviour Lemma 1.
Table 1. Impacts of vehicle’s behaviour Lemma 1.
I n t R V = m a x s i n
i T = X i T X i o ± W i μ i T
where T = 1,2 , , m a x s S a f e D
Based on X i T Based on i T
Define X i T = i = 1 T X i o , W i μ i T = 0
i T = 0 unless W i μ i T = 0
I n t R V = m a x s and I n t R V = m i n s
m a x s , if E r d = 0 , then
μ i T = Y i T = m a x s , m i n s , O V s , L V s
Impact 1: μ i T = X i T Y i T //low
where I n t R V = i = 1 T W i × Y i t c
I n t R V = m i n s , c = W i × Y i T X i T
Impact 2: μ i T = X i T i T //high
i T = X i μ i T i = 0 T c = I n t R V
where I n t R V = i = 1 T W i × Y i t c
I n t R V = m i n s , c = W i × Y i T X i T
Impact 2: μ i T = X i T i T //high
i T = X i μ i T i = 0 T c = I n t R V
Define μ i = X i T i T
For the S a f e D = R V s 2 E r d , the
μ i impacts as follows
E r d = R V o m a x s T
such that
X i T = Y i T + W i × i T
and
Impact 1: c = S a f e D Y i T X i T //low
Impact 2: c = 0 i > Y i T
Machines 12 00798 i001Machines 12 00798 i002Machines 12 00798 i003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Almutairi, A.; Asmari, A.F.A.; Alqubaysi, T.; Alanazi, F.; Armghan, A. Ensuring Driving and Road Safety of Autonomous Vehicles Using a Control Optimiser Interaction Framework Through Smart “Thing” Information Sensing and Actuation. Machines 2024, 12, 798. https://doi.org/10.3390/machines12110798

AMA Style

Almutairi A, Asmari AFA, Alqubaysi T, Alanazi F, Armghan A. Ensuring Driving and Road Safety of Autonomous Vehicles Using a Control Optimiser Interaction Framework Through Smart “Thing” Information Sensing and Actuation. Machines. 2024; 12(11):798. https://doi.org/10.3390/machines12110798

Chicago/Turabian Style

Almutairi, Ahmed, Abdullah Faiz Al Asmari, Tariq Alqubaysi, Fayez Alanazi, and Ammar Armghan. 2024. "Ensuring Driving and Road Safety of Autonomous Vehicles Using a Control Optimiser Interaction Framework Through Smart “Thing” Information Sensing and Actuation" Machines 12, no. 11: 798. https://doi.org/10.3390/machines12110798

APA Style

Almutairi, A., Asmari, A. F. A., Alqubaysi, T., Alanazi, F., & Armghan, A. (2024). Ensuring Driving and Road Safety of Autonomous Vehicles Using a Control Optimiser Interaction Framework Through Smart “Thing” Information Sensing and Actuation. Machines, 12(11), 798. https://doi.org/10.3390/machines12110798

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop