Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (105)

Search Parameters:
Keywords = road lane marking

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
34 pages, 4042 KB  
Article
Perceptual Elements and Sensitivity Analysis of Urban Tunnel Portals for Autonomous Driving
by Mengdie Xu, Bo Liang, Haonan Long, Chun Chen, Hongyi Zhou and Shuangkai Zhu
Appl. Sci. 2026, 16(1), 453; https://doi.org/10.3390/app16010453 - 31 Dec 2025
Viewed by 250
Abstract
Urban tunnel portals constitute critical safety zones for autonomous vehicles, where abrupt luminance transitions, shortened sight distances, and densely distributed structural and traffic elements pose considerable challenges to perception reliability. Existing driving scenario datasets are rarely tailored to tunnel environments and have not [...] Read more.
Urban tunnel portals constitute critical safety zones for autonomous vehicles, where abrupt luminance transitions, shortened sight distances, and densely distributed structural and traffic elements pose considerable challenges to perception reliability. Existing driving scenario datasets are rarely tailored to tunnel environments and have not quantitatively evaluated how specific infrastructure components influence perception latency in autonomous systems. This study develops a requirement-driven framework for the identification and sensitivity ranking of information perception elements within urban tunnel portals. Based on expert evaluations and a combined function–safety scoring system, nine key elements—including road surfaces, tunnel portals, lane markings, and vehicles—were identified as perception-critical. A “mandatory–optional” combination rule was then applied to generate 48 logical scene types, and 376 images after brightness (30–220 px), blur (Laplacian variance ≥ 100), and occlusion filtering (≤0.5% pixel error) were obtained after luminance and occlusion screening. A ResNet50–PSPNet convolutional neural network was trained to perform pixel-level segmentation, with inference rate adopted as a quantitative proxy for perceptual sensitivity. Field experiments across ten urban tunnels in China indicate that the model consistently recognized road surfaces, lane markings, cars, and motorcycles with the shortest inference times (<6.5 ms), whereas portal structures and vegetation required longer recognition times (>7.5 ms). This sensitivity ranking is statistically stable under clear, daytime conditions (p < 0.01). The findings provide engineering insights for optimizing tunnel lighting design, signage placement, and V2X configuration, and offers a pilot dataset to support perception-oriented design and evaluation of urban tunnel portals in semi-enclosed environments. Unlike generic segmentation datasets, this study quantifies element-specific CNN latency at tunnel portals for the first time. Full article
(This article belongs to the Section Civil Engineering)
Show Figures

Figure 1

17 pages, 498 KB  
Article
Developing Region-Specific Safety Performance Functions for Intercity Roads in Saudi Arabia
by Saif Alarifi and Khalid Alkahtani
Appl. Sci. 2026, 16(1), 227; https://doi.org/10.3390/app16010227 - 25 Dec 2025
Viewed by 246
Abstract
This study develops comprehensive Safety Performance Functions (SPFs) for various intercity road types in Saudi Arabia, including freeways, multilane highways, and two-lane two-way roads. Data spanning 2017–2019 were analyzed for five regions—Riyadh, Makkah, Eastern, Aseer, and Tabuk—using Negative Binomial (NB) regression models aligned [...] Read more.
This study develops comprehensive Safety Performance Functions (SPFs) for various intercity road types in Saudi Arabia, including freeways, multilane highways, and two-lane two-way roads. Data spanning 2017–2019 were analyzed for five regions—Riyadh, Makkah, Eastern, Aseer, and Tabuk—using Negative Binomial (NB) regression models aligned with the Highway Safety Manual (HSM). A total of 26 SPFs were developed to predict total and fatal and injury (FI) crashes, incorporating contextual variables (e.g., tunnel density, U-turn frequency, high-speed vehicle proportion) and developing models separately for each region. It was found that as the median width increases on freeway roads in the Riyadh region, the predicted number of total and fatal and injury crashes decreases. Also, as the percentage of heavy vehicles and U-turn density increases, the number of total and fatal crashes increases on multilane roads in the Makkah region. Moreover, as the degree of curvature increases, the predicted number of total and fatal andinjury crashes increase on multilane and two-lane two-way roads in Tabuk. Lastly, in Aseer, median double marking and tunnel density along curves were significantly affecting crashes on two-lane two-way roads. This study is useful to enhance the methodology used to identify hotspots on the intercity roads in KSA. Full article
(This article belongs to the Section Civil Engineering)
Show Figures

Figure 1

21 pages, 6826 KB  
Article
Research on a Road Crack Detection Method Based on YOLO11-MBC
by Jinhui Li, Xiaowei Jiang and Hui Peng
Sensors 2025, 25(24), 7435; https://doi.org/10.3390/s25247435 - 6 Dec 2025
Viewed by 728
Abstract
To address the issues of low accuracy and high rates of false detection and missed detection in existing methods for pavement crack identification under complex road conditions, this paper proposes a novel approach named YOLO11-MBC, based on the YOLO11 model. A Multi-scale Feature [...] Read more.
To address the issues of low accuracy and high rates of false detection and missed detection in existing methods for pavement crack identification under complex road conditions, this paper proposes a novel approach named YOLO11-MBC, based on the YOLO11 model. A Multi-scale Feature Fusion Backbone Network (MFFBN) is designed to enhance the model’s capability to recognize and extract crack features in complex environments. Considering that pavement cracks often exhibit elongated topologies and are susceptible to interference from similar features like tree roots or lane markings, we combine the Bidirectional Feature Pyramid Network (BiFPN) with a Multimodal Cross-Attention (MCA) mechanism, constructing a novel BiMCNet to replace the Concat layer in the original network, thereby optimizing the detection of minute cracks. The CGeoCIoU loss function replaces the original CIoU, employing three distinct penalty terms to better reflect the alignment between predicted and ground-truth boxes. The effectiveness of the proposed method is validated through comparative and ablation experiments on the public RDD2022 dataset. Results demonstrate the following: (1) Compared to the baseline YOLO11, YOLO11-MBC achieves a 22.5% improvement in F1-score and an 8% increase in mAP50 by integrating the three proposed modules, significantly enhancing performance for complex pavement crack detection. (2) The improved algorithm demonstrates superior performance. Compared to YOLOv8, YOLOv10, and YOLO11, it achieves precision, recall, F1-score, mAP50, and mAP50-95 of 61%, 70%, 72%, 75%, and 66%, respectively, validating the correctness of our approach. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

20 pages, 9080 KB  
Article
Integration of Multi-Sensor Fusion and Decision-Making Architecture for Autonomous Vehicles in Multi-Object Traffic Conditions
by Hai Ngoc Nguyen, Thien Nguyen Luong, Tuan Pham Minh, Nguyen Mai Thi Hong, Kiet Tran Anh, Quan Bui Hong and Ngoc Pham Van Bach
Sensors 2025, 25(22), 7083; https://doi.org/10.3390/s25227083 - 20 Nov 2025
Viewed by 1074
Abstract
Autonomous vehicles represent a transformative technology in modern transportation, promising enhanced safety, efficiency, and accessibility in mobility systems. This paper presents a comprehensive autonomous vehicle system designed specifically for Vietnam’s traffic conditions, featuring a multi-layered approach to perception, decision-making, and control. The system [...] Read more.
Autonomous vehicles represent a transformative technology in modern transportation, promising enhanced safety, efficiency, and accessibility in mobility systems. This paper presents a comprehensive autonomous vehicle system designed specifically for Vietnam’s traffic conditions, featuring a multi-layered approach to perception, decision-making, and control. The system utilizes dual 2D LiDARs, camera vision, and GPS sensing to navigate complex urban environments. A key contribution is the development of a specialized segmentation model that accurately identifies Vietnam-specific traffic signs, lane markings, road features, and pedestrians. The system implements a hierarchical decision-making architecture, combining long-term planning based on GPS and map data with short-term reactive planning derived from a bird’s-eye view transformation of segmentation and LiDAR data. The control system modulates the speed and steering angle through a validated model that ensures stable vehicle operation across various traffic scenarios. Experimental results demonstrate the system’s effectiveness in real-world conditions, achieving a high accuracy rate in terms of segmentation and detection and an exact response in navigation tasks. The proposed system shows robust performance in Vietnam’s unique traffic environment, addressing challenges such as mixed traffic flow and country-specific road infrastructure. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

17 pages, 36077 KB  
Article
AI-Based Detection and Classification of Horizontal Road Markings in Digital Images Dedicated to Driver Assistance Systems
by Joanna Kulawik and Łukasz Kuczyński
Appl. Sci. 2025, 15(22), 12189; https://doi.org/10.3390/app152212189 - 17 Nov 2025
Viewed by 542
Abstract
Horizontal road markings are crucial for safe driving and for the operation of advanced driver-assistance systems (ADAS), but they have been investigated less thoroughly than vertical signs or lane boundaries. This paper focuses on the detection and classification of horizontal road markings in [...] Read more.
Horizontal road markings are crucial for safe driving and for the operation of advanced driver-assistance systems (ADAS), but they have been investigated less thoroughly than vertical signs or lane boundaries. This paper focuses on the detection and classification of horizontal road markings in digital images using modern deep learning techniques. Three YOLO models (YOLOv7, YOLOv8n, YOLOv9t) were trained and tested on a new dataset comprising 6250 images with 13,360 annotated horizontal road-marking objects across nine classes collected on Polish roads in sunny and cloudy conditions. The dataset covers nine classes of markings recorded on urban streets, rural roads and highways. It includes many difficult cases: small markings visible only from long distance or side entry roads, and markings with different levels of wear, from new and bright to faded, dirty or partially erased. YOLOv7 achieved Precision = 0.95, Recall = 0.91 and mAP@0.5 = 0.98. YOLOv8n and YOLOv9t obtained lower Recall but higher mAP@0.5:0.95 (>0.77). The results confirm that YOLO-based detectors can handle horizontal road markings under varied road conditions and degrees of visibility, and the dataset with baseline results may serve as a reference for further studies in intelligent transport systems. Full article
Show Figures

Figure 1

16 pages, 3531 KB  
Article
Research on Reliability of Vehicle Line Detection and Lane Keeping Systems
by Vytenis Surblys, Vidas Žuraulis and Tadas Tinginys
Sustainability 2025, 17(22), 10222; https://doi.org/10.3390/su172210222 - 15 Nov 2025
Viewed by 2790
Abstract
This research focuses on vehicle Advanced Driver Assistance Systems (ADAS), with particular emphasis on Lane Keeping Assist (LKA) systems which is designed to help drivers keep a vehicle centered within its lane and reduce the risk of unintentional lane departures. These kinds of [...] Read more.
This research focuses on vehicle Advanced Driver Assistance Systems (ADAS), with particular emphasis on Lane Keeping Assist (LKA) systems which is designed to help drivers keep a vehicle centered within its lane and reduce the risk of unintentional lane departures. These kinds of systems detect lane boundaries using computer vision algorithms applied to video data captured by a forward-facing camera and interpret this visual information to provide corrective steering inputs or driver alerts. The research investigates the performance, reliability, sustainability, and limitations of LKA systems under adverse road and environmental conditions, such as wet pavement and in the presence of degraded, partially visible, or missing horizontal road markings. Improving the reliability of lane detection and keeping systems enhances road safety, reducing traffic accidents caused by lane departures, which directly supports social sustainability. For the theoretical test, a modified road model using MATLAB software was used to simulate poor road markings and to investigate possible test outcomes. A series of field tests were conducted on multiple passenger vehicles equipped with LKA technologies to evaluate their response in real-world scenarios. The results show that it is very important to ensure high quality horizontal road markings as specified in UNECE Regulation No. 130, as lane keeping aids are not uniformly effective. Furthermore, the study highlights the need to develop more robust line detection algorithms capable of adapting to diverse road and weather conditions, thereby enhancing overall driving safety and system reliability. LKA system research supports sustainable mobility strategies promoted by international organizations—aiming to transition to safer, smarter, and less polluting transportation systems. Full article
Show Figures

Figure 1

33 pages, 66840 KB  
Article
VR Human-Centric Winter Lane Detection: Performance and Driving Experience Evaluation
by Tatiana Ortegon-Sarmiento, Patricia Paderewski, Sousso Kelouwani, Francisco Gutierrez-Vela and Alvaro Uribe-Quevedo
Sensors 2025, 25(20), 6312; https://doi.org/10.3390/s25206312 - 12 Oct 2025
Viewed by 1072
Abstract
Driving in snowy conditions challenges both human drivers and autonomous systems. Snowfall and ice accumulation impair vehicle control and affect driver perception and performance. Road markings are often obscured, forcing drivers to rely on intuition and memory to stay in their lane, which [...] Read more.
Driving in snowy conditions challenges both human drivers and autonomous systems. Snowfall and ice accumulation impair vehicle control and affect driver perception and performance. Road markings are often obscured, forcing drivers to rely on intuition and memory to stay in their lane, which can lead to encroachment into adjacent lanes or sidewalks. Current lane detectors assist in lane keeping, but their performance is compromised by visual disturbances such as ice reflection, snowflake movement, fog, and snow cover. Furthermore, testing these systems with users on actual snowy roads involves risks to driver safety, equipment integrity, and ethical compliance. This study presents a low-cost virtual reality simulation for evaluating winter lane detection in controlled and safe conditions from a human-in-the-loop perspective. Participants drove in a simulated snowy scenario with and without the detector while quantitative and qualitative variables were monitored. Results showed a 49.9% reduction in unintentional lane departures with the detector and significantly improved user experience, as measured by the UEQ-S (p = 0.023, Cohen’s d = 0.72). Participants also reported higher perceived safety, situational awareness, and confidence. These findings highlight the potential of vision-based lane detection systems adapted to winter environments and demonstrate the value of immersive simulations for user-centered testing of ADASs. Full article
(This article belongs to the Topic Extended Reality: Models and Applications)
Show Figures

Figure 1

19 pages, 2217 KB  
Article
Assessing Infrastructure Readiness of Controlled-Access Roads in West Bangkok for Autonomous Vehicle Deployment
by Vasin Kiattikomol, Laphisa Nuangrod, Arissara Rung-in and Vanchanok Chuathong
Infrastructures 2025, 10(10), 270; https://doi.org/10.3390/infrastructures10100270 - 10 Oct 2025
Viewed by 805
Abstract
The deployment of autonomous vehicles (AVs) depends on the readiness of both physical and digital infrastructure. However, existing national and city-level indices often overlook deficiencies along specific routes, particularly in developing contexts such as Thailand, where infrastructure conditions vary widely. This study develops [...] Read more.
The deployment of autonomous vehicles (AVs) depends on the readiness of both physical and digital infrastructure. However, existing national and city-level indices often overlook deficiencies along specific routes, particularly in developing contexts such as Thailand, where infrastructure conditions vary widely. This study develops and applies a corridor-level framework to assess AV readiness on five controlled-access roads in western Bangkok. The framework evaluates key infrastructure dimensions beyond conventional vehicle requirements. In this study, infrastructure readiness means the extent to which essential physical (EV charging capacity, traffic sign visibility, and lane marking retroreflectivity) and digital (5G speed and coverage) subsystems meet minimum operational thresholds required for AV deployment. Data were collected through field measurements and secondary sources, utilizing tools such as a retroreflectometer, a handheld spectrum analyzer, and the Ookla Speedtest application. The results reveal significant contrasts for physical infrastructure, showing that traffic signage is generally satisfactory, but EV charging capacity and road marking retroreflectivity are insufficient on most routes. On the digital side, 5G coverage was generally adequate, but network speeds remained less than half of the global benchmark. Kanchanaphisek Road demonstrated comparatively higher digital readiness, whereas Ratchaphruek Road exhibited the weakest road marking conditions. These findings point out the need for stepwise enhancements to EV charging infrastructure, lane marking maintenance, and digital connectivity to support safe and reliable AV operations. The proposed framework not only provides policymakers in Thailand with a practical tool for prioritizing corridor-level investments but also offers transferability to other rapidly developing urban regions experiencing similar infrastructure challenges for AV deployment. Full article
Show Figures

Figure 1

33 pages, 14767 KB  
Article
Night-to-Day Image Translation with Road Light Attention Training for Traffic Information Detection
by Ye-Jin Lee, Young-Ho Go, Seung-Hwan Lee, Dong-Min Son and Sung-Hak Lee
Mathematics 2025, 13(18), 2998; https://doi.org/10.3390/math13182998 - 16 Sep 2025
Viewed by 1466
Abstract
Generative adversarial networks (GANs)-based image deep learning methods are useful to improve object visibility in nighttime driving environments, but they often fail to preserve critical road information like traffic light colors and vehicle lighting. This paper proposes a method to address this by [...] Read more.
Generative adversarial networks (GANs)-based image deep learning methods are useful to improve object visibility in nighttime driving environments, but they often fail to preserve critical road information like traffic light colors and vehicle lighting. This paper proposes a method to address this by utilizing both unpaired and four-channel paired training modules. The unpaired module performs the primary night-to-day conversion, while the paired module, enhanced with a fourth channel, focuses on preserving road details. Our key contribution is an inverse road light attention (RLA) map, which acts as this fourth channel to explicitly guide the network’s learning. This map also facilitates a final cross-blending process, synthesizing the results from both modules to maximize their respective advantages. Experimental results demonstrate that our approach more accurately preserves lane markings and traffic light colors. Furthermore, quantitative analysis confirms that our method achieves superior performance across eight no-reference image quality metrics compared to existing techniques. Full article
(This article belongs to the Special Issue Machine Learning Applications in Image Processing and Computer Vision)
Show Figures

Figure 1

15 pages, 339 KB  
Article
Hybrid MambaVision and Transformer-Based Architecture for 3D Lane Detection
by Raul-Mihai Cap and Călin-Adrian Popa
Sensors 2025, 25(18), 5729; https://doi.org/10.3390/s25185729 - 14 Sep 2025
Viewed by 1994
Abstract
Lane detection is an essential task in the field of computer vision and autonomous driving. This involves identifying and locating road markings on the road surface. This capability not only helps drivers keep the vehicle in the correct lane, but also provides critical [...] Read more.
Lane detection is an essential task in the field of computer vision and autonomous driving. This involves identifying and locating road markings on the road surface. This capability not only helps drivers keep the vehicle in the correct lane, but also provides critical data for advanced driver assistance systems and autonomous vehicles. Traditional lane detection models work mainly on the 2D image plane and achieve remarkable results. However, these models often assume a flat-world scenario, which does not correspond to real-world conditions, where roads have elevation variations and road markings may be curved. Our approach solves this challenge by focusing on 3D lane detection without relying on the inverse perspective mapping technique. Instead, we introduce a new framework using the MambaVision-S-1K backbone, which combines Mamba-based processing with Transformer capabilities to capture both local detail and global contexts from monocular images. This hybrid approach allows accurate modeling of lane geometry in three dimensions, even in the presence of elevation variations. By replacing the traditional convolutional neural network backbone with MambaVision, our proposed model significantly improves the capability of 3D lane detection systems. Our method achieved state-of-the-art performance on the ONCE-3DLanes dataset, thus demonstrating its superiority in accurately capturing lane curvature and elevation variations. These results highlight the potential of integrating advanced backbones based on Vision Transformers in the field of autonomous driving for more robust and reliable lane detection. The code will be available online. Full article
(This article belongs to the Special Issue Intelligent Sensors for Smart and Autonomous Vehicles)
Show Figures

Figure 1

82 pages, 17076 KB  
Review
Advancements in Embedded Vision Systems for Automotive: A Comprehensive Study on Detection and Recognition Techniques
by Anass Barodi, Mohammed Benbrahim and Abdelkarim Zemmouri
Vehicles 2025, 7(3), 99; https://doi.org/10.3390/vehicles7030099 - 12 Sep 2025
Cited by 1 | Viewed by 2579
Abstract
Embedded vision systems play a crucial role in the advancement of intelligent transportation by supporting real-time perception tasks such as traffic sign recognition and lane detection. Despite significant progress, their performance remains sensitive to environmental variability, computational constraints, and scene complexity. This review [...] Read more.
Embedded vision systems play a crucial role in the advancement of intelligent transportation by supporting real-time perception tasks such as traffic sign recognition and lane detection. Despite significant progress, their performance remains sensitive to environmental variability, computational constraints, and scene complexity. This review examines the current state of the art in embedded vision approaches used for the detection and classification of traffic signs and lane markings. The literature is structured around three main stages, localization, detection, and recognition, highlighting how visual features like color, geometry, and road edges are processed through both traditional and learning-based methods. A major contribution of this work is the introduction of a practical taxonomy that organizes recognition techniques according to their computational load and real-time applicability in embedded contexts. In addition, the paper presents a critical synthesis of existing limitations, with attention to sensor fusion challenges, dataset diversity, and deployment in real-world conditions. By adopting the SALSA methodology, the review follows a transparent and systematic selection process, ensuring reproducibility and clarity. The study concludes by identifying specific research directions aimed at improving the robustness, scalability, and interpretability of embedded vision systems. These contributions position the review as a structured reference for researchers working on intelligent driving technologies and next-generation driver assistance systems. The findings are expected to inform future implementations of embedded vision systems in real-world driving environments. Full article
Show Figures

Figure 1

17 pages, 2128 KB  
Article
Vision-Based Highway Lane Extraction from UAV Imagery: A Deep Learning and Geometric Constraints Approach
by Jin Wang, Guangjun He, Xiuwang Dai, Feng Wang and Yanxin Zhang
Electronics 2025, 14(17), 3554; https://doi.org/10.3390/electronics14173554 - 6 Sep 2025
Cited by 1 | Viewed by 1096
Abstract
The rapid evolution of unmanned aerial vehicle (UAV) technology and low-altitude economic development have propelled drone applications in critical infrastructure monitoring, particularly in intelligent transportation systems where real-time aerial image processing has emerged as a pressing requirement. We address the pivotal challenge of [...] Read more.
The rapid evolution of unmanned aerial vehicle (UAV) technology and low-altitude economic development have propelled drone applications in critical infrastructure monitoring, particularly in intelligent transportation systems where real-time aerial image processing has emerged as a pressing requirement. We address the pivotal challenge of highway lane extraction from low-altitude UAV perspectives by applying a novel three-stage framework. This framework consists of (1) a deep learning-based semantic segmentation module that employs an enhanced STDC network with boundary-aware loss for precise detection of roads and lane markings; (2) an optimized polynomial fitting algorithm incorporating iterative classification and adaptive error thresholds to achieve robust lane marking consolidation; and (3) a global optimization module designed for context-aware lane generation. Our methodology demonstrates superior performance with 94.11% F1-score and 93.84% IoU, effectively bridging the technical gap in UAV-based lane extraction while establishing a reliable foundation for advanced traffic monitoring applications. Full article
Show Figures

Figure 1

4 pages, 1714 KB  
Proceeding Paper
A Study on High-Precision Vehicle Navigation for Autonomous Driving on an Ultra-Long Underground Expressway
by Kyoung-Soo Choi, Yui-Hwan Sa, Min-Gyeong Choi, Sung-Jin Kim and Won-Woo Lee
Eng. Proc. 2025, 102(1), 10; https://doi.org/10.3390/engproc2025102010 - 5 Aug 2025
Viewed by 617
Abstract
GPSs typically have an accuracy ranging from a few meters to several tens of meters. However, when corrected using various methods, they can achieve an accuracy of several tens of centimeters. In autonomous driving, a positioning accuracy of less than 50 cm is [...] Read more.
GPSs typically have an accuracy ranging from a few meters to several tens of meters. However, when corrected using various methods, they can achieve an accuracy of several tens of centimeters. In autonomous driving, a positioning accuracy of less than 50 cm is required for lane-level positioning, route generation, and navigation. However, in environments where GPS signals are blocked, such as tunnels and underground roads, absolute positioning is impossible. Instead, relative positioning methods integrating IMU, IVN, and cameras are used. These methods are influenced by numerous variables, however, such as vehicle speed and road conditions, resulting in lower accuracy. In this study, we conducted experiments on current vehicle navigation technologies using an autonomous driving simulation vehicle in the Suri–Suam Tunnel of the Seoul Metropolitan Area 1st Ring Expressway. To recognize objects (lane markings/2D/3D) for position correction inside the tunnel, data on tunnel and underground road infrastructure in Seoul and Gyeonggi Province was collected, processed, refined, and trained. Additionally, a Loosely Coupled-based Kalman Filter was designed and applied for the fusion of GPSs, IMUs, and IVNs. As a result, an error of 113.62 cm was observed in certain sections. This suggests that while the technology is applicable for general vehicle lane-level navigation in ultra-long tunnels spanning several kilometers for public service, it falls short of meeting the precision required for autonomous driving systems, which demand lane-level accuracy. Therefore, it was concluded that infrastructure-based absolute positioning technology is necessary to enable precise navigation inside tunnels. Full article
(This article belongs to the Proceedings of The 2025 Suwon ITS Asia Pacific Forum)
Show Figures

Figure 1

30 pages, 2282 KB  
Article
User Experience of Navigating Work Zones with Automated Vehicles: Insights from YouTube on Challenges and Strengths
by Melika Ansarinejad, Kian Ansarinejad, Pan Lu and Ying Huang
Smart Cities 2025, 8(4), 120; https://doi.org/10.3390/smartcities8040120 - 19 Jul 2025
Viewed by 2126
Abstract
Understanding automated vehicle (AV) behavior in complex road environments and user attitudes in such contexts is critical for their safe and effective integration into smart cities. Despite growing deployment, limited public data exist on AV performance in construction zones; highly dynamic settings marked [...] Read more.
Understanding automated vehicle (AV) behavior in complex road environments and user attitudes in such contexts is critical for their safe and effective integration into smart cities. Despite growing deployment, limited public data exist on AV performance in construction zones; highly dynamic settings marked by irregular lane markings, shifting detours, and unpredictable human presence. This study investigates AV behavior in these conditions through qualitative, video-based analysis of user-documented experiences on YouTube, focusing on Tesla’s supervised Full Self-Driving (FSD) and Waymo systems. Spoken narration, captions, and subtitles were examined to evaluate AV perception, decision-making, control, and interaction with humans. Findings reveal that while AVs excel in structured tasks such as obstacle detection, lane tracking, and cautious speed control, they face challenges in interpreting temporary infrastructure, responding to unpredictable human actions, and navigating low-visibility environments. These limitations not only impact performance but also influence user trust and acceptance. The study underscores the need for continued technological refinement, improved infrastructure design, and user-informed deployment strategies. By addressing current shortcomings, this research offers critical insights into AV readiness for real-world conditions and contributes to safer, more adaptive urban mobility systems. Full article
Show Figures

Figure 1

17 pages, 7477 KB  
Article
The Development of a Lane Identification and Assessment Framework for Maintenance Using AI Technology
by Hohyuk Na, Do Gyeong Kim, Ji Min Kang and Chungwon Lee
Appl. Sci. 2025, 15(13), 7410; https://doi.org/10.3390/app15137410 - 1 Jul 2025
Viewed by 1331
Abstract
This study proposes a vision-based framework to support AVs in maintaining stable lane-keeping by assessing the condition of lane markings. Unlike existing infrastructure standards focused on human visibility, this study addresses the need for criteria suited to sensor-based AV environments. Using real driving [...] Read more.
This study proposes a vision-based framework to support AVs in maintaining stable lane-keeping by assessing the condition of lane markings. Unlike existing infrastructure standards focused on human visibility, this study addresses the need for criteria suited to sensor-based AV environments. Using real driving data from urban expressways in Seoul, a YOLOv5-based lane detection algorithm was developed and enhanced through multi-label annotation and data augmentation. The model achieved a mean average precision (mAP) of 97.4% and demonstrated strong generalization on external datasets such as KITTI and TuSimple. For lane condition assessment, a pixel occupancy–based method was applied, combined with Canny edge detection and morphological operations. A threshold of 80-pixel occupancy was used to classify lanes as intact or worn. The proposed framework reliably detected lane degradation under various road and lighting conditions. These results suggest that quantitative, image-based indicators can complement traditional standards and guide AV-oriented infrastructure policy. Limitations include a lack of adverse weather data and dataset-specific threshold sensitivity. Full article
Show Figures

Figure 1

Back to TopTop