Abstract
Urban tunnel portals constitute critical safety zones for autonomous vehicles, where abrupt luminance transitions, shortened sight distances, and densely distributed structural and traffic elements pose considerable challenges to perception reliability. Existing driving scenario datasets are rarely tailored to tunnel environments and have not quantitatively evaluated how specific infrastructure components influence perception latency in autonomous systems. This study develops a requirement-driven framework for the identification and sensitivity ranking of information perception elements within urban tunnel portals. Based on expert evaluations and a combined function–safety scoring system, nine key elements—including road surfaces, tunnel portals, lane markings, and vehicles—were identified as perception-critical. A “mandatory–optional” combination rule was then applied to generate 48 logical scene types, and 376 images after brightness (30–220 px), blur (Laplacian variance ≥ 100), and occlusion filtering (≤0.5% pixel error) were obtained after luminance and occlusion screening. A ResNet50–PSPNet convolutional neural network was trained to perform pixel-level segmentation, with inference rate adopted as a quantitative proxy for perceptual sensitivity. Field experiments across ten urban tunnels in China indicate that the model consistently recognized road surfaces, lane markings, cars, and motorcycles with the shortest inference times (<6.5 ms), whereas portal structures and vegetation required longer recognition times (>7.5 ms). This sensitivity ranking is statistically stable under clear, daytime conditions (p < 0.01). The findings provide engineering insights for optimizing tunnel lighting design, signage placement, and V2X configuration, and offers a pilot dataset to support perception-oriented design and evaluation of urban tunnel portals in semi-enclosed environments. Unlike generic segmentation datasets, this study quantifies element-specific CNN latency at tunnel portals for the first time.