Real-Time Segmentation of Tactile Paving and Zebra Crossings for Visually Impaired Assistance Using Embedded Visual Sensors
Abstract
1. Introduction
- We designed and implemented an intelligent guide-assist vehicle equipped with a high-definition camera and an Atlas 310 embedded processor, and successfully deployed the proposed lightweight segmentation network on this platform to achieve real-time tactile paving and zebra crossing segmentation.
- We propose an improved G-GhostNet backbone that incorporates depthwise separable convolutions and a Mix operation, significantly enhancing segmentation inference speed.
- We enhanced the spatial pyramid pooling module by integrating an attention mechanism for multi-scale feature extraction, and designed a decoding module tailored to tactile paving and zebra crossing characteristics to improve segmentation accuracy and edge-preservation performance.
2. Method
2.1. Hardware System Architecture
2.2. Encoding Module
2.2.1. Improved G-GhostNet Module
2.2.2. Improved Coordinate Attention Module
2.2.3. Redesigned ASPP Module
2.3. Decoding Module
3. Experiments
3.1. Dataset and Evaluation Metrics
3.2. Experimental Details
3.3. Comparative Experimental Analysis
3.4. Ablation Studies
4. Discussion and Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Xu, S.; Chen, J.; Wang, X.; Zhuo, X.; Wang, Y.; Xu, J.; Xiang, C.; Lu, M.; Mu, J.; Yang, X. Trends and projections of the burden of visual impairment in Asia: Findings from the Global Burden of Disease Study 2021. Asia-Pac. J. Ophthalmol. 2025, 14, 100196. [Google Scholar] [CrossRef] [PubMed]
- Zhang, Y.; Chen, J.; Liao, L.; Ding, X. Global Trends and Projections for Avoidable Visual Impairment among Working-Age Individuals: A Population-Based Study. Am. J. Ophthalmol. 2025, 271, 304–315. [Google Scholar] [CrossRef] [PubMed]
- Bonsaksen, T.; Brunes, A.; Heir, T. Quality of life in people with visual impairment compared with the general population. J. Public Health 2025, 33, 23–31. [Google Scholar] [CrossRef]
- Chen, Z.; Yang, J.; Chen, L.; Li, F.; Feng, Z.; Jia, L.; Li, P. RailVoxelDet: A Lightweight 3-D Object Detection Method for Railway Transportation Driven by Onboard LiDAR Data. IEEE Internet Things J. 2025, 12, 37175–37189. [Google Scholar] [CrossRef]
- Mai, C.; Xie, D.; Zeng, L.; Li, Z.; Li, Z.; Qiao, Z.; Qu, Y.; Liu, G.; Li, L. Laser Sensing and Vision Sensing Smart Blind Cane: A Review. Sensors 2023, 23, 869. [Google Scholar] [CrossRef] [PubMed]
- Mocanu, A.; Sita, V.; Avram, C.; Astilean, A. Enhanced Cane for Blind People Mobility Assistance. In Proceedings of the International Conference on Automation, Quality and Testing, Robotics (AQTR), Cluj-Napoca, Romania, 16–18 May 2024; IEEE: New York, NY, USA, 2024; pp. 1–6. [Google Scholar]
- Badri, S.; Alzahrani, S.S.; Alshehri, G. GPS-enabled smart stick guide for campus accessibility at King Abdulaziz University. Int. J. Smart Sens. Intell. Syst. 2025, 18, 1–11. [Google Scholar] [CrossRef]
- Lopatin, S.; von Zabiensky, F.; Kreutzer, M.; Rinn, K.; Bienhaus, D. An Electronic Guide Dog for the Blind Based on Artificial Neural Networks. In Proceedings of the 23rd International Conference on Human-Computer Interaction (HCII), Copenhagen, Denmark, 23–28 July 2021; Stephanidis, C., Antona, M., Ntoa, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2021; Volume 1499, pp. 23–30. [Google Scholar]
- Wu, J.; Yang, P.; Chai, Y.; Zhang, K.; Liang, S.; Mo, J.; Liu, C. Smart Guide Trolley Mango: Smart Blind Guide Device Based on Raspberry Pi. In Proceedings of the IEEE International Conference on Electrical Engineering, Big Data and Algorithms (EEBDA), Changchun, China, 25–27 February 2022; IEEE: New York, NY, USA, 2022; pp. 936–941. [Google Scholar]
- Almurayziq, T.S.; Alotibi, N.; Alshammari, G.; Alshammari, A.; Alsaffar, M. Smart and Guide Hat for Blind Persons in Smart Cities Using Deep Learning. J. Adv. Inf. Technol. 2023, 14, 1214–1220. [Google Scholar] [CrossRef]
- Chen, J.-H.; Lin, C.-C. Vision Enhancement System with AR Wearable Aid for Low-Vision People. In Proceedings of the Genetic and Evolutionary Computing, Malaga, Spain, 14–18 July 2025; Pan, J.-S., Zin, T.T., Sung, T.-W., Lin, J.C.-W., Eds.; Springer Nature: Singapore, 2025; pp. 44–47. [Google Scholar]
- Xu, J.; Xu, S.; Ma, M.; Ma, J.; Li, C. Research and Implementation of Travel Aids for Blind and Visually Impaired People. Sensors 2025, 25, 4518. [Google Scholar] [CrossRef] [PubMed]
- Müller, K.; Engel, C.; Loitsch, C.; Stiefelhagen, R.; Weber, G. Traveling More Independently: A Study on the Diverse Needs and Challenges of People with Visual or Mobility Impairments in Unfamiliar Indoor Environments. ACM Trans. Access Comput. 2022, 15, 13:1–13:44. [Google Scholar] [CrossRef]
- Ye, Q.; Xia, Y.; Li, Y.; Dong, J. A method for automatic segmentation of tactile pavement in outdoor environments. In Proceedings of the 2022 IEEE 2nd International Conference on Power, Electronics and Computer Applications (ICPECA), Shenyang, China, 21–23 January 2022; pp. 253–258. [Google Scholar]
- Zhang, X.; Liang, L.; Zhao, S.; Wang, Z. GRFB-UNet: A new multi-scale attention network with group receptive field block for tactile paving segmentation. Expert Syst. Appl. 2024, 238, 122109. [Google Scholar] [CrossRef]
- Wan, Z.; Zhang, P.; Wang, Y.; Yong, S.; Stepputtis, S.; Sycara, K.; Xie, Y. Sigma: Siamese Mamba Network for Multi-Modal Semantic Segmentation. In Proceedings of the 2025 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Tucson, AZ, USA, 26 February–6 March 2025; pp. 1734–1744. [Google Scholar]
- Li, M.; Lang, X.; Gong, R.; Zhou, J.; Yang, X.; Sang, N. TPSegmentDiff: An Enhanced Diffusion Model for Tactile Paving Image Segmentation. In Proceedings of the 6th ACM International Conference on Multimedia in Asia Workshops, New York, NY, USA, 3–6 December 2024. [Google Scholar]
- Tokita, R.; Ikeda, Y.; Go, K.; Nakahara, E.; Miyata, A. Segmentation-Based Approach Towards Detecting Improperly Installed Tactile Paving. In Proceedings of the 27th International ACM SIGACCESS Conference on Computers and Accessibility, New York, NY, USA, 26–29 October 2025; Association for Computing Machinery: New York, NY, USA, 2025; pp. 1–5. [Google Scholar]
- Chen, F.; Li, S.; Han, J.; Ren, F.; Yang, Z. Review of Lightweight Deep Convolutional Neural Networks. Arch. Comput. Methods Eng. 2024, 31, 1915–1937. [Google Scholar] [CrossRef]
- Mittal, P. A comprehensive survey of deep learning-based lightweight object detection models for edge devices. Artif. Intell. Rev. 2024, 57, 242. [Google Scholar] [CrossRef]
- Liu, H.-I.; Galindo, M.; Xie, H.; Wong, L.-K.; Shuai, H.-H.; Li, Y.-H.; Cheng, W.-H. Lightweight Deep Learning for Resource-Constrained Environments: A Survey. ACM Comput. Surv. 2024, 56, 267. [Google Scholar] [CrossRef]
- Yang, Y.; Wang, H.; Kang, J.; Xu, Z. A Method for Surveying Road Pavement Distress Based on Front-View Image Data Using a Lightweight Segmentation Approach. J. Comput. Civ. Eng. 2024, 38, 04024026. [Google Scholar] [CrossRef]
- Niu, L.; Bao, H. Fast Tactile Paving Segmentation Model Based on Reparameterized Structure. In Proceedings of the 2024 International Conference on Generative Artificial Intelligence and Information Security, New York, NY, USA, 10–12 May 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 24–28. [Google Scholar]
- Shi, M.; Lin, S.; Yi, Q.; Weng, J.; Luo, A.; Zhou, Y. Lightweight Context-Aware Network Using Partial-Channel Transformation for Real-Time Semantic Segmentation. IEEE Trans. Intell. Transp. Syst. 2024, 25, 7401–7416. [Google Scholar] [CrossRef]
- Zhang, J.; Ding, L.; Wang, W.; Wang, H.; Brilakis, I.; Davletshina, D.; Heikkila, R.; Yang, X. Crack segmentation-guided measurement with lightweight distillation network on edge device. Comput.-Aided Civ. Infrastruct. Eng. 2025, 40, 2269–2286. [Google Scholar] [CrossRef]
- Xue, C.; Xu, J.; Bao, Y.; Wu, Y. Sidewalk Segment Dataset. 2023. Available online: https://www.kaggle.com/dsv/5952661 (accessed on 28 September 2025).
- Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep High-Resolution Representation Learning for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3349–3364. [Google Scholar] [CrossRef] [PubMed]
- Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6230–6239. [Google Scholar]
- Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the 18th International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer: Berlin/Heidelberg, Germany, 2015; Volume 9351, pp. 234–241. [Google Scholar]











| Training Set | Validation Set | Test Set | For Deployment | |
|---|---|---|---|---|
| Our dataset | 307 | 50 | 50 | Video Data (4 sequences) |
| TP-Dataset | 800 | 200 | 200 | -- |
| Sidewalk Segment Dataset | 893 | 350 | 350 | -- |
| Total | 2000 | 600 | 600 | Video Data (4 sequences) |
| Parameters | Initial Learning Rate | Batch Size | Epoch | Optimizer | Lr Decay |
| Value | 0.007 | 26 | 500 | SGD | cos |
| Methods | PSPNet | UNet | HRNet | Ours |
|---|---|---|---|---|
| Accuracy | 95.02% | 97.15% | 97.58% | 98.60% |
| mPA | 89.39% | 95.85% | 96.69% | 96.44% |
| mIoU | 78.52% | 93.67% | 92.82% | 92.90% |
| PA Tactile paving | 92.92% | 95.16% | 93.77% | 97.05% |
| PA Zebra crossing | 78.97% | 91.93% | 89.47% | 93.14% |
| IoU Tactile paving | 87.53% | 93.16% | 92.31% | 94.34% |
| IoU Zebra crossing | 53.80% | 85.17% | 85.82% | 86.01% |
| Params size | 5.34 MB | 24.89 MB | 29.55 MB | 19.51 MB |
| FPS | 75.9 | 52.1 | 41.8 | 59.2 |
| Methods | PSPNet | UNet | HRNet | Ours |
|---|---|---|---|---|
| Accuracy | 92.92% (↓ 2.21) | 96.19% (↓ 0.96) | 97.47% (↓ 0.11) | 97.15% (↓ 1.45) |
| mPA | 89.80% (↑ 0.41) | 92.87% (↓ 2.98) | 94.88% (↓ 1.81) | 95.01% (↓ 1.43) |
| mIoU | 76.43% (↓ 0.09) | 91.06% (↓ 2.61) | 89.51% (↓ 3.31) | 93.24% (↑ 0.34) |
| Latency | 14.3 ms | 21.2 ms | 30.6 ms | 17.3 ms |
| Improved G-GhostNet | Improved Coord ATT | Redesigned ASPP | 3D Weights ATT | mIoU | PA Tactile Paving | PA Zebra Crossing |
|---|---|---|---|---|---|---|
| ✓ | -- | ✓ | -- | 85.21% | 86.67% | 82.69% |
| ✓ | -- | ✓ | ✓ | 86.14% | 88.86% | 86.43% |
| ✓ | ✓ | -- | -- | 52.06% | 53.73% | 48.92% |
| ✓ | ✓ | -- | ✓ | 53.69% | 53.81% | 51.07% |
| ✓ | ✓ | ✓ | -- | 91.09% | 92.17% | 90.36% |
| ✓ | ✓ | ✓ | ✓ | 92.90% | 97.05% | 93.14% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Jiang, Y.; Yan, S.; Liu, J. Real-Time Segmentation of Tactile Paving and Zebra Crossings for Visually Impaired Assistance Using Embedded Visual Sensors. Sensors 2026, 26, 770. https://doi.org/10.3390/s26030770
Jiang Y, Yan S, Liu J. Real-Time Segmentation of Tactile Paving and Zebra Crossings for Visually Impaired Assistance Using Embedded Visual Sensors. Sensors. 2026; 26(3):770. https://doi.org/10.3390/s26030770
Chicago/Turabian StyleJiang, Yiqiang, Shicheng Yan, and Jianyang Liu. 2026. "Real-Time Segmentation of Tactile Paving and Zebra Crossings for Visually Impaired Assistance Using Embedded Visual Sensors" Sensors 26, no. 3: 770. https://doi.org/10.3390/s26030770
APA StyleJiang, Y., Yan, S., & Liu, J. (2026). Real-Time Segmentation of Tactile Paving and Zebra Crossings for Visually Impaired Assistance Using Embedded Visual Sensors. Sensors, 26(3), 770. https://doi.org/10.3390/s26030770
