Abstract
Bladder cancer represents a significant health problem worldwide, with it being a major cause of death and characterized by frequent recurrences. Effective treatment hinges on early and accurate diagnosis; however, traditional methods are invasive, time-consuming, and subjective. In this research, we propose a transparent deep learning model based on the YOLOv11 structure to not only enhance lesion detection but also give the visual support of the model’s predictions. Five versions of YOLOv11—nano, small, medium, large, and extra large—were trained and tested by us on a comprehensive dataset of hematoxylin and eosin-stained histopathology slides with the inflammation, urothelial cell carcinoma (UCC), and invalid tissue categories. The YOLOv11-large variant turned out to be the best-performing model at the forefront of technology, with an accuracy of 97.09%, precision and recall of 95.47% each, and balanced accuracy of 96.60%. Besides the precision–recall curves (AUPRC: inflammation = 0.935, invalid = 0.852, UCC = 0.958), ROC-AUC curves (overall AUC = 0.972) and risk–coverage analysis (AUC = 0.994) were also used for detailed assessment of the model to confirm its steadiness and trustworthiness. The confusion matrix displayed the highest true positive rates in all classes and a few misclassifications, which mainly happened between inflammation and invalid samples, indicating a possible morphological overlap. Moreover, as supported by a low Expected Calibration Error (ECE), the model was in great calibration. YOLOv11 reaches higher performance while still being computationally efficient by incorporating advanced architectural features like the C3k2 block and C2PSA spatial attention module. This is a step towards the realization of the AI-assisted bladder cancer diagnostic system that is not only reliable and transparent but also scalable, presented in this work.