Smart Tourism Landmark Recognition: A Multi-Threshold Enhancement and Selective Ensemble Approach Using YOLO11
Abstract
1. Introduction
- RQ1: To what extent can a multi-threshold visual preprocessing strategy combined with selective ensemble methods improve the accuracy and reliability of landmark recognition across different environmental contexts and architectural forms?
- RQ2: Compared to individual baseline architectures and traditional fixed ensemble approaches, what is the practical effectiveness of the proposed YOLO11-based selective ensemble framework in achieving superior classification performance while maintaining computational efficiency for real-world deployment?
- To develop a comprehensive multi-threshold pixel enhancement strategy that systematically emphasizes different architectural feature categories across varying intensity ranges (100, 150, 225), improving model sensitivity to diverse visual patterns in historical landmarks.
- To implement an intelligent selective ensemble mechanism that automatically evaluates all possible model combinations and determines optimal subsets based on empirical performance validation, avoiding the limitations of fixed ensemble approaches.
- To validate the proposed approach through extensive experimentation on the enhanced Samarkand_v2 dataset, demonstrating superior performance compared to established baseline architectures and individual model implementations.
- To establish the practical applicability of the framework for smart tourism deployment through comprehensive evaluation of classification accuracy, environmental robustness, and computational efficiency.
- Formulation of a novel multi-threshold pixel-level optimization method specifically designed for architectural landmark recognition, focusing on multiple intensity bands to derive complementary image features.
- Designing a selective ensemble-based architecture which intelligently integrates YOLO11 models via data-driven optimization, filtering out non-effective variants while maintaining computational efficiency.
- Comprehensive experimental validation demonstrating remarkable performance improvements—99.24% accuracy, 99.36% precision, 99.40% recall, and 99.36% F1-score—significantly outperforming baseline architectures and traditional ensemble methods.
- Statistical validation through paired t-tests confirming the significance of performance improvements and providing insights into the effectiveness of different enhancement strategies.
- Practical advancement in smart tourism technology supporting both automated landmark recognition and cultural heritage preservation through intelligent automation.
2. Related Works
2.1. Smart Tourism: Integrating AI, IoT, and Big Data
2.2. Deep Learning Methods and Image Enhancement Strategies
2.3. Advanced Ensemble Approaches in Tourism AI
3. Proposed Methodology
Algorithm 1 Smart Tourism Landmark Recognition: Multi-Threshold Image Refinement and Adaptive Ensemble with YOLO11 |
Require: Samarkand dataset , where are input images and are the corresponding labels. |
Ensure: Final trained ensemble model E for classification of historical landmarks. |
1: Preprocessing: |
2: Resize every image to pixels. |
3: Normalize images as: |
4: Apply pixel-level refinement: if intensity , then update . |
5: Training YOLO11 with Refined Data (Threshold 225): |
6: Initialize model . |
7: for to N do |
8: Train on dataset . |
9: Validate performance on . |
10: if current accuracy is the best so far then |
11: Save . |
12: Apply refinement: if , set . |
13: Training YOLO11 with Refined Data (Threshold 150): |
14: Initialize model . |
15: for to N do |
16: Train on dataset . |
17: Validate on . |
18: if current accuracy is the best so far then |
19: Save . |
20: Apply refinement: if , set . |
21: Training YOLO11 with Refined Data (Threshold 100): |
22: Initialize model . |
23: for to N do |
24: Train on dataset . |
25: Validate on . |
26: if current accuracy is the best so far then |
27: Save . |
28: Training YOLO11 with Original Images: |
29: Initialize model . |
30: for to N do |
31: Train on . |
32: Validate on . |
33: if current accuracy is the best so far then |
34: Save . |
35: Logit Extraction: |
36: for each x in validation set do |
37: Collect outputs: |
38: Performance-Guided Ensemble Selection: |
39: Measure accuracy of each individual model: . |
40: Evaluate ensemble combinations. |
41: for each subset where do |
42: for each image x in do |
43: Calculate ensemble output: |
44: Compute validation accuracy . |
45: Select the best-performing subset . |
Algorithm 2 Smart Tourism Landmark Recognition: Multi-Threshold Image Refinement and Adaptive Ensemble with YOLO11 (continued) |
1: Final Ensemble Prediction with Chosen Models |
2: for each image x in validation and test datasets do |
3: Derive the aggregated ensemble output using only the selected models: |
4: Assign predicted label as: |
5: Assessment of the Ensemble: |
6: Measure the performance of the final ensemble E on test set using: |
7: return Overall evaluation metrics of ensemble E. |
- Dataset Collection and Preparation: The foundational dataset employed in this research was systematically assembled through comprehensive documentation of historical landmarks within Samarkand, resulting in the enhanced Samarkand_v2 dataset. This extended collection represents a significant expansion of the original Samarkand dataset, incorporating additional images captured from diverse perspectives and under varying illumination conditions to establish a more representative and robust training foundation. The Samarkand_v2 dataset encompasses a broader range of environmental scenarios including different seasons, weather conditions, and time-of-day variations to enhance model generalization capabilities. Manual annotation was performed by domain experts to assign appropriate landmark categories with high precision, followed by systematic partitioning into training (80%), validation (10%), and testing (10%) subsets using stratified sampling to maintain class distribution balance and facilitate rigorous model assessment across all landmark categories.
- Multi-Level Image Enhancement Strategy: To optimize feature extraction capabilities and improve model sensitivity to architectural details, we implemented a graduated pixel enhancement strategy targeting three distinct intensity thresholds based on empirical analysis of architectural feature distributions. The enhancement transformation is mathematically defined as
- Parallel Architecture Training Framework: Four independent YOLO11n-cls models are trained concurrently on distinct image variants derived from the Samarkand_v2 dataset: (threshold 225 enhanced focusing on highly reflective surfaces), (threshold 150 enhanced targeting moderate brightness features), (threshold 100 enhanced capturing subtle intensity variations), and (original unmodified images preserving natural feature distributions). Each model employs identical architectural configurations including backbone network parameters, classification head design, and optimization settings to ensure comparative validity and eliminate architectural bias in performance evaluation. The training process incorporates advanced techniques including data augmentation (random horizontal flips, slight rotations), learning rate scheduling with cosine annealing, and early stopping mechanisms based on validation performance plateaus. Optimal epoch selection is performed through continuous monitoring of validation accuracy, with model checkpoints saved only when performance improvements are observed.
- Logit Extraction and Feature Representation: Following individual model training completion, raw classification logits are systematically extracted from the Samarkand_v2 validation dataset to capture the learned feature representations before softmax normalization. These logits represent pre-activation confidence scores that preserve the relative magnitude of class predictions, serving as the foundation for ensemble decision integration without information loss from probability normalization:
- Performance-Driven Ensemble Selection Mechanism: Rather than employing fixed ensemble combinations that assume equal contribution from all models, our methodology incorporates intelligent subset selection based on empirical performance validation using the extended Samarkand_v2 dataset. This approach recognizes that different enhancement strategies may not always contribute positively to overall performance, particularly when dealing with specific architectural styles or lighting conditions. For each possible model combination where , ensemble predictions are computed using arithmetic logit averaging:The exhaustive evaluation process tests all possible non-empty subsets, computing validation accuracy for each combination to identify the optimal configuration. This data-driven selection mechanism automatically adapts to the specific characteristics of the dataset and landmark types, ensuring maximum performance while avoiding negative contributions from suboptimal model combinations. The optimal configuration is determined through performance maximization:
- Final Prediction Generation and Decision-Making: The ultimate classification decision utilizes only the models identified in the optimal subset , eliminating potential negative contributions from underperforming enhancement strategies. Final predictions are generated through logit-level fusion of selected models:
- Comprehensive Evaluation Framework and Statistical Validation: Model performance assessment employs standard classification metrics to provide thorough evaluation of the proposed methodology on the Samarkand_v2 dataset, supplemented by statistical significance testing to validate performance improvements. The evaluation framework includes both individual model assessment and ensemble performance analysis:
4. Experiments
4.1. Dataset
4.2. Baseline Models
- MobileNetV3 [36]—Introduced by researchers at Google, the MobileNetV3 model represents a leading approach in mobile-optimized neural network design, specifically engineered to achieve an optimal trade-off of processing delay alongside classification performance. The network integrates advanced efficiency-oriented core modules, including lightweight depthwise separable convolutional layers, squeeze-and-excitation units, and configurations refined through neural architecture search. Further enhanced by the novel h-swish activation function and platform-aware optimization strategies, MobileNetV3 demonstrates strong effectiveness in resource-limited scenarios. This model functions as an essential baseline for examining our method’s suitability in real-time smart tourism systems where computational efficiency is critical.
- EfficientNetB0 [37]—As the foundational member of the EfficientNet family, EfficientNetB0 introduces the revolutionary compound scaling methodology that uniformly scales network depth, width, and input resolution using a carefully calibrated compound coefficient. This approach systematically addresses the scaling of convolutional neural networks to achieve superior efficiency and accuracy trade-offs compared to traditional scaling methods. EfficientNetB0’s architecture demonstrates remarkable capability in handling diverse and complex visual datasets, making it an ideal benchmark for evaluating our multi-threshold enhancement approach’s effectiveness on architecturally complex landmark imagery.
- ResNet50 [35]—A cornerstone architecture in the ResNet50, belonging to the broader class of Residual Networks incorporates the groundbreaking “skip connections” or residual connections which revolutionized deep network training by effectively addressing the vanishing gradient problem. This architectural innovation enables the successful training of significantly deeper networks while maintaining gradient flow throughout the network hierarchy. ResNet50’s proven track record in large-scale image classification tasks and its robust feature extraction capabilities make it an essential baseline for assessing our selective ensemble methodology’s performance against well-established deep learning paradigms.
- YOLO11N [38]—Representing the latest advancement in the You Only Look Once (YOLO) family of architectures, YOLO11N is specifically optimized for real-time object detection and classification tasks with particular emphasis on balancing computational speed and prediction accuracy. The architecture incorporates state-of-the-art feature extraction mechanisms, advanced anchor-free detection strategies, and optimized inference pipelines that enable deployment in time-critical applications. YOLO11N’s capability for real-time visual processing makes it particularly relevant for tourist destination recognition systems where immediate response times are crucial for enhancing user experience in smart tourism environments.
4.3. Training Setup
4.4. Experimental Results and Discussion
4.5. Limitations and Future Work
- Geographic Scope: This research concentrates on cultural heritage landmarks in the city of Samarkand, and although the dataset provides extensive coverage of diverse architectural styles, validation across other geographic regions would enhance the generalizability of claims for global smart tourism deployment.
- Enhancement Threshold Optimization: The current threshold values (100, 150, 225) were selected based on empirical analysis of the Samarkand dataset, and these fixed thresholds may not be optimal for different architectural styles or lighting conditions in other geographic regions, suggesting the need for adaptive threshold selection mechanisms.
- Limited Architectural Diversity: The focus on Islamic architectural heritage in Samarkand, while comprehensive within this domain, does not encompass other major architectural traditions such as Gothic, Baroque, or contemporary architectural styles that are prevalent in global tourism destinations.
5. Conclusions
- Digital Heritage Preservation: By systematically documenting and classifying Samarkand’s architectural landmarks with near-perfect accuracy, our system creates a comprehensive digital archive that preserves cultural heritage for future generations while reducing physical interaction with fragile historical structures.
- Resource-Efficient Tourism Management: The selective ensemble mechanism reduces computational requirements by 25% compared to traditional approaches, enabling deployment on standard mobile devices without requiring extensive infrastructure particularly important for developing tourism destinations with limited technical resources.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Guerrero-Rodríguez, R.; Álvarez-Carmona, M.Á.; Aranda, R.; Díaz-Pacheco, Á. Big data analytics of online news to explore destination image using a comprehensive deep-learning approach: A case from Mexico. Inf. Technol. Tour. 2024, 26, 147–182. [Google Scholar] [CrossRef]
- Pencarelli, T. The digital revolution in the travel and tourism industry. Inf. Technol. Tour. 2020, 22, 455–476. [Google Scholar] [CrossRef]
- Jabbari, M.; Amini, M.; Malekinezhad, H.; Berahmand, Z. Improving augmented reality with the help of deep learning methods in the tourism industry. Math. Comput. Sci. 2023, 4, 33–45. [Google Scholar]
- Melo, M.; Coelho, H.; Gonçalves, G.; Losada, N.; Jorge, F.; Teixeira, M.S.; Bessa, M. Immersive multisensory virtual reality technologies for virtual tourism. Multimed. Syst. 2022, 28, 1027–1037. [Google Scholar] [CrossRef]
- Bhosale, T.A.; Pushkar, S. IWF-ECTIC: Improved Wiener filtering and ensemble of classification model for tourism image classification. Multimed. Tools Appl. 2024, 84, 32027–32064. [Google Scholar] [CrossRef]
- Doğan, S.; Niyet, I.Z. Artificial intelligence (AI) in tourism. In Future Tourism Trends Volume 2: Technology Advancement, Trends and Innovations for the Future in Tourism; Emerald Publishing Limited: Bingley, UK, 2024; pp. 3–21. [Google Scholar]
- Patrichi, I.C. AI solutions for sustainable tourism management: A comprehensive review. J. Inf. Syst. Oper. Manag. 2024, 18, 172–185. [Google Scholar]
- Asrifan, A.; Murni, M.; Hermansyah, S.; Dewi, A.C. Eco-Smart Cities: Sustainable Tourism Practices Enabled by Smart Technologies. In Modern Management Science Practices in the Age of AI; IGI Global: Hershey, PA, USA, 2024; pp. 267–296. [Google Scholar]
- Suanpang, P.; Pothipassa, P. Integrating generative AI and IoT for sustainable smart tourism destinations. Sustainability 2024, 16, 7435. [Google Scholar] [CrossRef]
- Li, X.; Zhang, X.; Zhang, C.; Wang, S. Forecasting tourism demand with a novel robust decomposition and ensemble framework. Expert Syst. Appl. 2024, 236, 121388. [Google Scholar] [CrossRef]
- Viyanon, W. An interactive multiplayer mobile application using feature detection and matching for tourism promotion. In Proceedings of the 2nd International Conference on Control and Computer Vision, Jeju, Republic of Korea, 15–18 June 2019; pp. 82–86. [Google Scholar]
- Alaei, A.; Wang, Y.; Bui, V.; Stantic, B. Target-oriented data annotation for emotion and sentiment analysis in tourism related social media data. Future Internet 2023, 15, 150. [Google Scholar] [CrossRef]
- Savvopoulos, A.; Kalogeras, G.; Mariotti, A.; Bernini, C.; Alexakos, C.; Kalogeras, A. Data analytics supporting decision making in the tourism sector. In Proceedings of the 2019 First International Conference on Societal Automation (SA), Kraków, Poland, 4–6 September 2019; pp. 1–6. [Google Scholar]
- Carneiro, A.; Nascimento, L.S.; Noernberg, M.A.; Hara, C.S.; Pozo, A.T.R. Social media image classification for jellyfish monitoring. Aquat. Ecol. 2024, 58, 3–15. [Google Scholar]
- Bui, V.; Alaei, A. Virtual reality in training artificial intelligence-based systems: A case study of fall detection. Multimed. Tools Appl. 2022, 81, 32625–32642. [Google Scholar] [CrossRef]
- Yao, J.; Chu, Y.; Xiang, X.; Huang, B.; Xiaoli, W. Research on detection and classification of traffic signs with data augmentation. Multimed. Tools Appl. 2023, 82, 38875–38899. [Google Scholar]
- Ma, H. Development of a smart tourism service system based on the Internet of Things and machine learning. J. Supercomput. 2024, 80, 6725–6745. [Google Scholar] [CrossRef]
- Hao, Y.; Zheng, L. Application of SLAM method in big data rural tourism management in dynamic scenes. Soft Comput. 2023. [Google Scholar] [CrossRef]
- Gretzel, U.; Sigala, M.; Xiang, Z.; Koo, C. Smart tourism: Foundations and developments. Electron. Mark. 2015, 25, 179–188. [Google Scholar] [CrossRef]
- Buhalis, D. Technology in tourism-from ICTs to eTourism and smart tourism towards ambient intelligence tourism: A perspective article. Tour. Rev. 2020, 75, 267–272. [Google Scholar]
- Jovicic, D.Z. From the traditional understanding of tourism destination to the smart tourism destination. Curr. Issues Tour. 2019, 22, 276–282. [Google Scholar]
- Li, Y.; Hu, C.; Huang, C.; Duan, L. The concept of smart tourism in the context of tourism information services. Tour. Manag. 2017, 58, 293–300. [Google Scholar] [CrossRef]
- Chen, Y.-C.; Yu, K.-M.; Kao, T.-H.; Hsieh, H.-L. Deep learning based real-time tourist spots detection and recognition mechanism. Sci. Prog. 2021, 104, 00368504211044228. [Google Scholar] [CrossRef]
- Chauhan, J.S.; Sharma, A.; Sharma, A.; Ahalawat, K.; Agarwal, P. AI-Driven decision making in wellness tourism: Understanding consumer preferences and optimizing tourism strategies. In Proceedings of the 2025 First International Conference on Advances in Computer Science, Electrical, Electronics, and Communication Technologies (CE2CT), Greater Noida, India, 21–22 February 2025; pp. 663–669. [Google Scholar]
- Zekos, G.I. AI risk management. In Economics and Law of Artificial Intelligence: Finance, Economic Impacts, Risk Management and Governance; Springer: Cham, Switzerland, 2021; pp. 233–288. [Google Scholar]
- Bi, J.-W.; Han, T.-Y.; Yao, Y. Collaborative forecasting of tourism demand for multiple tourist attractions with spatial dependence: A combined deep learning model. Tour. Econ. 2024, 30, 361–388. [Google Scholar]
- Yuan, X. Evaluation of rural tourism development level using BERT-enhanced deep learning model and BP algorithm. Sci. Rep. 2024, 14, 25748. [Google Scholar] [CrossRef]
- Sihombing, I.H.H.; Suastini, N.M.; Puja, I.B.P. Sustainable cultural tourism in the era of sustainable development. Int. J. Sustain. Compet. Tour. 2024, 3, 100–115. [Google Scholar] [CrossRef]
- Mishra, S.; Anifa, M.; Naidu, J. AI-driven personalization in tourism services: Enhancing tourist experiences and business performance. In Integrating Architecture and Design Into Sustainable Tourism Development; IGI Global: Hershey, PA, USA, 2025; pp. 1–20. [Google Scholar]
- Xiang, Z.; Du, Q.; Ma, Y.; Fan, W. A comparative analysis of major online review platforms: Implications for social media analytics in hospitality and tourism. Tour. Manag. 2017, 58, 51–65. [Google Scholar] [CrossRef]
- Chiengkul, W.; Kumjorn, P.; Tantipanichkul, T.; Suphan, K. Engaging with AI in tourism: A key to enhancing smart experiences and emotional bonds. Asia-Pac. J. Bus. Adm. 2025. [Google Scholar] [CrossRef]
- Irshad, A.; Singh, A.K. AI and conflict resolution in tourism destinations. In Impact of AI and Tech-Driven Solutions in Hospitality and Tourism; IGI Global: Hershey, PA, USA, 2024; pp. 429–439. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2012, 60, 84–90. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, VA, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Howard, A.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for MobileNetV3. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
- Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. Int. Conf. Mach. Learn. 2019, 97, 6105–6114. [Google Scholar]
- Jocher, G.; Qiu, J. Ultralytics YOLO11 (Version 11.0.0). Ultralytics. 2024. Available online: https://github.com/ultralytics/ultralytics (accessed on 1 July 2025).
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 29th IEEE Computer Vision and Pattern Recognition Conference, Las Vegas, VA, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
- Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 7464–7475. [Google Scholar]
- Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vision, Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
- Singh, H.; Verma, M.; Cheruku, R. DMFNet: Geometric multi-scale pixel-level contrastive learning for video salient object detection. Int. J. Multimed. Inf. Retr. 2025, 14, 12. [Google Scholar] [CrossRef]
- Remondino, F.; El-Hakim, S. Image-based 3D modelling: A review. Photogramm. Rec. 2006, 21, 269–291. [Google Scholar] [CrossRef]
- Llamas, J.; Lerones, P.M.; Zalama, E.; Gómez-García-Bermejo, J. Applying deep learning techniques to cultural heritage images within the INCEPTION project. In Proceedings of the Euro-Mediterranean Conference, Nicosia, Cyprus, 31 October–5 November 2016; pp. 25–32. [Google Scholar]
- Huang, G.; Li, Y.; Pleiss, G.; Liu, Z.; Hopcroft, J.E.; Weinberger, K.Q. Snapshot Ensembles: Train 1, get M for free. arXiv 2017, arXiv:1704.00109. [Google Scholar] [CrossRef]
- Ganaie, M.A.; Hu, M.; Malik, A.K.; Tanveer, M.; Suganthan, P.N. Ensemble deep learning: A review. Eng. Appl. Artif. Intell. 2022, 115, 105151. [Google Scholar] [CrossRef]
- Wu, D.C.; Zhong, S.; Wu, J.; Song, H. Tourism and hospitality forecasting with big data: A systematic review of the literature. J. Hosp. Tour. Res. 2025, 49, 615–634. [Google Scholar] [CrossRef]
- Anorboev, A.; Anorboeva, S.; Musaev, J.; Usmanov, E.; Hwang, D.; Seo, Y.S.; Hong, J. Selective Intensity Ensemble Classifier (SIEC): A Triple-Threshold Strategy for Microscopic Malaria Cell Image Classification. IEEE Access 2025, 13, 101609–101623. [Google Scholar] [CrossRef]
- Hudayberdiev, U.; Lee, J. An Ensemble Deep Learning Framework for Smart Tourism Landmark Recognition Using Pixel-Enhanced YOLO11 Models. Sustainability 2025, 17, 5420. [Google Scholar] [CrossRef]
- Anorboev, A.; Musaev, J.; Hong, J.; Nguyen, N.T.; Hwang, D. An image pixel interval power (IPIP) method using deep learning classification models. In Intelligent Information and Database Systems, Proceedings of the Asian Conference on Intelligent Information and Database Systems, Ho Chi Minh City, Vietnam, 28–30 November 2022; Springer: Cham, Switzerland, 2022; pp. 196–208. [Google Scholar]
- Musaev, J.; Anorboev, A.; Anorboeva, S.; Seo, Y.S.; Nguyen, N.T.; Hwang, D. Hybrid Convolutional Network Fusion: Enhanced Medical Image Classification with Dual-Pathway Learning from Raw and Enhanced Visual Features. In Computational Collective Intelligence, Proceedings of the International Conference on Computational Collective Intelligence, Leipzig, Germany, 9–11 September 2024; Springer: Cham, Switzerland, 2024; pp. 120–132. [Google Scholar]
Model | Accuracy | Precision | Recall | F1-Score |
---|---|---|---|---|
MobileNet_V3 | 0.9021 | 0.9067 | 0.9005 | 0.9011 |
ResNet50 | 0.8836 | 0.8902 | 0.8810 | 0.8823 |
EfficientNet_B0 | 0.9273 | 0.9310 | 0.9264 | 0.9270 |
YOLO11n-cls (MO) | 0.9875 | 0.9899 | 0.9865 | 0.9870 |
YOLO11n-cls (ME100) | 0.9897 | 0.9912 | 0.9891 | 0.9895 |
YOLO11n-cls (ME150) | 0.9883 | 0.9905 | 0.9878 | 0.9884 |
YOLO11n-cls (ME225) | 0.9871 | 0.9898 | 0.9867 | 0.9873 |
Proposed Selective Ensemble | 0.9924 | 0.9936 | 0.9940 | 0.9936 |
Model Pair | T-Statistic | P-Value | Cohen’s d | Significant |
---|---|---|---|---|
MO vs. ME100 | 2.2713 | 0.0248 | 0.1984 | Yes * |
MO vs. ME150 | 2.0235 | 0.0451 | 0.1768 | Yes * |
MO vs. M225 | 1.0000 | 0.3192 | 0.0874 | No |
ME100 vs. ME150 | −0.4458 | 0.6565 | −0.0390 | No |
ME100 vs. M225 | −2.0235 | 0.0451 | −0.1768 | Yes * |
ME150 vs. M225 | −1.3458 | 0.1807 | −0.1176 | No |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Hudayberdiev, U.; Lee, J.; Fayzullaev, O. Smart Tourism Landmark Recognition: A Multi-Threshold Enhancement and Selective Ensemble Approach Using YOLO11. Sustainability 2025, 17, 8081. https://doi.org/10.3390/su17178081
Hudayberdiev U, Lee J, Fayzullaev O. Smart Tourism Landmark Recognition: A Multi-Threshold Enhancement and Selective Ensemble Approach Using YOLO11. Sustainability. 2025; 17(17):8081. https://doi.org/10.3390/su17178081
Chicago/Turabian StyleHudayberdiev, Ulugbek, Junyeong Lee, and Odil Fayzullaev. 2025. "Smart Tourism Landmark Recognition: A Multi-Threshold Enhancement and Selective Ensemble Approach Using YOLO11" Sustainability 17, no. 17: 8081. https://doi.org/10.3390/su17178081
APA StyleHudayberdiev, U., Lee, J., & Fayzullaev, O. (2025). Smart Tourism Landmark Recognition: A Multi-Threshold Enhancement and Selective Ensemble Approach Using YOLO11. Sustainability, 17(17), 8081. https://doi.org/10.3390/su17178081