Next Article in Journal
Adaptive Non-Singular Fast Terminal Sliding Mode Trajectory Tracking Control for Robotic Manipulator with Novel Configuration Based on TD3 Deep Reinforcement Learning and Nonlinear Disturbance Observer
Previous Article in Journal
Magnetic Circuit Design and Optimization of Tension–Compression Giant Magnetostrictive Force Sensor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Review

Knowledge Distillation in Object Detection: A Survey from CNN to Transformer

by
Tahira Shehzadi
1,2,3,*,
Rabya Noor
1,3,
Ifza Ifza
1,3,
Marcus Liwicki
4,
Didier Stricker
1,2,3 and
Muhammad Zeshan Afzal
1,2,3
1
Department of Computer Science, RPTU Kaiserslautern-Landau, 67663 Kaiserslautern, Germany
2
Mindgarage Lab, RPTU Kaiserslautern-Landau, 67663 Kaiserslautern, Germany
3
German Research Institute for Artificial Intelligence (DFKI), 67663 Kaiserslautern, Germany
4
Department of Computer Science, Electrical and Space Engineering Luleå University of Technology, 971 87 Luleå, Sweden
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(1), 292; https://doi.org/10.3390/s26010292
Submission received: 16 August 2025 / Revised: 16 November 2025 / Accepted: 27 November 2025 / Published: 2 January 2026
(This article belongs to the Section Sensing and Imaging)

Abstract

Deep learning models, especially for object detection have gained immense popularity in
computer vision. These models have demonstrated remarkable accuracy and performance,
driving advancements across various applications. However, the high computational
complexity and large storage requirements of state-of-the-art object detection models pose
significant challenges for deployment on resource-constrained devices like mobile phones
and embedded systems. Knowledge Distillation (KD) has emerged as a prominent solution
to these challenges, effectively compressing large, complex teacher models into smaller,
efficient student models. This technique maintains good accuracy while significantly
reducing model size and computational demands, making object detection models more
practical for real-world applications. This survey provides a comprehensive review of KDbased
object detection models developed in recent years. It offers an in-depth analysis of
existing techniques, highlighting their novelty and limitations, and explores future research
directions. The survey covers the different distillation algorithms used in object detection.
It also examines extended applications of knowledge distillation in object detection, such
as improvements for lightweight models, addressing catastrophic forgetting in incremental
learning, and enhancing small object detection. Furthermore, the survey also delves into
the application of knowledge distillation in other domains such as image classification,
semantic segmentation, 3D reconstruction, and document analysis.

Share and Cite

MDPI and ACS Style

Shehzadi, T.; Noor, R.; Ifza, I.; Liwicki, M.; Stricker, D.; Afzal, M.Z. Knowledge Distillation in Object Detection: A Survey from CNN to Transformer. Sensors 2026, 26, 292. https://doi.org/10.3390/s26010292

AMA Style

Shehzadi T, Noor R, Ifza I, Liwicki M, Stricker D, Afzal MZ. Knowledge Distillation in Object Detection: A Survey from CNN to Transformer. Sensors. 2026; 26(1):292. https://doi.org/10.3390/s26010292

Chicago/Turabian Style

Shehzadi, Tahira, Rabya Noor, Ifza Ifza, Marcus Liwicki, Didier Stricker, and Muhammad Zeshan Afzal. 2026. "Knowledge Distillation in Object Detection: A Survey from CNN to Transformer" Sensors 26, no. 1: 292. https://doi.org/10.3390/s26010292

APA Style

Shehzadi, T., Noor, R., Ifza, I., Liwicki, M., Stricker, D., & Afzal, M. Z. (2026). Knowledge Distillation in Object Detection: A Survey from CNN to Transformer. Sensors, 26(1), 292. https://doi.org/10.3390/s26010292

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop