Satellite Optical Target Edge Detection Based on Knowledge Distillation
Abstract
1. Introduction
- This paper proposes a novel knowledge distillation-based edge detection method for space targets. By designing a distillation strategy, the student model with fewer parameters and simpler structure can effectively learn the key features and decision boundaries from the teacher model.
- A shape-prior-guided module is embedded in the student branch. By constructing a shape-prior model, computing similarity, and reconstructing features, this module effectively incorporates the geometric shape information of spatial targets into the edge-detection pipeline, thereby enhancing the model’s adaptability to spatial targets and improving the accuracy of edge detection.
- This paper designs a curvature-guided loss function. By leveraging the characteristic of edge curvature changes, the model is guided to better restore the continuity of discontinuous edges, thus achieving natural edge closure and further enhancing the overall effect of edge detection.
2. Related Work
- (1)
- Intrinsic Geometric Shape Features of Spatial Targets: Spatial targets typically possess specific geometric shape features that are important identifiers distinguishing them from other objects or backgrounds. For example, satellite components often exhibit regular rectangular or circular structures, while the background is mostly random noise or irregular structures. This difference between targets and backgrounds provides a crucial entry point for accurately distinguishing spatial targets. By extracting geometric shape features, important prior information can be provided for edge detection, thereby enhancing the performance of the task. In this paper, the Shape Prior Guidance Module we designed is based on this concept. It acquires structural information consistent with the morphological features of spatial targets through template matching, dynamic feature enhancement, and channel dimension filtering techniques and effectively suppresses background responses conflicting with geometric priors, thus improving the accuracy and robustness of edge detection.
- (2)
- Geometric Distribution and Topological Structure Constraints of Spatial Targets: Spatial targets are complex systems composed of multiple components. The locations and arrangement of these components in space (i.e., geometric distribution) and their interconnections and relative positions (i.e., topological structure) contain rich prior information. For instance, the satellite body usually occupies the central position and has a regular geometric shape (cuboid or cylinder). Solar panels are typically installed symmetrically on both sides of the satellite body and exhibit rectangular or square structures. The body and solar panels usually have axial symmetry and fixed connection points. The specific locations and arrangement of these components in space are crucial for understanding the overall structure of the satellite. By analyzing geometric distribution information, the various components of the satellite can be more accurately identified and interpreted, thereby improving the precision of edge detection. Moreover, introducing topological structure constraints ensures that the generated edge map not only matches the geometric shape features of each component but also maintains consistency with the interconnections and relative positions between components. This is particularly important in complex scenarios, such as when there is occlusion or poor lighting conditions, to achieve fine-grained parsing and understanding of the target’s overall structure and maintain the accuracy and integrity of edge detection.
3. Methodology
3.1. Overview Architecture
3.2. Knowledge Distillation
3.2.1. Teacher Model
3.2.2. Student Model
3.2.3. Multi-Stage Distillation
3.3. Shape Prior Guidance
3.3.1. Shape Prior Model Construction
- (1)
- Data Collection and Edge Annotation. After clarifying the target object or scene, widely collect relevant images to ensure coverage of variations in different angles, lighting conditions, and background environments to enhance the generalization ability of the shape prior model. For space targets, this paper collects images of key components (including solar panels, satellite bodies, antennae, etc.) under different angles and lighting conditions. Data annotation tools are used to annotate the edges of these key components, and they are roughly categorized into two types based on shape: circular and quadrilateral, as shown in Figure 6; these images are all from public datasets.
- (2)
- Statistical Feature Extraction. Conduct feature analysis on the annotated images to extract statistical features of the shape of the target components, including calculating the curvature distribution of edges and analyzing the directional features of edges, etc. For example, solar panels usually have regular rectangular edges with a relatively concentrated curvature distribution and mainly linear directional features. The curvature distribution can reflect the smoothness and bending of the object’s contour, while the directional features help to understand the main direction of the edges. These statistical features can quantitatively describe the typical characteristics of the target shape and provide a key basis for target recognition and classification.
- (3)
- Construction of the Shape Prior Model. Based on the statistical features, further use principal component analysis (PCA) to establish the shape prior model for each component separately. Combine the annotated edge coordinates with the extracted statistical features to form the feature vector f, and perform PCA analysis on the feature vector to extract the main feature vectors (principal components) and corresponding eigenvalues, as shown in Equation (6). The main reason for using standard PCA [38] for feature extraction when constructing the shape prior model is as follows: The regular geometric structure of space targets needs to be characterized by a continuous and complete feature distribution. Standard PCA can effectively retain the main feature components of the shape and ensure feature continuity, which meets the requirement of structural integrity for edge detection. Meanwhile, its calculation process is compatible with the existing model framework, which can avoid introducing additional complexity while satisfying the feature extraction effect.
3.3.2. Similarity Calculation
3.3.3. Feature Reconstruction
- (1)
- Dynamic Convolutional Kernel Generation. Based on the similarity matrix , weight the base convolutional kernel to generate target-oriented dynamic convolutional kernels.
- (2)
- Feature Reconstruction. Use the dynamic convolutional kernel to guide feature reconstruction, as shown in Equation (10).
3.4. Loss Function
Algorithm 1 STED-KD | |||
Input: Image //RGB space target image | |||
Output: edge map | |||
1. Teacher model Pre-trained CNN | |||
teacher_features Extract_features(Teacher_model, ); | |||
teacher_probs ← Softmax(Teacher_model()); | |||
2. Multi-stage Knowledge Distillation Teacher model | |||
Stage-1: Intermediate-feature distillation | |||
student_features ← Extract_features(Student_model, ); | |||
Loss_mid MSE (teacher_features , student_features ); | |||
Stage-2: Output-level distillation | |||
student_probs ← Softmax(Student_model()); | |||
Loss_out ← KL (teacher_ probs , student_ probs ); | |||
Loss_kd ← alpha * Loss_mid + beta * Loss_out; | |||
return Pre-trained Student model. | |||
3. Shape Prior Augmented Training ground_truth, Pre-trained Student model | |||
Shape Prior Guided Module | |||
1. Build shape prior: | |||
Shape_prior ← PCA(Edge_templates); | |||
2. Similarity computation: | |||
student_features Extract_features(Pre-trained Student model, ); | |||
S ← Similarity(student_features , Shape_prior ); | |||
3. Feature reconstruction: | |||
dynamic_kernel ← Weighted_kernel(S, Base_kernel ); | |||
reconstructed feature ← DynamicConv(student_features , ); | |||
student_features’ ← reconstructed feature ; | |||
student_probs’ ← Pre-trained Student model (student_features’); | |||
Loss Function: | |||
Loss_ce ← cross_entropy(student_probs’, ground_truth); | |||
Loss_cur ← curvature_penalty(student_probs’, ground_truth); | |||
Loss_spa ← gamma * Loss_ce + delta * Loss_cur. |
4. Experiments
4.1. Datasets
4.2. Implementation Details
4.3. Comparison with State-of-the-Arts
4.4. Ablation Study
4.5. Parameter Analysis
5. Conclusions
- (1)
- In the Related Work section, this paper summarizes the geometric prior information of spatial targets, including geometric distribution and topological structure constraints. Subsequent research will delve deeper into how to effectively utilize geometric prior information to guide model learning. For example, using the topological structure constraints of spatial targets to optimize edge connectivity, ensuring that the detected edges conform to the actual characteristics of spatial targets in terms of topological relationships, and reducing unreasonable edge breakage and connection situations.
- (2)
- In the method proposed in this paper, simple and representative prior models are constructed to guide the model’s feature learning, providing effective prior knowledge for spatial target edge detection. However, this prior model performs poorly on the BSDS500 dataset. The main reason is that the geometric shapes of spatial targets are simple, while the BSDS500 dataset contains complex natural scenes and objects, and the simple prior model struggles to capture their complex features. In the future, we will further optimize the shape prior guidance module and explore more complex and effective prior models to enhance the model’s ability to handle complex structures and make it perform better on more types of images.
- (3)
- For the occlusion and illumination variation problem in complex scenes, although the current method has some ability to cope, there is still room for improvement. In the future, edge recovery strategies can be explored, such as edge completion methods based on Generative Adversarial Networks (GANs), using the adversarial training of generators and discriminators to predict the edges of occluded parts.
Author Contributions
Funding
Conflicts of Interest
References
- Zhao, Y.; Zhong, R.; Cui, L. Intelligent recognition of spacecraft components from photorealistic images based on Unreal Engine 4. Adv. Space Res. 2023, 71, 3761–3774. [Google Scholar] [CrossRef]
- Roberts, L. Machine Perception of Three-Dimensional Solids; Massachusetts Institute of Technology: Cambridge, MA, USA, 1963. [Google Scholar]
- Prewitt, J.M.S. Object enhancement and extraction. Pict. Process. Psychopictorics 1970, 10, 15–19. [Google Scholar]
- Kittler, J. On the accuracy of the Sobel edge detector. Image Vis. Comput. 1983, 1, 37–42. [Google Scholar] [CrossRef]
- Torre, V.; Poggio, T.A. On edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, PAMI-8, 147–163. [Google Scholar] [CrossRef] [PubMed]
- Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, PAMI-8, 679–698. [Google Scholar] [CrossRef]
- Arbelaez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour detection and hierarchical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 898–916. [Google Scholar] [CrossRef] [PubMed]
- Ren, X.; Bo, L. Discriminatively trained sparse code gradients for contour detection. In Proceedings of the 26th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–8 December 2012; Volume 1, pp. 584–592. [Google Scholar]
- Martin, R.; Fowlkes, C.C.; Malik, J. Learning to detect natural image boundaries using local brightness, color, and texture cues. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 530–549. [Google Scholar] [CrossRef] [PubMed]
- Zhang, Z.; Xing, F.; Shi, X.; Yang, L. Semicontour: A semi-supervised learning approach for contour detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 251–259. [Google Scholar]
- Hallman, S.; Fowlkes, C.C. Oriented edge forests for boundary detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1732–1740. [Google Scholar]
- Konishi, S.; Yuille, A.L.; Coughlan, J.M.; Zhu, S.C. Statistical edge detection: Learning and evaluating edge cues. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 57–74. [Google Scholar] [CrossRef]
- Dollár, P.; Zitnick, C.L. Fast edge detection using structured forests. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1558–1570. [Google Scholar] [CrossRef]
- Comaniciu, D.; Meer, P. Mean shift: A robust approach toward feature space analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 603–619. [Google Scholar] [CrossRef]
- Felzenszwalb, P.F.; Huttenlocher, P. Efficient graph-based image segmentation. Int. J. Comput. Vis. 2004, 59, 167–181. [Google Scholar] [CrossRef]
- Cheng, M.M.; Liu, Y.; Hou, Q.; Bian, J.; Torr, P.; Hu, S.M.; Tu, Z. HFS: Hierarchical feature selection for efficient image segmentation. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part III 14. Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 867–882. [Google Scholar]
- Arbeláez, P.; Pont-Tuset, J.; Barron, J.T.; Marques, F.; Malik, J. Multiscale combinatorial grouping. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 328–335. [Google Scholar]
- Plötz, T.; Roth, S. Neural nearest neighbors networks. Adv. Neural Inf. Process. Syst. 2018, 31, 1–12. [Google Scholar]
- Shen, W.; Wang, X.; Wang, Y.; Bai, X.; Zhang, Z. Deepcontour: A deep convolutional feature learned by positive-sharing loss for contour detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3982–3991. [Google Scholar]
- Bertasius, G.; Shi, J.; Torresani, L. Deepedge: A multi-scale bifurcated deep network for top-down contour detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4380–4389. [Google Scholar]
- Bertasius, G.; Shi, J.; Torresani, L. High-for-low and low-for-high: Efficient boundary detection from deep object features and its applications to high-level vision. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 504–512. [Google Scholar]
- Xie, S.; Tu, Z. Holistically-nested edge detection. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1395–1403. [Google Scholar]
- Liu, Y.; Cheng, M.M.; Hu, X.; Wang, K.; Bai, X. Richer convolutional features for edge detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3000–3009. [Google Scholar]
- He, J.; Zhang, S.; Yang, M.; Shan, Y.; Huang, T. Bi-directional cascade network for perceptual edge detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3828–3837. [Google Scholar]
- Poma, X.S.; Riba, E.; Sappa, A. Dense extreme inception network: Towards a robust cnn model for edge detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Aspen, CO, USA, 1–5 March 2020; pp. 1923–1932. [Google Scholar]
- Pu, M.; Huang, Y.; Liu, Y.; Guan, Q.; Ling, H. Edter: Edge detection with transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1402–1412. [Google Scholar]
- Zhou, C.; Huang, Y.; Pu, M.; Guan, Q.; Huang, L.; Ling, H. The treasure beneath multiple annotations: An uncertainty-aware edge detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 15507–15517. [Google Scholar]
- Ye, Y.; Xu, K.; Huang, Y.; Yi, R.; Cai, Z. Diffusionedge: Diffusion probabilistic model for crisp edge detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 6675–6683. [Google Scholar]
- Wibisono, J.K.; Hang, H.M. Traditional method inspired deep neural network for edge detection. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Virtually, 25–28 October 2020; IEEE: New York, NY, USA, 2020; pp. 678–682. [Google Scholar]
- Wibisono, J.K.; Hang, H.M. Fined: Fast inference network for edge detection. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Virtual, 5–9 July 2021; IEEE: New York, NY, USA, 2021; pp. 1–6. [Google Scholar]
- Su, Z.; Liu, W.; Yu, Z.; Hu, D.; Liao, Q.; Tian, Q.; Liu, L. Pixel difference networks for efficient edge detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 10–17 October 2021; pp. 5117–5127. [Google Scholar]
- Soria, X.; Li, Y.; Rouhani, M.; Sappa, A.D. Tiny and efficient model for the edge detection generalization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 1364–1373. [Google Scholar]
- Fuzhang, L.; Chuan, L. UHNet: An ultra-lightweight and high-speed edge detection network. arXiv 2024, arXiv:2408.04258. [Google Scholar]
- Liufu, X.; Tan, C.; Lin, X.; Qi, Y.; Li, J.; Hu, J.F. SAUGE: Taming SAM for Uncertainty-Aligned Multi-Granularity Edge Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 5766–5774. [Google Scholar]
- Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar] [CrossRef]
- Gou, J.; Yu, B.; Maybank, S.J.; Tao, D. Knowledge distillation: A survey. Int. J. Comput. Vis. 2021, 129, 1789–1819. [Google Scholar] [CrossRef]
- Moslemi, A.; Briskina, A.; Dang, Z.; Li, J. A survey on knowledge distillation: Recent advancements. Mach. Learn. Appl. 2024, 18, 100605. [Google Scholar] [CrossRef]
- Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
- Mao, A.; Mohri, M.; Zhong, Y. Cross-entropy loss functions: Theoretical analysis and applications. In Proceedings of the International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; pp. 23803–23828. [Google Scholar]
- Marin, D.; Zhong, Y.; Drangova, M.; Boykov, Y. Thin structure estimation with curvature regularization. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 397–405. [Google Scholar]
- Russell, B.C.; Torralba, A.; Murphy, K.P.; Freeman, W.T. LabelMe: A database and web-based tool for image annotation. Int. J. Comput. Vis. 2008, 77, 157–173. [Google Scholar] [CrossRef]
Encoder Layer | Output Shape | Param |
---|---|---|
Input | (w, h, 3) | - |
DSConv1 | (w, h, 64) | 219 |
pool | (w/2, h/2, 64) | - |
DSConv2 | (w/2, h/2, 128) | 8768 |
pool | (w/4, h/4, 128) | - |
DSConv3 | (w/4, h/4, 256) | 33,920 |
DSConv4 | (w/4, h/4, 256) | 67,840 |
pool | (w/8, h/8, 256) | - |
DSConv5 | (w/8, h/8, 512) | 133,376 |
DSConv6 | (w/8, h/8, 512) | 266,752 |
pool | (w/16, h/16, 512) | - |
DSConv7 | (w/16, h/16, 512) | 266,752 |
DSConv8 | (w/16, h/16, 512) | 266,752 |
1,005,589 |
Methods | ODS | OIS | AP | Param | FPS |
---|---|---|---|---|---|
Sobel [4] | .539 | .575 | .498 | 18 | 278.8 |
Canny [6] | .611 | .676 | .520 | 9/25 | 27.7 |
MShift [14] | .598 | .645 | .497 | - | - |
EGB [15] | .614 | .658 | .564 | - | - |
HFS [16] | .650 | .688 | .201 | - | - |
Pb [7] | .672 | .695 | .652 | - | - |
SE [13] | .743 | .764 | .800 | - | - |
MCG [17] | .744 | .777 | .760 | - | - |
OEF [11] | .746 | .770 | .815 | - | - |
DeepEdge [20] | .753 | .772 | .807 | - | - |
DeepContour [19] | .757 | .776 | .790 | - | - |
HFL [21] | .767 | .788 | .795 | - | - |
DexiNed [25] | .729 | .745 | .583 | 35M | 1.43 |
TIN [29] | .772 | .795 | - | 244K | 10.1 |
HED [22] | .788 | .808 | .840 | 14.7M | 2.91 |
PiDiNet [31] | .787 | .804 | .817 | 710K | 3.59 |
BDCN [24] | .806 | .826 | .847 | 16.3M | 2.88 |
RCF [23] | .811 | .830 | .846 | 14.8M | 2.35 |
EDTER [26] | .824 | .841 | .880 | 468.8M | 0.67 |
UAED [27] | .828 | .847 | .892 | 72.5M | 1.02 |
DiffusionEdge [28] | .834 | .848 | .815 | 225M | 1.54 |
ours | .818 | .829 | .850 | 4.91M | 3.05 |
Methods | ODS | OIS | AP | Param | FPS |
---|---|---|---|---|---|
Canny [6] | .444 | .477 | .341 | 9/25 | 58.8 |
MCG [17] | .489 | .535 | .355 | - | - |
OEF [11] | .493 | .524 | .389 | - | - |
DexiNed [25] | .611 | .676 | .564 | 35M | 0.35 |
HED [22] | .635 | .702 | .583 | 14.7M | 1.5 |
PiDiNet [21] | .566 | .634 | .452 | 710K | 2.59 |
BDCN [24] | .557 | .676 | .440 | 16.3M | 1.57 |
DiffusionEdge [28] | .594 | .628 | .492 | 225M | 0.19 |
ours | .659 | .715 | .596 | 4.91M | 2.15 |
Method | ODS | OIS | AP |
---|---|---|---|
Baseline-T | .649 | .683 | .582 |
Baseline-S | .598 | .662 | .551 |
Baseline-S + KD | .630 | .686 | .571 |
Baseline-S + SPGM | .628 | .688 | .575 |
Baseline-S + | .608 | .674 | .559 |
Baseline-S + KD + SPGM | .655 | .709 | .591 |
Baseline-S + KD + | .638 | .694 | .574 |
Baseline-S + SPGM + | .635 | .695 | .579 |
Ours | .659 | .715 | .596 |
ODS | OIS | AP | |
---|---|---|---|
.648 | .703 | .582 | |
.635 | .692 | .559 | |
.647 | .707 | .589 | |
.659 | .715 | .596 | |
.641 | .705 | .579 | |
.652 | .708 | .590 | |
.638 | .699 | .573 | |
.647 | .692 | .582 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Meng, Y.; Zhang, L.; Zhang, Y.; Hu, M.; Zhao, F.; Shen, X. Satellite Optical Target Edge Detection Based on Knowledge Distillation. Remote Sens. 2025, 17, 3008. https://doi.org/10.3390/rs17173008
Meng Y, Zhang L, Zhang Y, Hu M, Zhao F, Shen X. Satellite Optical Target Edge Detection Based on Knowledge Distillation. Remote Sensing. 2025; 17(17):3008. https://doi.org/10.3390/rs17173008
Chicago/Turabian StyleMeng, Ying, Luping Zhang, Yan Zhang, Moufa Hu, Fei Zhao, and Xinglin Shen. 2025. "Satellite Optical Target Edge Detection Based on Knowledge Distillation" Remote Sensing 17, no. 17: 3008. https://doi.org/10.3390/rs17173008
APA StyleMeng, Y., Zhang, L., Zhang, Y., Hu, M., Zhao, F., & Shen, X. (2025). Satellite Optical Target Edge Detection Based on Knowledge Distillation. Remote Sensing, 17(17), 3008. https://doi.org/10.3390/rs17173008