计算机科学
比例(比率)
特征(语言学)
推论
跳跃式监视
对象(语法)
蒸馏
目标检测
嵌入
人工智能
代表(政治)
探测器
机器学习
特征提取
模式识别(心理学)
数据挖掘
有机化学
政治学
语言学
法学
化学
量子力学
哲学
物理
政治
电信
作者
Yichen Zhu,Qiqi Zhou,Ning Liu,Zhiyuan Xu,Zhicai Ou,Xiaofeng Mou,Jian Tang
标识
DOI:10.1109/cvpr52729.2023.01889
摘要
Despite the prominent success of general object detection, the performance and efficiency of Small Object Detection (SOD) are still unsatisfactory. Unlike existing works that struggle to balance the tradeoff between inference speed and SOD performance, in this paper, we propose a novel Scale-aware Knowledge Distillation (ScaleKD), which transfers knowledge of a complex teacher model to a compact student model. We design two novel modules to boost the quality of knowledge transfer in distillation for SOD: 1) a scale-decoupled feature distillation module that disentangled teacher's feature representation into multi-scale embedding that enables explicit feature mimicking of the student model on small objects. 2) a cross-scale assistant to refine the noisy and uninformative bounding boxes prediction student models, which can mislead the student model and impair the efficacy of knowledge distillation. A multi-scale cross-attention layer is established to capture the multi-scale semantic information to improve the student model. We conduct experiments on COCO and VisDrone datasets with diverse types of models, i.e., two-stage and one-stage detectors, to evaluate our proposed method. Our ScaleKD achieves superior performance on general detection performance and obtains spectacular improvement regarding the SOD performance.
科研通智能强力驱动
Strongly Powered by AbleSci AI