计算机科学
推论
修剪
目标检测
人工智能
机器学习
正规化(语言学)
数据挖掘
模式识别(心理学)
农学
生物
作者
Huaiyuan Sun,Shuili Zhang,X. Cindy Tian,Yuanyuan Zou
标识
DOI:10.1007/s11760-023-02719-4
摘要
Deep learning methods in the field of object detection have made significant progress in terms of performance, but end-to-end implementations still face challenges. Recently, the transformer-based DETR model successfully introduced the attention mechanism into object detection tasks, achieving end-to-end object detection. However, despite its competitive accuracy, DETR still falls short in terms of inference speed and computational costs. To address this issue, this paper proposes an optimization of the DETR model using structured pruning through sparsity-induced pruning, aiming to improve its inference speed and reduce computational costs. We adjust the importance of module outputs through parameter scaling factors and sparse regularization terms, and optimize the parameter scaling factors using an improved Accelerated Proximal Gradient (APG) method. Experimental results on the COCO dataset demonstrate that our approach achieves a computational cost reduction of over 32% while maintaining an AP value of 42.7%, resulting in an inference speed improvement of over 17% to reach 32.7 FPS. This study provides an effective solution for further enhancing the computational efficiency of transformer-based object detection models.
科研通智能强力驱动
Strongly Powered by AbleSci AI