投票
对象(语法)
计算机科学
编码(集合论)
霍夫变换
目标检测
任务(项目管理)
班级(哲学)
人工智能
航程(航空)
模式识别(心理学)
自上而下和自下而上的设计
图像(数学)
多数决原则
视觉对象识别的认知神经科学
计算机视觉
集合(抽象数据类型)
工程类
程序设计语言
系统工程
法学
航空航天工程
政治
政治学
作者
Nermin Samet,Samet Hiçsönmez,Emre Akbaş
出处
期刊:Cornell University - arXiv
日期:2020-07-05
被引量:2
摘要
This paper presents HoughNet, a one-stage, anchor-free, voting-based, bottom-up object detection method. Inspired by the Generalized Hough Transform, HoughNet determines the presence of an object at a certain location by the sum of the votes cast on that location. Votes are collected from both near and long-distance locations based on a log-polar vote field. Thanks to this voting mechanism, HoughNet is able to integrate both near and long-range, class-conditional evidence for visual recognition, thereby generalizing and enhancing current object detection methodology, which typically relies on only local evidence. On the COCO dataset, HoughNet's best model achieves 46.4 $AP$ (and 65.1 $AP_{50}$), performing on par with the state-of-the-art in bottom-up object detection and outperforming most major one-stage and two-stage methods. We further validate the effectiveness of our proposal in another task, namely, labels to photo image generation by integrating the voting module of HoughNet to two different GAN models and showing that the accuracy is significantly improved in both cases. Code is available at this https URL.
科研通智能强力驱动
Strongly Powered by AbleSci AI