计算机科学
计算机视觉
地标
人工智能
主管(地质)
活动识别
地貌学
地质学
作者
Eshed Ohn-Bar,Sujitha Martin,Ashish Tawari,Mohan M. Trivedi
标识
DOI:10.1109/icpr.2014.124
摘要
In this paper, a multiview, multimodal vision framework is proposed in order to characterize driver activity based on head, eye, and hand cues. Leveraging the three types of cues allows for a richer description of the driver's state and for improved activity detection performance. First, regions of interest are extracted from two videos, one observing the driver's hands and one the driver's head. Next, hand location hypotheses are generated and integrated with a head pose and facial landmark module in order to classify driver activity into three states: wheel region interaction with two hands on the wheel, gear region activity, or instrument cluster region activity. The method is evaluated on a video dataset captured in on-road settings.
科研通智能强力驱动
Strongly Powered by AbleSci AI