Back to Search Start Over

Discovering Primary Objects in Videos by Saliency Fusion and Iterative Appearance Estimation.

Authors :
Yang, Jiong
Zhao, Gangqiang
Yuan, Junsong
Shen, Xiaohui
Lin, Zhe
Price, Brian
Brandt, Jonathan
Source :
IEEE Transactions on Circuits & Systems for Video Technology. 6/1/2016, Vol. 26 Issue 6, p1070-1083. 14p.
Publication Year :
2016

Abstract

In this paper, we propose a new method for detecting primary objects in unconstrained videos in a completely automatic setting. Here, we define the primary object in a video as the object that presents saliently in most of the frames. Unlike previous works considering only local saliency detection or common pattern discovery, the proposed method integrates the local visual/motion saliency extracted from each frame, global appearance consistency throughout the video, and spatiotemporal smoothness constraint on object trajectories. We first identify a temporal coherent salient region throughout the whole video, and then explicitly learn a global appearance model to distinguish the primary object against the background. In order to obtain high-quality saliency estimations from both appearance and motion cues, we propose a novel self-adaptive saliency map fusion method by learning the reliability of saliency maps from labeled data. As a whole, our method can robustly localize and track primary objects in diverse video content, and handle the challenges such as fast object and camera motion, large scale and appearance variation, background clutter, and pose deformation. Moreover, compared with some existing approaches that assume the object is present in all the frames, our approach can naturally handle the case where the object is present only in part of the frames, e.g., the object enters the scene in the middle of the video or leaves the scene before the video ends. We also propose a new video data set containing 51 videos for primary object detection with per-frame ground-truth labeling. Quantitative experiments on several challenging video data sets demonstrate the superiority of our method compared with the recent state of the arts. [ABSTRACT FROM PUBLISHER]

Details

Language :
English
ISSN :
10518215
Volume :
26
Issue :
6
Database :
Academic Search Index
Journal :
IEEE Transactions on Circuits & Systems for Video Technology
Publication Type :
Academic Journal
Accession number :
116115833
Full Text :
https://doi.org/10.1109/TCSVT.2015.2433171