Back to Search Start Over

Learning dynamic background for weakly supervised moving object detection.

Authors :
Zhang, Zhijun
Chang, Yi
Zhong, Sheng
Yan, Luxin
Zou, Xu
Source :
Image & Vision Computing. May2022, Vol. 121, pN.PAG-N.PAG. 1p.
Publication Year :
2022

Abstract

Moving Object Detection (MOD) aims at extracting foreground moving objects in videos from static cameras. While low-rank based approaches have achieved impressive success in the MOD task, their performance remains limited on dynamics background scenes. The main reason is that dynamic clutters, e.g. , swaying leaves and rippers , are easy to mix up with moving objects in the decomposition model which simply classify the sparse noise as foregrounds. In order to improve the generalization ability of low-rank based moving object detectors, we suggest adding an explicit dynamic clutter component in the decomposition framework with realistic dynamic background modeling. Then the dynamic clutter can be learned through object-free video data due to their self-similarity across time and space. Thus, the moving objects can be naturally separated by a tensor-based decomposition model which formulates the static background by a unidirectional low-rank tensor, learns the dynamic clutter by a two-stream neural network, and constrains moving objects with spatiotemporal continuity. To further provide a more accurate object detection result, an objectness prior is embedded into our model in an attention manner. Extensive experimental results on the challenging datasets of dynamic background clearly demonstrate the superior performance of our model over the state-of-the-art in terms of quantitative metrics and visual quality. [Display omitted] • A novel video decomposition framework, additionally modeling the dynamic clutter explicitly. • A unified tensor-based framework to capture the appearance and motion information simultaneously for three components. • A self-supervised strategy to learn the dynamic clutter well by considering the video sequence itself as a dataset. • Bridge the gap between discriminative prior and the low-rank MAP framework. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
02628856
Volume :
121
Database :
Academic Search Index
Journal :
Image & Vision Computing
Publication Type :
Academic Journal
Accession number :
156373653
Full Text :
https://doi.org/10.1016/j.imavis.2022.104425