Proceedings of the 21st ACM international conference on Multimedia
Recently visual saliency has attracted wide attention of researchers in the computer vision and multimedia field. However, most of the visual saliency-related research was conducted on still images for studying static saliency. In this paper, we give a comprehensive comparative study for the first time of dynamic saliency (video shots) and static saliency (key frames of the corresponding video shots), and two key observations are obtained: 1) video saliency is often different from, yet quite related with, image saliency, and 2) camera motions, such as tilting, panning or zooming, affect dynamic saliency significantly.
Motivated by these observations, we propose a novel camera motion and image saliency aware model for dynamic saliency prediction.
The extensive experiments on two static-vs-dynamic saliency datasets collected by us show that our proposed method outperforms the state-of-the-art methods for dynamic saliency prediction. Finally, we also introduce the application of dynamic saliency prediction for dynamic video captioning, assisting people with hearing impairments to better entertain videos with only off-screen voices, e.g., documentary films, news videos and sports videos.
Copyright © 2013, Association for Computing Machinery
Association for Computing Machinery
Place of Publication
Nguyen, Tam; Xu, Mengdi; Gao, Guangyu; Kankanhalli, Mohan; Tian, Qi; and Yan, Shuicheng, "Static Saliency vs. Dynamic Saliency: A Comparative Study" (2013). Computer Science Faculty Publications. 69.