DocumentCode :
254286
Title :
Patch to the Future: Unsupervised Visual Prediction
Author :
Walker, Julian ; Gupta, Arpan ; Hebert, Martial
Author_Institution :
Robot. Inst., Carnegie Mellon Univ., Pittsburgh, PA, USA
fYear :
2014
fDate :
23-28 June 2014
Firstpage :
3302
Lastpage :
3309
Abstract :
In this paper we present a conceptually simple but surprisingly powerful method for visual prediction which combines the effectiveness of mid-level visual elements with temporal modeling. Our framework can be learned in a completely unsupervised manner from a large collection of videos. However, more importantly, because our approach models the prediction framework on these mid-level elements, we can not only predict the possible motion in the scene but also predict visual appearances - how are appearances going to change with time. This yields a visual "hallucination" of probable events on top of the scene. We show that our method is able to accurately predict and visualize simple future events, we also show that our approach is comparable to supervised methods for event prediction.
Keywords :
data visualisation; image motion analysis; learning (artificial intelligence); video signal processing; future event prediction; future event visualization; leraning; mid-level visual elements; motion prediction; temporal modeling; unsupervised visual prediction; video collection; visual appearance prediction; visual hallucination; Feature extraction; Prediction algorithms; Predictive models; Tracking; Training data; Videos; Visualization; Activity Forecasting; Prediction;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Computer Vision and Pattern Recognition (CVPR), 2014 IEEE Conference on
Conference_Location :
Columbus, OH
Type :
conf
DOI :
10.1109/CVPR.2014.416
Filename :
6909818
Link To Document :
بازگشت