DocumentCode :
3672562
Title :
Aligning 3D models to RGB-D images of cluttered scenes
Author :
Saurabh Gupta;Pablo Arbeláez;Ross Girshick;Jitendra Malik
Author_Institution :
UC Berkeley, USA
fYear :
2015
fDate :
6/1/2015 12:00:00 AM
Firstpage :
4731
Lastpage :
4740
Abstract :
The goal of this work is to represent objects in an RGB-D scene with corresponding 3D models from a library. We approach this problem by first detecting and segmenting object instances in the scene and then using a convolutional neural network (CNN) to predict the pose of the object. This CNN is trained using pixel surface normals in images containing renderings of synthetic objects. When tested on real data, our method outperforms alternative algorithms trained on real data. We then use this coarse pose estimate along with the inferred pixel support to align a small number of prototypical models to the data, and place into the scene the model that fits best. We observe a 48% relative improvement in performance at the task of 3D detection over the current state-of-the-art [34], while being an order of magnitude faster.
Keywords :
"Solid modeling","Three-dimensional displays","Data models","Image segmentation","Computational modeling","Feature extraction","Training"
Publisher :
ieee
Conference_Titel :
Computer Vision and Pattern Recognition (CVPR), 2015 IEEE Conference on
Electronic_ISBN :
1063-6919
Type :
conf
DOI :
10.1109/CVPR.2015.7299105
Filename :
7299105
Link To Document :
بازگشت