Title :
Relative attributes
Author :
Parikh, Devi ; Grauman, Kristen
Author_Institution :
Toyota Technol. Inst. Chicago (TTIC), Chicago, IL, USA
Abstract :
Human-nameable visual “attributes” can benefit various recognition tasks. However, existing techniques restrict these properties to categorical labels (for example, a person is `smiling´ or not, a scene is `dry´ or not), and thus fail to capture more general semantic relationships. We propose to model relative attributes. Given training data stating how object/scene categories relate according to different attributes, we learn a ranking function per attribute. The learned ranking functions predict the relative strength of each property in novel images. We then build a generative model over the joint space of attribute ranking outputs, and propose a novel form of zero-shot learning in which the supervisor relates the unseen object category to previously seen objects via attributes (for example, `bears are furrier than giraffes´). We further show how the proposed relative attributes enable richer textual descriptions for new images, which in practice are more precise for human interpretation. We demonstrate the approach on datasets of faces and natural scenes, and show its clear advantages over traditional binary attribute prediction for these new tasks.
Keywords :
face recognition; learning (artificial intelligence); natural scenes; binary attribute prediction; categorical label; face dataset; human interpretation; human-name visual attribute; image textual description; learned ranking function; natural scene; object category; ranking function per attribute; recognition task; scene category; training data; zero-shot learning; Humans; Image recognition; Machine learning; Support vector machines; Training; Visualization; Vocabulary;
Conference_Titel :
Computer Vision (ICCV), 2011 IEEE International Conference on
Conference_Location :
Barcelona
Print_ISBN :
978-1-4577-1101-5
DOI :
10.1109/ICCV.2011.6126281