DocumentCode
114326
Title
BiModal latent dirichlet allocation for text and image
Author
Xiaofeng Liao ; Qingshan Jiang ; Wei Zhang ; Kai Zhang
Author_Institution
Shenzhen Inst. of Adv. Technol., Shenzhen, China
fYear
2014
fDate
26-28 April 2014
Firstpage
736
Lastpage
739
Abstract
A BiModal Latent Dirichlet Allocation Model(BM-LDA) is proposed to learn a unified representation of data that comes from both the textual and visual modalities together. The model is able to form a unified representation that mixs both the textual and visual modalities. Based on the assumption, that the images and its surrounding text share a same topic, the model learns a posterior probability density in the space of latent variable of topics that bridging over the observed multi modality inputs. It maps the high dimensional space consist of the observed variables from both modalities to a low dimensional space of topcis. Experimental result on ImageCLEF data set, which consists of bi-modality data of images and surrounding text, shows our new BM-LDA model can get a fine representation for the multi-modality data, which is useful for tasks such as retrieval and classification.
Keywords
Internet; image classification; image representation; probability; text analysis; BM-LDA; BiModal latent Dirichlet allocation model; ImageCLEF data set; bimodality image data; bimodality text data; classification; observed multimodality inputs; posterior probability density; retrieval; textual modalities; unified data representation; visual modalities; Data models; Image classification; Kernel; Mathematical model; Resource management; Standards; Visualization; BiModal Latent Dirichlet Allocation; Image; Multiple Modality; Text;
fLanguage
English
Publisher
ieee
Conference_Titel
Information Science and Technology (ICIST), 2014 4th IEEE International Conference on
Conference_Location
Shenzhen
Type
conf
DOI
10.1109/ICIST.2014.6920582
Filename
6920582
Link To Document