DocumentCode
3348391
Title
Capturing Contextual Relationship for Effective Media Search
Author
Cha, Guang-Ho
Author_Institution
Dept. of Comput. Eng., Seoul Nat. Univ. of Technol., Seoul, South Korea
fYear
2009
fDate
10-12 Dec. 2009
Firstpage
1
Lastpage
6
Abstract
One of the central problems regarding media search is the semantic gap between the low-level features computed automatically from media data and the human interpretation of them. This is because the notion of similarity is usually based on high-level abstraction but the low-level features do not sometimes reflect the human perception. In this paper, we assume the semantics of a media is determined by the contextual relationship in a dataset, and introduce the method to capture the contextual information from a large media (especially image) dataset for effective search. Similarity search in an image database based on this contextual information shows encouraging experimental results.
Keywords
image retrieval; query formulation; visual databases; contextual relationship; high-level abstraction; human perception; image database; media search; semantic gap; similarity search; Data engineering; Euclidean distance; Humans; Image databases; Image retrieval; Information retrieval; Measurement standards; Nearest neighbor searches;
fLanguage
English
Publisher
ieee
Conference_Titel
Embedded and Multimedia Computing, 2009. EM-Com 2009. 4th International Conference on
Conference_Location
Jeju
Print_ISBN
978-1-4244-4995-8
Type
conf
DOI
10.1109/EM-COM.2009.5402975
Filename
5402975
Link To Document