Title :
Visual Phraselet: Refining Spatial Constraints for Large Scale Image Search
Author :
Liang Zheng ; Shengjin Wang
Author_Institution :
Dept. of Electron. Eng., Tsinghua Univ., Beijing, China
Abstract :
The Bag-of-Words (BoW) model is prone to the deficiency of spatial constraints among visual words. The state of the art methods encode spatial information via visual phrases. However, these methods discard the spatial context among visual phrases instead. To address the problem, this letter introduces a novel visual concept, the Visual Phraselet, as a kind of similarity measurement between images. The visual phraselet refers to the spatial consistent group of visual phrases. In a simple yet effective manner, visual phraselet filters out false visual phrase matches, and is much more discriminative than both visual word and visual phrase. To boost the discovery of visual phraselets, we apply the soft quantization scheme. Our method is evaluated through extensive experiments on three benchmark datasets (Oxford 5 K, Paris 6 K and Flickr 1 M). We report significant improvements as large as 54.6% over the baseline approach, thus validating the concept of visual phraselet.
Keywords :
image coding; image matching; image retrieval; quantisation (signal); Visual Phraselet; bag-of-words model; large scale image search; refining spatial constraints; similarity measurement; soft quantization scheme; spatial context; spatial information encoding; visual phraselet filters; visual phraselets; visual phrases; visual words; Image search; spatial constraint; visual phrase; visual phraselet;
Journal_Title :
Signal Processing Letters, IEEE
DOI :
10.1109/LSP.2013.2249513