DocumentCode
3300481
Title
Measuring word polysemousness and sense granularity at a language level
Author
Zhu, Hong ; Liu, Yang ; Yu, Shiwen
Author_Institution
Inst. of Comput. Linguistics, Peking Univ., Beijing
fYear
2008
fDate
19-22 Oct. 2008
Firstpage
1
Lastpage
7
Abstract
Word sense acquisition and distinction are key issues for both lexicography and lexical semantic processing. However, it is quite difficult to automatically acquire word senses and to further evaluate the results with the lexica, which more likely bear the different findings of word sense distinction and granularity. In this paper, we´d like to put forward the idea of measuring word polysemousness and sense granularity at a language level. Two methods, viz. MECBC and TIEM, are at first employed as attempts to extract Chinese word senses from the corpora. Automatic word senses mapping to the lexica and evaluation of the results are devised and realized afterwards. Our experiments shows a rather fine fitness of Chinese word polysemousness between the results and the lexica at the whole language level. Comaprison of sense granularity between different lexical semantic resources can hence be made on a sound judgment.
Keywords
natural language processing; word processing; Chinese word sense extraction; MECBC; TIEM; language level; lexical semantic processing; lexicography; sense granularity; word polysemousness measurement; word sense acquisition; Clustering algorithms; Clustering methods; Computational linguistics; Dictionaries; Humans; Iterative algorithms; Large-scale systems; Natural languages; Size control; Thesauri; CBC; EM; clustering; word sense discrimination;
fLanguage
English
Publisher
ieee
Conference_Titel
Natural Language Processing and Knowledge Engineering, 2008. NLP-KE '08. International Conference on
Conference_Location
Beijing
Print_ISBN
978-1-4244-4515-8
Electronic_ISBN
978-1-4244-2780-2
Type
conf
DOI
10.1109/NLPKE.2008.4906762
Filename
4906762
Link To Document