DocumentCode
2805100
Title
Contextual Entropy and Text Categorization
Author
García, Moisés ; Hidalgo, Hugo ; Chávez, Edgar
Author_Institution
Centro de Investigacion y de Educacion, Superior de Ensenada
fYear
2006
fDate
Oct. 2006
Firstpage
147
Lastpage
153
Abstract
In this paper we describe a new approach to text categorization, our focus is in the amount of information (the entropy) in the text. The entropy is computed with the empirical distribution of words in the text. We provide the system with a manually segmented collection of documents in different categories. For each category a separate empirical distribution of words is computed, we use this empirical distribution for categorization purposes. If we compute the entropy of the test document for each empirical distribution the correct category shows as a maximum. For example, if we compute the entropy of a sports document using the politics or the sports empirical word distributions then the computed entropy is higher in sports than in politics. Our text categorization approach is simple, easy to code and needs no training time (aside from histogram computations). The classification time is linear on the size of the document and the number of document categories. We support our claims with extensive experimentation
Keywords
classification; entropy; text analysis; contextual entropy; document processing; empirical word distribution; text categorization; Acceleration; Distributed computing; Entropy; Histograms; Internet; Support vector machine classification; Support vector machines; Taxonomy; Testing; Text categorization;
fLanguage
English
Publisher
ieee
Conference_Titel
Web Congress, 2006. LA-Web '06. Fourth Latin American
Conference_Location
Cholula
Print_ISBN
0-7695-2693-4
Type
conf
DOI
10.1109/LA-WEB.2006.11
Filename
4022104
Link To Document