• DocumentCode
    2805100
  • Title

    Contextual Entropy and Text Categorization

  • Author

    García, Moisés ; Hidalgo, Hugo ; Chávez, Edgar

  • Author_Institution
    Centro de Investigacion y de Educacion, Superior de Ensenada
  • fYear
    2006
  • fDate
    Oct. 2006
  • Firstpage
    147
  • Lastpage
    153
  • Abstract
    In this paper we describe a new approach to text categorization, our focus is in the amount of information (the entropy) in the text. The entropy is computed with the empirical distribution of words in the text. We provide the system with a manually segmented collection of documents in different categories. For each category a separate empirical distribution of words is computed, we use this empirical distribution for categorization purposes. If we compute the entropy of the test document for each empirical distribution the correct category shows as a maximum. For example, if we compute the entropy of a sports document using the politics or the sports empirical word distributions then the computed entropy is higher in sports than in politics. Our text categorization approach is simple, easy to code and needs no training time (aside from histogram computations). The classification time is linear on the size of the document and the number of document categories. We support our claims with extensive experimentation
  • Keywords
    classification; entropy; text analysis; contextual entropy; document processing; empirical word distribution; text categorization; Acceleration; Distributed computing; Entropy; Histograms; Internet; Support vector machine classification; Support vector machines; Taxonomy; Testing; Text categorization;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Web Congress, 2006. LA-Web '06. Fourth Latin American
  • Conference_Location
    Cholula
  • Print_ISBN
    0-7695-2693-4
  • Type

    conf

  • DOI
    10.1109/LA-WEB.2006.11
  • Filename
    4022104