DocumentCode
3325219
Title
A proposed metric for the evaluation of natural language systems
Author
Facemire, Jon
Author_Institution
Dept. of Comput. Sci., Alabama A&M Univ., Normal, AL, USA
fYear
1989
fDate
9-12 Apr 1989
Firstpage
1111
Abstract
It is proposed that a simple number scale be established for the evaluation of natural-language text-processing systems. This metric is to be based on human linguistic performance, taken as 1.0, and is the simple average of four subcomponents, i.e. the size of the lexicon, the speed and accuracy of the parse, and the overall experience of the system. The metric would represent the ability to parse English text at the rate of 12 sentences per minute with perfect accuracy based on a vocabulary of 25000 words and an experience base of 40 million sentences. It is noted that, if such a metric can be successfully implemented for natural-language projects it can then serve as a model for the development of validation techniques for other hard-to-evaluate stress in artificial intelligence such as expert system, learning, or pattern-recognition programs
Keywords
artificial intelligence; grammars; natural languages; English text; artificial intelligence; expert system; human linguistic performance; learning; natural language systems evaluation; number scale; parse; pattern-recognition programs; text-processing systems; validation techniques; Artificial intelligence; Computer science; Expert systems; Face; Humans; Natural language processing; Natural languages; Pattern recognition; Robustness; Text processing;
fLanguage
English
Publisher
ieee
Conference_Titel
Southeastcon '89. Proceedings. Energy and Information Technologies in the Southeast., IEEE
Conference_Location
Columbia, SC
Type
conf
DOI
10.1109/SECON.1989.132582
Filename
132582
Link To Document