Title :
A proposed metric for the evaluation of natural language systems
Author_Institution :
Dept. of Comput. Sci., Alabama A&M Univ., Normal, AL, USA
Abstract :
It is proposed that a simple number scale be established for the evaluation of natural-language text-processing systems. This metric is to be based on human linguistic performance, taken as 1.0, and is the simple average of four subcomponents, i.e. the size of the lexicon, the speed and accuracy of the parse, and the overall experience of the system. The metric would represent the ability to parse English text at the rate of 12 sentences per minute with perfect accuracy based on a vocabulary of 25000 words and an experience base of 40 million sentences. It is noted that, if such a metric can be successfully implemented for natural-language projects it can then serve as a model for the development of validation techniques for other hard-to-evaluate stress in artificial intelligence such as expert system, learning, or pattern-recognition programs
Keywords :
artificial intelligence; grammars; natural languages; English text; artificial intelligence; expert system; human linguistic performance; learning; natural language systems evaluation; number scale; parse; pattern-recognition programs; text-processing systems; validation techniques; Artificial intelligence; Computer science; Expert systems; Face; Humans; Natural language processing; Natural languages; Pattern recognition; Robustness; Text processing;
Conference_Titel :
Southeastcon '89. Proceedings. Energy and Information Technologies in the Southeast., IEEE
Conference_Location :
Columbia, SC
DOI :
10.1109/SECON.1989.132582