DocumentCode :
2138122
Title :
Exploring error-sensitive attributes and branches of decision trees
Author :
Wu, Wenchuan
Author_Institution :
Fac. of Eng. & Inf. Technol., Univ. of Technol., Sydney, NSW, Australia
fYear :
2013
fDate :
23-25 July 2013
Firstpage :
929
Lastpage :
934
Abstract :
Decision trees have a reputation of being efficient and illustrative in classification learning, and majority of the research effort has been focused on making classification improvement in a head-on style with wide-range research topics, such as tree algorithm development and refinement, attribute selection and prioritization, sampling technique improvement, and the addition of cost matrix and other performance-enhancing factors. One less commonly studied topic is about the characteristics of classification errors and how they may be associated with specific attributes due to correlation or causation, and within what value ranges on such attributes when pattern are most likely. This research intends to study this dim area in a sort-of reverse and forensic style as part of post-classification analysis, to analyze the patterns and relationship between errors and attributes, to explore how attributes´ risk level in error may play a role in leading to more risky, more error-prone decision tree branches or decision paths. Possible benefits from this study would include raising data stakeholders´ awareness of such specific error-sensitive attributes and decision paths, to facilitate better understanding of possible causes and impact of errors and the development of more effective error-reduction measures customized to suit the specific patterns and individual datasets. Such emphasis on highlighting the specific error-sensitive attributes and decision branches within individual datasets is a reflection of our observation which shared by others - “additional domain-specific knowledge, external to the training set, must be employed to estimate the noise level (... and) the underlying model´s complexity ... (because) knowledge-poor tree induction algorithms do not exploit such information.” [2].
Keywords :
decision trees; learning (artificial intelligence); matrix algebra; pattern classification; attribute selection; classification learning; cost matrix; decision path; decision trees; error-reduction measures; error-sensitive attribute; forensic style; knowledge-poor tree induction algorithm; noise level; post-classification analysis; reverse style; sampling technique improvement; tree algorithm development; tree algorithm refinement; Accuracy; Data models; Decision trees; Error analysis; Magnesium; Measurement uncertainty; Radiation detectors; decision tree; error-sensitive attribute; error-sensitive tree branch; feature selection; post-classification analysis;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Natural Computation (ICNC), 2013 Ninth International Conference on
Conference_Location :
Shenyang
Type :
conf
DOI :
10.1109/ICNC.2013.6818109
Filename :
6818109
Link To Document :
بازگشت