DocumentCode :
993487
Title :
Improving data quality: eliminating dupes & I-D-ing those spurious links
Author :
Lee, Mong Li ; Hsu, Wynne
Author_Institution :
Sch. of Comput., Nat. Univ. of Singapore, Singapore
Volume :
24
Issue :
2
fYear :
2005
Firstpage :
35
Lastpage :
38
Abstract :
Dirty data arise as a result of abbreviations, data entry mistakes, duplicate records, missing fields and so forth. This problem is aggravated when multiple data sources need to be integrated. Data cleaning refers to a series of processes employed to deal with detecting and removing errors and inconsistencies from data. Given the "garbage in, garbage out" principle, clean data is crucial for database integration, data warehousing and data mining. Data cleaning is a necessary step prior to the knowledge discovery process. We have reviewed a knowledge-based framework that provides for the definition of duplicate identification rules. We have described a context-based approach to identify these spurious links in the data.
Keywords :
data integrity; data mining; database management systems; knowledge based systems; abbreviations; data cleaning; data entry mistakes; data integrity; data mining; data quality improvement; dirty data; duplicate records; garbage collection; knowledge discovery; knowledge-based system; spurious link identification; Association rules; Cleaning; Communication networks; Computer networks; Data mining; Database systems; Humans; Intelligent systems; Performance analysis; Sorting;
fLanguage :
English
Journal_Title :
Potentials, IEEE
Publisher :
ieee
ISSN :
0278-6648
Type :
jour
DOI :
10.1109/MP.2005.1462465
Filename :
1462465
Link To Document :
بازگشت