DocumentCode
2398545
Title
Cheap, Fast, and Good Enough for the Non-biomedical Domain but is It Usable for Clinical Natural Language Processing? Evaluating Crowdsourcing for Clinical Trial Announcement Named Entity Annotations
Author
Zhai, Haijun ; Lingren, Todd ; Deleger, Louise ; Li, Qi ; Kaiser, Megan ; Stoutenborough, Laura ; Solti, Imre
Author_Institution
Div. of Biomed. Inf., Cincinnati Children´´s Hosp. Med. Center, Cincinnati, OH, USA
fYear
2012
fDate
27-28 Sept. 2012
Firstpage
106
Lastpage
106
Abstract
Building upon previous work from the general crowdsourcing research, this study investigates the usability of crowdsourcing in the clinical NLP domain for annotating medical named entities and entity linkages in a clinical trial announcement (CTA) corpus. The results indicate that crowdsourcing is a feasible, inexpensive, fast, and practical approach to annotate clinical text (without PHI) on large scale for medical named entities. The crowdsourcing program code was released publicly.
Keywords
information retrieval; medical computing; natural language processing; outsourcing; text analysis; CTA corpus; clinical NLP domain; clinical natural language processing; clinical text annotation; clinical trial announcement corpus; crowdsourcing evaluation; crowdsourcing program code; crowdsourcing usability; entity linkages; medical named entity annotation; Biomedical imaging; Clinical trials; Hospitals; Joining processes; Natural language processing; Pediatrics; Usability;
fLanguage
English
Publisher
ieee
Conference_Titel
Healthcare Informatics, Imaging and Systems Biology (HISB), 2012 IEEE Second International Conference on
Conference_Location
San Diego, CA
Print_ISBN
978-1-4673-4803-4
Type
conf
DOI
10.1109/HISB.2012.31
Filename
6366196
Link To Document