• DocumentCode
    3077527
  • Title

    Explanatory Debugging: Supporting End-User Debugging of Machine-Learned Programs

  • Author

    Kulesza, Todd ; Stumpf, Simone ; Burnett, Margaret ; Wong, Weng-Keen ; Riche, Yann ; Moore, Travis ; Oberst, Ian ; Shinsel, Amber ; McIntosh, Kevin

  • Author_Institution
    Oregon State Univ., Corvallis, OR, USA
  • fYear
    2010
  • fDate
    21-25 Sept. 2010
  • Firstpage
    41
  • Lastpage
    48
  • Abstract
    Many machine-learning algorithms learn rules of behavior from individual end users, such as task-oriented desktop organizers and handwriting recognizers. These rules form a “program” that tells the computer what to do when future inputs arrive. Little research has explored how an end user can debug these programs when they make mistakes. We present our progress toward enabling end users to debug these learned programs via a Natural Programming methodology. We began with a formative study exploring how users reason about and correct a text-classification program. From the results, we derived and prototyped a concept based on “explanatory debugging”, then empirically evaluated it. Our results contribute methods for exposing a learned program´s logic to end users and for eliciting user corrections to improve the program´s predictions.
  • Keywords
    learning (artificial intelligence); program debugging; end-user debugging; explanatory debugging; handwriting recognizers; machine-learning algorithms; natural programming methodology; task-oriented desktop organizers; text-classification program; Channel hot electron injection; Gallium nitride; High definition video; Visualization; HCI; end-user debugging; machine learning;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Visual Languages and Human-Centric Computing (VL/HCC), 2010 IEEE Symposium on
  • Conference_Location
    Leganes
  • ISSN
    1943-6092
  • Print_ISBN
    978-1-4244-8485-0
  • Type

    conf

  • DOI
    10.1109/VLHCC.2010.15
  • Filename
    5635185