• DocumentCode
    731496
  • Title

    Will They Like This? Evaluating Code Contributions with Language Models

  • Author

    Hellendoorn, Vincent J. ; Devanbu, Premkumar T. ; Bacchelli, Alberto

  • Author_Institution
    SORCERERS @ Software Eng. Res. Group, Delft Univ. of Technol., Delft, Netherlands
  • fYear
    2015
  • fDate
    16-17 May 2015
  • Firstpage
    157
  • Lastpage
    167
  • Abstract
    Popular open-source software projects receive and review contributions from a diverse array of developers, many of whom have little to no prior involvement with the project. A recent survey reported that reviewers consider conformance to the project´s code style to be one of the top priorities when evaluating code contributions on Github. We propose to quantitatively evaluate the existence and effects of this phenomenon. To this aim we use language models, which were shown to accurately capture stylistic aspects of code. We find that rejected change sets do contain code significantly less similar to the project than accepted ones, furthermore, the less similar change sets are more likely to be subject to thorough review. Armed with these results we further investigate whether new contributors learn to conform to the project style and find that experience is positively correlated with conformance to the project´s code style.
  • Keywords
    public domain software; software engineering; Github; code contributions; language models; open-source software projects; software development; Context; Context modeling; Data mining; Entropy; Java; Mathematical model; Software; code review; language model; pull request;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Mining Software Repositories (MSR), 2015 IEEE/ACM 12th Working Conference on
  • Conference_Location
    Florence
  • Type

    conf

  • DOI
    10.1109/MSR.2015.22
  • Filename
    7180076