Issue No. 04 - July/August (2010 vol. 36)
DOI Bookmark: http://doi.ieeecomputersociety.org/10.1109/TSE.2009.70
Raymond P.L. Buse , University of Virginia, Charlottesville
Westley R. Weimer , University of Virginia, Charlottesville
In this paper, we explore the concept of code readability and investigate its relation to software quality. With data collected from 120 human annotators, we derive associations between a simple set of local code features and human notions of readability. Using those features, we construct an automated readability measure and show that it can be 80 percent effective and better than a human, on average, at predicting readability judgments. Furthermore, we show that this metric correlates strongly with three measures of software quality: code changes, automated defect reports, and defect log messages. We measure these correlations on over 2.2 million lines of code, as well as longitudinally, over many releases of selected projects. Finally, we discuss the implications of this study on programming language design and engineering practice. For example, our data suggest that comments, in and of themselves, are less important than simple blank lines to local judgments of readability.
Software readability, program understanding, machine learning, software maintenance, code metrics, FindBugs.
W. R. Weimer and R. P. Buse, "Learning a Metric for Code Readability," in IEEE Transactions on Software Engineering, vol. 36, no. , pp. 546-558, 2009.