Publication Type
Journal Article
Version
acceptedVersion
Publication Date
8-2025
Abstract
Modern code review is a ubiquitous software quality assurance process aimed at identifying and resolving potential issues (e.g., functional, evolvability) within newly written code. Despite its effectiveness, the process demands large amounts of effort from the human reviewers involved. To help alleviate this workload, researchers have trained various deep learning based language models to imitate human reviewers in providing natural language code reviews for submitted code. Formally, this automation task is known as code review comment generation. Prior work has demonstrated improvements in code review comment generation by leveraging machine learning techniques and neural models, such as transfer learning and the transformer architecture. However, the quality of the model generated reviews remain sub-optimal due to the quality of the open-source code review data used in model training. This is in part due to the data obtained from open-source projects where code reviews are conducted in a public forum, and reviewers possess varying levels of software development experience, potentially affecting the quality of their feedback. To accommodate for this variation, we propose a suite of experience-aware training methods that utilise the reviewers’ past authoring and reviewing experiences as signals for review quality. Specifically, we propose experience-aware loss functions (ELF), which use the reviewers’ authoring and reviewing ownership of a project as weights in the model’s loss function. Through this method, experienced reviewers’ code reviews yield larger influence over the model’s behaviour. Compared to the SOTA model, ELF was able to generate higher quality reviews in terms of accuracy (e.g., +29% applicable comments), informativeness (e.g., +56% suggestions), and issue types discussed (e.g., +129% functional issues identified). The key contribution of this work is the demonstration of how traditional software engineering concepts such as reviewer experience can be integrated into the design of AI-based automated code review models.
Keywords
Code Review, Review Comments, Neural Machine Translation, Natural Language Generation
Discipline
Programming Languages and Compilers | Software Engineering
Research Areas
Intelligent Systems and Optimization
Areas of Excellence
Digital transformation
Publication
ACM Transactions on Software Engineering and Methodology
First Page
1
Last Page
34
ISSN
1049-331X
Identifier
10.1145/3762183
Publisher
Association for Computing Machinery (ACM)
Citation
LIN, Hong Yi; THONGTANUNAM, Patanamon; TREUDE, Christoph; GODFREY, Michael W.; LIU, Chunhua; and CHAROENWET, Wachiraphan.
Leveraging reviewer experience in code review comment generation. (2025). ACM Transactions on Software Engineering and Methodology. 1-34.
Available at: https://ink.library.smu.edu.sg/sis_research/10516
Creative Commons License

This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.1145/3762183