Detail View

DC Field Value Language
dc.contributor.author Tang, Xunzhu -
dc.contributor.author Tian, Haoye -
dc.contributor.author Pian, Weiguo -
dc.contributor.author Ezzini, Saad -
dc.contributor.author Kabore, Abdoul Kader -
dc.contributor.author Habib, Andrew -
dc.contributor.author Klein, Jacques -
dc.contributor.author Bissyande, Tegawende F. -
dc.contributor.author Kim, Kisub -
dc.date.accessioned 2026-02-10T10:40:13Z -
dc.date.available 2026-02-10T10:40:13Z -
dc.date.created 2025-12-26 -
dc.date.issued ACCEPT -
dc.identifier.issn 1382-3256 -
dc.identifier.uri https://scholar.dgist.ac.kr/handle/20.500.11750/60013 -
dc.description.abstract Code change representation plays a pivotal role in automating numerous software engineering tasks, such as classifying code change correctness or generating natural language summaries of code changes. Recent studies have leveraged deep learning to derive effective code change representation, primarily focusing on capturing changes in token sequences or Abstract Syntax Trees (ASTs). However, these current state-of-the-art representations do not explicitly calculate the intention semantic induced by the change on the AST, nor do they effectively explore the surrounding contextual information of the modified lines. To address this, we propose a new code change representation methodology, Patcherizer, which we refer to as our tool. This innovative approach explores the intention features of the context and structure, combining the context around the code change along with two novel representations. These new representations capture the sequence intention inside the code changes in the code change and the graph intention inside the structural changes of AST graphs before and after the code change. This comprehensive representation allows us to better capture the intentions underlying a code change. Patcherizer builds on graph convolutional neural networks for the structural input representation of the intention graph and on transformers for the intention sequence representation. We assess the generalizability of Patcherizer 's learned embeddings on three tasks: (1) Generating code change description in NL, (2) Predicting code change correctness in program repair, and (3) Code change intention detection. Experimental results show that the learned code change representation is effective for all three tasks and achieves superior performance to the state-of-the-art (SOTA) approaches. For instance, on the popular task of code change description generation (a.k.a. commit message generation), Patcherizer achieves an average improvement of 19.39%, 8.71%, and 34.03% in terms of BLEU, ROUGE-L, and METEOR metrics, respectively. -
dc.language English -
dc.publisher Springer Nature -
dc.title Learning to represent code changes -
dc.type Article -
dc.identifier.doi 10.1007/s10664-025-10763-6 -
dc.identifier.wosid 001641067500001 -
dc.identifier.scopusid 2-s2.0-105025197401 -
dc.identifier.bibliographicCitation Empirical Software Engineering -
dc.description.isOpenAccess FALSE -
dc.subject.keywordAuthor Code change correctness -
dc.subject.keywordAuthor Message generation -
dc.subject.keywordAuthor Code change representation -
dc.subject.keywordPlus COMMIT MESSAGES -
dc.subject.keywordPlus MODELS -
dc.citation.title Empirical Software Engineering -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.relation.journalResearchArea Computer Science -
dc.relation.journalWebOfScienceCategory Computer Science, Software Engineering -
dc.type.docType Article -
Show Simple Item Record

File Downloads

  • There are no files associated with this item.

공유

qrcode
공유하기

Related Researcher

김기섭
Kim, Kisub김기섭

Department of Electrical Engineering and Computer Science

read more

Total Views & Downloads

???jsp.display-item.statistics.view???: , ???jsp.display-item.statistics.download???: