Training Language Models for Programming Feedback Using Automated Repair Tools

Loading...
Thumbnail Image

Access rights

openAccess
acceptedVersion

URL

Journal Title

Journal ISSN

Volume Title

A4 Artikkeli konferenssijulkaisussa

Date

Major/Subject

Mcode

Degree programme

Language

en

Pages

Series

Artificial Intelligence in Education: 24th International Conference, AIED 2023, Tokyo, Japan, July 3–7, 2023, Proceedings, pp. 830–835, Lecture Notes in Computer Science ; Volume 13916

Abstract

In introductory programming courses, automated repair tools (ARTs) are used to provide feedback to students struggling with debugging. Most successful ARTs take advantage of context-specific educational data to construct repairs to students’ buggy codes. Recent work in student program repair using large language models (LLMs) has also started to utilize such data. An underexplored area in this field is the use of ARTs in combination with LLMs. In this paper, we propose to transfer the repairing capabilities of existing ARTs to open large language models by finetuning LLMs on ART corrections to buggy codes. We experiment with this approach using three large datasets of Python programs written by novices. Our results suggest that a finetuned LLM provides more reliable and higher-quality repairs than the repair tool used for finetuning the model. This opens venues for further deploying and using educational LLM-based repair techniques.

Description

Keywords

Other note

Citation

Koutcheme, C 2023, Training Language Models for Programming Feedback Using Automated Repair Tools. in N Wang, G Rebolledo-Mendez, N Matsuda, O C Santos & V Dimitrova (eds), Artificial Intelligence in Education : 24th International Conference, AIED 2023, Tokyo, Japan, July 3–7, 2023, Proceedings. Lecture Notes in Computer Science, vol. 13916, Springer, pp. 830–835, International Conference on Artificial Intelligence in Education, Tokyo, Japan, 03/07/2023. https://doi.org/10.1007/978-3-031-36272-9_79