Large language models for source code generation and editing
Zapiski Nauchnykh Seminarov POMI, Investigations on applied mathematics and informatics. Part IV, Tome 540 (2024), pp. 276-350
Voir la notice de l'article provenant de la source Math-Net.Ru
Lomshakov V. M., Nikolenko S. I. Large Language Models for Source Code Generation and Editing. In recent years, large language models (LLMs) have significantly transformed approaches to the automation of software development, providing powerful tools for code generation, correction, and optimization. In this survey, we examine methods for adapting LLMs to programming tasks, including reinforcement learning from human feedback (RLHF), instruction tuning, parameter-efficient fine-tuning (PEFT), and effective prompting strategies. We review modern approaches for fine-tuning and LLM applications, discuss their advantages and limitations, consider relevant datasets for code generation and correction tasks and the corresponding evaluation metrics. Additionally, we describe state of the art open weight models for working with source code.
@article{ZNSL_2024_540_a14,
author = {V. M. Lomshakov and S. I. Nikolenko},
title = {Large language models for source code generation and editing},
journal = {Zapiski Nauchnykh Seminarov POMI},
pages = {276--350},
publisher = {mathdoc},
volume = {540},
year = {2024},
language = {ru},
url = {http://geodesic.mathdoc.fr/item/ZNSL_2024_540_a14/}
}
V. M. Lomshakov; S. I. Nikolenko. Large language models for source code generation and editing. Zapiski Nauchnykh Seminarov POMI, Investigations on applied mathematics and informatics. Part IV, Tome 540 (2024), pp. 276-350. http://geodesic.mathdoc.fr/item/ZNSL_2024_540_a14/