Automatic Code Generation using Pre-Trained Language Models

02/21/2021
by   Luis Perez, et al.
0

Recent advancements in natural language processing <cit.> <cit.> have led to near-human performance in multiple natural language tasks. In this paper, we seek to understand whether similar techniques can be applied to a highly structured environment with strict syntax rules. Specifically, we propose an end-to-end machine learning model for code generation in the Python language built on-top of pre-trained language models. We demonstrate that a fine-tuned model can perform well in code generation tasks, achieving a BLEU score of 0.22, an improvement of 46% over a reasonable sequence-to-sequence baseline. All results and related code used for training and data processing are available on GitHub.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset