Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

CoTexT: Multi-task Learning with Code-Text Transformer

About

We present CoTexT, a pre-trained, transformer-based encoder-decoder model that learns the representative context between natural language (NL) and programming language (PL). Using self-supervision, CoTexT is pre-trained on large programming language corpora to learn a general understanding of language and code. CoTexT supports downstream NL-PL tasks such as code summarizing/documentation, code generation, defect detection, and code debugging. We train CoTexT on different combinations of available PL corpus including both "bimodal" and "unimodal" data. Here, bimodal data is the combination of text and corresponding code snippets, whereas unimodal data is merely code snippets. We first evaluate CoTexT with multi-task learning: we perform Code Summarization on 6 different programming languages and Code Refinement on both small and medium size featured in the CodeXGLUE dataset. We further conduct extensive experiments to investigate CoTexT on other tasks within the CodeXGlue dataset, including Code Generation and Defect Detection. We consistently achieve SOTA results in these tasks, demonstrating the versatility of our models.

Long Phan, Hieu Tran, Daniel Le, Hieu Nguyen, James Anibal, Alec Peltekian, Yanfang Ye• 2021

Related benchmarks

TaskDatasetResultRank
Code SummarizationCodeXGLUE (test)
Overall Score18.55
7
Code RepairTufano Medium Abstract 2019 (test)
Top-1 Exact Match Accuracy15.36
6
Code RepairTufano Small Abstract 2019 (test)
Top-1 Accuracy22.64
6
Showing 3 of 3 rows

Other info

Follow for update