CodeTrans: Towards Cracking the Language of Silicone's Code Through
Self-Supervised Deep Learning and High Performance Computing
- URL: http://arxiv.org/abs/2104.02443v1
- Date: Tue, 6 Apr 2021 11:57:12 GMT
- Title: CodeTrans: Towards Cracking the Language of Silicone's Code Through
Self-Supervised Deep Learning and High Performance Computing
- Authors: Ahmed Elnaggar, Wei Ding, Llion Jones, Tom Gibbs, Tamas Feher,
Christoph Angerer, Silvia Severini, Florian Matthes and Burkhard Rost
- Abstract summary: This paper describes CodeTrans - an encoder-decoder transformer model for tasks in the software engineering domain.
It explores the effectiveness of encoder-decoder transformer models for six software engineering tasks, including thirteen sub-tasks.
CodeTrans outperforms the state-of-the-art models on all the tasks.
- Score: 4.111243115567736
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Currently, a growing number of mature natural language processing
applications make people's life more convenient. Such applications are built by
source code - the language in software engineering. However, the applications
for understanding source code language to ease the software engineering process
are under-researched. Simultaneously, the transformer model, especially its
combination with transfer learning, has been proven to be a powerful technique
for natural language processing tasks. These breakthroughs point out a
promising direction for process source code and crack software engineering
tasks. This paper describes CodeTrans - an encoder-decoder transformer model
for tasks in the software engineering domain, that explores the effectiveness
of encoder-decoder transformer models for six software engineering tasks,
including thirteen sub-tasks. Moreover, we have investigated the effect of
different training strategies, including single-task learning, transfer
learning, multi-task learning, and multi-task learning with fine-tuning.
CodeTrans outperforms the state-of-the-art models on all the tasks. To expedite
future works in the software engineering domain, we have published our
pre-trained models of CodeTrans.
https://github.com/agemagician/CodeTrans
Related papers
- Linguacodus: A Synergistic Framework for Transformative Code Generation in Machine Learning Pipelines [0.0]
We introduce a dynamic pipeline that transforms natural language task descriptions into code through high-level data-shaping instructions.
This paper details the fine-tuning process, and sheds light on how natural language descriptions can be translated into functional code.
We propose an algorithm capable of transforming a natural description of an ML task into code with minimal human interaction.
arXiv Detail & Related papers (2024-03-18T08:58:47Z) - TransformCode: A Contrastive Learning Framework for Code Embedding via Subtree Transformation [9.477734501499274]
We present TransformCode, a novel framework that learns code embeddings in a contrastive learning manner.
Our framework is encoder-agnostic and language-agnostic, which means that it can leverage any encoder model and handle any programming language.
arXiv Detail & Related papers (2023-11-10T09:05:23Z) - ChatDev: Communicative Agents for Software Development [84.90400377131962]
ChatDev is a chat-powered software development framework in which specialized agents are guided in what to communicate.
These agents actively contribute to the design, coding, and testing phases through unified language-based communication.
arXiv Detail & Related papers (2023-07-16T02:11:34Z) - Learning Transformer Programs [78.9509560355733]
We introduce a procedure for training Transformers that are mechanistically interpretable by design.
Instead of compiling human-written programs into Transformers, we design a modified Transformer that can be trained using gradient-based optimization.
The Transformer Programs can automatically find reasonable solutions, performing on par with standard Transformers of comparable size.
arXiv Detail & Related papers (2023-06-01T20:27:01Z) - CodeTF: One-stop Transformer Library for State-of-the-art Code LLM [72.1638273937025]
We present CodeTF, an open-source Transformer-based library for state-of-the-art Code LLMs and code intelligence.
Our library supports a collection of pretrained Code LLM models and popular code benchmarks.
We hope CodeTF is able to bridge the gap between machine learning/generative AI and software engineering.
arXiv Detail & Related papers (2023-05-31T05:24:48Z) - TransCoder: Towards Unified Transferable Code Representation Learning Inspired by Human Skills [31.75121546422898]
We present TransCoder, a unified Transferable fine-tuning strategy for Code representation learning.
We employ a tunable prefix encoder as the meta-learner to capture cross-task and cross-language transferable knowledge.
Our method can lead to superior performance on various code-related tasks and encourage mutual reinforcement.
arXiv Detail & Related papers (2023-05-23T06:59:22Z) - Using Document Similarity Methods to create Parallel Datasets for Code
Translation [60.36392618065203]
Translating source code from one programming language to another is a critical, time-consuming task.
We propose to use document similarity methods to create noisy parallel datasets of code.
We show that these models perform comparably to models trained on ground truth for reasonable levels of noise.
arXiv Detail & Related papers (2021-10-11T17:07:58Z) - On the validity of pre-trained transformers for natural language
processing in the software engineering domain [78.32146765053318]
We compare BERT transformer models trained with software engineering data with transformers based on general domain data.
Our results show that for tasks that require understanding of the software engineering context, pre-training with software engineering data is valuable.
arXiv Detail & Related papers (2021-09-10T08:46:31Z) - Thinking Like Transformers [64.96770952820691]
We propose a computational model for the transformer-encoder in the form of a programming language.
We show how RASP can be used to program solutions to tasks that could conceivably be learned by a Transformer.
We provide RASP programs for histograms, sorting, and Dyck-languages.
arXiv Detail & Related papers (2021-06-13T13:04:46Z) - Empirical Study of Transformers for Source Code [14.904366372190943]
We study the capabilities of Transformers to utilize syntactic information in different tasks.
We show that Transformers are able to make meaningful predictions based purely on syntactic information.
arXiv Detail & Related papers (2020-10-15T19:09:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.