TrojanedCM: A Repository of Trojaned Large Language Models of Code
- URL: http://arxiv.org/abs/2311.14850v2
- Date: Mon, 11 Dec 2023 20:07:35 GMT
- Title: TrojanedCM: A Repository of Trojaned Large Language Models of Code
- Authors: Aftab Hussain, Md Rafiqul Islam Rabin, Mohammad Amin Alipour
- Abstract summary: TrojanedCM is a publicly available repository of clean and poisoned models of source code.
We provide poisoned models for two code classification tasks (defect detection and clone detection) and a code generation task.
The repository also provides full access to the architecture and parameters of the models, allowing practitioners to investigate different white-box analysis techniques.
- Score: 4.838807847761728
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the rapid growth of research in trojaning deep neural models of source
code, we observe that there is a need of developing a benchmark trojaned models
for testing various trojan detection and unlearning techniques. In this work,
we aim to provide the scientific community with diverse trojaned code models,
that cover a variety of state-of-the-art architectures, on which they can
examine such techniques. We thus present TrojanedCM, a publicly available
repository of clean and poisoned models of source code. We provide poisoned
models for two code classification tasks (defect detection and clone detection)
and a code generation task (text-to-code generation). We finetuned popular
pretrained code models such as CodeBERT, PLBART, CodeT5, CodeT5+, on poisoned
datasets that we generated from benchmark datasets (Devign, BigCloneBench,
CONCODE) for the above mentioned tasks. The repository also provides full
access to the architecture and parameters of the models, allowing practitioners
to investigate different white-box analysis techniques. In addition to the
poisoned models, we also provide a poisoning framework using which
practitioners can deploy various poisoning strategies for the different tasks
and models of source code. All the material are accessible via this link:
https://github.com/UH-SERG/TrojanedCM.
Related papers
- Trojans in Large Language Models of Code: A Critical Review through a Trigger-Based Taxonomy [11.075592348442225]
Large language models (LLMs) have provided a lot of exciting new capabilities in software development.
The opaque nature of these models makes them difficult to reason about and inspect.
This work presents an overview of the current state-of-the-art trojan attacks on large language models of code.
arXiv Detail & Related papers (2024-05-05T06:43:52Z) - Does Your Neural Code Completion Model Use My Code? A Membership Inference Approach [66.51005288743153]
We investigate the legal and ethical issues of current neural code completion models.
We tailor a membership inference approach (termed CodeMI) that was originally crafted for classification tasks.
We evaluate the effectiveness of this adapted approach across a diverse array of neural code completion models.
arXiv Detail & Related papers (2024-04-22T15:54:53Z) - On Trojan Signatures in Large Language Models of Code [4.838807847761728]
Trojan signatures are noticeable differences in the distribution of the trojaned class parameters (weights) and the non-trojaned class parameters of the trojaned model.
Our results suggest that trojan signatures could not generalize to LLMs of code.
This is the first work to examine weight-based trojan signature revelation techniques for large-language models of code.
arXiv Detail & Related papers (2024-02-23T22:48:29Z) - INSPECT: Intrinsic and Systematic Probing Evaluation for Code
Transformers [7.255653248042546]
We use a framework to define 15 probing tasks that exercise surface, syntactic, structural and semantic characteristics of source code.
We probe 8 pre-trained source code models, as well as a natural language model (BERT) as our baseline.
We find that models that incorporate some structural information (such as GraphCodeBERT) have a better representation of source code characteristics.
arXiv Detail & Related papers (2023-12-08T15:21:54Z) - A Survey of Trojans in Neural Models of Source Code: Taxonomy and Techniques [10.810570716212542]
We study literature in Explainable AI and Safe AI to understand poisoning of neural models of code.
We first establish a novel taxonomy for Trojan AI for code, and present a new aspect-based classification of triggers in neural models of code.
arXiv Detail & Related papers (2023-05-05T19:07:09Z) - CodeLMSec Benchmark: Systematically Evaluating and Finding Security
Vulnerabilities in Black-Box Code Language Models [58.27254444280376]
Large language models (LLMs) for automatic code generation have achieved breakthroughs in several programming tasks.
Training data for these models is usually collected from the Internet (e.g., from open-source repositories) and is likely to contain faults and security vulnerabilities.
This unsanitized training data can cause the language models to learn these vulnerabilities and propagate them during the code generation procedure.
arXiv Detail & Related papers (2023-02-08T11:54:07Z) - Poison Attack and Defense on Deep Source Code Processing Models [38.32413592143839]
We present a poison attack framework for source code named CodePoisoner as a strong imaginary enemy.
CodePoisoner can produce compilable even human-imperceptible poison samples and attack models by poisoning the training data.
We propose an effective defense approach named CodeDetector to detect poison samples in the training data.
arXiv Detail & Related papers (2022-10-31T03:06:40Z) - Defending against Model Stealing via Verifying Embedded External
Features [90.29429679125508]
adversaries can steal' deployed models even when they have no training samples and can not get access to the model parameters or structures.
We explore the defense from another angle by verifying whether a suspicious model contains the knowledge of defender-specified emphexternal features.
Our method is effective in detecting different types of model stealing simultaneously, even if the stolen model is obtained via a multi-stage stealing process.
arXiv Detail & Related papers (2021-12-07T03:51:54Z) - COSEA: Convolutional Code Search with Layer-wise Attention [90.35777733464354]
We propose a new deep learning architecture, COSEA, which leverages convolutional neural networks with layer-wise attention to capture the code's intrinsic structural logic.
COSEA can achieve significant improvements over state-of-the-art methods on code search tasks.
arXiv Detail & Related papers (2020-10-19T13:53:38Z) - Cassandra: Detecting Trojaned Networks from Adversarial Perturbations [92.43879594465422]
In many cases, pre-trained models are sourced from vendors who may have disrupted the training pipeline to insert Trojan behaviors into the models.
We propose a method to verify if a pre-trained model is Trojaned or benign.
Our method captures fingerprints of neural networks in the form of adversarial perturbations learned from the network gradients.
arXiv Detail & Related papers (2020-07-28T19:00:40Z) - Odyssey: Creation, Analysis and Detection of Trojan Models [91.13959405645959]
Trojan attacks interfere with the training pipeline by inserting triggers into some of the training samples and trains the model to act maliciously only for samples that contain the trigger.
Existing Trojan detectors make strong assumptions about the types of triggers and attacks.
We propose a detector that is based on the analysis of the intrinsic properties; that are affected due to the Trojaning process.
arXiv Detail & Related papers (2020-07-16T06:55:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.