A Survey of Trojans in Neural Models of Source Code: Taxonomy and Techniques
- URL: http://arxiv.org/abs/2305.03803v5
- Date: Thu, 18 Apr 2024 19:41:54 GMT
- Title: A Survey of Trojans in Neural Models of Source Code: Taxonomy and Techniques
- Authors: Aftab Hussain, Md Rafiqul Islam Rabin, Toufique Ahmed, Navid Ayoobi, Bowen Xu, Prem Devanbu, Mohammad Amin Alipour,
- Abstract summary: We study literature in Explainable AI and Safe AI to understand poisoning of neural models of code.
We first establish a novel taxonomy for Trojan AI for code, and present a new aspect-based classification of triggers in neural models of code.
- Score: 10.810570716212542
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we study literature in Explainable AI and Safe AI to understand poisoning of neural models of code. In order to do so, we first establish a novel taxonomy for Trojan AI for code, and present a new aspect-based classification of triggers in neural models of code. Next, we highlight recent works that help us deepen our conception of how these models understand software code. Then we pick some of the recent, state-of-art poisoning strategies that can be used to manipulate such models. The insights we draw can potentially help to foster future research in the area of Trojan AI for code.
Related papers
- AI-Aided Kalman Filters [65.35350122917914]
The Kalman filter (KF) and its variants are among the most celebrated algorithms in signal processing.
Recent developments illustrate the possibility of fusing deep neural networks (DNNs) with classic Kalman-type filtering.
This article provides a tutorial-style overview of design approaches for incorporating AI in aiding KF-type algorithms.
arXiv Detail & Related papers (2024-10-16T06:47:53Z) - Trojans in Large Language Models of Code: A Critical Review through a Trigger-Based Taxonomy [11.075592348442225]
Large language models (LLMs) have provided a lot of exciting new capabilities in software development.
The opaque nature of these models makes them difficult to reason about and inspect.
This work presents an overview of the current state-of-the-art trojan attacks on large language models of code.
arXiv Detail & Related papers (2024-05-05T06:43:52Z) - TrojanedCM: A Repository of Trojaned Large Language Models of Code [4.838807847761728]
TrojanedCM is a publicly available repository of clean and poisoned models of source code.
We provide poisoned models for two code classification tasks (defect detection and clone detection) and a code generation task.
The repository also provides full access to the architecture and parameters of the models, allowing practitioners to investigate different white-box analysis techniques.
arXiv Detail & Related papers (2023-11-24T21:58:06Z) - Attention-Enhancing Backdoor Attacks Against BERT-based Models [54.070555070629105]
Investigating the strategies of backdoor attacks will help to understand the model's vulnerability.
We propose a novel Trojan Attention Loss (TAL) which enhances the Trojan behavior by directly manipulating the attention patterns.
arXiv Detail & Related papers (2023-10-23T01:24:56Z) - Neurosymbolic AI and its Taxonomy: a survey [48.7576911714538]
Neurosymbolic AI deals with models that combine symbolic processing, like classic AI, and neural networks.
This survey investigates research papers in this area during recent years and brings classification and comparison between the presented models as well as applications.
arXiv Detail & Related papers (2023-05-12T19:51:13Z) - Implementing engrams from a machine learning perspective: matching for
prediction [0.0]
We propose how we might design a computer system to implement engrams using neural networks.
Building on autoencoders, we propose latent neural spaces as indexes for storing and retrieving information in a compressed format.
We consider how different states in latent neural spaces corresponding to different types of sensory input could be linked by synchronous activation.
arXiv Detail & Related papers (2023-03-01T10:05:40Z) - Explainable AI for Pre-Trained Code Models: What Do They Learn? When
They Do Not Work? [4.573310303307945]
We study two recent large language models (LLMs) for code on a set of software engineering downstream tasks.
We identify what CodeBERT and GraphCodeBERT learn (put the highest attention on, in terms of source code token types) on these tasks.
We show some of the common patterns when the model does not work as expected and suggest recommendations.
arXiv Detail & Related papers (2022-11-23T10:07:20Z) - Neuromorphic Artificial Intelligence Systems [58.1806704582023]
Modern AI systems, based on von Neumann architecture and classical neural networks, have a number of fundamental limitations in comparison with the brain.
This article discusses such limitations and the ways they can be mitigated.
It presents an overview of currently available neuromorphic AI projects in which these limitations are overcome.
arXiv Detail & Related papers (2022-05-25T20:16:05Z) - Learning to map source code to software vulnerability using
code-as-a-graph [67.62847721118142]
We explore the applicability of Graph Neural Networks in learning the nuances of source code from a security perspective.
We show that a code-as-graph encoding is more meaningful for vulnerability detection than existing code-as-photo and linear sequence encoding approaches.
arXiv Detail & Related papers (2020-06-15T16:05:27Z) - Scalable Backdoor Detection in Neural Networks [61.39635364047679]
Deep learning models are vulnerable to Trojan attacks, where an attacker can install a backdoor during training time to make the resultant model misidentify samples contaminated with a small trigger patch.
We propose a novel trigger reverse-engineering based approach whose computational complexity does not scale with the number of labels, and is based on a measure that is both interpretable and universal across different network and patch types.
In experiments, we observe that our method achieves a perfect score in separating Trojaned models from pure models, which is an improvement over the current state-of-the art method.
arXiv Detail & Related papers (2020-06-10T04:12:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.