FAIR: Flow Type-Aware Pre-Training of Compiler Intermediate
Representations
- URL: http://arxiv.org/abs/2309.04828v1
- Date: Sat, 9 Sep 2023 15:51:49 GMT
- Title: FAIR: Flow Type-Aware Pre-Training of Compiler Intermediate
Representations
- Authors: Changan Niu, Chuanyi Li, Vincent Ng, David Lo, Bin Luo
- Abstract summary: We propose a Flow type-Aware pre-trained model for compiler intermediate representations (IRs)
We specifically propose to enable FAIR to learn the semantics of IR tokens, flow type information, and the overall representation of IR.
Experimental results show that FAIR can achieve state-of-the-art results on four code-related downstream tasks.
- Score: 36.030609139210426
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: While the majority of existing pre-trained models from code learn source code
features such as code tokens and abstract syntax trees, there are some other
works that focus on learning from compiler intermediate representations (IRs).
Existing IR-based models typically utilize IR features such as instructions,
control and data flow graphs (CDFGs), call graphs, etc. However, these methods
confuse variable nodes and instruction nodes in a CDFG and fail to distinguish
different types of flows, and the neural networks they use fail to capture
long-distance dependencies and have over-smoothing and over-squashing problems.
To address these weaknesses, we propose FAIR, a Flow type-Aware pre-trained
model for IR that involves employing (1) a novel input representation of IR
programs; (2) Graph Transformer to address over-smoothing, over-squashing and
long-dependencies problems; and (3) five pre-training tasks that we
specifically propose to enable FAIR to learn the semantics of IR tokens, flow
type information, and the overall representation of IR. Experimental results
show that FAIR can achieve state-of-the-art results on four code-related
downstream tasks.
Related papers
- Can Large Language Models Understand Intermediate Representations? [17.033963652676164]
This paper investigates the capabilities of Large Language Models (LLMs) in understanding Intermediate Representations (IRs)
We analyze their performance across four tasks: Control Flow Graph (CFG) reconstruction, decompilation, code summarization, and execution reasoning.
The study recommends fine-tuning on structured IR datasets and integration of explicit control flow models to augment their comprehension and handling of IR-related tasks.
arXiv Detail & Related papers (2025-02-07T17:23:48Z) - Combining Induction and Transduction for Abstract Reasoning [13.399370315305408]
We train neural models for induction (inferring latent functions) and transduction (directly predicting the test output for a given test input) on ARC.
We find inductive and transductive models solve different kinds of test problems, despite having the same training problems and sharing the same neural architecture.
arXiv Detail & Related papers (2024-11-04T17:03:55Z) - Robust Neural Information Retrieval: An Adversarial and Out-of-distribution Perspective [111.58315434849047]
robustness of neural information retrieval models (IR) models has garnered significant attention.
We view the robustness of IR to be a multifaceted concept, emphasizing its necessity against adversarial attacks, out-of-distribution (OOD) scenarios and performance variance.
We provide an in-depth discussion of existing methods, datasets, and evaluation metrics, shedding light on challenges and future directions in the era of large language models.
arXiv Detail & Related papers (2024-07-09T16:07:01Z) - Towards Learning Abductive Reasoning using VSA Distributed Representations [56.31867341825068]
We introduce the Abductive Rule Learner with Context-awareness (ARLC) model.
ARLC features a novel and more broadly applicable training objective for abductive reasoning.
We show ARLC's robustness to post-programming training by incrementally learning from examples on top of programmed knowledge.
arXiv Detail & Related papers (2024-06-27T12:05:55Z) - Visual Delta Generator with Large Multi-modal Models for Semi-supervised Composed Image Retrieval [50.72924579220149]
Composed Image Retrieval (CIR) is a task that retrieves images similar to a query, based on a provided textual modification.
Current techniques rely on supervised learning for CIR models using labeled triplets of the reference image, text, target image.
We propose a new semi-supervised CIR approach where we search for a reference and its related target images in auxiliary data.
arXiv Detail & Related papers (2024-04-23T21:00:22Z) - Probabilistic Abduction for Visual Abstract Reasoning via Learning Rules
in Vector-symbolic Architectures [22.12114509953737]
Abstract reasoning is a cornerstone of human intelligence, and replicating it with artificial intelligence (AI) presents an ongoing challenge.
This study focuses on efficiently solving Raven's progressive matrices (RPM), a visual test for assessing abstract reasoning abilities.
Instead of hard-coding the rule formulations associated with RPMs, our approach can learn the VSA rule formulations with just one pass through the training data.
arXiv Detail & Related papers (2024-01-29T10:17:18Z) - DiffIR: Efficient Diffusion Model for Image Restoration [108.82579440308267]
Diffusion model (DM) has achieved SOTA performance by modeling the image synthesis process into a sequential application of a denoising network.
Traditional DMs running massive iterations on a large model to estimate whole images or feature maps is inefficient for image restoration.
We propose DiffIR, which consists of a compact IR prior extraction network (CPEN), dynamic IR transformer (DIRformer), and denoising network.
arXiv Detail & Related papers (2023-03-16T16:47:14Z) - Generalization on the Unseen, Logic Reasoning and Degree Curriculum [25.7378861650474]
This paper considers the learning of logical (Boolean) functions with a focus on the generalization on the unseen (GOTU) setting.
We study how different network architectures trained by (S)GD perform under GOTU.
More specifically, this means an interpolator of the training data that has minimal Fourier mass on the higher degree basis elements.
arXiv Detail & Related papers (2023-01-30T17:44:05Z) - Entity-Conditioned Question Generation for Robust Attention Distribution
in Neural Information Retrieval [51.53892300802014]
We show that supervised neural information retrieval models are prone to learning sparse attention patterns over passage tokens.
Using a novel targeted synthetic data generation method, we teach neural IR to attend more uniformly and robustly to all entities in a given passage.
arXiv Detail & Related papers (2022-04-24T22:36:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.