GREPO: A Benchmark for Graph Neural Networks on Repository-Level Bug Localization
- URL: http://arxiv.org/abs/2602.13921v1
- Date: Sat, 14 Feb 2026 23:22:15 GMT
- Title: GREPO: A Benchmark for Graph Neural Networks on Repository-Level Bug Localization
- Authors: Juntong Wang, Libin Chen, Xiyuan Wang, Shijia Kang, Haotong Yang, Da Zheng, Muhan Zhang,
- Abstract summary: Repository-level bug localization is a critical software engineering challenge.<n>GNNs offer a promising alternative due to their ability to model complex, repository-wide dependencies.<n>We introduce GREPO, the first GNN benchmark for repository-scale bug localization tasks.
- Score: 50.009407518866965
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Repository-level bug localization-the task of identifying where code must be modified to fix a bug-is a critical software engineering challenge. Standard Large Language Modles (LLMs) are often unsuitable for this task due to context window limitations that prevent them from processing entire code repositories. As a result, various retrieval methods are commonly used, including keyword matching, text similarity, and simple graph-based heuristics such as Breadth-First Search. Graph Neural Networks (GNNs) offer a promising alternative due to their ability to model complex, repository-wide dependencies; however, their application has been hindered by the lack of a dedicated benchmark. To address this gap, we introduce GREPO, the first GNN benchmark for repository-scale bug localization tasks. GREPO comprises 86 Python repositories and 47294 bug-fixing tasks, providing graph-based data structures ready for direct GNN processing. Our evaluation of various GNN architectures shows outstanding performance compared to established information retrieval baselines. This work highlights the potential of GNNs for bug localization and established GREPO as a foundation resource for future research, The code is available at https://github.com/qingpingmo/GREPO.
Related papers
- Natural Language Summarization Enables Multi-Repository Bug Localization by LLMs in Microservice Architectures [0.23332469289621782]
This work shows that engineered natural language representations can be more effective than raw source code for scalable bug localization.<n> Evaluated on DNext, an industrial system with 46 repositories and 1.1M lines of code, our method achieves Pass@10 of 0.82 and MRR of 0.50.
arXiv Detail & Related papers (2025-12-05T17:42:09Z) - Enhancing Fairness in Autoencoders for Node-Level Graph Anomaly Detection [3.487370856323828]
Graph anomaly detection (GAD) has become an increasingly important task across various domains.<n>However, fairness considerations in GAD remain largely underexplored.<n>We propose textbfDistextbfEntangled textbfCounterfactual textbfAdversarial textbfFair (DECAF)-GAD, a framework that alleviates bias while preserving GAD performance.
arXiv Detail & Related papers (2025-08-14T16:12:15Z) - Efficient Environmental Claim Detection with Hyperbolic Graph Neural Networks [1.7259898169307608]
We explore Graph Neural Networks (GNNs) and Hyperbolic Graph Neural Networks (HGNNs) as lightweight yet effective alternatives to transformer-based models.<n>Our results show that our graph-based models, particularly HGNNs in the poincar'e space (P-HGNNs), achieve performance superior to the state-of-the-art on environmental claim detection.
arXiv Detail & Related papers (2025-02-19T11:04:59Z) - RelGNN: Composite Message Passing for Relational Deep Learning [56.48834369525997]
We introduce RelGNN, a novel GNN framework specifically designed to leverage the unique structural characteristics of the graphs built from relational databases.<n>RelGNN is evaluated on 30 diverse real-world tasks from Relbench (Fey et al., 2024), and achieves state-of-the-art performance on the vast majority tasks, with improvements of up to 25%.
arXiv Detail & Related papers (2025-02-10T18:58:40Z) - Graph Neural Networks Gone Hogwild [14.665528337423249]
Graph neural networks (GNNs) appear to be powerful tools to learn state representations for agents in distributed, decentralized multi-agent systems.<n>GNNs generate catastrophically incorrect predictions when nodes update asynchronously during inference.<n>We identify "implicitly-defined" GNNs as a class of architectures which is provably robust to asynchronous "hogwild" inference.<n>We propose a novel implicitly-defined GNN architecture, which we call an 'energy GNN'
arXiv Detail & Related papers (2024-06-29T17:11:09Z) - Uncertainty in GNN Learning Evaluations: The Importance of a Consistent
Benchmark for Community Detection [4.358468367889626]
We propose a framework to establish a common evaluation protocol for Graph Neural Networks (GNNs)
We motivate and justify it by demonstrating the differences with and without the protocol.
We find that by ensuring the same evaluation criteria is followed, there may be significant differences from the reported performance of methods at this task.
arXiv Detail & Related papers (2023-05-10T10:22:28Z) - A Comprehensive Study on Large-Scale Graph Training: Benchmarking and
Rethinking [124.21408098724551]
Large-scale graph training is a notoriously challenging problem for graph neural networks (GNNs)
We present a new ensembling training manner, named EnGCN, to address the existing issues.
Our proposed method has achieved new state-of-the-art (SOTA) performance on large-scale datasets.
arXiv Detail & Related papers (2022-10-14T03:43:05Z) - RU-Net: Regularized Unrolling Network for Scene Graph Generation [92.95032610978511]
Scene graph generation (SGG) aims to detect objects and predict the relationships between each pair of objects.
Existing SGG methods usually suffer from several issues, including 1) ambiguous object representations, and 2) low diversity in relationship predictions.
We propose a regularized unrolling network (RU-Net) to address both problems.
arXiv Detail & Related papers (2022-05-03T04:21:15Z) - VQ-GNN: A Universal Framework to Scale up Graph Neural Networks using
Vector Quantization [70.8567058758375]
VQ-GNN is a universal framework to scale up any convolution-based GNNs using Vector Quantization (VQ) without compromising the performance.
Our framework avoids the "neighbor explosion" problem of GNNs using quantized representations combined with a low-rank version of the graph convolution matrix.
arXiv Detail & Related papers (2021-10-27T11:48:50Z) - A Biased Graph Neural Network Sampler with Near-Optimal Regret [57.70126763759996]
Graph neural networks (GNN) have emerged as a vehicle for applying deep network architectures to graph and relational data.
In this paper, we build upon existing work and treat GNN neighbor sampling as a multi-armed bandit problem.
We introduce a newly-designed reward function that introduces some degree of bias designed to reduce variance and avoid unstable, possibly-unbounded payouts.
arXiv Detail & Related papers (2021-03-01T15:55:58Z) - Learning Semantic Program Embeddings with Graph Interval Neural Network [7.747173589929493]
We present a new graph neural architecture, called Graph Interval Neural Network (GINN) to tackle the weaknesses of the existing GNN.
GINN generalizes from a curated graph representation obtained through an abstraction method designed to aid models to learn.
We have created a neural bug detector based on GINN to catch null pointer deference bugs in Java code.
arXiv Detail & Related papers (2020-05-18T02:38:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.