From Performance to Understanding: A Vision for Explainable Automated Algorithm Design
- URL: http://arxiv.org/abs/2511.16201v1
- Date: Thu, 20 Nov 2025 10:08:40 GMT
- Title: From Performance to Understanding: A Vision for Explainable Automated Algorithm Design
- Authors: Niki van Stein, Anna V. Kononova, Thomas Bäck,
- Abstract summary: This paper argues that the next breakthrough will come not from more automation, but from coupling automation with understanding.<n>We outline a vision for explainable automated algorithm design, built on three pillars.<n>These elements form a closed knowledge loop in which discovery, explanation and generalisation reinforce each other.
- Score: 2.780126715305372
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automated algorithm design is entering a new phase: Large Language Models can now generate full optimisation (meta)heuristics, explore vast design spaces and adapt through iterative feedback. Yet this rapid progress is largely performance-driven and opaque. Current LLM-based approaches rarely reveal why a generated algorithm works, which components matter or how design choices relate to underlying problem structures. This paper argues that the next breakthrough will come not from more automation, but from coupling automation with understanding from systematic benchmarking. We outline a vision for explainable automated algorithm design, built on three pillars: (i) LLM-driven discovery of algorithmic variants, (ii) explainable benchmarking that attributes performance to components and hyperparameters and (iii) problem-class descriptors that connect algorithm behaviour to landscape structure. Together, these elements form a closed knowledge loop in which discovery, explanation and generalisation reinforce each other. We argue that this integration will shift the field from blind search to interpretable, class-specific algorithm design, accelerating progress while producing reusable scientific insight into when and why optimisation strategies succeed.
Related papers
- Landscape-aware Automated Algorithm Design: An Efficient Framework for Real-world Optimization [32.203665659052845]
Large Language Models (LLMs) have opened new frontiers in automated algorithm design.<n>LLMs require extensive evaluation of the target problem to guide the search process.<n>This research proposes an innovative and efficient framework that decouples algorithm discovery from high-cost evaluation.
arXiv Detail & Related papers (2026-02-04T13:18:45Z) - LLaMEA-SAGE: Guiding Automated Algorithm Design with Structural Feedback from Explainable AI [4.440668887299803]
Large language models have enabled automated algorithm design (AAD) by generating optimization algorithms directly from natural-language prompts.<n>We propose a mechanism for guiding AAD using feedback constructed from graph-theoretic and complexity features extracted from the abstract syntax trees of the generated algorithms.<n>We show that the proposed structured guidance achieves the same performance faster than vanilla LLaMEA in a small controlled experiment.
arXiv Detail & Related papers (2026-01-29T10:27:29Z) - Experience-Guided Reflective Co-Evolution of Prompts and Heuristics for Automatic Algorithm Design [124.54166764570972]
Combinatorial optimization problems are traditionally tackled with handcrafted algorithms.<n>Recent progress has highlighted the potential of automatics design powered by large language models.<n>We propose the Experience-Evolution Reflective Co-Guided of Prompt and Heuristics (EvoPH) for automatic algorithm design.
arXiv Detail & Related papers (2025-09-29T09:24:09Z) - Evolving Prompts In-Context: An Open-ended, Self-replicating Perspective [65.12150411762273]
We show that pruning random demonstrations into seemingly incoherent "gibberish" can remarkably improve performance across diverse tasks.<n>We propose a self-discover prompt optimization framework, PromptQuine, that automatically searches for the pruning strategy by itself using only low-data regimes.
arXiv Detail & Related papers (2025-06-22T07:53:07Z) - From Understanding to Excelling: Template-Free Algorithm Design through Structural-Functional Co-Evolution [39.42526347710991]
Large language models (LLMs) have greatly accelerated the automation of algorithm generation and optimization.<n>We introduce an end-to-end algorithm generation and optimization framework based on LLMs.<n>Our approach utilizes the deep semantic understanding of LLMs to convert natural language requirements or human-authored papers into code solutions.
arXiv Detail & Related papers (2025-03-13T08:26:18Z) - Designing Algorithms Empowered by Language Models: An Analytical Framework, Case Studies, and Insights [86.06371692309972]
This work presents an analytical framework for the design and analysis of large language models (LLMs)-based algorithms.<n>Our proposed framework serves as an attempt to mitigate such headaches.
arXiv Detail & Related papers (2024-07-20T07:39:07Z) - From Decoding to Meta-Generation: Inference-time Algorithms for Large Language Models [63.188607839223046]
This survey focuses on the benefits of scaling compute during inference.
We explore three areas under a unified mathematical formalism: token-level generation algorithms, meta-generation algorithms, and efficient generation.
arXiv Detail & Related papers (2024-06-24T17:45:59Z) - LLaMEA: A Large Language Model Evolutionary Algorithm for Automatically Generating Metaheuristics [0.023020018305241332]
This paper introduces a novel Large Language Model Evolutionary Algorithm (LLaMEA) framework.<n>Given a set of criteria and a task definition (the search space), LLaMEA iteratively generates, mutates and selects algorithms.<n>We show how this framework can be used to generate novel black-box metaheuristic optimization algorithms automatically.
arXiv Detail & Related papers (2024-05-30T15:10:59Z) - AutoOpt: A General Framework for Automatically Designing Metaheuristic Optimization Algorithms with Diverse Structures [22.624811044236516]
This paper proposes a general framework, AutoOpt, for automatically designing metaheuristic algorithms with diverse structures.<n>A general algorithm prototype dedicated to covering the metaheuristic family as widely as possible.<n>A directed acyclic graph algorithm representation to fit the proposed prototype.<n>A graph representation embedding method offering an alternative compact form of the graph to be manipulated.
arXiv Detail & Related papers (2022-04-03T05:31:56Z) - Investigating Bi-Level Optimization for Learning and Vision from a
Unified Perspective: A Survey and Beyond [114.39616146985001]
In machine learning and computer vision fields, despite the different motivations and mechanisms, a lot of complex problems contain a series of closely related subproblms.
In this paper, we first uniformly express these complex learning and vision problems from the perspective of Bi-Level Optimization (BLO)
Then we construct a value-function-based single-level reformulation and establish a unified algorithmic framework to understand and formulate mainstream gradient-based BLO methodologies.
arXiv Detail & Related papers (2021-01-27T16:20:23Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.