Learning compositional programs with arguments and sampling
- URL: http://arxiv.org/abs/2109.00619v1
- Date: Wed, 1 Sep 2021 21:27:41 GMT
- Title: Learning compositional programs with arguments and sampling
- Authors: Giovanni De Toni, Luca Erculiani, Andrea Passerini
- Abstract summary: We train a machine learning model to discover a program that satisfies specific requirements.
We extend upon a state of the art model, AlphaNPI, by learning to generate functions that can accept arguments.
- Score: 12.790055619773565
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: One of the most challenging goals in designing intelligent systems is
empowering them with the ability to synthesize programs from data. Namely,
given specific requirements in the form of input/output pairs, the goal is to
train a machine learning model to discover a program that satisfies those
requirements. A recent class of methods exploits combinatorial search
procedures and deep learning to learn compositional programs. However, they
usually generate only toy programs using a domain-specific language that does
not provide any high-level feature, such as function arguments, which reduces
their applicability in real-world settings. We extend upon a state of the art
model, AlphaNPI, by learning to generate functions that can accept arguments.
This improvement will enable us to move closer to real computer programs.
Moreover, we investigate employing an Approximate version of Monte Carlo Tree
Search (A-MCTS) to speed up convergence. We showcase the potential of our
approach by learning the Quicksort algorithm, showing how the ability to deal
with arguments is crucial for learning and generalization.
Related papers
- Learning to Reason via Program Generation, Emulation, and Search [33.11955431589091]
Program synthesis with language models (LMs) has unlocked a large set of reasoning abilities.
Not all reasoning tasks are easily expressible as code, e.g. tasks involving commonsense reasoning, moral decision-making, and sarcasm understanding.
We propose Code Generation and Emulated EXecution (CoGEX) to extend an LM's program synthesis skills to such tasks.
arXiv Detail & Related papers (2024-05-25T19:40:50Z) - Learning a Hierarchical Planner from Humans in Multiple Generations [21.045112705349222]
We present natural programming, a library learning system that combines programmatic learning with a hierarchical planner.
A user teaches the system via curriculum building, by identifying a challenging yet not impossible goal.
The system solves for the goal via hierarchical planning, using the linguistic hints to guide its probability distribution.
arXiv Detail & Related papers (2023-10-17T22:28:13Z) - When Do Program-of-Thoughts Work for Reasoning? [51.2699797837818]
We propose complexity-impacted reasoning score (CIRS) to measure correlation between code and reasoning abilities.
Specifically, we use the abstract syntax tree to encode the structural information and calculate logical complexity.
Code will be integrated into the EasyInstruct framework at https://github.com/zjunlp/EasyInstruct.
arXiv Detail & Related papers (2023-08-29T17:22:39Z) - Leveraging Language to Learn Program Abstractions and Search Heuristics [66.28391181268645]
We introduce LAPS (Language for Abstraction and Program Search), a technique for using natural language annotations to guide joint learning of libraries and neurally-guided search models for synthesis.
When integrated into a state-of-the-art library learning system (DreamCoder), LAPS produces higher-quality libraries and improves search efficiency and generalization.
arXiv Detail & Related papers (2021-06-18T15:08:47Z) - Representing Partial Programs with Blended Abstract Semantics [62.20775388513027]
We introduce a technique for representing partially written programs in a program synthesis engine.
We learn an approximate execution model implemented as a modular neural network.
We show that these hybrid neuro-symbolic representations enable execution-guided synthesizers to use more powerful language constructs.
arXiv Detail & Related papers (2020-12-23T20:40:18Z) - Latent Programmer: Discrete Latent Codes for Program Synthesis [56.37993487589351]
In many sequence learning tasks, such as program synthesis and document summarization, a key problem is searching over a large space of possible output sequences.
We propose to learn representations of the outputs that are specifically meant for search: rich enough to specify the desired output but compact enough to make search more efficient.
We introduce the emphLatent Programmer, a program synthesis method that first predicts a discrete latent code from input/output examples, and then generates the program in the target language.
arXiv Detail & Related papers (2020-12-01T10:11:35Z) - BUSTLE: Bottom-Up Program Synthesis Through Learning-Guided Exploration [72.88493072196094]
We present a new synthesis approach that leverages learning to guide a bottom-up search over programs.
In particular, we train a model to prioritize compositions of intermediate values during search conditioned on a set of input-output examples.
We show that the combination of learning and bottom-up search is remarkably effective, even with simple supervised learning approaches.
arXiv Detail & Related papers (2020-07-28T17:46:18Z) - Learning Differentiable Programs with Admissible Neural Heuristics [43.54820901841979]
We study the problem of learning differentiable functions expressed as programs in a domain-specific language.
We frame this optimization problem as a search in a weighted graph whose paths encode top-down derivations of program syntax.
Our key innovation is to view various classes of neural networks as continuous relaxations over the space of programs.
arXiv Detail & Related papers (2020-07-23T16:07:39Z) - Can We Learn Heuristics For Graphical Model Inference Using
Reinforcement Learning? [114.24881214319048]
We show that we can learn programs, i.e., policies, for solving inference in higher order Conditional Random Fields (CRFs) using reinforcement learning.
Our method solves inference tasks efficiently without imposing any constraints on the form of the potentials.
arXiv Detail & Related papers (2020-04-27T19:24:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.