Mastering Symbolic Operations: Augmenting Language Models with Compiled
Neural Networks
- URL: http://arxiv.org/abs/2304.01665v3
- Date: Sat, 9 Mar 2024 15:27:33 GMT
- Title: Mastering Symbolic Operations: Augmenting Language Models with Compiled
Neural Networks
- Authors: Yixuan Weng, Minjun Zhu, Fei Xia, Bin Li, Shizhu He, Kang Liu, Jun
Zhao
- Abstract summary: "Neural architecture" integrates compiled neural networks (CoNNs) into a standard transformer.
CoNNs are neural modules designed to explicitly encode rules through artificially generated attention weights.
Experiments demonstrate superiority of our approach over existing techniques in terms of length generalization, efficiency, and interpretability for symbolic operations.
- Score: 48.14324895100478
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Language models' (LMs) proficiency in handling deterministic symbolic
reasoning and rule-based tasks remains limited due to their dependency implicit
learning on textual data. To endow LMs with genuine rule comprehension
abilities, we propose "Neural Comprehension" - a framework that synergistically
integrates compiled neural networks (CoNNs) into the standard transformer
architecture. CoNNs are neural modules designed to explicitly encode rules
through artificially generated attention weights. By incorporating CoNN
modules, the Neural Comprehension framework enables LMs to accurately and
robustly execute rule-intensive symbolic tasks. Extensive experiments
demonstrate the superiority of our approach over existing techniques in terms
of length generalization, efficiency, and interpretability for symbolic
operations. Furthermore, it can be applied to LMs across different model
scales, outperforming tool-calling methods in arithmetic reasoning tasks while
maintaining superior inference efficiency. Our work highlights the potential of
seamlessly unifying explicit rule learning via CoNNs and implicit pattern
learning in LMs, paving the way for true symbolic comprehension capabilities.
Related papers
- Large Language Models are Interpretable Learners [53.56735770834617]
In this paper, we show a combination of Large Language Models (LLMs) and symbolic programs can bridge the gap between expressiveness and interpretability.
The pretrained LLM with natural language prompts provides a massive set of interpretable modules that can transform raw input into natural language concepts.
As the knowledge learned by LSP is a combination of natural language descriptions and symbolic rules, it is easily transferable to humans (interpretable) and other LLMs.
arXiv Detail & Related papers (2024-06-25T02:18:15Z) - Deep Inductive Logic Programming meets Reinforcement Learning [0.0]
Differentiable Neural Logic (dNL) networks are able to learn functions as their neural architecture includes symbolic reasoning.
We propose an application of dNL in the field ofReinforcement Learning (RRL) to address dynamic continuous environments.
arXiv Detail & Related papers (2023-08-30T09:08:46Z) - Emergent Linguistic Structures in Neural Networks are Fragile [20.692540987792732]
Large Language Models (LLMs) have been reported to have strong performance on natural language processing tasks.
We propose a framework to assess the consistency and robustness of linguistic representations.
arXiv Detail & Related papers (2022-10-31T15:43:57Z) - Neural-Symbolic Recursive Machine for Systematic Generalization [113.22455566135757]
We introduce the Neural-Symbolic Recursive Machine (NSR), whose core is a Grounded Symbol System (GSS)
NSR integrates neural perception, syntactic parsing, and semantic reasoning.
We evaluate NSR's efficacy across four challenging benchmarks designed to probe systematic generalization capabilities.
arXiv Detail & Related papers (2022-10-04T13:27:38Z) - Neuro-Symbolic Inductive Logic Programming with Logical Neural Networks [65.23508422635862]
We propose learning rules with the recently proposed logical neural networks (LNN)
Compared to others, LNNs offer strong connection to classical Boolean logic.
Our experiments on standard benchmarking tasks confirm that LNN rules are highly interpretable.
arXiv Detail & Related papers (2021-12-06T19:38:30Z) - Reinforcement Learning with External Knowledge by using Logical Neural
Networks [67.46162586940905]
A recent neuro-symbolic framework called the Logical Neural Networks (LNNs) can simultaneously provide key-properties of both neural networks and symbolic logic.
We propose an integrated method that enables model-free reinforcement learning from external knowledge sources.
arXiv Detail & Related papers (2021-03-03T12:34:59Z) - NSL: Hybrid Interpretable Learning From Noisy Raw Data [66.15862011405882]
This paper introduces a hybrid neural-symbolic learning framework, called NSL, that learns interpretable rules from labelled unstructured data.
NSL combines pre-trained neural networks for feature extraction with FastLAS, a state-of-the-art ILP system for rule learning under the answer set semantics.
We demonstrate that NSL is able to learn robust rules from MNIST data and achieve comparable or superior accuracy when compared to neural network and random forest baselines.
arXiv Detail & Related papers (2020-12-09T13:02:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.