Neural Interactive Proofs
- URL: http://arxiv.org/abs/2412.08897v1
- Date: Thu, 12 Dec 2024 03:21:53 GMT
- Title: Neural Interactive Proofs
- Authors: Lewis Hammond, Sam Adam-Day,
- Abstract summary: We study the case in which agents are represented using neural networks.
We introduce a unifying framework based on prover-verifier games.
We describe several new protocols for generating neural interactive proofs.
- Score: 1.519321208145928
- License:
- Abstract: We consider the problem of how a trusted, but computationally bounded agent (a 'verifier') can learn to interact with one or more powerful but untrusted agents ('provers') in order to solve a given task. More specifically, we study the case in which agents are represented using neural networks and refer to solutions of this problem as neural interactive proofs. First we introduce a unifying framework based on prover-verifier games, which generalises previously proposed interaction protocols. We then describe several new protocols for generating neural interactive proofs, and provide a theoretical comparison of both new and existing approaches. Finally, we support this theory with experiments in two domains: a toy graph isomorphism problem that illustrates the key ideas, and a code validation task using large language models. In so doing, we aim to create a foundation for future work on neural interactive proofs and their application in building safer AI systems.
Related papers
- Implementation and Application of an Intelligibility Protocol for Interaction with an LLM [0.9187505256430948]
Our interest is in constructing interactive systems involving a human-expert interacting with a machine learning engine.
This is of relevance when addressing complex problems arising in areas of science, the environment, medicine and so on.
We present an algorithmic description of general-purpose implementation, and conduct preliminary experiments on its use in two different areas.
arXiv Detail & Related papers (2024-10-27T21:20:18Z) - Scaling Large-Language-Model-based Multi-Agent Collaboration [75.5241464256688]
Pioneering advancements in large language model-powered agents have underscored the design pattern of multi-agent collaboration.
Inspired by the neural scaling law, this study investigates whether a similar principle applies to increasing agents in multi-agent collaboration.
arXiv Detail & Related papers (2024-06-11T11:02:04Z) - Deception Detection from Linguistic and Physiological Data Streams Using Bimodal Convolutional Neural Networks [19.639533220155965]
This paper explores the application of convolutional neural networks for the purpose of multimodal deception detection.
We use a dataset built by interviewing 104 subjects about two topics, with one truthful and one falsified response from each subject about each topic.
arXiv Detail & Related papers (2023-11-18T02:44:33Z) - Neural Amortized Inference for Nested Multi-agent Reasoning [54.39127942041582]
We propose a novel approach to bridge the gap between human-like inference capabilities and computational limitations.
We evaluate our method in two challenging multi-agent interaction domains.
arXiv Detail & Related papers (2023-08-21T22:40:36Z) - All Roads Lead to Rome? Exploring the Invariance of Transformers'
Representations [69.3461199976959]
We propose a model based on invertible neural networks, BERT-INN, to learn the Bijection Hypothesis.
We show the advantage of BERT-INN both theoretically and through extensive experiments.
arXiv Detail & Related papers (2023-05-23T22:30:43Z) - Zonotope Domains for Lagrangian Neural Network Verification [102.13346781220383]
We decompose the problem of verifying a deep neural network into the verification of many 2-layer neural networks.
Our technique yields bounds that improve upon both linear programming and Lagrangian-based verification techniques.
arXiv Detail & Related papers (2022-10-14T19:31:39Z) - Neural Causal Models for Counterfactual Identification and Estimation [62.30444687707919]
We study the evaluation of counterfactual statements through neural models.
First, we show that neural causal models (NCMs) are expressive enough.
Second, we develop an algorithm for simultaneously identifying and estimating counterfactual distributions.
arXiv Detail & Related papers (2022-09-30T18:29:09Z) - Dependable Neural Networks Through Redundancy, A Comparison of Redundant
Architectures [0.0]
We discuss the issue of dependability in general implementation terms before examining lockstep solutions.
We intuit that it is not necessarily a given that two similar neural networks generate results at precisely the same time and that synchronization between the platforms will be required.
arXiv Detail & Related papers (2021-07-30T12:17:36Z) - Backprop-Free Reinforcement Learning with Active Neural Generative
Coding [84.11376568625353]
We propose a computational framework for learning action-driven generative models without backpropagation of errors (backprop) in dynamic environments.
We develop an intelligent agent that operates even with sparse rewards, drawing inspiration from the cognitive theory of planning as inference.
The robust performance of our agent offers promising evidence that a backprop-free approach for neural inference and learning can drive goal-directed behavior.
arXiv Detail & Related papers (2021-07-10T19:02:27Z) - Explaining Away Attacks Against Neural Networks [3.658164271285286]
We investigate the problem of identifying adversarial attacks on image-based neural networks.
We present intriguing experimental results showing significant discrepancies between the explanations generated for the predictions of a model on clean and adversarial data.
We propose a framework which can identify whether a given input is adversarial based on the explanations given by the model.
arXiv Detail & Related papers (2020-03-06T15:32:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.