Interactive Visualization for Debugging RL
- URL: http://arxiv.org/abs/2008.07331v2
- Date: Tue, 18 Aug 2020 22:27:29 GMT
- Title: Interactive Visualization for Debugging RL
- Authors: Shuby Deshpande, Benjamin Eysenbach, Jeff Schneider
- Abstract summary: Our system addresses many features missing from previous tools such as tools for supervised learning often are not interactive.
We provide an example workflow of how this system could be used, along with ideas for future extensions.
- Score: 11.6341132172284
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visualization tools for supervised learning allow users to interpret,
introspect, and gain an intuition for the successes and failures of their
models. While reinforcement learning practitioners ask many of the same
questions, existing tools are not applicable to the RL setting as these tools
address challenges typically found in the supervised learning regime. In this
work, we design and implement an interactive visualization tool for debugging
and interpreting RL algorithms. Our system addresses many features missing from
previous tools such as (1) tools for supervised learning often are not
interactive; (2) while debugging RL policies researchers use state
representations that are different from those seen by the agent; (3) a
framework designed to make the debugging RL policies more conducive. We provide
an example workflow of how this system could be used, along with ideas for
future extensions.
Related papers
- DMC-VB: A Benchmark for Representation Learning for Control with Visual Distractors [13.700885996266457]
Learning from previously collected data via behavioral cloning or offline reinforcement learning (RL) is a powerful recipe for scaling generalist agents.
We present theDeepMind Control Visual Benchmark (DMC-VB), a dataset collected in the DeepMind Control Suite to evaluate the robustness of offline RL agents.
Accompanying our dataset, we propose three benchmarks to evaluate representation learning methods for pretraining, and carry out experiments on several recently proposed methods.
arXiv Detail & Related papers (2024-09-26T23:07:01Z) - Enhancing Tool Retrieval with Iterative Feedback from Large Language Models [9.588592185027455]
Large language models (LLMs) can effectively handle a certain amount of tools through in-context learning or fine-tuning.
In real-world scenarios, the number of tools is typically extensive and irregularly updated, emphasizing the necessity for a dedicated tool retrieval component.
We propose to enhance tool retrieval with iterative feedback from the large language model.
arXiv Detail & Related papers (2024-06-25T11:12:01Z) - Tool Learning with Large Language Models: A Survey [60.733557487886635]
Tool learning with large language models (LLMs) has emerged as a promising paradigm for augmenting the capabilities of LLMs to tackle highly complex problems.
Despite growing attention and rapid advancements in this field, the existing literature remains fragmented and lacks systematic organization.
arXiv Detail & Related papers (2024-05-28T08:01:26Z) - Chain of Tools: Large Language Model is an Automatic Multi-tool Learner [54.992464510992605]
Automatic Tool Chain (ATC) is a framework that enables the large language models (LLMs) to act as a multi-tool user.
To scale up the scope of the tools, we next propose a black-box probing method.
For a comprehensive evaluation, we build a challenging benchmark named ToolFlow.
arXiv Detail & Related papers (2024-05-26T11:40:58Z) - Towards Completeness-Oriented Tool Retrieval for Large Language Models [60.733557487886635]
Real-world systems often incorporate a wide array of tools, making it impractical to input all tools into Large Language Models.
Existing tool retrieval methods primarily focus on semantic matching between user queries and tool descriptions.
We propose a novel modelagnostic COllaborative Learning-based Tool Retrieval approach, COLT, which captures not only the semantic similarities between user queries and tool descriptions but also takes into account the collaborative information of tools.
arXiv Detail & Related papers (2024-05-25T06:41:23Z) - CLOVA: A Closed-Loop Visual Assistant with Tool Usage and Update [69.59482029810198]
CLOVA is a Closed-Loop Visual Assistant that operates within a framework encompassing inference, reflection, and learning phases.
Results demonstrate that CLOVA surpasses existing tool-usage methods by 5% in visual question answering and multiple-image reasoning, by 10% in knowledge tagging, and by 20% in image editing.
arXiv Detail & Related papers (2023-12-18T03:34:07Z) - CausalVLR: A Toolbox and Benchmark for Visual-Linguistic Causal
Reasoning [107.81733977430517]
CausalVLR (Causal Visual-Linguistic Reasoning) is an open-source toolbox containing a rich set of state-of-the-art causal relation discovery and causal inference methods.
These methods have been included in the toolbox with PyTorch implementations under NVIDIA computing system.
arXiv Detail & Related papers (2023-06-30T08:17:38Z) - Making Language Models Better Tool Learners with Execution Feedback [36.30542737293863]
Tools serve as pivotal interfaces that enable humans to understand and reshape the environment.
Existing tool learning methodologies induce large language models to utilize tools indiscriminately.
We propose Tool leaRning wIth exeCution fEedback (TRICE), a two-stage end-to-end framework that enables the model to continually learn through feedback derived from tool execution.
arXiv Detail & Related papers (2023-05-22T14:37:05Z) - Mastering the Unsupervised Reinforcement Learning Benchmark from Pixels [112.63440666617494]
Reinforcement learning algorithms can succeed but require large amounts of interactions between the agent and the environment.
We propose a new method to solve it, using unsupervised model-based RL, for pre-training the agent.
We show robust performance on the Real-Word RL benchmark, hinting at resiliency to environment perturbations during adaptation.
arXiv Detail & Related papers (2022-09-24T14:22:29Z) - Vizarel: A System to Help Better Understand RL Agents [4.009038499050246]
We describe our initial attempt at constructing a prototype of these ideas.
Our design is motivated by envisioning the system to be a platform on which to experiment with interpretable reinforcement learning.
arXiv Detail & Related papers (2020-07-10T19:19:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.