pyRDDLGym: From RDDL to Gym Environments
- URL: http://arxiv.org/abs/2211.05939v5
- Date: Tue, 6 Feb 2024 00:25:23 GMT
- Title: pyRDDLGym: From RDDL to Gym Environments
- Authors: Ayal Taitler, Michael Gimelfarb, Jihwan Jeong, Sriram Gopalakrishnan,
Martin Mladenov, Xiaotian Liu, Scott Sanner
- Abstract summary: pyRDDLGym is a Python framework for auto-generation of OpenAI Gym environments from RDDL declerative description.
We present the design and built-in examples of pyRDDLGym, and the additions made to the RDDL language that were incorporated into the framework.
- Score: 22.439740618373346
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present pyRDDLGym, a Python framework for auto-generation of OpenAI Gym
environments from RDDL declerative description. The discrete time step
evolution of variables in RDDL is described by conditional probability
functions, which fits naturally into the Gym step scheme. Furthermore, since
RDDL is a lifted description, the modification and scaling up of environments
to support multiple entities and different configurations becomes trivial
rather than a tedious process prone to errors. We hope that pyRDDLGym will
serve as a new wind in the reinforcement learning community by enabling easy
and rapid development of benchmarks due to the unique expressive power of RDDL.
By providing explicit access to the model in the RDDL description, pyRDDLGym
can also facilitate research on hybrid approaches for learning from interaction
while leveraging model knowledge. We present the design and built-in examples
of pyRDDLGym, and the additions made to the RDDL language that were
incorporated into the framework.
Related papers
- DeepSieve: Information Sieving via LLM-as-a-Knowledge-Router [57.28685457991806]
DeepSieve is an agentic RAG framework that incorporates information sieving via LLM-as-a-knowledge-router.<n>Our design emphasizes modularity, transparency, and adaptability, leveraging recent advances in agentic system design.
arXiv Detail & Related papers (2025-07-29T17:55:23Z) - HDDLGym: A Tool for Studying Multi-Agent Hierarchical Problems Defined in HDDL with OpenAI Gym [7.365229047837842]
Reinforcement learning (RL) methods have been widely tested using tools like OpenAI Gym.<n>There is a lack of a tool that enables seamless integration of hierarchical planning with RL.<n>We introduce HDDLGym, a Python-based tool that automatically generates OpenAI Gym environments from HDDL domains and problems.
arXiv Detail & Related papers (2025-05-28T17:10:43Z) - MLE-Dojo: Interactive Environments for Empowering LLM Agents in Machine Learning Engineering [57.156093929365255]
Gym-style framework for systematically reinforcement learning, evaluating, and improving autonomous large language model (LLM) agents.<n>MLE-Dojo covers diverse, open-ended MLE tasks carefully curated to reflect realistic engineering scenarios.<n>Its fully executable environment supports comprehensive agent training via both supervised fine-tuning and reinforcement learning.
arXiv Detail & Related papers (2025-05-12T17:35:43Z) - LoRA-IR: Taming Low-Rank Experts for Efficient All-in-One Image Restoration [62.3751291442432]
We propose LoRA-IR, a flexible framework that dynamically leverages compact low-rank experts to facilitate efficient all-in-one image restoration.
LoRA-IR consists of two training stages: degradation-guided pre-training and parameter-efficient fine-tuning.
Experiments demonstrate that LoRA-IR achieves SOTA performance across 14 IR tasks and 29 benchmarks, while maintaining computational efficiency.
arXiv Detail & Related papers (2024-10-20T13:00:24Z) - Reference Trustable Decoding: A Training-Free Augmentation Paradigm for Large Language Models [79.41139393080736]
Large language models (LLMs) have rapidly advanced and demonstrated impressive capabilities.
In-Context Learning (ICL) and.
Efficient Fine-Tuning (PEFT) are currently two mainstream methods for augmenting.
LLMs to downstream tasks.
We propose Reference Trustable Decoding (RTD), a paradigm that allows models to quickly adapt to new tasks without fine-tuning.
arXiv Detail & Related papers (2024-09-30T10:48:20Z) - Leveraging Environment Interaction for Automated PDDL Translation and Planning with Large Language Models [7.3238629831871735]
Large Language Models (LLMs) have shown remarkable performance in various natural language tasks.
Planning problems into the Planning Domain Definition Language (PDDL) has been proposed as a potential solution.
We propose a novel approach that leverages LLMs and environment feedback to automatically generate PDDL domain and problem description files.
arXiv Detail & Related papers (2024-07-17T19:50:51Z) - LLAVIDAL: A Large LAnguage VIsion Model for Daily Activities of Living [15.062325300795257]
Current Large Language Models (LLVMs) trained on web perform well in general understanding but struggle with fine-temporal details.
This limitation stems from a lack of specialized ADL video instruction-tuning and insufficient modality integration.
We propose a semi-grained framework for curating ADL datasets, creating ADL-X, a multiview, multimodal RGBS instruction-tuning dataset.
We also introduce LLAVIDAL, an LLVM integrating videos, 3D skeletons, and HOIs to model ADL's complex relationships.
arXiv Detail & Related papers (2024-06-13T17:59:05Z) - MMA-DFER: MultiModal Adaptation of unimodal models for Dynamic Facial Expression Recognition in-the-wild [81.32127423981426]
Multimodal emotion recognition based on audio and video data is important for real-world applications.
Recent methods have focused on exploiting advances of self-supervised learning (SSL) for pre-training of strong multimodal encoders.
We propose a different perspective on the problem and investigate the advancement of multimodal DFER performance by adapting SSL-pre-trained disjoint unimodal encoders.
arXiv Detail & Related papers (2024-04-13T13:39:26Z) - Serving Deep Learning Model in Relational Databases [70.53282490832189]
Serving deep learning (DL) models on relational data has become a critical requirement across diverse commercial and scientific domains.
We highlight three pivotal paradigms: The state-of-the-art DL-centric architecture offloads DL computations to dedicated DL frameworks.
The potential UDF-centric architecture encapsulates one or more tensor computations into User Defined Functions (UDFs) within the relational database management system (RDBMS)
arXiv Detail & Related papers (2023-10-07T06:01:35Z) - Enabling Resource-efficient AIoT System with Cross-level Optimization: A
survey [20.360136850102833]
This survey aims to provide a broader optimization space for more free resource-performance tradeoffs.
By consolidating problems and techniques scattered over diverse levels, we aim to help readers understand their connections and stimulate further discussions.
arXiv Detail & Related papers (2023-09-27T08:04:24Z) - DIETERpy: a Python framework for The Dispatch and Investment Evaluation
Tool with Endogenous Renewables [62.997667081978825]
DIETER is an open-source power sector model designed to analyze future settings with very high shares of variable renewable energy sources.
It minimizes overall system costs, including fixed and variable costs of various generation, flexibility and sector coupling options.
We introduce DIETERpy that builds on the existing model version, written in the General Algebraic Modeling System (GAMS) and enhances it with a Python framework.
arXiv Detail & Related papers (2020-10-02T09:27:33Z) - Unsupervised Controllable Generation with Self-Training [90.04287577605723]
controllable generation with GANs remains a challenging research problem.
We propose an unsupervised framework to learn a distribution of latent codes that control the generator through self-training.
Our framework exhibits better disentanglement compared to other variants such as the variational autoencoder.
arXiv Detail & Related papers (2020-07-17T21:50:35Z) - PDDLGym: Gym Environments from PDDL Problems [13.630185187102413]
We present PDDLGym, a framework that automatically constructs OpenAI Gym environments from PDDL domains and problems.
Observations and actions in PDDLGym are relational, making the framework particularly well-suited for research in relational reinforcement learning and relational sequential decision-making.
arXiv Detail & Related papers (2020-02-15T19:10:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.