Optimizing Indoor Navigation Policies For Spatial Distancing
- URL: http://arxiv.org/abs/2207.08860v1
- Date: Sat, 4 Jun 2022 21:57:22 GMT
- Title: Optimizing Indoor Navigation Policies For Spatial Distancing
- Authors: Xun Zhang, Mathew Schwartz, Muhammad Usman, Petros Faloutsos, Mubbasir
Kapadia
- Abstract summary: In this paper, we focus on the modification of policies that can lead to movement patterns and directional guidance of occupants.
We show that within our framework, the simulation-optimization process can help to improve spatial distancing between agents.
- Score: 8.635212273689273
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper, we focus on the modification of policies that can lead to
movement patterns and directional guidance of occupants, which are represented
as agents in a 3D simulation engine. We demonstrate an optimization method that
improves a spatial distancing metric by modifying the navigation graph by
introducing a measure of spatial distancing of agents as a function of agent
density (i.e., occupancy). Our optimization framework utilizes such metrics as
the target function, using a hybrid approach of combining genetic algorithm and
simulated annealing. We show that within our framework, the
simulation-optimization process can help to improve spatial distancing between
agents by optimizing the navigation policies for a given indoor environment.
Related papers
- Co-Optimization of Environment and Policies for Decentralized Multi-Agent Navigation [14.533605727697775]
This work views the multi-agent system and its surrounding environment as a co-evolving system, where the behavior of one affects the other.
We develop an algorithm that alternates between sub-objectives to search for an optimal of agent actions and obstacle configurations in the environment.
arXiv Detail & Related papers (2024-03-21T17:37:43Z) - Dynamic Kernel-Based Adaptive Spatial Aggregation for Learned Image
Compression [63.56922682378755]
We focus on extending spatial aggregation capability and propose a dynamic kernel-based transform coding.
The proposed adaptive aggregation generates kernel offsets to capture valid information in the content-conditioned range to help transform.
Experimental results demonstrate that our method achieves superior rate-distortion performance on three benchmarks compared to the state-of-the-art learning-based methods.
arXiv Detail & Related papers (2023-08-17T01:34:51Z) - Reparameterized Policy Learning for Multimodal Trajectory Optimization [61.13228961771765]
We investigate the challenge of parametrizing policies for reinforcement learning in high-dimensional continuous action spaces.
We propose a principled framework that models the continuous RL policy as a generative model of optimal trajectories.
We present a practical model-based RL method, which leverages the multimodal policy parameterization and learned world model.
arXiv Detail & Related papers (2023-07-20T09:05:46Z) - Acceleration in Policy Optimization [50.323182853069184]
We work towards a unifying paradigm for accelerating policy optimization methods in reinforcement learning (RL) by integrating foresight in the policy improvement step via optimistic and adaptive updates.
We define optimism as predictive modelling of the future behavior of a policy, and adaptivity as taking immediate and anticipatory corrective actions to mitigate errors from overshooting predictions or delayed responses to change.
We design an optimistic policy gradient algorithm, adaptive via meta-gradient learning, and empirically highlight several design choices pertaining to acceleration, in an illustrative task.
arXiv Detail & Related papers (2023-06-18T15:50:57Z) - Wasserstein Gradient Flows for Optimizing Gaussian Mixture Policies [0.0]
Policy optimization is the emphde facto paradigm to adapt robot policies as a function of task-specific objectives.
We propose to leverage the structure of probabilistic policies by casting the policy optimization as an optimal transport problem.
We evaluate our approach on common robotic settings: reaching motions, collision-avoidance behaviors, and multi-goal tasks.
arXiv Detail & Related papers (2023-05-17T17:48:24Z) - Accelerated Federated Learning with Decoupled Adaptive Optimization [53.230515878096426]
federated learning (FL) framework enables clients to collaboratively learn a shared model while keeping privacy of training data on clients.
Recently, many iterations efforts have been made to generalize centralized adaptive optimization methods, such as SGDM, Adam, AdaGrad, etc., to federated settings.
This work aims to develop novel adaptive optimization methods for FL from the perspective of dynamics of ordinary differential equations (ODEs)
arXiv Detail & Related papers (2022-07-14T22:46:43Z) - MetaAlign: Coordinating Domain Alignment and Classification for
Unsupervised Domain Adaptation [84.90801699807426]
This paper proposes an effective meta-optimization based strategy dubbed MetaAlign.
We treat the domain alignment objective and the classification objective as the meta-train and meta-test tasks in a meta-learning scheme.
Experimental results demonstrate the effectiveness of our proposed method on top of various alignment-based baseline approaches.
arXiv Detail & Related papers (2021-03-25T03:16:05Z) - Distributionally-Constrained Policy Optimization via Unbalanced Optimal
Transport [15.294456568539148]
We formulate policy optimization as unbalanced optimal transport over the space of occupancy measures.
We propose a general purpose RL objective based on Bregman divergence and optimize it using Dykstra's algorithm.
arXiv Detail & Related papers (2021-02-15T23:04:37Z) - Distributed Multi-agent Meta Learning for Trajectory Design in Wireless
Drone Networks [151.27147513363502]
This paper studies the problem of the trajectory design for a group of energyconstrained drones operating in dynamic wireless network environments.
A value based reinforcement learning (VDRL) solution and a metatraining mechanism is proposed.
arXiv Detail & Related papers (2020-12-06T01:30:12Z) - Learning Task-Agnostic Action Spaces for Movement Optimization [18.37812596641983]
We propose a novel method for exploring the dynamics of physically based animated characters.
We parameterize actions as target states, and learn a short-horizon goal-conditioned low-level control policy that drives the agent's state towards the targets.
arXiv Detail & Related papers (2020-09-22T06:18:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.