Model-based Validation as Probabilistic Inference
- URL: http://arxiv.org/abs/2305.09930v1
- Date: Wed, 17 May 2023 03:27:36 GMT
- Title: Model-based Validation as Probabilistic Inference
- Authors: Harrison Delecki, Anthony Corso, Mykel J. Kochenderfer
- Abstract summary: Estimating the distribution over failures is a key step in validating autonomous systems.
We frame estimating the distribution over failure trajectories for sequential systems as Bayesian inference.
Our approach is demonstrated in an inverted pendulum control system, an autonomous vehicle driving scenario, and a partially observable lunar lander.
- Score: 37.61747231296097
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Estimating the distribution over failures is a key step in validating
autonomous systems. Existing approaches focus on finding failures for a small
range of initial conditions or make restrictive assumptions about the
properties of the system under test. We frame estimating the distribution over
failure trajectories for sequential systems as Bayesian inference. Our
model-based approach represents the distribution over failure trajectories
using rollouts of system dynamics and computes trajectory gradients using
automatic differentiation. Our approach is demonstrated in an inverted pendulum
control system, an autonomous vehicle driving scenario, and a partially
observable lunar lander. Sampling is performed using an off-the-shelf
implementation of Hamiltonian Monte Carlo with multiple chains to capture
multimodality and gradient smoothing for safe trajectories. In all experiments,
we observed improvements in sample efficiency and parameter space coverage
compared to black-box baseline approaches. This work is open sourced.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Diffusion-Based Failure Sampling for Cyber-Physical Systems [33.25641206808538]
We propose a conditional denoising diffusion model, which has shown success in complex high-dimensional problems such as robotic task planning.
We demonstrate the effectiveness of our approach on high-dimensional robotic validation tasks, improving sample efficiency and mode coverage compared to existing black-box techniques.
arXiv Detail & Related papers (2024-06-20T22:22:28Z) - Amortizing intractable inference in diffusion models for vision, language, and control [89.65631572949702]
This paper studies amortized sampling of the posterior over data, $mathbfxsim prm post(mathbfx)propto p(mathbfx)r(mathbfx)$, in a model that consists of a diffusion generative model prior $p(mathbfx)$ and a black-box constraint or function $r(mathbfx)$.
We prove the correctness of a data-free learning objective, relative trajectory balance, for training a diffusion model that samples from
arXiv Detail & Related papers (2024-05-31T16:18:46Z) - Denoising Diffusion-Based Control of Nonlinear Systems [3.4530027457862]
We propose a novel approach based on Denoising Diffusion Probabilistic Models (DDPMs) to control nonlinear dynamical systems.
DDPMs are the state-of-art of generative models that have achieved success in a wide variety of sampling tasks.
We numerically study our approach on various nonlinear systems and verify our theoretical results.
arXiv Detail & Related papers (2024-02-03T23:19:26Z) - Observation-Guided Diffusion Probabilistic Models [41.749374023639156]
We propose a novel diffusion-based image generation method called the observation-guided diffusion probabilistic model (OGDM)
Our approach reestablishes the training objective by integrating the guidance of the observation process with the Markov chain.
We demonstrate the effectiveness of our training algorithm using diverse inference techniques on strong diffusion model baselines.
arXiv Detail & Related papers (2023-10-06T06:29:06Z) - Practical Probabilistic Model-based Deep Reinforcement Learning by
Integrating Dropout Uncertainty and Trajectory Sampling [7.179313063022576]
This paper addresses the prediction stability, prediction accuracy and control capability of the current probabilistic model-based reinforcement learning (MBRL) built on neural networks.
A novel approach dropout-based probabilistic ensembles with trajectory sampling (DPETS) is proposed.
arXiv Detail & Related papers (2023-09-20T06:39:19Z) - Score-based Source Separation with Applications to Digital Communication
Signals [72.6570125649502]
We propose a new method for separating superimposed sources using diffusion-based generative models.
Motivated by applications in radio-frequency (RF) systems, we are interested in sources with underlying discrete nature.
Our method can be viewed as a multi-source extension to the recently proposed score distillation sampling scheme.
arXiv Detail & Related papers (2023-06-26T04:12:40Z) - Unrolling Particles: Unsupervised Learning of Sampling Distributions [102.72972137287728]
Particle filtering is used to compute good nonlinear estimates of complex systems.
We show in simulations that the resulting particle filter yields good estimates in a wide range of scenarios.
arXiv Detail & Related papers (2021-10-06T16:58:34Z) - Certainty Equivalent Perception-Based Control [29.216967322052785]
We show a uniform error bound on non kernel regression under a dynamically-achievable dense sampling scheme.
This allows for a finite-time convergence rate on the sub-optimality of using the regressor in closed-loop for waypoint tracking.
arXiv Detail & Related papers (2020-08-27T18:45:40Z) - Batch Stationary Distribution Estimation [98.18201132095066]
We consider the problem of approximating the stationary distribution of an ergodic Markov chain given a set of sampled transitions.
We propose a consistent estimator that is based on recovering a correction ratio function over the given data.
arXiv Detail & Related papers (2020-03-02T09:10:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.