Sparsely Changing Latent States for Prediction and Planning in Partially
Observable Domains
- URL: http://arxiv.org/abs/2110.15949v1
- Date: Fri, 29 Oct 2021 17:50:44 GMT
- Title: Sparsely Changing Latent States for Prediction and Planning in Partially
Observable Domains
- Authors: Christian Gumbsch and Martin V. Butz and Georg Martius
- Abstract summary: GateL0RD is a novel recurrent architecture that incorporates the inductive bias to maintain stable, sparsely changing latent states.
We demonstrate that GateL0RD can compete with or outperform state-of-the-art RNNs in a variety of partially observable prediction and control tasks.
- Score: 11.371889042789219
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A common approach to prediction and planning in partially observable domains
is to use recurrent neural networks (RNNs), which ideally develop and maintain
a latent memory about hidden, task-relevant factors. We hypothesize that many
of these hidden factors in the physical world are constant over time, changing
only sparsely. Accordingly, we propose Gated $L_0$ Regularized Dynamics
(GateL0RD), a novel recurrent architecture that incorporates the inductive bias
to maintain stable, sparsely changing latent states. The bias is implemented by
means of a novel internal gating function and a penalty on the $L_0$ norm of
latent state changes. We demonstrate that GateL0RD can compete with or
outperform state-of-the-art RNNs in a variety of partially observable
prediction and control tasks. GateL0RD tends to encode the underlying
generative factors of the environment, ignores spurious temporal dependencies,
and generalizes better, improving sampling efficiency and prediction accuracy
as well as behavior in model-based planning and reinforcement learning tasks.
Moreover, we show that the developing latent states can be easily interpreted,
which is a step towards better explainability in RNNs.
Related papers
- Deep neural networks for choice analysis: Enhancing behavioral regularity with gradient regularization [9.07183008265166]
This study proposes strong and weak behavioral regularities as novel metrics to evaluate the monotonicity of individual demand functions.
It further designs a constrained optimization framework with six gradient regularizers to enhance DNNs' behavioral regularity.
The proposed framework is applicable to other NN-based choice models such as TasteNets.
arXiv Detail & Related papers (2024-04-23T03:01:09Z) - Learning Low Dimensional State Spaces with Overparameterized Recurrent
Neural Nets [57.06026574261203]
We provide theoretical evidence for learning low-dimensional state spaces, which can also model long-term memory.
Experiments corroborate our theory, demonstrating extrapolation via learning low-dimensional state spaces with both linear and non-linear RNNs.
arXiv Detail & Related papers (2022-10-25T14:45:15Z) - Stability and Generalization Analysis of Gradient Methods for Shallow
Neural Networks [59.142826407441106]
We study the generalization behavior of shallow neural networks (SNNs) by leveraging the concept of algorithmic stability.
We consider gradient descent (GD) and gradient descent (SGD) to train SNNs, for both of which we develop consistent excess bounds.
arXiv Detail & Related papers (2022-09-19T18:48:00Z) - Domain-Adjusted Regression or: ERM May Already Learn Features Sufficient
for Out-of-Distribution Generalization [52.7137956951533]
We argue that devising simpler methods for learning predictors on existing features is a promising direction for future research.
We introduce Domain-Adjusted Regression (DARE), a convex objective for learning a linear predictor that is provably robust under a new model of distribution shift.
Under a natural model, we prove that the DARE solution is the minimax-optimal predictor for a constrained set of test distributions.
arXiv Detail & Related papers (2022-02-14T16:42:16Z) - Discovering Invariant Rationales for Graph Neural Networks [104.61908788639052]
Intrinsic interpretability of graph neural networks (GNNs) is to find a small subset of the input graph's features.
We propose a new strategy of discovering invariant rationale (DIR) to construct intrinsically interpretable GNNs.
arXiv Detail & Related papers (2022-01-30T16:43:40Z) - Recurrent Neural Networks for Learning Long-term Temporal Dependencies
with Reanalysis of Time Scale Representation [16.32068729107421]
We argue that the interpretation of a forget gate as a temporal representation is valid when the gradient of loss with respect to the state decreases exponentially as time goes back.
We propose an approach to construct new RNNs that can represent a longer time scale than conventional models.
arXiv Detail & Related papers (2021-11-05T06:22:58Z) - Interpretable Additive Recurrent Neural Networks For Multivariate
Clinical Time Series [4.125698836261585]
We present the Interpretable-RNN (I-RNN) that balances model complexity and accuracy by forcing the relationship between variables in the model to be additive.
I-RNN specifically captures the unique characteristics of clinical time series, which are unevenly sampled in time, asynchronously acquired, and have missing data.
We evaluate the I-RNN model on the Physionet 2012 Challenge dataset to predict in-hospital mortality, and on a real-world clinical decision support task: predicting hemodynamic interventions in the intensive care unit.
arXiv Detail & Related papers (2021-09-15T22:30:19Z) - A Biased Graph Neural Network Sampler with Near-Optimal Regret [57.70126763759996]
Graph neural networks (GNN) have emerged as a vehicle for applying deep network architectures to graph and relational data.
In this paper, we build upon existing work and treat GNN neighbor sampling as a multi-armed bandit problem.
We introduce a newly-designed reward function that introduces some degree of bias designed to reduce variance and avoid unstable, possibly-unbounded payouts.
arXiv Detail & Related papers (2021-03-01T15:55:58Z) - Towards Understanding the Regularization of Adversarial Robustness on
Neural Networks [46.54437309608066]
We study the degradation through the regularization perspective.
We find that AR is achieved by regularizing/biasing NNs towards less confident solutions.
arXiv Detail & Related papers (2020-11-15T08:32:09Z) - Stochastic Graph Neural Networks [123.39024384275054]
Graph neural networks (GNNs) model nonlinear representations in graph data with applications in distributed agent coordination, control, and planning.
Current GNN architectures assume ideal scenarios and ignore link fluctuations that occur due to environment, human factors, or external attacks.
In these situations, the GNN fails to address its distributed task if the topological randomness is not considered accordingly.
arXiv Detail & Related papers (2020-06-04T08:00:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.