Online Bayesian Meta-Learning for Cognitive Tracking Radar
- URL: http://arxiv.org/abs/2207.06917v1
- Date: Thu, 7 Jul 2022 20:21:54 GMT
- Title: Online Bayesian Meta-Learning for Cognitive Tracking Radar
- Authors: Charles E. Thornton, R. Michael Buehrer, Anthony F. Martone
- Abstract summary: We develop an online meta-learning approach for waveform-agile tracking.
We exploit inherent similarity across tracking scenes, attributed to common physical elements such as target type or clutter.
- Score: 9.805913930878
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: A key component of cognitive radar is the ability to generalize, or achieve
consistent performance across a broad class of sensing environments, since
aspects of the physical scene may vary over time. This presents a challenge for
learning-based waveform selection approaches, since transmission policies which
are effective in one scene may be highly suboptimal in another. One way to
address this problem is to bias a learning algorithm strategically by
exploiting high-level structure across tracking instances, referred to as
meta-learning. In this work, we develop an online meta-learning approach for
waveform-agile tracking. This approach uses information gained from previous
target tracks to speed up and enhance learning in new tracking instances. This
results in sample-efficient learning across a class of finite state target
channels by exploiting inherent similarity across tracking scenes, attributed
to common physical elements such as target type or clutter. We formulate the
online waveform selection problem in the framework of Bayesian learning, and
provide prior-dependent performance bounds for the meta-learning problem using
PAC-Bayes theory. We present a computationally feasible posterior sampling
algorithm and study the performance in a simulation study consisting of diverse
scenes. Finally, we examine the potential performance benefits and practical
challenges associated with online meta-learning for waveform-agile tracking.
Related papers
- Online waveform selection for cognitive radar [8.187445866881637]
We propose adaptive algorithms that select waveform parameters in an online fashion.
We propose three reinforcement learning algorithms: bandwidth scaling, Q-learning, and Q-learning lookahead.
Our proposed algorithms achieve the dual objectives of minimizing range error and maintaining continuous tracking without losing the target.
arXiv Detail & Related papers (2024-10-14T15:01:41Z) - Causality-Aware Transformer Networks for Robotic Navigation [13.719643934968367]
Current research in Visual Navigation reveals opportunities for improvement.
Direct adoption of RNNs and Transformers often overlooks the specific differences between Embodied AI and traditional sequential data modelling.
We propose Causality-Aware Transformer (CAT) Networks for Navigation, featuring a Causal Understanding Module.
arXiv Detail & Related papers (2024-09-04T12:53:26Z) - RLIF: Interactive Imitation Learning as Reinforcement Learning [56.997263135104504]
We show how off-policy reinforcement learning can enable improved performance under assumptions that are similar but potentially even more practical than those of interactive imitation learning.
Our proposed method uses reinforcement learning with user intervention signals themselves as rewards.
This relaxes the assumption that intervening experts in interactive imitation learning should be near-optimal and enables the algorithm to learn behaviors that improve over the potential suboptimal human expert.
arXiv Detail & Related papers (2023-11-21T21:05:21Z) - Diffusion Generative Flow Samplers: Improving learning signals through
partial trajectory optimization [87.21285093582446]
Diffusion Generative Flow Samplers (DGFS) is a sampling-based framework where the learning process can be tractably broken down into short partial trajectory segments.
Our method takes inspiration from the theory developed for generative flow networks (GFlowNets)
arXiv Detail & Related papers (2023-10-04T09:39:05Z) - ALP: Action-Aware Embodied Learning for Perception [60.64801970249279]
We introduce Action-Aware Embodied Learning for Perception (ALP)
ALP incorporates action information into representation learning through a combination of optimizing a reinforcement learning policy and an inverse dynamics prediction objective.
We show that ALP outperforms existing baselines in several downstream perception tasks.
arXiv Detail & Related papers (2023-06-16T21:51:04Z) - Learning to Continuously Optimize Wireless Resource in a Dynamic
Environment: A Bilevel Optimization Perspective [52.497514255040514]
This work develops a new approach that enables data-driven methods to continuously learn and optimize resource allocation strategies in a dynamic environment.
We propose to build the notion of continual learning into wireless system design, so that the learning model can incrementally adapt to the new episodes.
Our design is based on a novel bilevel optimization formulation which ensures certain fairness" across different data samples.
arXiv Detail & Related papers (2021-05-03T07:23:39Z) - Meta-learning the Learning Trends Shared Across Tasks [123.10294801296926]
Gradient-based meta-learning algorithms excel at quick adaptation to new tasks with limited data.
Existing meta-learning approaches only depend on the current task information during the adaptation.
We propose a 'Path-aware' model-agnostic meta-learning approach.
arXiv Detail & Related papers (2020-10-19T08:06:47Z) - Meta-Gradient Reinforcement Learning with an Objective Discovered Online [54.15180335046361]
We propose an algorithm based on meta-gradient descent that discovers its own objective, flexibly parameterised by a deep neural network.
Because the objective is discovered online, it can adapt to changes over time.
On the Atari Learning Environment, the meta-gradient algorithm adapts over time to learn with greater efficiency.
arXiv Detail & Related papers (2020-07-16T16:17:09Z) - Neuromodulated Neural Architectures with Local Error Signals for
Memory-Constrained Online Continual Learning [4.2903672492917755]
We develop a biologically-inspired light weight neural network architecture that incorporates local learning and neuromodulation.
We demonstrate the efficacy of our approach on both single task and continual learning setting.
arXiv Detail & Related papers (2020-07-16T07:41:23Z) - Efficient Exploration in Constrained Environments with Goal-Oriented
Reference Path [15.679210057474922]
We train a deep convolutional network that can predict collision-free paths based on a map of the environment.
This is then used by a reinforcement learning algorithm to learn to closely follow the path.
We show that our method consistently improves the sample efficiency and generalization capability to novel environments.
arXiv Detail & Related papers (2020-03-03T17:07:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.