Active Representation Learning for General Task Space with Applications
in Robotics
- URL: http://arxiv.org/abs/2306.08942v1
- Date: Thu, 15 Jun 2023 08:27:50 GMT
- Title: Active Representation Learning for General Task Space with Applications
in Robotics
- Authors: Yifang Chen, Yingbing Huang, Simon S. Du, Kevin Jamieson, Guanya Shi
- Abstract summary: We propose an algorithmic framework for textitactive representation learning, where the learner optimally chooses which source tasks to sample from.
We provide several instantiations under this framework, from bilinear and feature-based nonlinear to general nonlinear cases.
Our algorithms outperform baselines by $20%-70%$ on average.
- Score: 44.36398212117328
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Representation learning based on multi-task pretraining has become a powerful
approach in many domains. In particular, task-aware representation learning
aims to learn an optimal representation for a specific target task by sampling
data from a set of source tasks, while task-agnostic representation learning
seeks to learn a universal representation for a class of tasks. In this paper,
we propose a general and versatile algorithmic and theoretic framework for
\textit{active representation learning}, where the learner optimally chooses
which source tasks to sample from. This framework, along with a tractable meta
algorithm, allows most arbitrary target and source task spaces (from discrete
to continuous), covers both task-aware and task-agnostic settings, and is
compatible with deep representation learning practices. We provide several
instantiations under this framework, from bilinear and feature-based nonlinear
to general nonlinear cases. In the bilinear case, by leveraging the non-uniform
spectrum of the task representation and the calibrated source-target relevance,
we prove that the sample complexity to achieve $\varepsilon$-excess risk on
target scales with $ (k^*)^2 \|v^*\|_2^2 \varepsilon^{-2}$ where $k^*$ is the
effective dimension of the target and $\|v^*\|_2^2 \in (0,1]$ represents the
connection between source and target space. Compared to the passive one, this
can save up to $\frac{1}{d_W}$ of sample complexity, where $d_W$ is the task
space dimension. Finally, we demonstrate different instantiations of our meta
algorithm in synthetic datasets and robotics problems, from pendulum
simulations to real-world drone flight datasets. On average, our algorithms
outperform baselines by $20\%-70\%$.
Related papers
- Metalearning with Very Few Samples Per Task [19.78398372660794]
We consider a binary classification setting where tasks are related by a shared representation.
Here, the amount of data is measured in terms of the number of tasks $t$ that we need to see and the number of samples $n$ per task.
Our work also yields a characterization of distribution-free multitask learning and reductions between meta and multitask learning.
arXiv Detail & Related papers (2023-12-21T16:06:44Z) - Improved Active Multi-Task Representation Learning via Lasso [44.607652031235716]
In this paper, we show the dominance of the L1-regularized-relevance-based ($nu1$) strategy by giving a lower bound for the $nu2$-based strategy.
We also characterize the potential of our $nu1$-based strategy in sample-cost-sensitive settings.
arXiv Detail & Related papers (2023-06-05T03:08:29Z) - Multi-Task Imitation Learning for Linear Dynamical Systems [50.124394757116605]
We study representation learning for efficient imitation learning over linear systems.
We find that the imitation gap over trajectories generated by the learned target policy is bounded by $tildeOleft( frack n_xHN_mathrmshared + frack n_uN_mathrmtargetright)$.
arXiv Detail & Related papers (2022-12-01T00:14:35Z) - On the Sample Complexity of Representation Learning in Multi-task
Bandits with Global and Local structure [77.60508571062958]
We investigate the sample complexity of learning the optimal arm for multi-task bandit problems.
Arms consist of two components: one that is shared across tasks (that we call representation) and one that is task-specific (that we call predictor)
We devise an algorithm OSRL-SC whose sample complexity approaches the lower bound, and scales at most as $H(Glog(delta_G)+ Xlog(delta_H))$, with $X,G,H$ being, respectively, the number of tasks, representations and predictors.
arXiv Detail & Related papers (2022-11-28T08:40:12Z) - Active Multi-Task Representation Learning [50.13453053304159]
We give the first formal study on resource task sampling by leveraging the techniques from active learning.
We propose an algorithm that iteratively estimates the relevance of each source task to the target task and samples from each source task based on the estimated relevance.
arXiv Detail & Related papers (2022-02-02T08:23:24Z) - Exploring Relational Context for Multi-Task Dense Prediction [76.86090370115]
We consider a multi-task environment for dense prediction tasks, represented by a common backbone and independent task-specific heads.
We explore various attention-based contexts, such as global and local, in the multi-task setting.
We propose an Adaptive Task-Relational Context module, which samples the pool of all available contexts for each task pair.
arXiv Detail & Related papers (2021-04-28T16:45:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.