Fast Line Search for Multi-Task Learning
- URL: http://arxiv.org/abs/2110.00874v1
- Date: Sat, 2 Oct 2021 21:02:29 GMT
- Title: Fast Line Search for Multi-Task Learning
- Authors: Andrey Filatov and Daniil Merkulov
- Abstract summary: We propose a novel idea for line search algorithms in multi-task learning.
The idea is to use latent representation space instead of parameter space for finding step size.
We compare this idea with classical backtracking and gradient methods with a constant learning rate on MNIST, CIFAR-10, Cityscapes tasks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Multi-task learning is a powerful method for solving several tasks jointly by
learning robust representation. Optimization of the multi-task learning model
is a more complex task than a single-task due to task conflict. Based on
theoretical results, convergence to the optimal point is guaranteed when step
size is chosen through line search. But, usually, line search for the step size
is not the best choice due to the large computational time overhead. We propose
a novel idea for line search algorithms in multi-task learning. The idea is to
use latent representation space instead of parameter space for finding step
size. We examined this idea with backtracking line search. We compare this fast
backtracking algorithm with classical backtracking and gradient methods with a
constant learning rate on MNIST, CIFAR-10, Cityscapes tasks. The systematic
empirical study showed that the proposed method leads to more accurate and fast
solution, than the traditional backtracking approach and keep competitive
computational time and performance compared to the constant learning rate
method.
Related papers
- Learning Dual-arm Object Rearrangement for Cartesian Robots [28.329845378085054]
This work focuses on the dual-arm object rearrangement problem abstracted from a realistic industrial scenario of Cartesian robots.
The goal of this problem is to transfer all the objects from sources to targets with the minimum total completion time.
We develop an effective object-to-arm task assignment strategy for minimizing the cumulative task execution time and maximizing the dual-arm cooperation efficiency.
arXiv Detail & Related papers (2024-02-21T09:13:08Z) - Tree-of-Mixed-Thought: Combining Fast and Slow Thinking for Multi-hop
Visual Reasoning [16.495754104540605]
Large language models (LLMs) can generate code-like plans for complex inference tasks such as visual reasoning.
We propose a hierarchical plan-searching algorithm that integrates the one-stop reasoning (fast) and the Tree-of-thought (slow)
arXiv Detail & Related papers (2023-08-18T16:21:40Z) - Multitask Learning with No Regret: from Improved Confidence Bounds to
Active Learning [79.07658065326592]
Quantifying uncertainty in the estimated tasks is of pivotal importance for many downstream applications, such as online or active learning.
We provide novel multitask confidence intervals in the challenging setting when neither the similarity between tasks nor the tasks' features are available to the learner.
We propose a novel online learning algorithm that achieves such improved regret without knowing this parameter in advance.
arXiv Detail & Related papers (2023-08-03T13:08:09Z) - NEVIS'22: A Stream of 100 Tasks Sampled from 30 Years of Computer Vision
Research [96.53307645791179]
We introduce the Never-Ending VIsual-classification Stream (NEVIS'22), a benchmark consisting of a stream of over 100 visual classification tasks.
Despite being limited to classification, the resulting stream has a rich diversity of tasks from OCR, to texture analysis, scene recognition, and so forth.
Overall, NEVIS'22 poses an unprecedented challenge for current sequential learning approaches due to the scale and diversity of tasks.
arXiv Detail & Related papers (2022-11-15T18:57:46Z) - CrossBeam: Learning to Search in Bottom-Up Program Synthesis [51.37514793318815]
We propose training a neural model to learn a hands-on search policy for bottom-up synthesis.
Our approach, called CrossBeam, uses the neural model to choose how to combine previously-explored programs into new programs.
We observe that CrossBeam learns to search efficiently, exploring much smaller portions of the program space compared to the state-of-the-art.
arXiv Detail & Related papers (2022-03-20T04:41:05Z) - In Defense of the Unitary Scalarization for Deep Multi-Task Learning [121.76421174107463]
We present a theoretical analysis suggesting that many specialized multi-tasks can be interpreted as forms of regularization.
We show that, when coupled with standard regularization and stabilization techniques, unitary scalarization matches or improves upon the performance of complex multitasks.
arXiv Detail & Related papers (2022-01-11T18:44:17Z) - Leveraging Experience in Lazy Search [37.75223642505171]
Lazy graph search algorithms are efficient at solving motion planning problems where edge evaluation is the computational bottleneck.
We formulate this problem as a Markov Decision Process (MDP) on the state of the search problem.
We show that we can compute oracular selectors that can solve the MDP during training.
arXiv Detail & Related papers (2021-10-10T00:46:44Z) - Efficient Active Search for Combinatorial Optimization Problems [1.6543719822033436]
We show that (efficient) active search enables learned models to effectively solve instances that are much larger than those seen during training.
The proposed methods offer a simple way to significantly improve the search performance of a given model and outperform state-of-the-art machine learning based methods on routing problems.
arXiv Detail & Related papers (2021-06-09T15:08:03Z) - Towards Optimally Efficient Tree Search with Deep Learning [76.64632985696237]
This paper investigates the classical integer least-squares problem which estimates signals integer from linear models.
The problem is NP-hard and often arises in diverse applications such as signal processing, bioinformatics, communications and machine learning.
We propose a general hyper-accelerated tree search (HATS) algorithm by employing a deep neural network to estimate the optimal estimation for the underlying simplified memory-bounded A* algorithm.
arXiv Detail & Related papers (2021-01-07T08:00:02Z) - Linear Mode Connectivity in Multitask and Continual Learning [46.98656798573886]
We investigate whether multitask and continual solutions are similarly connected.
We propose an effective algorithm that constrains the sequentially learned minima to behave as the multitask solution.
arXiv Detail & Related papers (2020-10-09T10:53:25Z) - Meta-learning with Stochastic Linear Bandits [120.43000970418939]
We consider a class of bandit algorithms that implement a regularized version of the well-known OFUL algorithm, where the regularization is a square euclidean distance to a bias vector.
We show both theoretically and experimentally, that when the number of tasks grows and the variance of the task-distribution is small, our strategies have a significant advantage over learning the tasks in isolation.
arXiv Detail & Related papers (2020-05-18T08:41:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.