Open Problems in Applied Deep Learning
- URL: http://arxiv.org/abs/2301.11316v1
- Date: Thu, 26 Jan 2023 18:55:43 GMT
- Title: Open Problems in Applied Deep Learning
- Authors: Maziar Raissi
- Abstract summary: This work formulates the machine learning mechanism as a bi-level optimization problem.
The inner level optimization loop entails minimizing a properly chosen loss function evaluated on the training data.
The outer level optimization loop is less well-studied and involves maximizing a properly chosen performance metric evaluated on the validation data.
- Score: 2.1320960069210475
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This work formulates the machine learning mechanism as a bi-level
optimization problem. The inner level optimization loop entails minimizing a
properly chosen loss function evaluated on the training data. This is nothing
but the well-studied training process in pursuit of optimal model parameters.
The outer level optimization loop is less well-studied and involves maximizing
a properly chosen performance metric evaluated on the validation data. This is
what we call the "iteration process", pursuing optimal model hyper-parameters.
Among many other degrees of freedom, this process entails model engineering
(e.g., neural network architecture design) and management, experiment tracking,
dataset versioning and augmentation. The iteration process could be automated
via Automatic Machine Learning (AutoML) or left to the intuitions of machine
learning students, engineers, and researchers. Regardless of the route we take,
there is a need to reduce the computational cost of the iteration step and as a
direct consequence reduce the carbon footprint of developing artificial
intelligence algorithms. Despite the clean and unified mathematical formulation
of the iteration step as a bi-level optimization problem, its solutions are
case specific and complex. This work will consider such cases while increasing
the level of complexity from supervised learning to semi-supervised,
self-supervised, unsupervised, few-shot, federated, reinforcement, and
physics-informed learning. As a consequence of this exercise, this proposal
surfaces a plethora of open problems in the field, many of which can be
addressed in parallel.
Related papers
- Gradual Optimization Learning for Conformational Energy Minimization [69.36925478047682]
Gradual Optimization Learning Framework (GOLF) for energy minimization with neural networks significantly reduces the required additional data.
Our results demonstrate that the neural network trained with GOLF performs on par with the oracle on a benchmark of diverse drug-like molecules.
arXiv Detail & Related papers (2023-11-05T11:48:08Z) - Delayed Geometric Discounts: An Alternative Criterion for Reinforcement
Learning [1.52292571922932]
reinforcement learning (RL) proposes a theoretical background to learn optimal behaviors.
In practice, RL algorithms rely on geometric discounts to evaluate this optimality.
In this paper, we tackle these issues by generalizing the discounted problem formulation with a family of delayed objective functions.
arXiv Detail & Related papers (2022-09-26T07:49:38Z) - Learning to Optimize: A Primer and A Benchmark [94.29436694770953]
Learning to optimize (L2O) is an emerging approach that leverages machine learning to develop optimization methods.
This article is poised to be the first comprehensive survey and benchmark of L2O for continuous optimization.
arXiv Detail & Related papers (2021-03-23T20:46:20Z) - Application of an automated machine learning-genetic algorithm
(AutoML-GA) coupled with computational fluid dynamics simulations for rapid
engine design optimization [0.0]
The present work describes and validates an automated active learning approach, AutoML-GA, for surrogate-based optimization of internal combustion engines.
A genetic algorithm is employed to locate the design optimum on the machine learning surrogate surface.
It is demonstrated that AutoML-GA leads to a better optimum with a lower number of CFD simulations.
arXiv Detail & Related papers (2021-01-07T17:50:52Z) - Optimization for Supervised Machine Learning: Randomized Algorithms for
Data and Parameters [10.279748604797911]
Key problems in machine learning and data science are routinely modeled as optimization problems and solved via optimization algorithms.
With the increase of the volume of data and the size and complexity of the statistical models used to formulate these often ill-conditioned optimization tasks, there is a need for new efficient algorithms able to cope with these challenges.
In this thesis, we deal with each of these sources of difficulty in a different way. To efficiently address the big data issue, we develop new methods which in each iteration examine a small random subset of the training data only.
To handle the big model issue, we develop methods which in each iteration update
arXiv Detail & Related papers (2020-08-26T21:15:18Z) - Automatically Learning Compact Quality-aware Surrogates for Optimization
Problems [55.94450542785096]
Solving optimization problems with unknown parameters requires learning a predictive model to predict the values of the unknown parameters and then solving the problem using these values.
Recent work has shown that including the optimization problem as a layer in a complex training model pipeline results in predictions of iteration of unobserved decision making.
We show that we can improve solution quality by learning a low-dimensional surrogate model of a large optimization problem.
arXiv Detail & Related papers (2020-06-18T19:11:54Z) - A Primer on Zeroth-Order Optimization in Signal Processing and Machine
Learning [95.85269649177336]
ZO optimization iteratively performs three major steps: gradient estimation, descent direction, and solution update.
We demonstrate promising applications of ZO optimization, such as evaluating and generating explanations from black-box deep learning models, and efficient online sensor management.
arXiv Detail & Related papers (2020-06-11T06:50:35Z) - Global Optimization of Gaussian processes [52.77024349608834]
We propose a reduced-space formulation with trained Gaussian processes trained on few data points.
The approach also leads to significantly smaller and computationally cheaper sub solver for lower bounding.
In total, we reduce time convergence by orders of orders of the proposed method.
arXiv Detail & Related papers (2020-05-21T20:59:11Z) - Self-Directed Online Machine Learning for Topology Optimization [58.920693413667216]
Self-directed Online Learning Optimization integrates Deep Neural Network (DNN) with Finite Element Method (FEM) calculations.
Our algorithm was tested by four types of problems including compliance minimization, fluid-structure optimization, heat transfer enhancement and truss optimization.
It reduced the computational time by 2 5 orders of magnitude compared with directly using methods, and outperformed all state-of-the-art algorithms tested in our experiments.
arXiv Detail & Related papers (2020-02-04T20:00:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.