Multi-Task Learning for Multi-Dimensional Regression: Application to
Luminescence Sensing
- URL: http://arxiv.org/abs/2007.13875v1
- Date: Mon, 27 Jul 2020 21:23:51 GMT
- Title: Multi-Task Learning for Multi-Dimensional Regression: Application to
Luminescence Sensing
- Authors: Umberto, Michelucci, Francesca Venturini
- Abstract summary: A new approach to non-linear regression is to use neural networks, particularly feed-forward architectures with a sufficient number of hidden layers and an appropriate number of output neurons.
We propose multi-task learning (MTL) architectures. These are characterized by multiple branches of task-specific layers, which have as input the output of a common set of layers.
To demonstrate the power of this approach for multi-dimensional regression, the method is applied to luminescence sensing.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The classical approach to non-linear regression in physics, is to take a
mathematical model describing the functional dependence of the dependent
variable from a set of independent variables, and then, using non-linear
fitting algorithms, extract the parameters used in the modeling. Particularly
challenging are real systems, characterized by several additional influencing
factors related to specific components, like electronics or optical parts. In
such cases, to make the model reproduce the data, empirically determined terms
are built-in the models to compensate for the impossibility of modeling things
that are, by construction, impossible to model. A new approach to solve this
issue is to use neural networks, particularly feed-forward architectures with a
sufficient number of hidden layers and an appropriate number of output neurons,
each responsible for predicting the desired variables. Unfortunately,
feed-forward neural networks (FFNNs) usually perform less efficiently when
applied to multi-dimensional regression problems, that is when they are
required to predict simultaneously multiple variables that depend from the
input dataset in fundamentally different ways. To address this problem, we
propose multi-task learning (MTL) architectures. These are characterized by
multiple branches of task-specific layers, which have as input the output of a
common set of layers. To demonstrate the power of this approach for
multi-dimensional regression, the method is applied to luminescence sensing.
Here the MTL architecture allows predicting multiple parameters, the oxygen
concentration and the temperature, from a single set of measurements.
Related papers
- A least distance estimator for a multivariate regression model using
deep neural networks [1.8876415010297893]
(A)GDNN-LD estimator enjoys variable selection and model estimation simultaneously, by applying the (adaptive) group Lasso penalty to weight parameters in the DNN structure.
For the computation, we propose a quadratic smoothing approximation method to facilitate optimizing the non-smooth objective function based on the least distance loss.
arXiv Detail & Related papers (2024-01-06T04:36:00Z) - A statistical approach to latent dynamic modeling with differential
equations [0.0]
Ordinary differential equations (ODEs) can provide mechanistic models of temporally local changes of processes.
We propose to use each observation in the course of time as the initial value to obtain multiple local ODE solutions.
We illustrate the proposed approach in an application with spinal muscular atrophy patients and a corresponding simulation study.
arXiv Detail & Related papers (2023-11-27T20:02:55Z) - Deep Learning-based surrogate models for parametrized PDEs: handling
geometric variability through graph neural networks [0.0]
This work explores the potential usage of graph neural networks (GNNs) for the simulation of time-dependent PDEs.
We propose a systematic strategy to build surrogate models based on a data-driven time-stepping scheme.
We show that GNNs can provide a valid alternative to traditional surrogate models in terms of computational efficiency and generalization to new scenarios.
arXiv Detail & Related papers (2023-08-03T08:14:28Z) - Neural network analysis of neutron and X-ray reflectivity data:
Incorporating prior knowledge for tackling the phase problem [141.5628276096321]
We present an approach that utilizes prior knowledge to regularize the training process over larger parameter spaces.
We demonstrate the effectiveness of our method in various scenarios, including multilayer structures with box model parameterization.
In contrast to previous methods, our approach scales favorably when increasing the complexity of the inverse problem.
arXiv Detail & Related papers (2023-06-28T11:15:53Z) - A predictive physics-aware hybrid reduced order model for reacting flows [65.73506571113623]
A new hybrid predictive Reduced Order Model (ROM) is proposed to solve reacting flow problems.
The number of degrees of freedom is reduced from thousands of temporal points to a few POD modes with their corresponding temporal coefficients.
Two different deep learning architectures have been tested to predict the temporal coefficients.
arXiv Detail & Related papers (2023-01-24T08:39:20Z) - Dynamic Latent Separation for Deep Learning [67.62190501599176]
A core problem in machine learning is to learn expressive latent variables for model prediction on complex data.
Here, we develop an approach that improves expressiveness, provides partial interpretation, and is not restricted to specific applications.
arXiv Detail & Related papers (2022-10-07T17:56:53Z) - Multi-fidelity Hierarchical Neural Processes [79.0284780825048]
Multi-fidelity surrogate modeling reduces the computational cost by fusing different simulation outputs.
We propose Multi-fidelity Hierarchical Neural Processes (MF-HNP), a unified neural latent variable model for multi-fidelity surrogate modeling.
We evaluate MF-HNP on epidemiology and climate modeling tasks, achieving competitive performance in terms of accuracy and uncertainty estimation.
arXiv Detail & Related papers (2022-06-10T04:54:13Z) - Scalable Gaussian Processes for Data-Driven Design using Big Data with
Categorical Factors [14.337297795182181]
Gaussian processes (GP) have difficulties in accommodating big datasets, categorical inputs, and multiple responses.
We propose a GP model that utilizes latent variables and functions obtained through variational inference to address the aforementioned challenges simultaneously.
Our approach is demonstrated for machine learning of ternary oxide materials and topology optimization of a multiscale compliant mechanism.
arXiv Detail & Related papers (2021-06-26T02:17:23Z) - Large-scale Neural Solvers for Partial Differential Equations [48.7576911714538]
Solving partial differential equations (PDE) is an indispensable part of many branches of science as many processes can be modelled in terms of PDEs.
Recent numerical solvers require manual discretization of the underlying equation as well as sophisticated, tailored code for distributed computing.
We examine the applicability of continuous, mesh-free neural solvers for partial differential equations, physics-informed neural networks (PINNs)
We discuss the accuracy of GatedPINN with respect to analytical solutions -- as well as state-of-the-art numerical solvers, such as spectral solvers.
arXiv Detail & Related papers (2020-09-08T13:26:51Z) - DynNet: Physics-based neural architecture design for linear and
nonlinear structural response modeling and prediction [2.572404739180802]
In this study, a physics-based recurrent neural network model is designed that is able to learn the dynamics of linear and nonlinear multiple degrees of freedom systems.
The model is able to estimate a complete set of responses, including displacement, velocity, acceleration, and internal forces.
arXiv Detail & Related papers (2020-07-03T17:05:35Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.