Invertible Fourier Neural Operators for Tackling Both Forward and
Inverse Problems
- URL: http://arxiv.org/abs/2402.11722v1
- Date: Sun, 18 Feb 2024 22:16:43 GMT
- Title: Invertible Fourier Neural Operators for Tackling Both Forward and
Inverse Problems
- Authors: Da Long and Shandian Zhe
- Abstract summary: We propose an invertible Fourier Neural Operator (iFNO) that tackles both the forward and inverse problems.
We integrated a variational auto-encoder to capture the intrinsic structures within the input space and to enable posterior inference.
The evaluations on five benchmark problems have demonstrated the effectiveness of our approach.
- Score: 18.48295539583625
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fourier Neural Operator (FNO) is a popular operator learning method, which
has demonstrated state-of-the-art performance across many tasks. However, FNO
is mainly used in forward prediction, yet a large family of applications rely
on solving inverse problems. In this paper, we propose an invertible Fourier
Neural Operator (iFNO) that tackles both the forward and inverse problems. We
designed a series of invertible Fourier blocks in the latent channel space to
share the model parameters, efficiently exchange the information, and mutually
regularize the learning for the bi-directional tasks. We integrated a
variational auto-encoder to capture the intrinsic structures within the input
space and to enable posterior inference so as to overcome challenges of
illposedness, data shortage, noises, etc. We developed a three-step process for
pre-training and fine tuning for efficient training. The evaluations on five
benchmark problems have demonstrated the effectiveness of our approach.
Related papers
- Hilbert Neural Operator: Operator Learning in the Analytic Signal Domain [0.0]
We introduce the textbfHilbert Neural Operator (HNO), a new neural operator architecture to address some advantages.<n>HNO operates by first mapping the input signal to its analytic representation via the Hilbert transform.<n>We hypothesize that this architecture enables HNO to model operators more effectively for causal, phase-sensitive, and non-stationary systems.
arXiv Detail & Related papers (2025-08-06T21:12:15Z) - Robustifying Fourier Features Embeddings for Implicit Neural Representations [25.725097757343367]
Implicit Neural Representations (INRs) employ neural networks to represent continuous functions by mapping coordinates to the corresponding values of the target function.<n>INRs face a challenge known as spectral bias when dealing with scenes containing varying frequencies.<n>We propose the use of multi-layer perceptrons (MLPs) without additive.
arXiv Detail & Related papers (2025-02-08T07:43:37Z) - Half-VAE: An Encoder-Free VAE to Bypass Explicit Inverse Mapping [5.212606755867746]
Inference and inverse problems are closely related concepts, both fundamentally involving the deduction of unknown causes or parameters from observed data.
This study explores the potential of VAEs for solving inverse problems, such as Independent Component Analysis (ICA)
Unlike other VAE-based ICA methods, this approach discards the encoder in the VAE architecture, directly setting the latent variables as trainable parameters.
arXiv Detail & Related papers (2024-09-06T09:11:15Z) - Spectral-Refiner: Accurate Fine-Tuning of Spatiotemporal Fourier Neural Operator for Turbulent Flows [6.961408873053586]
Recent in operator-type neural networks have shown promising results in approximating Partial Differential Equations (PDEs)
These neural networks entail considerable training expenses, and may not always achieve the desired accuracy required in many scientific and engineering disciplines.
arXiv Detail & Related papers (2024-05-27T14:33:06Z) - End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes [52.818579746354665]
This paper proposes the first end-to-end differentiable meta-BO framework that generalises neural processes to learn acquisition functions via transformer architectures.
We enable this end-to-end framework with reinforcement learning (RL) to tackle the lack of labelled acquisition data.
arXiv Detail & Related papers (2023-05-25T10:58:46Z) - Fourier Continuation for Exact Derivative Computation in
Physics-Informed Neural Operators [53.087564562565774]
PINO is a machine learning architecture that has shown promising empirical results for learning partial differential equations.
We present an architecture that leverages Fourier continuation (FC) to apply the exact gradient method to PINO for nonperiodic problems.
arXiv Detail & Related papers (2022-11-29T06:37:54Z) - Incremental Spatial and Spectral Learning of Neural Operators for
Solving Large-Scale PDEs [86.35471039808023]
We introduce the Incremental Fourier Neural Operator (iFNO), which progressively increases the number of frequency modes used by the model.
We show that iFNO reduces total training time while maintaining or improving generalization performance across various datasets.
Our method demonstrates a 10% lower testing error, using 20% fewer frequency modes compared to the existing Fourier Neural Operator, while also achieving a 30% faster training.
arXiv Detail & Related papers (2022-11-28T09:57:15Z) - Improving the Robustness of Neural Multiplication Units with Reversible
Stochasticity [2.4278445972594525]
Multilayer Perceptrons struggle to learn certain simple arithmetic tasks.
Specialist neural NMU (sNMU) is proposed to apply reversibleity, encouraging avoidance of such optima.
arXiv Detail & Related papers (2022-11-10T14:56:37Z) - Deep neural networks can stably solve high-dimensional, noisy,
non-linear inverse problems [2.6651200086513107]
We study the problem of reconstructing solutions of inverse problems when only noisy measurements are available.
For the inverse operator, we demonstrate that there exists a neural network which is a robust-to-noise approximation of the operator.
arXiv Detail & Related papers (2022-06-02T08:51:46Z) - Factorized Fourier Neural Operators [77.47313102926017]
The Factorized Fourier Neural Operator (F-FNO) is a learning-based method for simulating partial differential equations.
We show that our model maintains an error rate of 2% while still running an order of magnitude faster than a numerical solver.
arXiv Detail & Related papers (2021-11-27T03:34:13Z) - Adaptive Fourier Neural Operators: Efficient Token Mixers for
Transformers [55.90468016961356]
We propose an efficient token mixer that learns to mix in the Fourier domain.
AFNO is based on a principled foundation of operator learning.
It can handle a sequence size of 65k and outperforms other efficient self-attention mechanisms.
arXiv Detail & Related papers (2021-11-24T05:44:31Z) - Choose a Transformer: Fourier or Galerkin [0.0]
We apply the self-attention from the state-of-the-art Transformer in Attention Is All You Need to a data-driven operator learning problem.
We show that softmax normalization in the scaled dot-product attention is sufficient but not necessary, and have proved the approximation capacity of a linear variant as a Petrov-Galerkin projection.
We present three operator learning experiments, including the viscid Burgers' equation, an interface Darcy flow, and an inverse interface coefficient identification problem.
arXiv Detail & Related papers (2021-05-31T14:30:53Z) - Deep Feedback Inverse Problem Solver [141.26041463617963]
We present an efficient, effective, and generic approach towards solving inverse problems.
We leverage the feedback signal provided by the forward process and learn an iterative update model.
Our approach does not have any restrictions on the forward process; it does not require any prior knowledge either.
arXiv Detail & Related papers (2021-01-19T16:49:06Z) - Learning Set Functions that are Sparse in Non-Orthogonal Fourier Bases [73.53227696624306]
We present a new family of algorithms for learning Fourier-sparse set functions.
In contrast to other work that focused on the Walsh-Hadamard transform, our novel algorithms operate with recently introduced non-orthogonal Fourier transforms.
We demonstrate effectiveness on several real-world applications.
arXiv Detail & Related papers (2020-10-01T14:31:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.