Inducing Point Operator Transformer: A Flexible and Scalable
Architecture for Solving PDEs
- URL: http://arxiv.org/abs/2312.10975v1
- Date: Mon, 18 Dec 2023 06:57:31 GMT
- Title: Inducing Point Operator Transformer: A Flexible and Scalable
Architecture for Solving PDEs
- Authors: Seungjun Lee, Taeil Oh
- Abstract summary: We introduce an attention-based model called an inducing-point operator transformer (IPOT)
IPOT is designed to handle any input function and output query while capturing global interactions in a computationally efficient way.
By detaching the inputs/outputs discretizations from the processor with a smaller latent bottleneck, IPOT offers flexibility in processing arbitrary discretizations.
- Score: 7.152311859951986
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Solving partial differential equations (PDEs) by learning the solution
operators has emerged as an attractive alternative to traditional numerical
methods. However, implementing such architectures presents two main challenges:
flexibility in handling irregular and arbitrary input and output formats and
scalability to large discretizations. Most existing architectures are limited
by their desired structure or infeasible to scale large inputs and outputs. To
address these issues, we introduce an attention-based model called an
inducing-point operator transformer (IPOT). Inspired by inducing points
methods, IPOT is designed to handle any input function and output query while
capturing global interactions in a computationally efficient way. By detaching
the inputs/outputs discretizations from the processor with a smaller latent
bottleneck, IPOT offers flexibility in processing arbitrary discretizations and
scales linearly with the size of inputs/outputs. Our experimental results
demonstrate that IPOT achieves strong performances with manageable
computational complexity on an extensive range of PDE benchmarks and real-world
weather forecasting scenarios, compared to state-of-the-art methods.
Related papers
- Resource Management for Low-latency Cooperative Fine-tuning of Foundation Models at the Network Edge [35.40849522296486]
Large-scale foundation models (FoMos) can perform human-like intelligence.
FoMos need to be adapted to specialized downstream tasks through fine-tuning techniques.
We advocate multi-device cooperation within the device-edge cooperative fine-tuning paradigm.
arXiv Detail & Related papers (2024-07-13T12:47:14Z) - Tender: Accelerating Large Language Models via Tensor Decomposition and Runtime Requantization [0.6445087473595953]
Large language models (LLMs) demonstrate outstanding performance in various tasks in machine learning.
deploying LLM inference poses challenges due to the high compute and memory requirements.
We present Tender, an algorithm-hardware co-design solution that enables efficient deployment of LLM inference at low precision.
arXiv Detail & Related papers (2024-06-16T09:51:55Z) - Using AI libraries for Incompressible Computational Fluid Dynamics [0.7734726150561089]
We present a novel methodology to bring the power of both AI software and hardware into the field of numerical modelling.
We use the proposed methodology to solve the advection-diffusion equation, the non-linear Burgers equation and incompressible flow past a bluff body.
arXiv Detail & Related papers (2024-02-27T22:00:50Z) - HAMLET: Graph Transformer Neural Operator for Partial Differential Equations [13.970458554623939]
We present a novel graph transformer framework, HAMLET, designed to address the challenges in solving partial differential equations (PDEs) using neural networks.
The framework uses graph transformers with modular input encoders to directly incorporate differential equation information into the solution process.
Notably, HAMLET scales effectively with increasing data complexity and noise, showcasing its robustness.
arXiv Detail & Related papers (2024-02-05T21:55:24Z) - A Multi-Head Ensemble Multi-Task Learning Approach for Dynamical
Computation Offloading [62.34538208323411]
We propose a multi-head ensemble multi-task learning (MEMTL) approach with a shared backbone and multiple prediction heads (PHs)
MEMTL outperforms benchmark methods in both the inference accuracy and mean square error without requiring additional training data.
arXiv Detail & Related papers (2023-09-02T11:01:16Z) - Energy-efficient Task Adaptation for NLP Edge Inference Leveraging
Heterogeneous Memory Architectures [68.91874045918112]
adapter-ALBERT is an efficient model optimization for maximal data reuse across different tasks.
We demonstrate the advantage of mapping the model to a heterogeneous on-chip memory architecture by performing simulations on a validated NLP edge accelerator.
arXiv Detail & Related papers (2023-03-25T14:40:59Z) - Solving High-Dimensional PDEs with Latent Spectral Models [74.1011309005488]
We present Latent Spectral Models (LSM) toward an efficient and precise solver for high-dimensional PDEs.
Inspired by classical spectral methods in numerical analysis, we design a neural spectral block to solve PDEs in the latent space.
LSM achieves consistent state-of-the-art and yields a relative gain of 11.5% averaged on seven benchmarks.
arXiv Detail & Related papers (2023-01-30T04:58:40Z) - Neural Stochastic Dual Dynamic Programming [99.80617899593526]
We introduce a trainable neural model that learns to map problem instances to a piece-wise linear value function.
$nu$-SDDP can significantly reduce problem solving cost without sacrificing solution quality.
arXiv Detail & Related papers (2021-12-01T22:55:23Z) - Adaptive Fourier Neural Operators: Efficient Token Mixers for
Transformers [55.90468016961356]
We propose an efficient token mixer that learns to mix in the Fourier domain.
AFNO is based on a principled foundation of operator learning.
It can handle a sequence size of 65k and outperforms other efficient self-attention mechanisms.
arXiv Detail & Related papers (2021-11-24T05:44:31Z) - Neural TMDlayer: Modeling Instantaneous flow of features via SDE
Generators [37.92379202320938]
We study how differential equation (SDE) based ideas can inspire new modifications to existing algorithms for a set of problems in computer vision.
We show promising experiments on a number of vision tasks including few shot learning, point cloud transformers and deep variational segmentation.
arXiv Detail & Related papers (2021-08-19T19:54:04Z) - Efficient Learning of Generative Models via Finite-Difference Score
Matching [111.55998083406134]
We present a generic strategy to efficiently approximate any-order directional derivative with finite difference.
Our approximation only involves function evaluations, which can be executed in parallel, and no gradient computations.
arXiv Detail & Related papers (2020-07-07T10:05:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.