A Unified Framework for U-Net Design and Analysis
- URL: http://arxiv.org/abs/2305.19638v2
- Date: Wed, 10 Jan 2024 14:55:22 GMT
- Title: A Unified Framework for U-Net Design and Analysis
- Authors: Christopher Williams, Fabian Falck, George Deligiannidis, Chris
Holmes, Arnaud Doucet, Saifuddin Syed
- Abstract summary: We provide a framework for designing and analysing general U-Net architectures.
We propose Multi-ResNets, U-Nets with a simplified, wavelet-based encoder without learnable parameters.
In diffusion models, our framework enables us to identify that high-frequency information is dominated by noise exponentially faster.
- Score: 32.48954087522557
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: U-Nets are a go-to, state-of-the-art neural architecture across numerous
tasks for continuous signals on a square such as images and Partial
Differential Equations (PDE), however their design and architecture is
understudied. In this paper, we provide a framework for designing and analysing
general U-Net architectures. We present theoretical results which characterise
the role of the encoder and decoder in a U-Net, their high-resolution scaling
limits and their conjugacy to ResNets via preconditioning. We propose
Multi-ResNets, U-Nets with a simplified, wavelet-based encoder without
learnable parameters. Further, we show how to design novel U-Net architectures
which encode function constraints, natural bases, or the geometry of the data.
In diffusion models, our framework enables us to identify that high-frequency
information is dominated by noise exponentially faster, and show how U-Nets
with average pooling exploit this. In our experiments, we demonstrate how
Multi-ResNets achieve competitive and often superior performance compared to
classical U-Nets in image segmentation, PDE surrogate modelling, and generative
modelling with diffusion models. Our U-Net framework paves the way to study the
theoretical properties of U-Nets and design natural, scalable neural
architectures for a multitude of problems beyond the square.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - U-Nets as Belief Propagation: Efficient Classification, Denoising, and Diffusion in Generative Hierarchical Models [21.212179660694098]
We study certain generative hierarchical models, which are tree-structured graphical models extensively utilized in both language and image domains.
With their encoder-decoder structure, long skip connections, and pooling and up-sampling layers, we demonstrate how U-Nets can naturally implement the belief propagation denoising algorithm.
We also demonstrate that the conventional architecture of convolutional neural networks (ConvNets) is ideally suited for classification tasks within these models.
arXiv Detail & Related papers (2024-04-29T05:57:03Z) - TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - A Multi-Resolution Framework for U-Nets with Applications to
Hierarchical VAEs [29.995904718691204]
We formulate a multi-resolution framework which identifies U-Nets as finite-dimensional truncations of models on an infinite-dimensional function space.
We then leverage our framework to identify state-of-the-art hierarchical VAEs (HVAEs) which have a U-Net architecture.
arXiv Detail & Related papers (2023-01-19T17:33:48Z) - Towards Multi-spatiotemporal-scale Generalized PDE Modeling [4.924631198058705]
We make a comparison between various FNO and U-Net like approaches on fluid mechanics problems in both vorticity-stream and velocity function form.
We show promising results on generalization to different PDE parameters and time-scales with a single surrogate model.
arXiv Detail & Related papers (2022-09-30T17:40:05Z) - Universal approximation property of invertible neural networks [76.95927093274392]
Invertible neural networks (INNs) are neural network architectures with invertibility by design.
Thanks to their invertibility and the tractability of Jacobian, INNs have various machine learning applications such as probabilistic modeling, generative modeling, and representation learning.
arXiv Detail & Related papers (2022-04-15T10:45:26Z) - Learning Deep Context-Sensitive Decomposition for Low-Light Image
Enhancement [58.72667941107544]
A typical framework is to simultaneously estimate the illumination and reflectance, but they disregard the scene-level contextual information encapsulated in feature spaces.
We develop a new context-sensitive decomposition network architecture to exploit the scene-level contextual dependencies on spatial scales.
We develop a lightweight CSDNet (named LiteCSDNet) by reducing the number of channels.
arXiv Detail & Related papers (2021-12-09T06:25:30Z) - Dynamic Graph: Learning Instance-aware Connectivity for Neural Networks [78.65792427542672]
Dynamic Graph Network (DG-Net) is a complete directed acyclic graph, where the nodes represent convolutional blocks and the edges represent connection paths.
Instead of using the same path of the network, DG-Net aggregates features dynamically in each node, which allows the network to have more representation ability.
arXiv Detail & Related papers (2020-10-02T16:50:26Z) - BiO-Net: Learning Recurrent Bi-directional Connections for
Encoder-Decoder Architecture [82.64881585566825]
We present a novel Bi-directional O-shape network (BiO-Net) that reuses the building blocks in a recurrent manner without introducing any extra parameters.
Our method significantly outperforms the vanilla U-Net as well as other state-of-the-art methods.
arXiv Detail & Related papers (2020-07-01T05:07:49Z) - Tidying Deep Saliency Prediction Architectures [6.613005108411055]
In this paper, we identify four key components of saliency models, i.e., input features, multi-level integration, readout architecture, and loss functions.
We propose two novel end-to-end architectures called SimpleNet and MDNSal, which are neater, minimal, more interpretable and achieve state of the art performance on public saliency benchmarks.
arXiv Detail & Related papers (2020-03-10T19:34:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.