Spectral Analysis for Semantic Segmentation with Applications on Feature
Truncation and Weak Annotation
- URL: http://arxiv.org/abs/2012.14123v5
- Date: Thu, 11 May 2023 12:21:33 GMT
- Title: Spectral Analysis for Semantic Segmentation with Applications on Feature
Truncation and Weak Annotation
- Authors: Li-Wei Chen, Wei-Chen Chiu, Chin-Tien Wu
- Abstract summary: A striking balance between the accuracy and the training cost of the SSNNs such as U-Net exists.
We propose a spectral analysis to investigate the correlations among the resolution of the down sampled grid, the loss function and the accuracy of the SSNNs.
- Score: 19.967870811543737
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: It is well known that semantic segmentation neural networks (SSNNs) produce
dense segmentation maps to resolve the objects' boundaries while restrict the
prediction on down-sampled grids to alleviate the computational cost. A
striking balance between the accuracy and the training cost of the SSNNs such
as U-Net exists. We propose a spectral analysis to investigate the correlations
among the resolution of the down sampled grid, the loss function and the
accuracy of the SSNNs. By analyzing the network back-propagation process in
frequency domain, we discover that the traditional loss function,
cross-entropy, and the key features of CNN are mainly affected by the
low-frequency components of segmentation labels. Our discoveries can be applied
to SSNNs in several ways including (i) determining an efficient low resolution
grid for resolving the segmentation maps (ii) pruning the networks by
truncating the high frequency decoder features for saving computation costs,
and (iii) using block-wise weak annotation for saving the labeling time.
Experimental results shown in this paper agree with our spectral analysis for
the networks such as DeepLab V3+ and Deep Aggregation Net (DAN).
Related papers
- Data-Driven Deep MIMO Detection:Network Architectures and Generalization Analysis [50.20709408241935]
This paper proposes inspecting the fully data-driven DeepSIC detection within a Network-of-MLPs architecture.<n>Within such an architecture, DeepSIC can be upgraded as a graph-based message-passing process using Graph Neural Networks (GNNs)<n>GNNSIC achieves excellent expressivity comparable to DeepSIC with substantially fewer trainable parameters.
arXiv Detail & Related papers (2026-02-13T04:38:51Z) - Preserving instance continuity and length in segmentation through connectivity-aware loss computation [0.29656637520758655]
We propose two novel loss functions, Negative Centerline Loss and Simplified Topology Loss, that help preserve connectivity of output instances.<n>We evaluate our approach on a 3D light-sheet fluorescence microscopy dataset of axon initial segments (AIS), a task prone to discontinuity due to signal dropout.
arXiv Detail & Related papers (2025-09-03T09:08:47Z) - Deeper or Wider: A Perspective from Optimal Generalization Error with Sobolev Loss [2.07180164747172]
We compare deeper neural networks (DeNNs) with a flexible number of layers and wider neural networks (WeNNs) with limited hidden layers.
We find that a higher number of parameters tends to favor WeNNs, while an increased number of sample points and greater regularity in the loss function lean towards the adoption of DeNNs.
arXiv Detail & Related papers (2024-01-31T20:10:10Z) - Geometric Learning-Based Transformer Network for Estimation of
Segmentation Errors [1.376408511310322]
We propose an approach to identify and measure erroneous regions in the segmentation map.
Our method can estimate error at any point or node in a 3D mesh generated from a possibly erroneous volumetric segmentation map.
We have evaluated our network on a high-resolution micro-CT dataset of the human inner-ear bony labyrinth structure.
arXiv Detail & Related papers (2023-08-09T16:58:03Z) - Joint Edge-Model Sparse Learning is Provably Efficient for Graph Neural
Networks [89.28881869440433]
This paper provides the first theoretical characterization of joint edge-model sparse learning for graph neural networks (GNNs)
It proves analytically that both sampling important nodes and pruning neurons with the lowest-magnitude can reduce the sample complexity and improve convergence without compromising the test accuracy.
arXiv Detail & Related papers (2023-02-06T16:54:20Z) - NAF: Neural Attenuation Fields for Sparse-View CBCT Reconstruction [79.13750275141139]
This paper proposes a novel and fast self-supervised solution for sparse-view CBCT reconstruction.
The desired attenuation coefficients are represented as a continuous function of 3D spatial coordinates, parameterized by a fully-connected deep neural network.
A learning-based encoder entailing hash coding is adopted to help the network capture high-frequency details.
arXiv Detail & Related papers (2022-09-29T04:06:00Z) - Mean-field Analysis of Piecewise Linear Solutions for Wide ReLU Networks [83.58049517083138]
We consider a two-layer ReLU network trained via gradient descent.
We show that SGD is biased towards a simple solution.
We also provide empirical evidence that knots at locations distinct from the data points might occur.
arXiv Detail & Related papers (2021-11-03T15:14:20Z) - Locally Sparse Networks for Interpretable Predictions [7.362415721170984]
We propose a framework for training locally sparse neural networks where the local sparsity is learned via a sample-specific gating mechanism.
The sample-specific sparsity is predicted via a textitgating network, which is trained in tandem with the textitprediction network.
We demonstrate that our method outperforms state-of-the-art models when predicting the target function with far fewer features per instance.
arXiv Detail & Related papers (2021-06-11T15:46:50Z) - SignalNet: A Low Resolution Sinusoid Decomposition and Estimation
Network [79.04274563889548]
We propose SignalNet, a neural network architecture that detects the number of sinusoids and estimates their parameters from quantized in-phase and quadrature samples.
We introduce a worst-case learning threshold for comparing the results of our network relative to the underlying data distributions.
In simulation, we find that our algorithm is always able to surpass the threshold for three-bit data but often cannot exceed the threshold for one-bit data.
arXiv Detail & Related papers (2021-06-10T04:21:20Z) - Topological obstructions in neural networks learning [67.8848058842671]
We study global properties of the loss gradient function flow.
We use topological data analysis of the loss function and its Morse complex to relate local behavior along gradient trajectories with global properties of the loss surface.
arXiv Detail & Related papers (2020-12-31T18:53:25Z) - Encoding the latent posterior of Bayesian Neural Networks for
uncertainty quantification [10.727102755903616]
We aim for efficient deep BNNs amenable to complex computer vision architectures.
We achieve this by leveraging variational autoencoders (VAEs) to learn the interaction and the latent distribution of the parameters at each network layer.
Our approach, Latent-Posterior BNN (LP-BNN), is compatible with the recent BatchEnsemble method, leading to highly efficient (in terms of computation and memory during both training and testing) ensembles.
arXiv Detail & Related papers (2020-12-04T19:50:09Z) - Binary Segmentation of Seismic Facies Using Encoder-Decoder Neural
Networks [0.0]
This work presents a Deep Neural Network for Facies (DNFS) to obtain state-of-the-art results for seismic facies segmentation.
DNFS is trained using a combination of cross-entropy and Jaccard loss functions.
Our results show that DNFS obtains highly detailed predictions for seismic facies segmentation using fewer parameters than StNet and U-Net.
arXiv Detail & Related papers (2020-11-15T01:36:52Z) - Beyond Dropout: Feature Map Distortion to Regularize Deep Neural
Networks [107.77595511218429]
In this paper, we investigate the empirical Rademacher complexity related to intermediate layers of deep neural networks.
We propose a feature distortion method (Disout) for addressing the aforementioned problem.
The superiority of the proposed feature map distortion for producing deep neural network with higher testing performance is analyzed and demonstrated.
arXiv Detail & Related papers (2020-02-23T13:59:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.