Differentiable VQ-VAE's for Robust White Matter Streamline Encodings
- URL: http://arxiv.org/abs/2311.06212v2
- Date: Sat, 18 Nov 2023 17:49:26 GMT
- Title: Differentiable VQ-VAE's for Robust White Matter Streamline Encodings
- Authors: Andrew Lizarraga, Brandon Taraku, Edouardo Honig, Ying Nian Wu,
Shantanu H. Joshi
- Abstract summary: Autoencoders have been proposed as a dimension-reduction tool to simplify the analysis streamlines in a low-dimensional latent spaces.
We propose a novel Differentiable Vector Quantized Variational Autoencoder, which ingests entire bundles of streamlines as single data-point.
- Score: 33.936125620525
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Given the complex geometry of white matter streamlines, Autoencoders have
been proposed as a dimension-reduction tool to simplify the analysis
streamlines in a low-dimensional latent spaces. However, despite these recent
successes, the majority of encoder architectures only perform dimension
reduction on single streamlines as opposed to a full bundle of streamlines.
This is a severe limitation of the encoder architecture that completely
disregards the global geometric structure of streamlines at the expense of
individual fibers. Moreover, the latent space may not be well structured which
leads to doubt into their interpretability. In this paper we propose a novel
Differentiable Vector Quantized Variational Autoencoder, which are engineered
to ingest entire bundles of streamlines as single data-point and provides
reliable trustworthy encodings that can then be later used to analyze
streamlines in the latent space. Comparisons with several state of the art
Autoencoders demonstrate superior performance in both encoding and synthesis.
Related papers
- PMR-Net: Parallel Multi-Resolution Encoder-Decoder Network Framework for Medical Image Segmentation [5.554987043653931]
We propose a novel parallel multi-resolution encoder-decoder network, namely PMR-Net.
The proposed PMR-Net can achieve more accurate segmentation results than state-of-the-art methods on five public available datasets.
arXiv Detail & Related papers (2024-09-19T11:45:08Z) - Decoder Decomposition for the Analysis of the Latent Space of Nonlinear Autoencoders With Wind-Tunnel Experimental Data [3.7960472831772765]
The goal of this paper is to propose a method to aid the interpretability of autoencoders.
We propose the decoder decomposition, which is a post-processing method to connect the latent variables to the coherent structures of flows.
The ability to rank and select latent variables will help users design and interpret nonlinear autoencoders.
arXiv Detail & Related papers (2024-04-25T10:09:37Z) - Triple-Encoders: Representations That Fire Together, Wire Together [51.15206713482718]
Contrastive Learning is a representation learning method that encodes relative distances between utterances into the embedding space via a bi-encoder.
This study introduces triple-encoders, which efficiently compute distributed utterance mixtures from these independently encoded utterances.
We find that triple-encoders lead to a substantial improvement over bi-encoders, and even to better zero-shot generalization than single-vector representation models.
arXiv Detail & Related papers (2024-02-19T18:06:02Z) - Interpretable Spectral Variational AutoEncoder (ISVAE) for time series
clustering [48.0650332513417]
We introduce a novel model that incorporates an interpretable bottleneck-termed the Filter Bank (FB)-at the outset of a Variational Autoencoder (VAE)
This arrangement compels the VAE to attend on the most informative segments of the input signal.
By deliberately constraining the VAE with this FB, we promote the development of an encoding that is discernible, separable, and of reduced dimensionality.
arXiv Detail & Related papers (2023-10-18T13:06:05Z) - Decoder-Only or Encoder-Decoder? Interpreting Language Model as a
Regularized Encoder-Decoder [75.03283861464365]
The seq2seq task aims at generating the target sequence based on the given input source sequence.
Traditionally, most of the seq2seq task is resolved by an encoder to encode the source sequence and a decoder to generate the target text.
Recently, a bunch of new approaches have emerged that apply decoder-only language models directly to the seq2seq task.
arXiv Detail & Related papers (2023-04-08T15:44:29Z) - On the Suitability of Representations for Quality Diversity Optimization
of Shapes [77.34726150561087]
The representation, or encoding, utilized in evolutionary algorithms has a substantial effect on their performance.
This study compares the impact of several representations, including direct encoding, a dictionary-based representation, parametric encoding, compositional pattern producing networks, and cellular automata, on the generation of voxelized meshes.
arXiv Detail & Related papers (2023-04-07T07:34:23Z) - Rethinking Text Line Recognition Models [57.47147190119394]
We consider two decoder families (Connectionist Temporal Classification and Transformer) and three encoder modules (Bidirectional LSTMs, Self-Attention, and GRCLs)
We compare their accuracy and performance on widely used public datasets of scene and handwritten text.
Unlike the more common Transformer-based models, this architecture can handle inputs of arbitrary length.
arXiv Detail & Related papers (2021-04-15T21:43:13Z) - Non-linear, Sparse Dimensionality Reduction via Path Lasso Penalized
Autoencoders [0.0]
We present path lasso penalized autoencoders for complex data structures.
Our algorithm uses a group lasso penalty and non-negative matrix factorization to construct a sparse, non-linear latent representation.
We show that the algorithm exhibits much lower reconstruction errors than sparse PCA and parameter-wise lasso regularized autoencoders for low-dimensional representations.
arXiv Detail & Related papers (2021-02-22T10:14:46Z) - Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective
with Transformers [149.78470371525754]
We treat semantic segmentation as a sequence-to-sequence prediction task. Specifically, we deploy a pure transformer to encode an image as a sequence of patches.
With the global context modeled in every layer of the transformer, this encoder can be combined with a simple decoder to provide a powerful segmentation model, termed SEgmentation TRansformer (SETR)
SETR achieves new state of the art on ADE20K (50.28% mIoU), Pascal Context (55.83% mIoU) and competitive results on Cityscapes.
arXiv Detail & Related papers (2020-12-31T18:55:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.