Prion-ViT: Prions-Inspired Vision Transformers for Temperature prediction with Specklegrams
- URL: http://arxiv.org/abs/2411.05836v2
- Date: Thu, 14 Nov 2024 03:21:12 GMT
- Title: Prion-ViT: Prions-Inspired Vision Transformers for Temperature prediction with Specklegrams
- Authors: Abhishek Sebastian, Pragna R,
- Abstract summary: Prion-ViT is a vision transformer inspired by biological prion memory mecha-nisms.
It reduces mean absolute error (MAE) to 0.52degC and outperforming models like ResNet, Inception Net V2, and standard vision transformers.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Fiber Specklegram Sensors (FSS) are vital for environmental monitoring due to their high temperature sensitivity, but their complex data poses challeng-es for predictive models. This study introduces Prion-ViT, a prion-inspired Vision Transformer model, inspired by biological prion memory mecha-nisms, to improve long-term dependency modeling and temperature prediction accuracy using FSS data. Prion-ViT leverages a persistent memory state to retain and propagate key features across layers, reducing mean absolute error (MAE) to 0.52{\deg}C and outperforming models like ResNet, Inception Net V2, and standard vision transformers. This work highlights Prion-ViT's potential for real-time industrial temperature monitoring and broader optical sensing applications.
Related papers
- A Sensorimotor Vision Transformer [0.0]
Sensorimotor Transformer (SMT) is a vision model inspired by human saccadic eye movements.
SMT identifies and selects the most salient patches based on intrinsic two-dimensional (i2D) features.
arXiv Detail & Related papers (2025-04-03T12:37:44Z) - Residual-based Attention Physics-informed Neural Networks for Spatio-Temporal Ageing Assessment of Transformers Operated in Renewable Power Plants [0.6223528900192875]
This article introduces a-temporal model for transformer winding temperature and ageing.
It uses physics-based partial differential equations with data-driven Neural Networks.
Results are validated with a distribution transformer operating on a floating photovoltaic power plant.
arXiv Detail & Related papers (2024-05-10T12:48:57Z) - Denoising Vision Transformers [43.03068202384091]
We propose a two-stage denoising approach, termed Denoising Vision Transformers (DVT)
In the first stage, we separate the clean features from those contaminated by positional artifacts by enforcing cross-view feature consistency with neural fields on a per-image basis.
In the second stage, we train a lightweight transformer block to predict clean features from raw ViT outputs, leveraging the derived estimates of the clean features as supervision.
arXiv Detail & Related papers (2024-01-05T18:59:52Z) - TSViT: A Time Series Vision Transformer for Fault Diagnosis [2.710064390178205]
This paper presents the Time Series Vision Transformer (TSViT) for effective fault diagnosis.
TSViT incorporates a convolutional layer to extract local features from vibration signals, alongside a transformer encoder to discern long-term temporal patterns.
Remarkably, TSViT achieves an unprecedented 100% average accuracy on two test sets and 99.99% on another.
arXiv Detail & Related papers (2023-11-12T18:16:48Z) - Generalized Face Forgery Detection via Adaptive Learning for Pre-trained Vision Transformer [54.32283739486781]
We present a textbfForgery-aware textbfAdaptive textbfVision textbfTransformer (FA-ViT) under the adaptive learning paradigm.
FA-ViT achieves 93.83% and 78.32% AUC scores on Celeb-DF and DFDC datasets in the cross-dataset evaluation.
arXiv Detail & Related papers (2023-09-20T06:51:11Z) - 2-D SSM: A General Spatial Layer for Visual Transformers [79.4957965474334]
A central objective in computer vision is to design models with appropriate 2-D inductive bias.
We leverage an expressive variation of the multidimensional State Space Model.
Our approach introduces efficient parameterization, accelerated computation, and a suitable normalization scheme.
arXiv Detail & Related papers (2023-06-11T09:41:37Z) - V1T: large-scale mouse V1 response prediction using a Vision Transformer [1.5703073293718952]
We introduce V1T, a novel Vision Transformer based architecture that learns a shared visual and behavioral representation across animals.
We evaluate our model on two large datasets recorded from mouse primary visual cortex and outperform previous convolution-based models by more than 12.7% in prediction performance.
arXiv Detail & Related papers (2023-02-06T18:58:38Z) - A Close Look at Spatial Modeling: From Attention to Convolution [70.5571582194057]
Vision Transformers have shown great promise recently for many vision tasks due to the insightful architecture design and attention mechanism.
We generalize self-attention formulation to abstract a queryirrelevant global context directly and integrate the global context into convolutions.
With less than 14M parameters, our FCViT-S12 outperforms related work ResT-Lite by 3.7% top1 accuracy on ImageNet-1K.
arXiv Detail & Related papers (2022-12-23T19:13:43Z) - Environmental Sensor Placement with Convolutional Gaussian Neural
Processes [65.13973319334625]
It is challenging to place sensors in a way that maximises the informativeness of their measurements, particularly in remote regions like Antarctica.
Probabilistic machine learning models can suggest informative sensor placements by finding sites that maximally reduce prediction uncertainty.
This paper proposes using a convolutional Gaussian neural process (ConvGNP) to address these issues.
arXiv Detail & Related papers (2022-11-18T17:25:14Z) - Surface Vision Transformers: Flexible Attention-Based Modelling of
Biomedical Surfaces [9.425082767553935]
Recent state-of-the-art performances of Vision Transformers (ViT) in computer vision tasks demonstrate that ViT could replace local feature learning operations of convolutional neural networks.
We extend ViTs to surfaces by reformulating the task of surface learning as a sequence-to-sequence learning problem.
We validate our method on a range of different biomedical surface domains and tasks.
arXiv Detail & Related papers (2022-04-07T12:45:54Z) - Improving Vision Transformers by Revisiting High-frequency Components [106.7140968644414]
We show that Vision Transformer (ViT) models are less effective in capturing the high-frequency components of images than CNN models.
To compensate, we propose HAT, which directly augments high-frequency components of images via adversarial training.
We show that HAT can consistently boost the performance of various ViT models.
arXiv Detail & Related papers (2022-04-03T05:16:51Z) - Learning Generative Vision Transformer with Energy-Based Latent Space
for Saliency Prediction [51.80191416661064]
We propose a novel vision transformer with latent variables following an informative energy-based prior for salient object detection.
Both the vision transformer network and the energy-based prior model are jointly trained via Markov chain Monte Carlo-based maximum likelihood estimation.
With the generative vision transformer, we can easily obtain a pixel-wise uncertainty map from an image, which indicates the model confidence in predicting saliency from the image.
arXiv Detail & Related papers (2021-12-27T06:04:33Z) - PhysFormer: Facial Video-based Physiological Measurement with Temporal
Difference Transformer [55.936527926778695]
Recent deep learning approaches focus on mining subtle r clues using convolutional neural networks with limited-temporal receptive fields.
In this paper, we propose the PhysFormer, an end-to-end video transformer based architecture.
arXiv Detail & Related papers (2021-11-23T18:57:11Z) - Fine-tuning Vision Transformers for the Prediction of State Variables in
Ising Models [2.9005223064604078]
Transformers are state-of-the-art deep learning models that are composed of stacked attention and point-wise, fully connected layers.
In this work, a Vision Transformer (ViT) is applied to predict the state variables of 2-dimensional Ising model simulations.
arXiv Detail & Related papers (2021-09-28T00:23:31Z) - Vision Transformers are Robust Learners [65.91359312429147]
We study the robustness of the Vision Transformer (ViT) against common corruptions and perturbations, distribution shifts, and natural adversarial examples.
We present analyses that provide both quantitative and qualitative indications to explain why ViTs are indeed more robust learners.
arXiv Detail & Related papers (2021-05-17T02:39:22Z) - Transformers Solve the Limited Receptive Field for Monocular Depth
Prediction [82.90445525977904]
We propose TransDepth, an architecture which benefits from both convolutional neural networks and transformers.
This is the first paper which applies transformers into pixel-wise prediction problems involving continuous labels.
arXiv Detail & Related papers (2021-03-22T18:00:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.