Deep Learning for Multi-View Ultrasonic Image Fusion
- URL: http://arxiv.org/abs/2109.03616v1
- Date: Wed, 8 Sep 2021 13:04:07 GMT
- Title: Deep Learning for Multi-View Ultrasonic Image Fusion
- Authors: Georgios Pilikos, Lars Horchens, Tristan van Leeuwen, Felix Lucka
- Abstract summary: Delay-And-Sum (DAS) algorithm forms images using the main path on which reflected signals travel back to transducers.
Traditional image fusion techniques typically use ad-hoc combinations of pre-defined image transforms, pooling operations and thresholding.
We propose a deep neural network architecture that directly maps all available data to a segmentation map while explicitly incorporating the DAS image formation for the different insonification paths as network layers.
- Score: 2.1410799064827226
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Ultrasonic imaging is being used to obtain information about the acoustic
properties of a medium by emitting waves into it and recording their
interaction using ultrasonic transducer arrays. The Delay-And-Sum (DAS)
algorithm forms images using the main path on which reflected signals travel
back to the transducers. In some applications, different insonification paths
can be considered, for instance by placing the transducers at different
locations or if strong reflectors inside the medium are known a-priori. These
different modes give rise to multiple DAS images reflecting different geometric
information about the scatterers and the challenge is to either fuse them into
one image or to directly extract higher-level information regarding the
materials of the medium, e.g., a segmentation map. Traditional image fusion
techniques typically use ad-hoc combinations of pre-defined image transforms,
pooling operations and thresholding. In this work, we propose a deep neural
network (DNN) architecture that directly maps all available data to a
segmentation map while explicitly incorporating the DAS image formation for the
different insonification paths as network layers. This enables information flow
between data pre-processing and image post-processing DNNs, trained end-to-end.
We compare our proposed method to a traditional image fusion technique using
simulated data experiments, mimicking a non-destructive testing application
with four image modes, i.e., two transducer locations and two internal
reflection boundaries. Using our approach, it is possible to obtain much more
accurate segmentation of defects.
Related papers
- Deep Learning Based Speckle Filtering for Polarimetric SAR Images. Application to Sentinel-1 [51.404644401997736]
We propose a complete framework to remove speckle in polarimetric SAR images using a convolutional neural network.
Experiments show that the proposed approach offers exceptional results in both speckle reduction and resolution preservation.
arXiv Detail & Related papers (2024-08-28T10:07:17Z) - DiAD: A Diffusion-based Framework for Multi-class Anomaly Detection [55.48770333927732]
We propose a Difusion-based Anomaly Detection (DiAD) framework for multi-class anomaly detection.
It consists of a pixel-space autoencoder, a latent-space Semantic-Guided (SG) network with a connection to the stable diffusion's denoising network, and a feature-space pre-trained feature extractor.
Experiments on MVTec-AD and VisA datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-12-11T18:38:28Z) - Unified Frequency-Assisted Transformer Framework for Detecting and
Grounding Multi-Modal Manipulation [109.1912721224697]
We present the Unified Frequency-Assisted transFormer framework, named UFAFormer, to address the DGM4 problem.
By leveraging the discrete wavelet transform, we decompose images into several frequency sub-bands, capturing rich face forgery artifacts.
Our proposed frequency encoder, incorporating intra-band and inter-band self-attentions, explicitly aggregates forgery features within and across diverse sub-bands.
arXiv Detail & Related papers (2023-09-18T11:06:42Z) - Scale-MAE: A Scale-Aware Masked Autoencoder for Multiscale Geospatial
Representation Learning [55.762840052788945]
We present Scale-MAE, a pretraining method that explicitly learns relationships between data at different, known scales.
We find that tasking the network with reconstructing both low/high frequency images leads to robust multiscale representations for remote sensing imagery.
arXiv Detail & Related papers (2022-12-30T03:15:34Z) - Cross-Modal Fusion Distillation for Fine-Grained Sketch-Based Image
Retrieval [55.21569389894215]
We propose a cross-attention framework for Vision Transformers (XModalViT) that fuses modality-specific information instead of discarding them.
Our framework first maps paired datapoints from the individual photo and sketch modalities to fused representations that unify information from both modalities.
We then decouple the input space of the aforementioned modality fusion network into independent encoders of the individual modalities via contrastive and relational cross-modal knowledge distillation.
arXiv Detail & Related papers (2022-10-19T11:50:14Z) - Deep Uncalibrated Photometric Stereo via Inter-Intra Image Feature
Fusion [17.686973510425172]
This paper presents a new method for deep uncalibrated photometric stereo.
It efficiently utilizes the inter-image representation to guide the normal estimation.
Our method produces significantly better results than the state-of-the-art methods on both synthetic and real data.
arXiv Detail & Related papers (2022-08-06T03:59:54Z) - Single Plane-Wave Imaging using Physics-Based Deep Learning [2.1410799064827226]
In plane-wave imaging, multiple unfocused ultrasound waves are transmitted into a medium of interest from different angles.
Deep learning methods have been proposed to improve ultrasound imaging.
We propose a data-to-image architecture that incorporates a wave-physics-based image formation algorithm in-between deep convolutional neural networks.
arXiv Detail & Related papers (2021-09-08T14:06:29Z) - M2TR: Multi-modal Multi-scale Transformers for Deepfake Detection [74.19291916812921]
forged images generated by Deepfake techniques pose a serious threat to the trustworthiness of digital information.
In this paper, we aim to capture the subtle manipulation artifacts at different scales for Deepfake detection.
We introduce a high-quality Deepfake dataset, SR-DF, which consists of 4,000 DeepFake videos generated by state-of-the-art face swapping and facial reenactment methods.
arXiv Detail & Related papers (2021-04-20T05:43:44Z) - AS-Net: Fast Photoacoustic Reconstruction with Multi-feature Fusion from
Sparse Data [1.7237160821929758]
Photoacoustic imaging is capable of acquiring high contrast images of optical absorption at depths much greater than traditional optical imaging techniques.
In this paper, we employ a novel signal processing method to make sparse PA raw data more suitable for the neural network.
We then propose Attention Steered Network (AS-Net) for PA reconstruction with multi-feature fusion.
arXiv Detail & Related papers (2021-01-22T03:49:30Z) - A deep learning pipeline for identification of motor units in
musculoskeletal ultrasound [0.5249805590164902]
It has been shown that ultrafast ultrasound can be used to record and analyze the mechanical response of individual MUs.
We present an alternative method - a deep learning pipeline - to identify active MUs in ultrasound image sequences.
We train evaluate the model using simulated data mimicking the complex activation pattern and overlapping territories.
arXiv Detail & Related papers (2020-09-23T20:44:29Z) - Fast ultrasonic imaging using end-to-end deep learning [1.0266286487433585]
Deep neural networks (DNNs) are being used for the data pre-processing and the image post-processing steps separately.
We propose a novel deep learning architecture that integrates all three steps to enable end-to-end training.
arXiv Detail & Related papers (2020-09-04T13:53:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.