Distributed optimization for nonrigid nano-tomography
- URL: http://arxiv.org/abs/2008.03375v2
- Date: Sun, 28 Feb 2021 17:14:31 GMT
- Title: Distributed optimization for nonrigid nano-tomography
- Authors: Viktor Nikitin, Vincent De Andrade, Azat Slyamov, Benjamin J. Gould,
Yuepeng Zhang, Vandana Sampathkumar, Narayanan Kasthuri, Doga Gursoy,
Francesco De Carlo
- Abstract summary: We propose a joint solver for imaging samples at the nanoscale with projection alignment, unwarping and regularization.
Projection data consistency is regulated by dense optical flow estimated by Farneback's algorithm, leading to sharp sample reconstructions with less artifacts.
- Score: 0.40631409309544836
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Resolution level and reconstruction quality in nano-computed tomography
(nano-CT) are in part limited by the stability of microscopes, because the
magnitude of mechanical vibrations during scanning becomes comparable to the
imaging resolution, and the ability of the samples to resist beam damage during
data acquisition. In such cases, there is no incentive in recovering the sample
state at different time steps like in time-resolved reconstruction methods, but
instead the goal is to retrieve a single reconstruction at the highest possible
spatial resolution and without any imaging artifacts. Here we propose a joint
solver for imaging samples at the nanoscale with projection alignment,
unwarping and regularization. Projection data consistency is regulated by dense
optical flow estimated by Farneback's algorithm, leading to sharp sample
reconstructions with less artifacts. Synthetic data tests show robustness of
the method to Poisson and low-frequency background noise. Applicability of the
method is demonstrated on two large-scale nano-imaging experimental data sets.
Related papers
- A Unified Model for Compressed Sensing MRI Across Undersampling Patterns [69.19631302047569]
Deep neural networks have shown great potential for reconstructing high-fidelity images from undersampled measurements.
Our model is based on neural operators, a discretization-agnostic architecture.
Our inference speed is also 1,400x faster than diffusion methods.
arXiv Detail & Related papers (2024-10-05T20:03:57Z) - Denoising diffusion models for high-resolution microscopy image restoration [34.82692226532414]
We train a denoising diffusion probabilistic model (DDPM) to predict high-resolution images by conditioning the model on low-resolution information.
We show that our model achieves a performance that is better or similar to the previously best-performing methods, across four highly diverse datasets.
arXiv Detail & Related papers (2024-09-18T15:53:45Z) - Rotational Augmented Noise2Inverse for Low-dose Computed Tomography
Reconstruction [83.73429628413773]
Supervised deep learning methods have shown the ability to remove noise in images but require accurate ground truth.
We propose a novel self-supervised framework for LDCT, in which ground truth is not required for training the convolutional neural network (CNN)
Numerical and experimental results show that the reconstruction accuracy of N2I with sparse views is degrading while the proposed rotational augmented Noise2Inverse (RAN2I) method keeps better image quality over a different range of sampling angles.
arXiv Detail & Related papers (2023-12-19T22:40:51Z) - Fluctuation-based deconvolution in fluorescence microscopy using
plug-and-play denoisers [2.236663830879273]
spatial resolution of images of living samples obtained by fluorescence microscopes is physically limited due to the diffraction of visible light.
Several deconvolution and super-resolution techniques have been proposed to overcome this limitation.
arXiv Detail & Related papers (2023-03-20T15:43:52Z) - Pixelated Reconstruction of Foreground Density and Background Surface
Brightness in Gravitational Lensing Systems using Recurrent Inference
Machines [116.33694183176617]
We use a neural network based on the Recurrent Inference Machine to reconstruct an undistorted image of the background source and the lens mass density distribution as pixelated maps.
When compared to more traditional parametric models, the proposed method is significantly more expressive and can reconstruct complex mass distributions.
arXiv Detail & Related papers (2023-01-10T19:00:12Z) - OADAT: Experimental and Synthetic Clinical Optoacoustic Data for
Standardized Image Processing [62.993663757843464]
Optoacoustic (OA) imaging is based on excitation of biological tissues with nanosecond-duration laser pulses followed by detection of ultrasound waves generated via light-absorption-mediated thermoelastic expansion.
OA imaging features a powerful combination between rich optical contrast and high resolution in deep tissues.
No standardized datasets generated with different types of experimental set-up and associated processing methods are available to facilitate advances in broader applications of OA in clinical settings.
arXiv Detail & Related papers (2022-06-17T08:11:26Z) - Compressive Ptychography using Deep Image and Generative Priors [9.658250977094562]
Ptychography is a well-established coherent diffraction imaging technique that enables non-invasive imaging of samples at a nanometer scale.
One major limitation of ptychography is the long data acquisition time due to mechanical scanning of the sample.
We propose a generative model combining deep image priors with deep generative priors.
arXiv Detail & Related papers (2022-05-05T02:18:26Z) - Regularization by Denoising Sub-sampled Newton Method for Spectral CT
Multi-Material Decomposition [78.37855832568569]
We propose to solve a model-based maximum-a-posterior problem to reconstruct multi-materials images with application to spectral CT.
In particular, we propose to solve a regularized optimization problem based on a plug-in image-denoising function.
We show numerical and experimental results for spectral CT materials decomposition.
arXiv Detail & Related papers (2021-03-25T15:20:10Z) - Deep Unfolded Recovery of Sub-Nyquist Sampled Ultrasound Image [94.42139459221784]
We propose a reconstruction method from sub-Nyquist samples in the time and spatial domain, that is based on unfolding the ISTA algorithm.
Our method allows reducing the number of array elements, sampling rate, and computational time while ensuring high quality imaging performance.
arXiv Detail & Related papers (2021-03-01T19:19:38Z) - Sampling possible reconstructions of undersampled acquisitions in MR
imaging [9.75702493778194]
Undersampling the k-space during MR saves time, however results in an ill-posed inversion problem, leading to an infinite set of images as possible solutions.
Traditionally, this is tackled as a reconstruction problem by searching for a single "best" image out of this solution set according to some chosen regularization or prior.
We propose a method that instead returns multiple images which are possible under the acquisition model and the chosen prior to capture the uncertainty in the inversion process.
arXiv Detail & Related papers (2020-09-30T18:20:06Z) - Real-time sparse-sampled Ptychographic imaging through deep neural
networks [3.3351024234383946]
A ptychography reconstruction is achieved by means of solving a complex inverse problem that imposes constraints both on the acquisition and on the analysis of the data.
We propose PtychoNN, a novel approach to solve the ptychography reconstruction problem based on deep convolutional neural networks.
arXiv Detail & Related papers (2020-04-15T23:43:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.