A Physics-Inspired Deep Learning Framework with Polar Coordinate Attention for Ptychographic Imaging
- URL: http://arxiv.org/abs/2412.06806v2
- Date: Fri, 02 May 2025 05:24:06 GMT
- Title: A Physics-Inspired Deep Learning Framework with Polar Coordinate Attention for Ptychographic Imaging
- Authors: Han Yue, Jun Cheng, Yu-Xuan Ren, Chien-Chun Chen, Grant A. van Riessen, Philip Heng Wai Leong, Steve Feng Shu,
- Abstract summary: Ptychographic imaging confronts inherent challenges in applying deep learning for phase retrieval from diffraction patterns.<n>We present PPN, a physics-inspired deep learning network with Polar Coordinate Attention (PoCA) for ptychographic imaging.
- Score: 8.436077464774755
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Ptychographic imaging confronts inherent challenges in applying deep learning for phase retrieval from diffraction patterns. Conventional neural architectures, both convolutional neural networks and Transformer-based methods, are optimized for natural images with Euclidean spatial neighborhood-based inductive biases that exhibit geometric mismatch with the concentric coherent patterns characteristic of diffraction data in reciprocal space. In this paper, we present PPN, a physics-inspired deep learning network with Polar Coordinate Attention (PoCA) for ptychographic imaging, that aligns neural inductive biases with diffraction physics through a dual-branch architecture separating local feature extraction from non-local coherence modeling. It consists of a PoCA mechanism that replaces Euclidean spatial priors with physically consistent radial-angular correlations. PPN outperforms existing end-to-end models, with spectral and spatial analysis confirming its greater preservation of high-frequency details. Notably, PPN maintains robust performance compared to iterative methods even at low overlap ratios, making it well suited for high-throughput imaging in real-world acquisition scenarios for samples with consistent structural characteristics.
Related papers
- Muographic Image Upsampling with Machine Learning for Built Infrastructure Applications [2.983520467199724]
Muography, a non-invasive imaging technique, constructs three-dimensional density maps by detecting interactions of cosmic-ray muons.
Cosmic-ray muons provide deep penetration and inherent safety due to their high momenta and natural source.
However, the technology's reliance on this source results in constrained muon flux, leading to prolonged acquisition times.
We developed a two-model deep learning approach to address these limitations.
arXiv Detail & Related papers (2025-02-04T14:37:37Z) - Understanding and Improving Training-Free AI-Generated Image Detections with Vision Foundation Models [68.90917438865078]
Deepfake techniques for facial synthesis and editing pose serious risks for generative models.<n>In this paper, we investigate how detection performance varies across model backbones, types, and datasets.<n>We introduce Contrastive Blur, which enhances performance on facial images, and MINDER, which addresses noise type bias, balancing performance across domains.
arXiv Detail & Related papers (2024-11-28T13:04:45Z) - How to Bridge Spatial and Temporal Heterogeneity in Link Prediction? A Contrastive Method [11.719027225797037]
We propose a novel textbfContrastive Learning-based textbfLink textbfPrediction model, textbfCLP.<n>Our mymodel consistently outperforms the state-of-the-art models, demonstrating an average improvement of 10.10%, 13.44% in terms of AUC and AP.
arXiv Detail & Related papers (2024-11-01T14:20:53Z) - ND-SDF: Learning Normal Deflection Fields for High-Fidelity Indoor Reconstruction [50.07671826433922]
It is non-trivial to simultaneously recover meticulous geometry and preserve smoothness across regions with differing characteristics.<n>We propose ND-SDF, which learns a Normal Deflection field to represent the angular deviation between the scene normal and the prior normal.<n>Our method not only obtains smooth weakly textured regions such as walls and floors but also preserves the geometric details of complex structures.
arXiv Detail & Related papers (2024-08-22T17:59:01Z) - Invertible Neural Warp for NeRF [29.00183106905031]
This paper tackles the simultaneous optimization of pose and Neural Radiance Fields (NeRF)
We propose a novel over parameterized representation that models camera poses as learnable rigid warp functions.
We present results on synthetic and real-world datasets, and demonstrate that our approach outperforms existing baselines in terms of pose estimation and high-fidelity reconstruction.
arXiv Detail & Related papers (2024-07-17T07:14:08Z) - PASTA: Towards Flexible and Efficient HDR Imaging Via Progressively Aggregated Spatio-Temporal Alignment [91.38256332633544]
PASTA is a Progressively Aggregated Spatio-Temporal Alignment framework for HDR deghosting.
Our approach achieves effectiveness and efficiency by harnessing hierarchical representation during feature distanglement.
Experimental results showcase PASTA's superiority over current SOTA methods in both visual quality and performance metrics.
arXiv Detail & Related papers (2024-03-15T15:05:29Z) - Hallmarks of Optimization Trajectories in Neural Networks: Directional Exploration and Redundancy [75.15685966213832]
We analyze the rich directional structure of optimization trajectories represented by their pointwise parameters.
We show that training only scalar batchnorm parameters some while into training matches the performance of training the entire network.
arXiv Detail & Related papers (2024-03-12T07:32:47Z) - DGNet: Dynamic Gradient-Guided Network for Water-Related Optics Image
Enhancement [77.0360085530701]
Underwater image enhancement (UIE) is a challenging task due to the complex degradation caused by underwater environments.
Previous methods often idealize the degradation process, and neglect the impact of medium noise and object motion on the distribution of image features.
Our approach utilizes predicted images to dynamically update pseudo-labels, adding a dynamic gradient to optimize the network's gradient space.
arXiv Detail & Related papers (2023-12-12T06:07:21Z) - Leveraging Neural Radiance Fields for Uncertainty-Aware Visual
Localization [56.95046107046027]
We propose to leverage Neural Radiance Fields (NeRF) to generate training samples for scene coordinate regression.
Despite NeRF's efficiency in rendering, many of the rendered data are polluted by artifacts or only contain minimal information gain.
arXiv Detail & Related papers (2023-10-10T20:11:13Z) - Exploring Geometric Deep Learning For Precipitation Nowcasting [28.44612565923532]
We propose a geometric deep learning-based temporal Graph Convolutional Network (GCN) for precipitation nowcasting.
The adjacency matrix that simulates the interactions among grid cells is learned automatically by minimizing the L1 loss between prediction and ground truth pixel value.
We test the model on sequences of radar reflectivity maps over the Trento/Italy area.
arXiv Detail & Related papers (2023-09-11T21:14:55Z) - Physics-Driven Turbulence Image Restoration with Stochastic Refinement [80.79900297089176]
Image distortion by atmospheric turbulence is a critical problem in long-range optical imaging systems.
Fast and physics-grounded simulation tools have been introduced to help the deep-learning models adapt to real-world turbulence conditions.
This paper proposes the Physics-integrated Restoration Network (PiRN) to help the network to disentangle theity from the degradation and the underlying image.
arXiv Detail & Related papers (2023-07-20T05:49:21Z) - Combining multi-spectral data with statistical and deep-learning models
for improved exoplanet detection in direct imaging at high contrast [39.90150176899222]
Exoplanet signals can only be identified when combining several observations with dedicated detection algorithms.
We learn a model of the spatial, temporal and spectral characteristics of the nuisance, directly from the observations.
A convolutional neural network (CNN) is then trained in a supervised fashion to detect the residual signature of synthetic sources.
arXiv Detail & Related papers (2023-06-21T13:42:07Z) - Parents and Children: Distinguishing Multimodal DeepFakes from Natural Images [60.34381768479834]
Recent advancements in diffusion models have enabled the generation of realistic deepfakes from textual prompts in natural language.
We pioneer a systematic study on deepfake detection generated by state-of-the-art diffusion models.
arXiv Detail & Related papers (2023-04-02T10:25:09Z) - Unsupervised Domain Transfer with Conditional Invertible Neural Networks [83.90291882730925]
We propose a domain transfer approach based on conditional invertible neural networks (cINNs)
Our method inherently guarantees cycle consistency through its invertible architecture, and network training can efficiently be conducted with maximum likelihood.
Our method enables the generation of realistic spectral data and outperforms the state of the art on two downstream classification tasks.
arXiv Detail & Related papers (2023-03-17T18:00:27Z) - Pixelated Reconstruction of Foreground Density and Background Surface
Brightness in Gravitational Lensing Systems using Recurrent Inference
Machines [116.33694183176617]
We use a neural network based on the Recurrent Inference Machine to reconstruct an undistorted image of the background source and the lens mass density distribution as pixelated maps.
When compared to more traditional parametric models, the proposed method is significantly more expressive and can reconstruct complex mass distributions.
arXiv Detail & Related papers (2023-01-10T19:00:12Z) - DeepDC: Deep Distance Correlation as a Perceptual Image Quality
Evaluator [53.57431705309919]
ImageNet pre-trained deep neural networks (DNNs) show notable transferability for building effective image quality assessment (IQA) models.
We develop a novel full-reference IQA (FR-IQA) model based exclusively on pre-trained DNN features.
We conduct comprehensive experiments to demonstrate the superiority of the proposed quality model on five standard IQA datasets.
arXiv Detail & Related papers (2022-11-09T14:57:27Z) - Lossy compression of multidimensional medical images using sinusoidal
activation networks: an evaluation study [0.0]
We evaluate how neural networks with periodic activation functions can be leveraged to reliably compress large multidimensional medical image datasets.
We show how any given 4D dMRI dataset can be accurately represented through the parameters of a sinusoidal activation network.
Our results show that the proposed approach outperforms benchmark ReLU and Tanh activation perceptron architectures in terms of mean squared error, peak signal-to-noise ratio and structural similarity index.
arXiv Detail & Related papers (2022-08-02T17:16:33Z) - Deep Domain Adversarial Adaptation for Photon-efficient Imaging Based on
Spatiotemporal Inception Network [11.58898808789911]
In single-photon LiDAR, photon-efficient imaging captures the 3D structure of a scene by only several signal detected per pixel.
Existing deep learning models for this task are trained on simulated datasets, which poses the domain shift challenge when applied to realistic scenarios.
We propose a network (STIN) for photon-efficient imaging, which is able to precisely predict the depth from a sparse and high-noise photon counting histogram by fully exploiting spatial and temporal information.
arXiv Detail & Related papers (2022-01-07T14:51:48Z) - Uncovering the Over-smoothing Challenge in Image Super-Resolution: Entropy-based Quantification and Contrastive Optimization [67.99082021804145]
We propose an explicit solution to the COO problem, called Detail Enhanced Contrastive Loss (DECLoss)
DECLoss utilizes the clustering property of contrastive learning to directly reduce the variance of the potential high-resolution distribution.
We evaluate DECLoss on multiple super-resolution benchmarks and demonstrate that it improves the perceptual quality of PSNR-oriented models.
arXiv Detail & Related papers (2022-01-04T08:30:09Z) - PAS-MEF: Multi-exposure image fusion based on principal component
analysis, adaptive well-exposedness and saliency map [0.0]
With regular low dynamic range (LDR) capture/display devices, significant details may not be preserved in images due to the huge dynamic range of natural scenes.
This study proposes an efficient multi-exposure fusion (MEF) approach with a simple yet effective weight extraction method.
Experimental comparisons with existing techniques demonstrate that the proposed method produces very strong statistical and visual results.
arXiv Detail & Related papers (2021-05-25T10:22:43Z) - Fusing the Old with the New: Learning Relative Camera Pose with
Geometry-Guided Uncertainty [91.0564497403256]
We present a novel framework that involves probabilistic fusion between the two families of predictions during network training.
Our network features a self-attention graph neural network, which drives the learning by enforcing strong interactions between different correspondences.
We propose motion parmeterizations suitable for learning and show that our method achieves state-of-the-art performance on the challenging DeMoN and ScanNet datasets.
arXiv Detail & Related papers (2021-04-16T17:59:06Z) - Noise Reduction in X-ray Photon Correlation Spectroscopy with
Convolutional Neural Networks Encoder-Decoder Models [0.0]
We propose a computational approach for improving the signal-to-noise ratio in two-time correlation functions.
CNN-ED models are based on Convolutional Neural Network-Decoder (CNN-ED) models.
We demonstrate that the CNN-ED models trained on real-world experimental data help to effectively extract equilibrium dynamics parameters from two-time correlation functions.
arXiv Detail & Related papers (2021-02-07T18:38:59Z) - CNN-Based Image Reconstruction Method for Ultrafast Ultrasound Imaging [9.659642285903418]
Ultrafast ultrasound (US) revolutionized biomedical imaging with its capability of acquiring full-view frames at over 1 kHz.
It suffers from strong diffraction artifacts, mainly caused by grating lobes, side lobes, or edge waves.
We propose a two-step convolutional neural network (CNN)-based image reconstruction method, compatible with real-time imaging.
arXiv Detail & Related papers (2020-08-28T17:15:37Z) - Phase space learning with neural networks [0.0]
This work proposes an autoencoder neural network as a non-linear generalization of projection-based methods for solving Partial Differential Equations (PDEs)
The proposed deep learning architecture is capable of generating the dynamics of PDEs by integrating them completely in a very reduced latent space without intermediate reconstructions, to then decode the latent solution back to the original space.
It is shown the reliability of properly regularized neural networks to learn the global characteristics of a dynamical system's phase space from the sample data of a single path, as well as its ability to predict unseen bifurcations.
arXiv Detail & Related papers (2020-06-22T20:28:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.