Controllable diffusion-based generation for multi-channel biological data
- URL: http://arxiv.org/abs/2507.02902v1
- Date: Tue, 24 Jun 2025 00:56:21 GMT
- Title: Controllable diffusion-based generation for multi-channel biological data
- Authors: Haoran Zhang, Mingyuan Zhou, Wesley Tansey,
- Abstract summary: This work proposes a unified diffusion framework for controllable generation over structured and spatial biological data.<n>We show state-of-the-art performance across both spatial and non-spatial prediction tasks, including protein imputation in IMC and gene-to-protein prediction in single-cell datasets.
- Score: 66.44042377817074
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Spatial profiling technologies in biology, such as imaging mass cytometry (IMC) and spatial transcriptomics (ST), generate high-dimensional, multi-channel data with strong spatial alignment and complex inter-channel relationships. Generative modeling of such data requires jointly capturing intra- and inter-channel structure, while also generalizing across arbitrary combinations of observed and missing channels for practical application. Existing diffusion-based models generally assume low-dimensional inputs (e.g., RGB images) and rely on simple conditioning mechanisms that break spatial correspondence and ignore inter-channel dependencies. This work proposes a unified diffusion framework for controllable generation over structured and spatial biological data. Our model contains two key innovations: (1) a hierarchical feature injection mechanism that enables multi-resolution conditioning on spatially aligned channels, and (2) a combination of latent-space and output-space channel-wise attention to capture inter-channel relationships. To support flexible conditioning and generalization to arbitrary subsets of observed channels, we train the model using a random masking strategy, enabling it to reconstruct missing channels from any combination of inputs. We demonstrate state-of-the-art performance across both spatial and non-spatial prediction tasks, including protein imputation in IMC and gene-to-protein prediction in single-cell datasets, and show strong generalization to unseen conditional configurations.
Related papers
- Few-Shot Object Detection via Spatial-Channel State Space Model [9.644454618045133]
Current methods may struggle to accurately extract effective features from each channel.<n>We propose a Spatial-Channel State Space Modeling (SCSM) module for spatial-channel state modeling.<n>SCSM module highlights the effective patterns and rectifies those ineffective ones in feature channels.
arXiv Detail & Related papers (2025-07-21T07:08:19Z) - DFYP: A Dynamic Fusion Framework with Spectral Channel Attention and Adaptive Operator learning for Crop Yield Prediction [18.24061967822792]
DFYP is a novel Dynamic Fusion framework for crop Yield Prediction.<n>It combines spectral channel attention, edge-adaptive spatial modeling and a learnable fusion mechanism.<n> DFYP consistently outperforms current state-of-the-art baselines in RMSE, MAE, and R2.
arXiv Detail & Related papers (2025-07-08T10:24:04Z) - Knowledge-guided Complex Diffusion Model for PolSAR Image Classification in Contourlet Domain [58.46450049579116]
We propose a knowledge-guided complex diffusion model for PolSAR image classification in the Contourlet domain.<n> Specifically, the Contourlet transform is first applied to decompose the data into low- and high-frequency subbands.<n>A knowledge-guided complex diffusion network is then designed to model the statistical properties of the low-frequency components.
arXiv Detail & Related papers (2025-07-08T04:50:28Z) - Biaxialformer: Leveraging Channel Independence and Inter-Channel Correlations in EEG Signal Decoding for Predicting Neurological Outcomes [1.7656489005289302]
Accurate decoding of EEG signals requires modeling of both temporal dynamics within individual channels and spatial dependencies across channels.<n> Transformer-based models utilizing channel-independence (CI) strategies have demonstrated strong performance in various time series tasks.<n>We propose Biformer, characterized by a meticulously engineered two-stage attention-based framework.
arXiv Detail & Related papers (2025-06-18T11:47:26Z) - Enhanced High-Dimensional Data Visualization through Adaptive Multi-Scale Manifold Embedding [0.7705234721762716]
We propose an Adaptive Multi-Scale Manifold Embedding (AMSME) algorithm.<n>By introducing ordinal distance, we demonstrate that ordinal distance overcomes the constraints of the curse of dimensionality in high-dimensional spaces.<n> Experimental results demonstrate that AMSME significantly preserves intra-cluster topological structures and improves inter-cluster separation on real-world datasets.
arXiv Detail & Related papers (2025-03-18T06:46:53Z) - CMamba: Channel Correlation Enhanced State Space Models for Multivariate Time Series Forecasting [18.50360049235537]
Mamba, a state space model, has emerged with robust sequence and feature mixing capabilities.
Capturing cross-channel dependencies is critical in enhancing performance of time series prediction.
We introduce a refined Mamba variant tailored for time series forecasting.
arXiv Detail & Related papers (2024-06-08T01:32:44Z) - DA-Flow: Dual Attention Normalizing Flow for Skeleton-based Video Anomaly Detection [52.74152717667157]
We propose a lightweight module called Dual Attention Module (DAM) for capturing cross-dimension interaction relationships in-temporal skeletal data.
It employs the frame attention mechanism to identify the most significant frames and the skeleton attention mechanism to capture broader relationships across fixed partitions with minimal parameters and flops.
arXiv Detail & Related papers (2024-06-05T06:18:03Z) - From Similarity to Superiority: Channel Clustering for Time Series Forecasting [61.96777031937871]
We develop a novel and adaptable Channel Clustering Module ( CCM)
CCM dynamically groups channels characterized by intrinsic similarities and leverages cluster information instead of individual channel identities.
CCM can boost the performance of CI and CD models by an average margin of 2.4% and 7.2% on long-term and short-term forecasting, respectively.
arXiv Detail & Related papers (2024-03-31T02:46:27Z) - DIFFormer: Scalable (Graph) Transformers Induced by Energy Constrained
Diffusion [66.21290235237808]
We introduce an energy constrained diffusion model which encodes a batch of instances from a dataset into evolutionary states.
We provide rigorous theory that implies closed-form optimal estimates for the pairwise diffusion strength among arbitrary instance pairs.
Experiments highlight the wide applicability of our model as a general-purpose encoder backbone with superior performance in various tasks.
arXiv Detail & Related papers (2023-01-23T15:18:54Z) - Channelized Axial Attention for Semantic Segmentation [70.14921019774793]
We propose the Channelized Axial Attention (CAA) to seamlessly integratechannel attention and axial attention with reduced computationalcomplexity.
Our CAA not onlyrequires much less computation resources compared with otherdual attention models such as DANet, but also outperforms the state-of-the-art ResNet-101-based segmentation models on alltested datasets.
arXiv Detail & Related papers (2021-01-19T03:08:03Z) - Effect of The Latent Structure on Clustering with GANs [13.970914037707724]
We focus on the problem of clustering in generated space of GANs and uncover its relationship with the characteristics of the latent space.
We derive from first principles, the necessary and sufficient conditions needed to achieve faithful clustering in the GAN framework.
We also describe a procedure to construct a multimodal latent space which facilitates learning of cluster priors with sparse supervision.
arXiv Detail & Related papers (2020-05-05T18:52:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.