Active Target Discovery under Uninformative Prior: The Power of Permanent and Transient Memory
- URL: http://arxiv.org/abs/2510.16676v1
- Date: Sun, 19 Oct 2025 00:42:56 GMT
- Title: Active Target Discovery under Uninformative Prior: The Power of Permanent and Transient Memory
- Authors: Anindya Sarkar, Binglin Ji, Yevgeniy Vorobeychik,
- Abstract summary: In many scientific and engineering fields, where acquiring high-quality data is expensive, strategic sampling of unobserved regions is crucial for maximizing discovery rates within a constrained budget.<n>We propose a novel approach that enables effective active target discovery even in settings with uninformative priors.<n>Unlike black-box policies, our approach is inherently interpretable, providing clear insights into decision-making.
- Score: 26.488250231429774
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In many scientific and engineering fields, where acquiring high-quality data is expensive--such as medical imaging, environmental monitoring, and remote sensing--strategic sampling of unobserved regions based on prior observations is crucial for maximizing discovery rates within a constrained budget. The rise of powerful generative models, such as diffusion models, has enabled active target discovery in partially observable environments by leveraging learned priors--probabilistic representations that capture underlying structure from data. With guidance from sequentially gathered task-specific observations, these models can progressively refine exploration and efficiently direct queries toward promising regions. However, in domains where learning a strong prior is infeasible due to extremely limited data or high sampling cost (such as rare species discovery, diagnostics for emerging diseases, etc.), these methods struggle to generalize. To overcome this limitation, we propose a novel approach that enables effective active target discovery even in settings with uninformative priors, ensuring robust exploration and adaptability in complex real-world scenarios. Our framework is theoretically principled and draws inspiration from neuroscience to guide its design. Unlike black-box policies, our approach is inherently interpretable, providing clear insights into decision-making. Furthermore, it guarantees a strong, monotonic improvement in prior estimates with each new observation, leading to increasingly accurate sampling and reinforcing both reliability and adaptability in dynamic settings. Through comprehensive experiments and ablation studies across various domains, including species distribution modeling and remote sensing, we demonstrate that our method substantially outperforms baseline approaches.
Related papers
- BATIS: Bayesian Approaches for Targeted Improvement of Species Distribution Models [15.029163153558533]
Species distribution models (SDMs) aim to predict species occurrence based on environmental variables.<n>Recent deep learning advances for SDMs have been shown to perform well on complex and heterogeneous datasets.<n>We introduce BATIS, a novel and practical framework wherein prior predictions are updated iteratively using limited observational data.
arXiv Detail & Related papers (2025-10-22T16:42:46Z) - On Efficient Bayesian Exploration in Model-Based Reinforcement Learning [0.24578723416255752]
We address the challenge of data-efficient exploration in reinforcement learning by examining existing principled, information-theoretic approaches to intrinsic motivation.<n>We prove that exploration bonuses naturally signal epistemic information gains and converge to zero once the agent becomes sufficiently certain about the environment's dynamics and rewards.<n>We then outline a general framework - Predictive Trajectory Sampling with Bayesian Exploration (PTS-BE) - which integrates model-based planning with information-theoretic bonuses to achieve sample-efficient deep exploration.
arXiv Detail & Related papers (2025-07-03T14:03:47Z) - Provable Maximum Entropy Manifold Exploration via Diffusion Models [58.89696361871563]
Exploration is critical for solving real-world decision-making problems such as scientific discovery.<n>We introduce a novel framework that casts exploration as entropy over approximate data manifold implicitly defined by a pre-trained diffusion model.<n>We develop an algorithm based on mirror descent that solves the exploration problem as sequential fine-tuning of a pre-trained diffusion model.
arXiv Detail & Related papers (2025-06-18T11:59:15Z) - Anomaly Detection and Generation with Diffusion Models: A Survey [51.61574868316922]
Anomaly detection (AD) plays a pivotal role across diverse domains, including cybersecurity, finance, healthcare, and industrial manufacturing.<n>Recent advancements in deep learning, specifically diffusion models (DMs), have sparked significant interest.<n>This survey aims to guide researchers and practitioners in leveraging DMs for innovative AD solutions across diverse applications.
arXiv Detail & Related papers (2025-06-11T03:29:18Z) - Consistent World Models via Foresight Diffusion [56.45012929930605]
We argue that a key bottleneck in learning consistent diffusion-based world models lies in the suboptimal predictive ability.<n>We propose Foresight Diffusion (ForeDiff), a diffusion-based world modeling framework that enhances consistency by decoupling condition understanding from target denoising.
arXiv Detail & Related papers (2025-05-22T10:01:59Z) - Online Feedback Efficient Active Target Discovery in Partially Observable Environments [26.488250231429774]
Diffusion-guided Active Target Discovery (DiffATD) is a novel method that leverages diffusion dynamics for active target discovery.<n>DiffATD enables efficient target discovery in a partially observable environment within a fixed sampling budget.<n>We show that DiffATD performs significantly better than baselines and competitively with supervised methods that operate under full environmental observability.
arXiv Detail & Related papers (2025-05-10T06:50:01Z) - Offline Model-Based Optimization: Comprehensive Review [61.91350077539443]
offline optimization is a fundamental challenge in science and engineering, where the goal is to optimize black-box functions using only offline datasets.<n>Recent advances in model-based optimization have harnessed the generalization capabilities of deep neural networks to develop offline-specific surrogate and generative models.<n>Despite its growing impact in accelerating scientific discovery, the field lacks a comprehensive review.
arXiv Detail & Related papers (2025-03-21T16:35:02Z) - Can Large Language Models Help Experimental Design for Causal Discovery? [94.66802142727883]
Large Language Model Guided Intervention Targeting (LeGIT) is a robust framework that effectively incorporates LLMs to augment existing numerical approaches for the intervention targeting in causal discovery.<n>LeGIT demonstrates significant improvements and robustness over existing methods and even surpasses humans.
arXiv Detail & Related papers (2025-03-03T03:43:05Z) - Enhancement of Approximation Spaces by the Use of Primals and Neighborhood [0.0]
We introduce four new generalized rough set models that draw inspiration from "neighborhoods and primals"
We claim that the current models can preserve nearly all significant aspects associated with the rough set model.
We also demonstrate that the new strategy we define for our everyday health-related problem yields more accurate findings.
arXiv Detail & Related papers (2024-10-23T18:49:13Z) - Seeing Unseen: Discover Novel Biomedical Concepts via
Geometry-Constrained Probabilistic Modeling [53.7117640028211]
We present a geometry-constrained probabilistic modeling treatment to resolve the identified issues.
We incorporate a suite of critical geometric properties to impose proper constraints on the layout of constructed embedding space.
A spectral graph-theoretic method is devised to estimate the number of potential novel classes.
arXiv Detail & Related papers (2024-03-02T00:56:05Z) - Drug Discovery under Covariate Shift with Domain-Informed Prior
Distributions over Functions [30.305418761024143]
Real-world drug discovery tasks are often characterized by a scarcity of labeled data and a significant range of data.
We present a principled way to encode explicit prior knowledge of the data-generating process into a prior distribution.
We demonstrate that using integrate Q-SAVI to contextualize prior knowledgelike chemical space into the modeling process affords substantial accuracy and calibration.
arXiv Detail & Related papers (2023-07-14T05:01:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.