NOCTA: Non-Greedy Objective Cost-Tradeoff Acquisition for Longitudinal Data
- URL: http://arxiv.org/abs/2507.12412v1
- Date: Wed, 16 Jul 2025 17:00:41 GMT
- Title: NOCTA: Non-Greedy Objective Cost-Tradeoff Acquisition for Longitudinal Data
- Authors: Dzung Dinh, Boqi Chen, Marc Niethammer, Junier Oliva,
- Abstract summary: We propose NOCTA, a Non-Greedy Objective Cost-Tradeoff Acquisition method.<n>We first introduce a cohesive estimation target for our NOCTA setting, and then develop two complementary estimators.<n>Experiments on synthetic and real-world medical datasets demonstrate that both NOCTA variants outperform existing baselines.
- Score: 23.75715594365611
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In many critical applications, resource constraints limit the amount of information that can be gathered to make predictions. For example, in healthcare, patient data often spans diverse features ranging from lab tests to imaging studies. Each feature may carry different information and must be acquired at a respective cost of time, money, or risk to the patient. Moreover, temporal prediction tasks, where both instance features and labels evolve over time, introduce additional complexity in deciding when or what information is important. In this work, we propose NOCTA, a Non-Greedy Objective Cost-Tradeoff Acquisition method that sequentially acquires the most informative features at inference time while accounting for both temporal dynamics and acquisition cost. We first introduce a cohesive estimation target for our NOCTA setting, and then develop two complementary estimators: 1) a non-parametric method based on nearest neighbors to guide the acquisition (NOCTA-NP), and 2) a parametric method that directly predicts the utility of potential acquisitions (NOCTA-P). Experiments on synthetic and real-world medical datasets demonstrate that both NOCTA variants outperform existing baselines.
Related papers
- How Much Data are Enough? Investigating Dataset Requirements for Patch-Based Brain MRI Segmentation Tasks [74.21484375019334]
Training deep neural networks reliably requires access to large-scale datasets.
To mitigate both the time and financial costs associated with model development, a clear understanding of the amount of data required to train a satisfactory model is crucial.
This paper proposes a strategic framework for estimating the amount of annotated data required to train patch-based segmentation networks.
arXiv Detail & Related papers (2024-04-04T13:55:06Z) - Clairvoyance: A Pipeline Toolkit for Medical Time Series [95.22483029602921]
Time-series learning is the bread and butter of data-driven *clinical decision support*
Clairvoyance proposes a unified, end-to-end, autoML-friendly pipeline that serves as a software toolkit.
Clairvoyance is the first to demonstrate viability of a comprehensive and automatable pipeline for clinical time-series ML.
arXiv Detail & Related papers (2023-10-28T12:08:03Z) - TrackFlow: Multi-Object Tracking with Normalizing Flows [36.86830078167583]
We aim at extending tracking-by-detection to multi-modal settings.
A rough estimate of 3D information is also available and must be merged with other traditional metrics.
Our approach consistently enhances the performance of several tracking-by-detection algorithms.
arXiv Detail & Related papers (2023-08-22T15:40:03Z) - Acquisition Conditioned Oracle for Nongreedy Active Feature Acquisition [16.350351668269415]
We develop methodology for active feature acquisition (AFA)
We show that we can bypass many challenges with a novel, nonparametric oracle based approach.
arXiv Detail & Related papers (2023-02-27T17:02:11Z) - Active Acquisition for Multimodal Temporal Data: A Challenging
Decision-Making Task [13.291343999247898]
We introduce a challenging decision-making task that we call active acquisition for multimodal temporal data (A2MT)
We aim to learn agents that actively select which modalities of an input to acquire, trading off acquisition cost and predictive performance.
Applications of A2MT may be impactful in domains like medicine, robotics, or finance, where modalities differ in acquisition cost and informativeness.
arXiv Detail & Related papers (2022-11-09T17:16:36Z) - Grouped self-attention mechanism for a memory-efficient Transformer [64.0125322353281]
Real-world tasks such as forecasting weather, electricity consumption, and stock market involve predicting data that vary over time.
Time-series data are generally recorded over a long period of observation with long sequences owing to their periodic characteristics and long-range dependencies over time.
We propose two novel modules, Grouped Self-Attention (GSA) and Compressed Cross-Attention (CCA)
Our proposed model efficiently exhibited reduced computational complexity and performance comparable to or better than existing methods.
arXiv Detail & Related papers (2022-10-02T06:58:49Z) - Micro-level Reserving for General Insurance Claims using a Long
Short-Term Memory Network [0.0]
We introduce a discrete-time individual reserving framework incorporating granular information in a deep learning approach named Long Short-Term Memory (LSTM) neural network.
At each time period, the network has two tasks: first, classifying whether there is a payment or a recovery, and second, predicting the corresponding non-zero amount, if any.
arXiv Detail & Related papers (2022-01-27T02:49:42Z) - Multitask Learning and Bandits via Robust Statistics [3.103098467546532]
Decision-makers often simultaneously face many related but heterogeneous learning problems.
We propose a novel two-stage multitask learning estimator that exploits this structure in a sample-efficient way.
Our estimator yields improved sample complexity bounds in the feature dimension $d$ relative to commonly-employed estimators.
arXiv Detail & Related papers (2021-12-28T17:37:08Z) - Brain Image Synthesis with Unsupervised Multivariate Canonical
CSC$\ell_4$Net [122.8907826672382]
We propose to learn dedicated features that cross both intre- and intra-modal variations using a novel CSC$ell_4$Net.
arXiv Detail & Related papers (2021-03-22T05:19:40Z) - BiteNet: Bidirectional Temporal Encoder Network to Predict Medical
Outcomes [53.163089893876645]
We propose a novel self-attention mechanism that captures the contextual dependency and temporal relationships within a patient's healthcare journey.
An end-to-end bidirectional temporal encoder network (BiteNet) then learns representations of the patient's journeys.
We have evaluated the effectiveness of our methods on two supervised prediction and two unsupervised clustering tasks with a real-world EHR dataset.
arXiv Detail & Related papers (2020-09-24T00:42:36Z) - Transformer Hawkes Process [79.16290557505211]
We propose a Transformer Hawkes Process (THP) model, which leverages the self-attention mechanism to capture long-term dependencies.
THP outperforms existing models in terms of both likelihood and event prediction accuracy by a notable margin.
We provide a concrete example, where THP achieves improved prediction performance for learning multiple point processes when incorporating their relational information.
arXiv Detail & Related papers (2020-02-21T13:48:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.