Improved clinical data imputation via classical and quantum
determinantal point processes
- URL: http://arxiv.org/abs/2303.17893v2
- Date: Tue, 12 Dec 2023 15:39:58 GMT
- Title: Improved clinical data imputation via classical and quantum
determinantal point processes
- Authors: Skander Kazdaghli, Iordanis Kerenidis, Jens Kieckbusch and Philip
Teare
- Abstract summary: Imputing data is a critical issue for machine learning practitioners.
Here we propose novel imputation methods based on determinantal point processes.
We demonstrate competitive results with up to ten qubits for small-scale imputation tasks.
- Score: 1.3749490831384268
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Imputing data is a critical issue for machine learning practitioners,
including in the life sciences domain, where missing clinical data is a typical
situation and the reliability of the imputation is of great importance.
Currently, there is no canonical approach for imputation of clinical data and
widely used algorithms introduce variance in the downstream classification.
Here we propose novel imputation methods based on determinantal point processes
that enhance popular techniques such as the Multivariate Imputation by Chained
Equations (MICE) and MissForest. Their advantages are two-fold: improving the
quality of the imputed data demonstrated by increased accuracy of the
downstream classification; and providing deterministic and reliable imputations
that remove the variance from the classification results. We experimentally
demonstrate the advantages of our methods by performing extensive imputations
on synthetic and real clinical data. We also perform quantum hardware
experiments by applying the quantum circuits for DPP sampling, since such
quantum algorithms provide a computational advantage with respect to classical
ones. We demonstrate competitive results with up to ten qubits for small-scale
imputation tasks on a state-of-the-art IBM quantum processor. Our classical and
quantum methods improve the effectiveness and robustness of clinical data
prediction modeling by providing better and more reliable data imputations.
These improvements can add significant value in settings demanding high
precision, such as in pharmaceutical drug trials where our approach can provide
higher confidence in the predictions made.
Related papers
- Quantum-Inspired Optimization Process for Data Imputation [1.5186937600119894]
This study introduces a novel quantum-inspired imputation framework evaluated on the UCI Diabetes dataset.<n>The method integrates Principal Component Analysis (PCA) with quantum-assisted rotations, optimized through gradient-free classicals.<n>It reconstructs missing values while preserving statistical fidelity.
arXiv Detail & Related papers (2025-05-07T22:37:07Z) - Precision Adaptive Imputation Network : An Unified Technique for Mixed Datasets [0.0]
This study introduces the Precision Adaptive Imputation Network (PAIN), a novel algorithm designed to enhance data reconstruction.
PAIN employs a tri-step process that integrates statistical methods, random forests, and autoencoders, ensuring balanced accuracy and efficiency in imputation.
The findings highlight PAIN's superior ability to preserve data distributions and maintain analytical integrity, particularly in complex scenarios where missingness is not completely at random.
arXiv Detail & Related papers (2025-01-18T06:22:27Z) - Leveraging Pre-Trained Neural Networks to Enhance Machine Learning with Variational Quantum Circuits [48.33631905972908]
We introduce an innovative approach that utilizes pre-trained neural networks to enhance Variational Quantum Circuits (VQC)
This technique effectively separates approximation error from qubit count and removes the need for restrictive conditions.
Our results extend to applications such as human genome analysis, demonstrating the broad applicability of our approach.
arXiv Detail & Related papers (2024-11-13T12:03:39Z) - Strategic Data Re-Uploads: A Pathway to Improved Quantum Classification Data Re-Uploading Strategies for Improved Quantum Classifier Performance [0.0]
Re-uploading classical information into quantum states multiple times can enhance the accuracy of quantum classifiers.
We demonstrate our approach to two classification patterns: a linear classification pattern (LCP) and a non-linear classification pattern (NLCP)
arXiv Detail & Related papers (2024-05-15T14:28:00Z) - Physical formula enhanced multi-task learning for pharmacokinetics prediction [54.13787789006417]
A major challenge for AI-driven drug discovery is the scarcity of high-quality data.
We develop a formula enhanced mul-ti-task learning (PEMAL) method that predicts four key parameters of pharmacokinetics simultaneously.
Our experiments reveal that PEMAL significantly lowers the data demand, compared to typical Graph Neural Networks.
arXiv Detail & Related papers (2024-04-16T07:42:55Z) - XAI for In-hospital Mortality Prediction via Multimodal ICU Data [57.73357047856416]
We propose an efficient, explainable AI solution for predicting in-hospital mortality via multimodal ICU data.
We employ multimodal learning in our framework, which can receive heterogeneous inputs from clinical data and make decisions.
Our framework can be easily transferred to other clinical tasks, which facilitates the discovery of crucial factors in healthcare research.
arXiv Detail & Related papers (2023-12-29T14:28:04Z) - Informative Priors Improve the Reliability of Multimodal Clinical Data
Classification [7.474271086307501]
We consider neural networks and design a tailor-made multimodal data-driven (M2D2) prior distribution over network parameters.
We use simple and scalable mean-field variational inference to train a Bayesian neural network using the M2D2 prior.
Our empirical results show that the proposed method produces a more reliable predictive model compared to deterministic and Bayesian neural network baselines.
arXiv Detail & Related papers (2023-11-17T03:44:15Z) - Mixed-Integer Projections for Automated Data Correction of EMRs Improve
Predictions of Sepsis among Hospitalized Patients [7.639610349097473]
We introduce an innovative projections-based method that seamlessly integrates clinical expertise as domain constraints.
We measure the distance of corrected data from the constraints defining a healthy range of patient data, resulting in a unique predictive metric we term as "trust-scores"
We show an AUROC of 0.865 and a precision of 0.922, that surpasses conventional ML models without such projections.
arXiv Detail & Related papers (2023-08-21T15:14:49Z) - FineEHR: Refine Clinical Note Representations to Improve Mortality
Prediction [3.9026461169566673]
Large-scale electronic health records provide machine learning models with an abundance of clinical text and vital sign data.
Despite the emergence of advanced Natural Language Processing (NLP) algorithms for clinical note analysis, the complex textual structure and noise present in raw clinical data have posed significant challenges.
We propose FINEEHR, a system that utilizes two representation learning techniques, namely metric learning and fine-tuning, to refine clinical note embeddings.
arXiv Detail & Related papers (2023-04-24T02:42:52Z) - Towards Reliable Medical Image Segmentation by utilizing Evidential Calibrated Uncertainty [52.03490691733464]
We introduce DEviS, an easily implementable foundational model that seamlessly integrates into various medical image segmentation networks.
By leveraging subjective logic theory, we explicitly model probability and uncertainty for the problem of medical image segmentation.
DeviS incorporates an uncertainty-aware filtering module, which utilizes the metric of uncertainty-calibrated error to filter reliable data.
arXiv Detail & Related papers (2023-01-01T05:02:46Z) - Benchmarking Heterogeneous Treatment Effect Models through the Lens of
Interpretability [82.29775890542967]
Estimating personalized effects of treatments is a complex, yet pervasive problem.
Recent developments in the machine learning literature on heterogeneous treatment effect estimation gave rise to many sophisticated, but opaque, tools.
We use post-hoc feature importance methods to identify features that influence the model's predictions.
arXiv Detail & Related papers (2022-06-16T17:59:05Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Efficient training of lightweight neural networks using Online
Self-Acquired Knowledge Distillation [51.66271681532262]
Online Self-Acquired Knowledge Distillation (OSAKD) is proposed, aiming to improve the performance of any deep neural model in an online manner.
We utilize k-nn non-parametric density estimation technique for estimating the unknown probability distributions of the data samples in the output feature space.
arXiv Detail & Related papers (2021-08-26T14:01:04Z) - Extracting Chemical-Protein Interactions via Calibrated Deep Neural
Network and Self-training [0.8376091455761261]
"calibration" techniques have been applied to deep learning models to estimate the data uncertainty and improve the reliability.
In this study, to extract chemical--protein interactions, we propose a DNN-based approach incorporating uncertainty information and calibration techniques.
Our approach has achieved state-of-the-art performance with regard to the Biocreative VI ChemProt task, while preserving higher calibration abilities than those of previous approaches.
arXiv Detail & Related papers (2020-11-04T10:14:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.