Blending Knowledge in Deep Recurrent Networks for Adverse Event
Prediction at Hospital Discharge
- URL: http://arxiv.org/abs/2104.04377v1
- Date: Fri, 9 Apr 2021 14:07:45 GMT
- Title: Blending Knowledge in Deep Recurrent Networks for Adverse Event
Prediction at Hospital Discharge
- Authors: Prithwish Chakraborty, James Codella, Piyush Madan, Ying Li, Hu Huang,
Yoonyoung Park, Chao Yan, Ziqi Zhang, Cheng Gao, Steve Nyemba, Xu Min, Sanjib
Basak, Mohamed Ghalwash, Zach Shahn, Parthasararathy Suryanarayanan, Italo
Buleje, Shannon Harrer, Sarah Miller, Amol Rajmane, Colin Walsh, Jonathan
Wanderer, Gigi Yuen Reed, Kenney Ng, Daby Sow, Bradley A. Malin
- Abstract summary: We introduce a learning architecture that fuses a representation of patient data computed by a self-attention based recurrent neural network, with clinically relevant features.
We conduct extensive experiments on a large claims dataset and show that the blended method outperforms the standard machine learning approaches.
- Score: 15.174501264797309
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning architectures have an extremely high-capacity for modeling
complex data in a wide variety of domains. However, these architectures have
been limited in their ability to support complex prediction problems using
insurance claims data, such as readmission at 30 days, mainly due to data
sparsity issue. Consequently, classical machine learning methods, especially
those that embed domain knowledge in handcrafted features, are often on par
with, and sometimes outperform, deep learning approaches. In this paper, we
illustrate how the potential of deep learning can be achieved by blending
domain knowledge within deep learning architectures to predict adverse events
at hospital discharge, including readmissions. More specifically, we introduce
a learning architecture that fuses a representation of patient data computed by
a self-attention based recurrent neural network, with clinically relevant
features. We conduct extensive experiments on a large claims dataset and show
that the blended method outperforms the standard machine learning approaches.
Related papers
- Breaking the Curse of Dimensionality in Deep Neural Networks by Learning
Invariant Representations [1.9580473532948401]
This thesis explores the theoretical foundations of deep learning by studying the relationship between the architecture of these models and the inherent structures found within the data they process.
We ask What drives the efficacy of deep learning algorithms and allows them to beat the so-called curse of dimensionality.
Our methodology takes an empirical approach to deep learning, combining experimental studies with physics-inspired toy models.
arXiv Detail & Related papers (2023-10-24T19:50:41Z) - Homological Neural Networks: A Sparse Architecture for Multivariate
Complexity [0.0]
We develop a novel deep neural network unit characterized by a sparse higher-order graphical architecture built over the homological structure of underlying data.
Results demonstrate the advantages of this novel design which can tie or overcome the results of state-of-the-art machine learning and deep learning models using only a fraction of parameters.
arXiv Detail & Related papers (2023-06-27T09:46:16Z) - Deep networks for system identification: a Survey [56.34005280792013]
System identification learns mathematical descriptions of dynamic systems from input-output data.
Main aim of the identified model is to predict new data from previous observations.
We discuss architectures commonly adopted in the literature, like feedforward, convolutional, and recurrent networks.
arXiv Detail & Related papers (2023-01-30T12:38:31Z) - Leveraging the structure of dynamical systems for data-driven modeling [111.45324708884813]
We consider the impact of the training set and its structure on the quality of the long-term prediction.
We show how an informed design of the training set, based on invariants of the system and the structure of the underlying attractor, significantly improves the resulting models.
arXiv Detail & Related papers (2021-12-15T20:09:20Z) - Reducing Catastrophic Forgetting in Self Organizing Maps with
Internally-Induced Generative Replay [67.50637511633212]
A lifelong learning agent is able to continually learn from potentially infinite streams of pattern sensory data.
One major historic difficulty in building agents that adapt is that neural systems struggle to retain previously-acquired knowledge when learning from new samples.
This problem is known as catastrophic forgetting (interference) and remains an unsolved problem in the domain of machine learning to this day.
arXiv Detail & Related papers (2021-12-09T07:11:14Z) - Model-Based Deep Learning [155.063817656602]
Signal processing, communications, and control have traditionally relied on classical statistical modeling techniques.
Deep neural networks (DNNs) use generic architectures which learn to operate from data, and demonstrate excellent performance.
We are interested in hybrid techniques that combine principled mathematical models with data-driven systems to benefit from the advantages of both approaches.
arXiv Detail & Related papers (2020-12-15T16:29:49Z) - Domain Generalization for Medical Imaging Classification with
Linear-Dependency Regularization [59.5104563755095]
We introduce a simple but effective approach to improve the generalization capability of deep neural networks in the field of medical imaging classification.
Motivated by the observation that the domain variability of the medical images is to some extent compact, we propose to learn a representative feature space through variational encoding.
arXiv Detail & Related papers (2020-09-27T12:30:30Z) - A Survey on Knowledge integration techniques with Artificial Neural
Networks for seq-2-seq/time series models [0.456877715768796]
Deep neural networks have enabled the exploration of uncharted areas in several domains.
But they under-perform due to insufficient data, poor data quality, data that might not be covering the domain broadly.
This paper explores techniques to integrate expert knowledge to the Deep Neural Networks for sequence-to-sequence and time series models.
arXiv Detail & Related papers (2020-08-13T15:40:38Z) - Uncovering the structure of clinical EEG signals with self-supervised
learning [64.4754948595556]
Supervised learning paradigms are often limited by the amount of labeled data that is available.
This phenomenon is particularly problematic in clinically-relevant data, such as electroencephalography (EEG)
By extracting information from unlabeled data, it might be possible to reach competitive performance with deep neural networks.
arXiv Detail & Related papers (2020-07-31T14:34:47Z) - Learning Cross-domain Generalizable Features by Representation
Disentanglement [11.74643883335152]
Deep learning models exhibit limited generalizability across different domains.
We propose Mutual-Information-based Disentangled Neural Networks (MIDNet) to extract generalizable features that enable transferring knowledge to unseen categorical features in target domains.
We demonstrate our method on handwritten digits datasets and a fetal ultrasound dataset for image classification tasks.
arXiv Detail & Related papers (2020-02-29T17:53:16Z) - Neural Architecture Search For Fault Diagnosis [6.226564415963648]
Deep learning is suitable for processing big data, and has a strong feature extraction ability to realize end-to-end fault diagnosis systems.
Neural architecture search (NAS) is developing rapidly, and is becoming one of the next directions for deep learning.
In this paper, we proposed a NAS method for fault diagnosis using reinforcement learning.
arXiv Detail & Related papers (2020-02-19T04:03:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.