HICH Image/Text (HICH-IT): Comprehensive Text and Image Datasets for
Hypertensive Intracerebral Hemorrhage Research
- URL: http://arxiv.org/abs/2401.15934v2
- Date: Mon, 5 Feb 2024 08:10:36 GMT
- Title: HICH Image/Text (HICH-IT): Comprehensive Text and Image Datasets for
Hypertensive Intracerebral Hemorrhage Research
- Authors: Jie Li and Yulong Xia and Tongxin Yang and Fenglin Cai and Miao Wei
and Zhiwei Zhang and Li Jiang
- Abstract summary: We introduce a new dataset in the medical field of hypertensive intracerebral hemorrhage (HICH) called HICH-IT.
This dataset is designed to enhance the accuracy of artificial intelligence in the diagnosis and treatment of HICH.
- Score: 12.479936404475803
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we introduce a new dataset in the medical field of
hypertensive intracerebral hemorrhage (HICH), called HICH-IT, which includes
both electronic medical records (EMRs) and head CT images. This dataset is
designed to enhance the accuracy of artificial intelligence in the diagnosis
and treatment of HICH. This dataset, built upon the foundation of standard text
and image data, incorporates specific annotations within the EMRs, extracting
key content from the text information, and categorizes the annotation content
of imaging data into four types: brain midline, hematoma, left and right
cerebral ventricle. HICH-IT aims to be a foundational dataset for feature
learning in image segmentation tasks and named entity recognition. To further
understand the dataset, we have trained deep learning algorithms to observe the
performance. The pretrained models have been released at both www.daip.club and
github.com/Deep-AI-Application-DAIP. The dataset has been uploaded to
https://github.com/CYBUS123456/HICH-IT-Datasets.
Index Terms-HICH, Deep learning, Intraparenchymal hemorrhage, named entity
recognition, novel dataset
Related papers
- MedicalNarratives: Connecting Medical Vision and Language with Localized Narratives [11.242775987217032]
MedicalNarratives is a dataset curated from medical pedagogical videos similar in nature to data collected in Think-Aloud studies.
Our dataset contains 4.7M image-text pairs from videos and articles, with 1M samples containing dense annotations in the form of traces and bounding boxes.
To evaluate the utility of MedicalNarratives, we train GenMedClip based on the CLIP architecture using our dataset spanning 12 medical domains.
arXiv Detail & Related papers (2025-01-07T23:32:05Z) - Medical Vision-Language Pre-Training for Brain Abnormalities [96.1408455065347]
We show how to automatically collect medical image-text aligned data for pretraining from public resources such as PubMed.
In particular, we present a pipeline that streamlines the pre-training process by initially collecting a large brain image-text dataset.
We also investigate the unique challenge of mapping subfigures to subcaptions in the medical domain.
arXiv Detail & Related papers (2024-04-27T05:03:42Z) - TBI Image/Text (TBI-IT): Comprehensive Text and Image Datasets for Traumatic Brain Injury Research [11.098732957191748]
TBI-IT is a new dataset in the medical field of Traumatic Brain Injury (TBI)
It includes both electronic medical records (EMRs) and head CT images.
This dataset is designed to enhance the accuracy of artificial intelligence in the diagnosis and treatment of TBI.
arXiv Detail & Related papers (2024-03-14T03:07:49Z) - DeepMediX: A Deep Learning-Driven Resource-Efficient Medical Diagnosis
Across the Spectrum [15.382184404673389]
This work presents textttDeepMediX, a groundbreaking, resource-efficient model that significantly addresses this challenge.
Built on top of the MobileNetV2 architecture, DeepMediX excels in classifying brain MRI scans and skin cancer images.
DeepMediX's design also includes the concept of Federated Learning, enabling a collaborative learning approach without compromising data privacy.
arXiv Detail & Related papers (2023-07-01T12:30:58Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Vision-Language Modelling For Radiological Imaging and Reports In The
Low Data Regime [70.04389979779195]
This paper explores training medical vision-language models (VLMs) where the visual and language inputs are embedded into a common space.
We explore several candidate methods to improve low-data performance, including adapting generic pre-trained models to novel image and text domains.
Using text-to-image retrieval as a benchmark, we evaluate the performance of these methods with variable sized training datasets of paired chest X-rays and radiological reports.
arXiv Detail & Related papers (2023-03-30T18:20:00Z) - Dynamic Graph Enhanced Contrastive Learning for Chest X-ray Report
Generation [92.73584302508907]
We propose a knowledge graph with Dynamic structure and nodes to facilitate medical report generation with Contrastive Learning.
In detail, the fundamental structure of our graph is pre-constructed from general knowledge.
Each image feature is integrated with its very own updated graph before being fed into the decoder module for report generation.
arXiv Detail & Related papers (2023-03-18T03:53:43Z) - Improving CT Image Segmentation Accuracy Using StyleGAN Driven Data
Augmentation [42.034896915716374]
This paper presents a StyleGAN-driven approach for segmenting publicly available large medical datasets.
Style transfer is used to augment the training dataset and generate new anatomically sound images.
The augmented dataset is then used to train a U-Net segmentation network which displays a significant improvement in the segmentation accuracy.
arXiv Detail & Related papers (2023-02-07T06:34:10Z) - Fader Networks for domain adaptation on fMRI: ABIDE-II study [68.5481471934606]
We use 3D convolutional autoencoders to build the domain irrelevant latent space image representation and demonstrate this method to outperform existing approaches on ABIDE data.
arXiv Detail & Related papers (2020-10-14T16:50:50Z) - Fed-Sim: Federated Simulation for Medical Imaging [131.56325440976207]
We introduce a physics-driven generative approach that consists of two learnable neural modules.
We show that our data synthesis framework improves the downstream segmentation performance on several datasets.
arXiv Detail & Related papers (2020-09-01T19:17:46Z) - IntrA: 3D Intracranial Aneurysm Dataset for Deep Learning [18.163031102785904]
We introduce an open-access 3D intracranial aneurysm dataset, IntrA, that makes the application of points-based and mesh-based classification and segmentation models available.
Our dataset can be used to diagnose intracranial aneurysms and to extract the neck for a clipping operation in medicine.
arXiv Detail & Related papers (2020-03-02T05:21:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.