Pathological Visual Question Answering
- URL: http://arxiv.org/abs/2010.12435v1
- Date: Tue, 6 Oct 2020 00:36:55 GMT
- Title: Pathological Visual Question Answering
- Authors: Xuehai He, Zhuo Cai, Wenlan Wei, Yichen Zhang, Luntian Mou, Eric Xing,
Pengtao Xie
- Abstract summary: We need to create a visual question answering (VQA) dataset where the AI agent is presented with a pathology image together with a question and is asked to give the correct answer.
Due to privacy concerns, pathology images are usually not publicly available.
It is difficult to hire highly experienced pathologists to create pathology visual questions and answers.
The medical concepts and knowledge covered in pathology question-answer (QA) pairs are very diverse.
- Score: 14.816825480418588
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Is it possible to develop an "AI Pathologist" to pass the board-certified
examination of the American Board of Pathology (ABP)? To build such a system,
three challenges need to be addressed. First, we need to create a visual
question answering (VQA) dataset where the AI agent is presented with a
pathology image together with a question and is asked to give the correct
answer. Due to privacy concerns, pathology images are usually not publicly
available. Besides, only well-trained pathologists can understand pathology
images, but they barely have time to help create datasets for AI research. The
second challenge is: since it is difficult to hire highly experienced
pathologists to create pathology visual questions and answers, the resulting
pathology VQA dataset may contain errors. Training pathology VQA models using
these noisy or even erroneous data will lead to problematic models that cannot
generalize well on unseen images. The third challenge is: the medical concepts
and knowledge covered in pathology question-answer (QA) pairs are very diverse
while the number of QA pairs available for modeling training is limited. How to
learn effective representations of diverse medical concepts based on limited
data is technically demanding. In this paper, we aim to address these three
challenges. To our best knowledge, our work represents the first one addressing
the pathology VQA problem. To deal with the issue that a publicly available
pathology VQA dataset is lacking, we create PathVQA dataset. To address the
second challenge, we propose a learning-by-ignoring approach. To address the
third challenge, we propose to use cross-modal self-supervised learning. We
perform experiments on our created PathVQA dataset and the results demonstrate
the effectiveness of our proposed learning-by-ignoring method and cross-modal
self-supervised learning methods.
Related papers
- Ask Questions with Double Hints: Visual Question Generation with Answer-awareness and Region-reference [107.53380946417003]
We propose a novel learning paradigm to generate visual questions with answer-awareness and region-reference.
We develop a simple methodology to self-learn the visual hints without introducing any additional human annotations.
arXiv Detail & Related papers (2024-07-06T15:07:32Z) - Robust and Interpretable Medical Image Classifiers via Concept
Bottleneck Models [49.95603725998561]
We propose a new paradigm to build robust and interpretable medical image classifiers with natural language concepts.
Specifically, we first query clinical concepts from GPT-4, then transform latent image features into explicit concepts with a vision-language model.
arXiv Detail & Related papers (2023-10-04T21:57:09Z) - DRAC: Diabetic Retinopathy Analysis Challenge with Ultra-Wide Optical
Coherence Tomography Angiography Images [51.27125547308154]
We organized a challenge named "DRAC - Diabetic Retinopathy Analysis Challenge" in conjunction with the 25th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI 2022)
The challenge consists of three tasks: segmentation of DR lesions, image quality assessment and DR grading.
This paper presents a summary and analysis of the top-performing solutions and results for each task of the challenge.
arXiv Detail & Related papers (2023-04-05T12:04:55Z) - Diagnosis of Paratuberculosis in Histopathological Images Based on
Explainable Artificial Intelligence and Deep Learning [0.0]
This study examines a new and original dataset using the deep learning algorithm, and visualizes the output with gradient-weighted class activation mapping (Grad-CAM)
Both the decision-making processes and the explanations were verified, and the accuracy of the output was tested.
The research results greatly help pathologists in the diagnosis of paratuberculosis.
arXiv Detail & Related papers (2022-08-02T18:05:26Z) - Consistency-preserving Visual Question Answering in Medical Imaging [2.005299372367689]
Visual Question Answering (VQA) models take an image and a natural-language question as input and infer the answer to the question.
We propose a novel loss function and corresponding training procedure that allows the inclusion of relations between questions into the training process.
Our experiments show that our method outperforms state-of-the-art baselines.
arXiv Detail & Related papers (2022-06-27T13:38:50Z) - Medical Visual Question Answering: A Survey [55.53205317089564]
Medical Visual Question Answering(VQA) is a combination of medical artificial intelligence and popular VQA challenges.
Given a medical image and a clinically relevant question in natural language, the medical VQA system is expected to predict a plausible and convincing answer.
arXiv Detail & Related papers (2021-11-19T05:55:15Z) - A Systematic Collection of Medical Image Datasets for Deep Learning [37.476768951211206]
Deep learning algorithms are data-dependent and require large datasets for training.
The lack of data in the medical imaging field creates a bottleneck for the application of deep learning to medical image analysis.
This paper provides a collection of medical image datasets with their associated challenges for deep learning research.
arXiv Detail & Related papers (2021-06-24T10:00:30Z) - Knowledge-Routed Visual Question Reasoning: Challenges for Deep
Representation Embedding [140.5911760063681]
We propose a novel dataset named Knowledge-Routed Visual Question Reasoning for VQA model evaluation.
We generate the question-answer pair based on both the Visual Genome scene graph and an external knowledge base with controlled programs.
arXiv Detail & Related papers (2020-12-14T00:33:44Z) - Interpretable Multi-Step Reasoning with Knowledge Extraction on Complex
Healthcare Question Answering [89.76059961309453]
HeadQA dataset contains multiple-choice questions authorized for the public healthcare specialization exam.
These questions are the most challenging for current QA systems.
We present a Multi-step reasoning with Knowledge extraction framework (MurKe)
We are striving to make full use of off-the-shelf pre-trained models.
arXiv Detail & Related papers (2020-08-06T02:47:46Z) - PathVQA: 30000+ Questions for Medical Visual Question Answering [15.343890121216335]
This is the first dataset for pathology VQA. To our best knowledge, this is the first dataset for pathology VQA. Our dataset will be released publicly to promote research in medical VQA.
arXiv Detail & Related papers (2020-03-07T17:55:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.