Intelligent Robotic Sonographer: Mutual Information-based Disentangled
Reward Learning from Few Demonstrations
- URL: http://arxiv.org/abs/2307.03705v2
- Date: Wed, 29 Nov 2023 10:11:49 GMT
- Title: Intelligent Robotic Sonographer: Mutual Information-based Disentangled
Reward Learning from Few Demonstrations
- Authors: Zhongliang Jiang, Yuan Bi, Mingchuan Zhou, Ying Hu, Michael Burke and
Nassir Navab
- Abstract summary: This work proposes an intelligent robotic sonographer to autonomously "explore" target anatomies and navigate a US probe to a relevant 2D plane by learning from the expert.
The underlying high-level physiological knowledge from experts is inferred by a neural reward function.
The proposed advanced framework can robustly work on a variety of seen and unseen phantoms as well as in-vivo human carotid data.
- Score: 42.731081399649916
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Ultrasound (US) imaging is widely used for biometric measurement and
diagnosis of internal organs due to the advantages of being real-time and
radiation-free. However, due to inter-operator variations, resulting images
highly depend on the experience of sonographers. This work proposes an
intelligent robotic sonographer to autonomously "explore" target anatomies and
navigate a US probe to a relevant 2D plane by learning from the expert. The
underlying high-level physiological knowledge from experts is inferred by a
neural reward function, using a ranked pairwise image comparisons approach in a
self-supervised fashion. This process can be referred to as understanding the
"language of sonography". Considering the generalization capability to overcome
inter-patient variations, mutual information is estimated by a network to
explicitly disentangle the task-related and domain features in latent space.
The robotic localization is carried out in coarse-to-fine mode based on the
predicted reward associated with B-mode images. To validate the effectiveness
of the proposed reward inference network, representative experiments were
performed on vascular phantoms ("line" target), two types of ex-vivo animal
organs (chicken heart and lamb kidney) phantoms ("point" target) and in-vivo
human carotids, respectively. To further validate the performance of the
autonomous acquisition framework, physical robotic acquisitions were performed
on three phantoms (vascular, chicken heart, and lamb kidney). The results
demonstrated that the proposed advanced framework can robustly work on a
variety of seen and unseen phantoms as well as in-vivo human carotid data.
Related papers
- Active inference and deep generative modeling for cognitive ultrasound [20.383444113659476]
We show that US imaging systems can be recast as information-seeking agents that engage in reciprocal interactions with their anatomical environment.
Such agents autonomously adapt their transmit-receive sequences to fully personalize imaging and actively maximize information gain in-situ.
We then equip systems with a mechanism to actively reduce uncertainty and maximize diagnostic value across a sequence of experiments.
arXiv Detail & Related papers (2024-10-17T08:09:14Z) - CathFlow: Self-Supervised Segmentation of Catheters in Interventional Ultrasound Using Optical Flow and Transformers [66.15847237150909]
We introduce a self-supervised deep learning architecture to segment catheters in longitudinal ultrasound images.
The network architecture builds upon AiAReSeg, a segmentation transformer built with the Attention in Attention mechanism.
We validated our model on a test dataset, consisting of unseen synthetic data and images collected from silicon aorta phantoms.
arXiv Detail & Related papers (2024-03-21T15:13:36Z) - AiAReSeg: Catheter Detection and Segmentation in Interventional
Ultrasound using Transformers [75.20925220246689]
endovascular surgeries are performed using the golden standard of Fluoroscopy, which uses ionising radiation to visualise catheters and vasculature.
This work proposes a solution using an adaptation of a state-of-the-art machine learning transformer architecture to detect and segment catheters in axial interventional Ultrasound image sequences.
arXiv Detail & Related papers (2023-09-25T19:34:12Z) - LOTUS: Learning to Optimize Task-based US representations [39.81131738128329]
Anatomical segmentation of organs in ultrasound images is essential to many clinical applications.
Existing deep neural networks require a large amount of labeled data for training in order to achieve clinically acceptable performance.
In this paper, we propose a novel approach for learning to optimize task-based ultra-sound image representations.
arXiv Detail & Related papers (2023-07-29T16:29:39Z) - Towards Autonomous Atlas-based Ultrasound Acquisitions in Presence of
Articulated Motion [48.52403516006036]
This paper proposes a vision-based approach allowing autonomous robotic US limb scanning.
To this end, an atlas MRI template of a human arm with annotated vascular structures is used to generate trajectories.
In all cases, the system can successfully acquire the planned vascular structure on volunteers' limbs.
arXiv Detail & Related papers (2022-08-10T15:39:20Z) - Image-Guided Navigation of a Robotic Ultrasound Probe for Autonomous
Spinal Sonography Using a Shadow-aware Dual-Agent Framework [35.17207004351791]
We propose a novel dual-agent framework that integrates a reinforcement learning agent and a deep learning agent.
Our method can effectively interpret the US images and navigate the probe to acquire multiple standard views of the spine.
arXiv Detail & Related papers (2021-11-03T12:11:27Z) - Voice-assisted Image Labelling for Endoscopic Ultrasound Classification
using Neural Networks [48.732863591145964]
We propose a multi-modal convolutional neural network architecture that labels endoscopic ultrasound (EUS) images from raw verbal comments provided by a clinician during the procedure.
Our results show a prediction accuracy of 76% at image level on a dataset with 5 different labels.
arXiv Detail & Related papers (2021-10-12T21:22:24Z) - Generalized Organ Segmentation by Imitating One-shot Reasoning using
Anatomical Correlation [55.1248480381153]
We propose OrganNet which learns a generalized organ concept from a set of annotated organ classes and then transfer this concept to unseen classes.
We show that OrganNet can effectively resist the wide variations in organ morphology and produce state-of-the-art results in one-shot segmentation task.
arXiv Detail & Related papers (2021-03-30T13:41:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.