MEMTRACK: A Deep Learning-Based Approach to Microrobot Tracking in Dense
and Low-Contrast Environments
- URL: http://arxiv.org/abs/2310.09441v1
- Date: Fri, 13 Oct 2023 23:21:32 GMT
- Title: MEMTRACK: A Deep Learning-Based Approach to Microrobot Tracking in Dense
and Low-Contrast Environments
- Authors: Medha Sawhney, Bhas Karmarkar, Eric J. Leaman, Arka Daw, Anuj
Karpatne, Bahareh Behkam
- Abstract summary: Motion Enhanced Multi-level Tracker (MEMTrack) is a robust pipeline for detecting and tracking microrobots.
We trained and validated our model using bacterial micro-motors in collagen (tissue phantom) and tested it in collagen and aqueous media.
MEMTrack can quantify average bacteria speed with no statistically significant difference from the laboriously-produced manual tracking data.
- Score: 4.638136711579875
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Tracking microrobots is challenging, considering their minute size and high
speed. As the field progresses towards developing microrobots for biomedical
applications and conducting mechanistic studies in physiologically relevant
media (e.g., collagen), this challenge is exacerbated by the dense surrounding
environments with feature size and shape comparable to microrobots. Herein, we
report Motion Enhanced Multi-level Tracker (MEMTrack), a robust pipeline for
detecting and tracking microrobots using synthetic motion features, deep
learning-based object detection, and a modified Simple Online and Real-time
Tracking (SORT) algorithm with interpolation for tracking. Our object detection
approach combines different models based on the object's motion pattern. We
trained and validated our model using bacterial micro-motors in collagen
(tissue phantom) and tested it in collagen and aqueous media. We demonstrate
that MEMTrack accurately tracks even the most challenging bacteria missed by
skilled human annotators, achieving precision and recall of 77% and 48% in
collagen and 94% and 35% in liquid media, respectively. Moreover, we show that
MEMTrack can quantify average bacteria speed with no statistically significant
difference from the laboriously-produced manual tracking data. MEMTrack
represents a significant contribution to microrobot localization and tracking,
and opens the potential for vision-based deep learning approaches to microrobot
control in dense and low-contrast settings. All source code for training and
testing MEMTrack and reproducing the results of the paper have been made
publicly available https://github.com/sawhney-medha/MEMTrack.
Related papers
- A Data-Centric Revisit of Pre-Trained Vision Models for Robot Learning [67.72413262980272]
Pre-trained vision models (PVMs) are fundamental to modern robotics, yet their optimal configuration remains unclear.
We develop SlotMIM, a method that induces object-centric representations by introducing a semantic bottleneck.
Our approach achieves significant improvements over prior work in image recognition, scene understanding, and robot learning evaluations.
arXiv Detail & Related papers (2025-03-10T06:18:31Z) - OneTrack-M: A multitask approach to transformer-based MOT models [0.0]
Multi-Object Tracking (MOT) is a critical problem in computer vision.
OneTrack-M is a transformer-based MOT model designed to enhance tracking computational efficiency and accuracy.
arXiv Detail & Related papers (2025-02-06T20:02:06Z) - Temporal Correlation Meets Embedding: Towards a 2nd Generation of JDE-based Real-Time Multi-Object Tracking [52.04679257903805]
Joint Detection and Embedding (JDE) trackers have demonstrated excellent performance in Multi-Object Tracking (MOT) tasks.
Our tracker, named TCBTrack, achieves state-of-the-art performance on multiple public benchmarks.
arXiv Detail & Related papers (2024-07-19T07:48:45Z) - Chemotaxis of sea urchin sperm cells through deep reinforcement learning [0.0]
In this work, we investigate how a model of sea urchin sperm cell can self-learn chemotactic motion in a chemoattractant concentration field.
We employ an artificial neural network to act as a decision-making agent and facilitate the sperm cell to discover efficient maneuver strategies.
Our results provide insights to the chemotactic process of sea urchin sperm cells and also prepare guidance for the intelligent maneuver of microrobots.
arXiv Detail & Related papers (2022-08-02T06:04:32Z) - Transformer Inertial Poser: Attention-based Real-time Human Motion
Reconstruction from Sparse IMUs [79.72586714047199]
We propose an attention-based deep learning method to reconstruct full-body motion from six IMU sensors in real-time.
Our method achieves new state-of-the-art results both quantitatively and qualitatively, while being simple to implement and smaller in size.
arXiv Detail & Related papers (2022-03-29T16:24:52Z) - Learnable Online Graph Representations for 3D Multi-Object Tracking [156.58876381318402]
We propose a unified and learning based approach to the 3D MOT problem.
We employ a Neural Message Passing network for data association that is fully trainable.
We show the merit of the proposed approach on the publicly available nuScenes dataset by achieving state-of-the-art performance of 65.6% AMOTA and 58% fewer ID-switches.
arXiv Detail & Related papers (2021-04-23T17:59:28Z) - Autonomous object harvesting using synchronized optoelectronic
microrobots [10.860767733334306]
Optoelectronic tweezer-driven microrobots (OETdMs) are a versatile micromanipulation technology.
We describe an approach to automated targeting and path planning to enable open-loop control of multiple microrobots.
arXiv Detail & Related papers (2021-03-08T17:24:15Z) - Domain Adaptive Robotic Gesture Recognition with Unsupervised
Kinematic-Visual Data Alignment [60.31418655784291]
We propose a novel unsupervised domain adaptation framework which can simultaneously transfer multi-modality knowledge, i.e., both kinematic and visual data, from simulator to real robot.
It remedies the domain gap with enhanced transferable features by using temporal cues in videos, and inherent correlations in multi-modal towards recognizing gesture.
Results show that our approach recovers the performance with great improvement gains, up to 12.91% in ACC and 20.16% in F1score without using any annotations in real robot.
arXiv Detail & Related papers (2021-03-06T09:10:03Z) - Online Body Schema Adaptation through Cost-Sensitive Active Learning [63.84207660737483]
The work was implemented in a simulation environment, using the 7DoF arm of the iCub robot simulator.
A cost-sensitive active learning approach is used to select optimal joint configurations.
The results show cost-sensitive active learning has similar accuracy to the standard active learning approach, while reducing in about half the executed movement.
arXiv Detail & Related papers (2021-01-26T16:01:02Z) - Towards an Automatic Analysis of CHO-K1 Suspension Growth in
Microfluidic Single-cell Cultivation [63.94623495501023]
We propose a novel Machine Learning architecture, which allows us to infuse a neural deep network with human-powered abstraction on the level of data.
Specifically, we train a generative model simultaneously on natural and synthetic data, so that it learns a shared representation, from which a target variable, such as the cell count, can be reliably estimated.
arXiv Detail & Related papers (2020-10-20T08:36:51Z) - DistNet: Deep Tracking by displacement regression: application to
bacteria growing in the Mother Machine [2.741266294612776]
We introduce a Deep Neural Network architecture taking advantage of a self-attention mechanism which yields extremely low tracking error rate and segmentation error rate.
Our method is named DiSTNet which stands for DISTance+DISplacement and Tracking Network.
arXiv Detail & Related papers (2020-03-17T16:13:44Z) - AP-MTL: Attention Pruned Multi-task Learning Model for Real-time
Instrument Detection and Segmentation in Robot-assisted Surgery [23.33984309289549]
Training a real-time robotic system for the detection and segmentation of high-resolution images provides a challenging problem with the limited computational resource.
We develop a novel end-to-end trainable real-time Multi-Task Learning model with weight-shared encoder and task-aware detection and segmentation decoders.
Our model significantly outperforms state-of-the-art segmentation and detection models, including best-performed models in the challenge.
arXiv Detail & Related papers (2020-03-10T14:24:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.