BrackishMOT: The Brackish Multi-Object Tracking Dataset
- URL: http://arxiv.org/abs/2302.10645v1
- Date: Tue, 21 Feb 2023 13:02:36 GMT
- Title: BrackishMOT: The Brackish Multi-Object Tracking Dataset
- Authors: Malte Pedersen, Daniel Lehotsk\'y, Ivan Nikolov, and Thomas B.
Moeslund
- Abstract summary: There exist no publicly available annotated underwater multi-object tracking (MOT) datasets captured in turbid environments.
BrackishMOT consists of 98 sequences captured in the wild. Alongside the novel dataset, we present baseline results by training a state-of-the-art tracker.
We analyse the effects of including synthetic data during training and show that a combination of real and synthetic underwater training data can enhance tracking performance.
- Score: 20.52569822945148
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: There exist no publicly available annotated underwater multi-object tracking
(MOT) datasets captured in turbid environments. To remedy this we propose the
BrackishMOT dataset with focus on tracking schools of small fish, which is a
notoriously difficult MOT task. BrackishMOT consists of 98 sequences captured
in the wild. Alongside the novel dataset, we present baseline results by
training a state-of-the-art tracker. Additionally, we propose a framework for
creating synthetic sequences in order to expand the dataset. The framework
consists of animated fish models and realistic underwater environments. We
analyse the effects of including synthetic data during training and show that a
combination of real and synthetic underwater training data can enhance tracking
performance. Links to code and data can be found at
https://www.vap.aau.dk/brackishmot
Related papers
- Improving Underwater Visual Tracking With a Large Scale Dataset and
Image Enhancement [70.2429155741593]
This paper presents a new dataset and general tracker enhancement method for Underwater Visual Object Tracking (UVOT)
It poses distinct challenges; the underwater environment exhibits non-uniform lighting conditions, low visibility, lack of sharpness, low contrast, camouflage, and reflections from suspended particles.
We propose a novel underwater image enhancement algorithm designed specifically to boost tracking quality.
The method has resulted in a significant performance improvement, of up to 5.0% AUC, of state-of-the-art (SOTA) visual trackers.
arXiv Detail & Related papers (2023-08-30T07:41:26Z) - Whale Detection Enhancement through Synthetic Satellite Images [13.842008598751445]
We show that we can achieve a 15% performance boost on whale detection compared to using the real data alone for training.
We open source the code of the simulation platform SeaDroneSim2 and the dataset generated through it.
arXiv Detail & Related papers (2023-08-15T13:35:29Z) - Synthetic Data-based Detection of Zebras in Drone Imagery [0.8249180979158817]
We present an approach for training an animal detector using only synthetic data.
The dataset includes RGB, depth, skeletal joint locations, pose, shape and instance segmentations for each subject.
We show that we can detect zebras by using only synthetic data during training.
arXiv Detail & Related papers (2023-04-30T09:24:31Z) - Large Scale Real-World Multi-Person Tracking [68.27438015329807]
This paper presents a new large scale multi-person tracking dataset -- textttPersonPath22.
It is over an order of magnitude larger than currently available high quality multi-object tracking datasets such as MOT17, HiEve, and MOT20.
arXiv Detail & Related papers (2022-11-03T23:03:13Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - Learning Dynamic View Synthesis With Few RGBD Cameras [60.36357774688289]
We propose to utilize RGBD cameras to synthesize free-viewpoint videos of dynamic indoor scenes.
We generate point clouds from RGBD frames and then render them into free-viewpoint videos via a neural feature.
We introduce a simple Regional Depth-Inpainting module that adaptively inpaints missing depth values to render complete novel views.
arXiv Detail & Related papers (2022-04-22T03:17:35Z) - Unified Transformer Tracker for Object Tracking [58.65901124158068]
We present the Unified Transformer Tracker (UTT) to address tracking problems in different scenarios with one paradigm.
A track transformer is developed in our UTT to track the target in both Single Object Tracking (SOT) and Multiple Object Tracking (MOT)
arXiv Detail & Related papers (2022-03-29T01:38:49Z) - A Realistic Fish-Habitat Dataset to Evaluate Algorithms for Underwater
Visual Analysis [2.6476746128312194]
We present DeepFish as a benchmark suite with a large-scale dataset to train and test methods for several computer vision tasks.
The dataset consists of approximately 40 thousand images collected underwater from 20 greenhabitats in the marine-environments of tropical Australia.
Our experiments provide an in-depth analysis of the dataset characteristics, and the performance evaluation of several state-of-the-art approaches.
arXiv Detail & Related papers (2020-08-28T12:20:59Z) - Meta-Sim2: Unsupervised Learning of Scene Structure for Synthetic Data
Generation [88.04759848307687]
In Meta-Sim2, we aim to learn the scene structure in addition to parameters, which is a challenging problem due to its discrete nature.
We use Reinforcement Learning to train our model, and design a feature space divergence between our synthesized and target images that is key to successful training.
We also show that this leads to downstream improvement in the performance of an object detector trained on our generated dataset as opposed to other baseline simulation methods.
arXiv Detail & Related papers (2020-08-20T17:28:45Z) - Exploring the Impacts from Datasets to Monocular Depth Estimation (MDE)
Models with MineNavi [5.689127984415125]
Current computer vision tasks based on deep learning require a huge amount of data with annotations for model training or testing.
In practice, manual labeling for dense estimation tasks is very difficult or even impossible, and the scenes of the dataset are often restricted to a small range.
We propose a synthetic dataset generation method to obtain the expandable dataset without burdensome manual workforce.
arXiv Detail & Related papers (2020-08-19T14:03:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.