MOTS R-CNN: Cosine-margin-triplet loss for multi-object tracking
- URL: http://arxiv.org/abs/2102.03512v1
- Date: Sat, 6 Feb 2021 05:03:29 GMT
- Title: MOTS R-CNN: Cosine-margin-triplet loss for multi-object tracking
- Authors: Amit Satish Unde and Renu M. Rameshan
- Abstract summary: One of the central tasks of multi-object tracking involves learning a distance metric consistent with the semantic similarities of objects.
In this paper, we propose cosine-margin-contrastive (CMC) and cosine-margin-triplet (CMT) loss by reformulating both contrastive and triplet loss functions.
We then propose the MOTS R-CNN framework for joint multi-object tracking and segmentation, particularly targeted at improving the tracking performance.
- Score: 2.8935588665357077
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: One of the central tasks of multi-object tracking involves learning a
distance metric that is consistent with the semantic similarities of objects.
The design of an appropriate loss function that encourages discriminative
feature learning is among the most crucial challenges in deep neural
network-based metric learning. Despite significant progress, slow convergence
and a poor local optimum of the existing contrastive and triplet loss based
deep metric learning methods necessitates a better solution. In this paper, we
propose cosine-margin-contrastive (CMC) and cosine-margin-triplet (CMT) loss by
reformulating both contrastive and triplet loss functions from the perspective
of cosine distance. The proposed reformulation as a cosine loss is achieved by
feature normalization which distributes the learned features on a hypersphere.
We then propose the MOTS R-CNN framework for joint multi-object tracking and
segmentation, particularly targeted at improving the tracking performance.
Specifically, the tracking problem is addressed through deep metric learning
based on the proposed loss functions. We propose a scale-invariant tracking by
using a multi-layer feature aggregation scheme to make the model robust against
object scale variations and occlusions. The MOTS R-CNN achieves the
state-of-the-art tracking performance on the KITTI MOTS dataset. We show that
the MOTS R-CNN reduces the identity switching by $62\%$ and $61\%$ on cars and
pedestrians, respectively in comparison to Track R-CNN.
Related papers
- DiTMoS: Delving into Diverse Tiny-Model Selection on Microcontrollers [34.282971510732736]
We introduce DiTMoS, a novel DNN training and inference framework with a selector-classifiers architecture.
A composition of weak models can exhibit high diversity and the union of them can significantly boost the accuracy upper bound.
We deploy DiTMoS on the Neucleo STM32F767ZI board and evaluate it based on three time-series datasets for human activity recognition, keywords spotting, and emotion recognition.
arXiv Detail & Related papers (2024-03-14T02:11:38Z) - A Multi-Head Ensemble Multi-Task Learning Approach for Dynamical
Computation Offloading [62.34538208323411]
We propose a multi-head ensemble multi-task learning (MEMTL) approach with a shared backbone and multiple prediction heads (PHs)
MEMTL outperforms benchmark methods in both the inference accuracy and mean square error without requiring additional training data.
arXiv Detail & Related papers (2023-09-02T11:01:16Z) - Class Anchor Margin Loss for Content-Based Image Retrieval [97.81742911657497]
We propose a novel repeller-attractor loss that falls in the metric learning paradigm, yet directly optimize for the L2 metric without the need of generating pairs.
We evaluate the proposed objective in the context of few-shot and full-set training on the CBIR task, by using both convolutional and transformer architectures.
arXiv Detail & Related papers (2023-06-01T12:53:10Z) - SuSana Distancia is all you need: Enforcing class separability in metric
learning via two novel distance-based loss functions for few-shot image
classification [0.9236074230806579]
We propose two loss functions which consider the importance of the embedding vectors by looking at the intra-class and inter-class distance between the few data.
Our results show a significant improvement in accuracy in the miniImagenNet benchmark compared to other metric-based few-shot learning methods by a margin of 2%.
arXiv Detail & Related papers (2023-05-15T23:12:09Z) - Improved Aggregating and Accelerating Training Methods for Spatial Graph
Neural Networks on Fraud Detection [0.0]
This work proposes an improved deep architecture to extend CAmouflage-REsistant GNN (CARE-GNN) to deep models named as Residual Layered CARE-GNN (RLC-GNN)
Three issues of RLC-GNN are the usage of neighboring information reaching limitation, the training difficulty and lack of comprehensive consideration about node features and external patterns.
Experiments are conducted on Yelp and Amazon datasets.
arXiv Detail & Related papers (2022-02-14T09:51:35Z) - InverseForm: A Loss Function for Structured Boundary-Aware Segmentation [80.39674800972182]
We present a novel boundary-aware loss term for semantic segmentation using an inverse-transformation network.
This plug-in loss term complements the cross-entropy loss in capturing boundary transformations.
We analyze the quantitative and qualitative effects of our loss function on three indoor and outdoor segmentation benchmarks.
arXiv Detail & Related papers (2021-04-06T18:52:45Z) - Object Tracking through Residual and Dense LSTMs [67.98948222599849]
Deep learning-based trackers based on LSTMs (Long Short-Term Memory) recurrent neural networks have emerged as a powerful alternative.
DenseLSTMs outperform Residual and regular LSTM, and offer a higher resilience to nuisances.
Our case study supports the adoption of residual-based RNNs for enhancing the robustness of other trackers.
arXiv Detail & Related papers (2020-06-22T08:20:17Z) - MOPS-Net: A Matrix Optimization-driven Network forTask-Oriented 3D Point
Cloud Downsampling [86.42733428762513]
MOPS-Net is a novel interpretable deep learning-based method for matrix optimization.
We show that MOPS-Net can achieve favorable performance against state-of-the-art deep learning-based methods over various tasks.
arXiv Detail & Related papers (2020-05-01T14:01:53Z) - When Residual Learning Meets Dense Aggregation: Rethinking the
Aggregation of Deep Neural Networks [57.0502745301132]
We propose Micro-Dense Nets, a novel architecture with global residual learning and local micro-dense aggregations.
Our micro-dense block can be integrated with neural architecture search based models to boost their performance.
arXiv Detail & Related papers (2020-04-19T08:34:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.