SQE: a Self Quality Evaluation Metric for Parameters Optimization in
Multi-Object Tracking
- URL: http://arxiv.org/abs/2004.07472v1
- Date: Thu, 16 Apr 2020 06:07:29 GMT
- Title: SQE: a Self Quality Evaluation Metric for Parameters Optimization in
Multi-Object Tracking
- Authors: Yanru Huang, Feiyu Zhu, Zheni Zeng, Xi Qiu, Yuan Shen, Jianan Wu
- Abstract summary: We present a novel self quality evaluation metric SQE for parameters optimization in the challenging yet critical multi-object tracking task.
By contrast, our metric reflects the internal characteristics of trajectory hypotheses and measures tracking performance without ground truth.
- Score: 25.723436561224297
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a novel self quality evaluation metric SQE for parameters
optimization in the challenging yet critical multi-object tracking task.
Current evaluation metrics all require annotated ground truth, thus will fail
in the test environment and realistic circumstances prohibiting further
optimization after training. By contrast, our metric reflects the internal
characteristics of trajectory hypotheses and measures tracking performance
without ground truth. We demonstrate that trajectories with different qualities
exhibit different single or multiple peaks over feature distance distribution,
inspiring us to design a simple yet effective method to assess the quality of
trajectories using a two-class Gaussian mixture model. Experiments mainly on
MOT16 Challenge data sets verify the effectiveness of our method in both
correlating with existing metrics and enabling parameters self-optimization to
achieve better performance. We believe that our conclusions and method are
inspiring for future multi-object tracking in practice.
Related papers
- Unlearning as multi-task optimization: A normalized gradient difference approach with an adaptive learning rate [105.86576388991713]
We introduce a normalized gradient difference (NGDiff) algorithm, enabling us to have better control over the trade-off between the objectives.
We provide a theoretical analysis and empirically demonstrate the superior performance of NGDiff among state-of-the-art unlearning methods on the TOFU and MUSE datasets.
arXiv Detail & Related papers (2024-10-29T14:41:44Z) - Trajectory-Based Multi-Objective Hyperparameter Optimization for Model Retraining [8.598456741786801]
We present a novel trajectory-based multi-objective Bayesian optimization algorithm.
Our algorithm outperforms the state-of-the-art multi-objectives in both locating better trade-offs and tuning efficiency.
arXiv Detail & Related papers (2024-05-24T07:43:45Z) - Are We Really Achieving Better Beyond-Accuracy Performance in Next Basket Recommendation? [57.91114305844153]
Next basket recommendation (NBR) is a special type of sequential recommendation that is increasingly receiving attention.
Recent studies into NBR have found a substantial performance difference between recommending repeat items and explore items.
We propose a plug-and-play two-step repetition-exploration framework that treats repeat items and explores items separately.
arXiv Detail & Related papers (2024-05-02T09:59:35Z) - On Task Performance and Model Calibration with Supervised and
Self-Ensembled In-Context Learning [71.44986275228747]
In-context learning (ICL) has become an efficient approach propelled by the recent advancements in large language models (LLMs)
However, both paradigms are prone to suffer from the critical problem of overconfidence (i.e., miscalibration)
arXiv Detail & Related papers (2023-12-21T11:55:10Z) - QualEval: Qualitative Evaluation for Model Improvement [82.73561470966658]
We propose QualEval, which augments quantitative scalar metrics with automated qualitative evaluation as a vehicle for model improvement.
QualEval uses a powerful LLM reasoner and our novel flexible linear programming solver to generate human-readable insights.
We demonstrate that leveraging its insights, for example, improves the absolute performance of the Llama 2 model by up to 15% points relative.
arXiv Detail & Related papers (2023-11-06T00:21:44Z) - A Study of Unsupervised Evaluation Metrics for Practical and Automatic
Domain Adaptation [15.728090002818963]
Unsupervised domain adaptation (UDA) methods facilitate the transfer of models to target domains without labels.
In this paper, we aim to find an evaluation metric capable of assessing the quality of a transferred model without access to target validation labels.
arXiv Detail & Related papers (2023-08-01T05:01:05Z) - A Perceptually-Validated Metric for Crowd Trajectory Quality Evaluation [3.0448872422956432]
We study the relation between parametric values for simulation techniques and the quality of the resulting trajectories.
A quality metric, QF, is proposed to abstract from reference data while capturing the most salient features that affect the perception of trajectory realism.
arXiv Detail & Related papers (2021-08-27T15:22:26Z) - MetricOpt: Learning to Optimize Black-Box Evaluation Metrics [21.608384691401238]
We study the problem of optimizing arbitrary non-differentiable task evaluation metrics such as misclassification rate and recall.
Our method, named MetricOpt, operates in a black-box setting where the computational details of the target metric are unknown.
We achieve this by learning a differentiable value function, which maps compact task-specific model parameters to metric observations.
arXiv Detail & Related papers (2021-04-21T16:50:01Z) - ReMP: Rectified Metric Propagation for Few-Shot Learning [67.96021109377809]
A rectified metric space is learned to maintain the metric consistency from training to testing.
Numerous analyses indicate that a simple modification of the objective can yield substantial performance gains.
The proposed ReMP is effective and efficient, and outperforms the state of the arts on various standard few-shot learning datasets.
arXiv Detail & Related papers (2020-12-02T00:07:53Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.