ERA: Expert Retrieval and Assembly for Early Action Prediction
- URL: http://arxiv.org/abs/2207.09675v3
- Date: Fri, 22 Jul 2022 06:20:42 GMT
- Title: ERA: Expert Retrieval and Assembly for Early Action Prediction
- Authors: Lin Geng Foo, Tianjiao Li, Hossein Rahmani, Qiuhong Ke, Jun Liu
- Abstract summary: Early action prediction aims to successfully predict the class label of an action before it is completely performed.
This is a challenging task because the beginning stages of different actions can be very similar.
We propose a novel Expert Retrieval and Assembly (ERA) module that retrieves and assembles a set of experts specialized at using subtle differences.
- Score: 13.721609856985376
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Early action prediction aims to successfully predict the class label of an
action before it is completely performed. This is a challenging task because
the beginning stages of different actions can be very similar, with only minor
subtle differences for discrimination. In this paper, we propose a novel Expert
Retrieval and Assembly (ERA) module that retrieves and assembles a set of
experts most specialized at using discriminative subtle differences, to
distinguish an input sample from other highly similar samples. To encourage our
model to effectively use subtle differences for early action prediction, we
push experts to discriminate exclusively between samples that are highly
similar, forcing these experts to learn to use subtle differences that exist
between those samples. Additionally, we design an effective Expert Learning
Rate Optimization method that balances the experts' optimization and leads to
better performance. We evaluate our ERA module on four public action datasets
and achieve state-of-the-art performance.
Related papers
- Domain-Specific Pruning of Large Mixture-of-Experts Models with Few-shot Demonstrations [48.890534958441016]
We investigate domain specialization and expert redundancy in large-scale MoE models.
We propose a simple yet effective pruning framework, EASY-EP, to identify and retain only the most relevant experts.
Our method can achieve comparable performances and $2.99times$ throughput under the same memory budget with full DeepSeek-R1 with only half the experts.
arXiv Detail & Related papers (2025-04-09T11:34:06Z) - Diverging Preferences: When do Annotators Disagree and do Models Know? [92.24651142187989]
We develop a taxonomy of disagreement sources spanning 10 categories across four high-level classes.
We find that the majority of disagreements are in opposition with standard reward modeling approaches.
We develop methods for identifying diverging preferences to mitigate their influence on evaluation and training.
arXiv Detail & Related papers (2024-10-18T17:32:22Z) - Evaluating Alternative Training Interventions Using Personalized Computational Models of Learning [0.0]
evaluating different training interventions to determine which produce the best learning outcomes is one of the main challenges faced by instructional designers.
We present an approach for automatically tuning models to specific individuals and show that personalized models make better predictions of students' behavior than generic ones.
Our approach makes predictions that align with previous human findings, as well as testable predictions that might be evaluated with future human experiments.
arXiv Detail & Related papers (2024-08-24T22:51:57Z) - Benchmarking for Deep Uplift Modeling in Online Marketing [17.70084353772874]
Deep uplift modeling (DUM) as a promising technique has attracted increased research from academia and industry.
Current DUM still lacks some standardized benchmarks and unified evaluation protocols.
We provide an open benchmark for DUM and present comparison results of existing models in a reproducible and uniform manner.
arXiv Detail & Related papers (2024-06-01T07:23:37Z) - MoCaE: Mixture of Calibrated Experts Significantly Improves Object
Detection [18.059899772411033]
We find that na"ively combining expert object detectors in a similar way to Deep Ensembles, can often lead to degraded performance.
We identify that the primary cause of this issue is that the predictions of the experts do not match their performance.
To address this, when constructing the Mixture of Experts, we propose to combine their predictions in a manner which reflects the individual performance of the experts.
arXiv Detail & Related papers (2023-09-26T14:52:51Z) - Incorporating Experts' Judgment into Machine Learning Models [2.5363839239628843]
In some cases, domain experts might have a judgment about the expected outcome that might conflict with the prediction of machine learning models.
We present a novel framework that aims at leveraging experts' judgment to mitigate the conflict.
arXiv Detail & Related papers (2023-04-24T07:32:49Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - Agree to Disagree: Diversity through Disagreement for Better
Transferability [54.308327969778155]
We propose D-BAT (Diversity-By-disAgreement Training), which enforces agreement among the models on the training data.
We show how D-BAT naturally emerges from the notion of generalized discrepancy.
arXiv Detail & Related papers (2022-02-09T12:03:02Z) - Revisiting Consistency Regularization for Semi-Supervised Learning [80.28461584135967]
We propose an improved consistency regularization framework by a simple yet effective technique, FeatDistLoss.
Experimental results show that our model defines a new state of the art for various datasets and settings.
arXiv Detail & Related papers (2021-12-10T20:46:13Z) - Contrastive Learning for Fair Representations [50.95604482330149]
Trained classification models can unintentionally lead to biased representations and predictions.
Existing debiasing methods for classification models, such as adversarial training, are often expensive to train and difficult to optimise.
We propose a method for mitigating bias by incorporating contrastive learning, in which instances sharing the same class label are encouraged to have similar representations.
arXiv Detail & Related papers (2021-09-22T10:47:51Z) - An Effective Baseline for Robustness to Distributional Shift [5.627346969563955]
Refraining from confidently predicting when faced with categories of inputs different from those seen during training is an important requirement for the safe deployment of deep learning systems.
We present a simple, but highly effective approach to deal with out-of-distribution detection that uses the principle of abstention.
arXiv Detail & Related papers (2021-05-15T00:46:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.