A Survey on Dropout Methods and Experimental Verification in
Recommendation
- URL: http://arxiv.org/abs/2204.02027v1
- Date: Tue, 5 Apr 2022 07:08:21 GMT
- Title: A Survey on Dropout Methods and Experimental Verification in
Recommendation
- Authors: Yangkun Li, Weizhi Ma, Chong Chen, Min Zhang, Yiqun Liu, Shaoping Ma,
Yuekui Yang
- Abstract summary: Overfitting is a common problem in machine learning, which means the model too closely fits the training data while performing poorly in the test data.
Among various methods of coping with overfitting, dropout is one of the representative ways.
From randomly dropping neurons to dropping neural structures, dropout has achieved great success in improving model performances.
- Score: 34.557554809126415
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Overfitting is a common problem in machine learning, which means the model
too closely fits the training data while performing poorly in the test data.
Among various methods of coping with overfitting, dropout is one of the
representative ways. From randomly dropping neurons to dropping neural
structures, dropout has achieved great success in improving model performances.
Although various dropout methods have been designed and widely applied in past
years, their effectiveness, application scenarios, and contributions have not
been comprehensively summarized and empirically compared by far. It is the
right time to make a comprehensive survey.
In this paper, we systematically review previous dropout methods and classify
them into three major categories according to the stage where dropout operation
is performed. Specifically, more than seventy dropout methods published in top
AI conferences or journals (e.g., TKDE, KDD, TheWebConf, SIGIR) are involved.
The designed taxonomy is easy to understand and capable of including new
dropout methods. Then, we further discuss their application scenarios,
connections, and contributions. To verify the effectiveness of distinct dropout
methods, extensive experiments are conducted on recommendation scenarios with
abundant heterogeneous information. Finally, we propose some open problems and
potential research directions about dropout that worth to be further explored.
Related papers
- FairDropout: Using Example-Tied Dropout to Enhance Generalization of Minority Groups [10.274236106456758]
We show that models trained with empirical risk minimization tend to generalize well for examples from the majority groups while memorizing instances from minority groups.
We apply example-tied dropout as a method we term FairDropout, aimed at redirecting this memorization to specific neurons that we subsequently drop out during inference.
We empirically evaluate FairDropout using the subpopulation benchmark suite encompassing vision, language, and healthcare tasks, demonstrating that it significantly reduces reliance on spurious correlations, and outperforms state-of-the-art methods.
arXiv Detail & Related papers (2025-02-10T17:18:54Z) - Breaking Fine-Grained Classification Barriers with Cost-Free Data in Few-Shot Class-Incremental Learning [13.805180905579832]
We propose a novel learning paradigm to break barriers in fine-grained classification.
It enables the model to learn beyond the standard training phase and benefit from cost-free data encountered during system operation.
arXiv Detail & Related papers (2024-12-29T07:11:44Z) - Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - Zero-shot Retrieval: Augmenting Pre-trained Models with Search Engines [83.65380507372483]
Large pre-trained models can dramatically reduce the amount of task-specific data required to solve a problem, but they often fail to capture domain-specific nuances out of the box.
This paper shows how to leverage recent advances in NLP and multi-modal learning to augment a pre-trained model with search engine retrieval.
arXiv Detail & Related papers (2023-11-29T05:33:28Z) - Systematic Evaluation of Predictive Fairness [60.0947291284978]
Mitigating bias in training on biased datasets is an important open problem.
We examine the performance of various debiasing methods across multiple tasks.
We find that data conditions have a strong influence on relative model performance.
arXiv Detail & Related papers (2022-10-17T05:40:13Z) - On Modality Bias Recognition and Reduction [70.69194431713825]
We study the modality bias problem in the context of multi-modal classification.
We propose a plug-and-play loss function method, whereby the feature space for each label is adaptively learned.
Our method yields remarkable performance improvements compared with the baselines.
arXiv Detail & Related papers (2022-02-25T13:47:09Z) - CODA: Constructivism Learning for Instance-Dependent Dropout
Architecture Construction [3.2238887070637805]
We propose Constructivism learning for instance-dependent Dropout Architecture (CODA)
Based on the theory we have designed a better drop out technique, Uniform Process Mixture Models.
We have evaluated our proposed method on 5 real-world datasets and compared the performance with other state-of-the-art dropout techniques.
arXiv Detail & Related papers (2021-06-15T21:32:28Z) - Contextual Dropout: An Efficient Sample-Dependent Dropout Module [60.63525456640462]
Dropout has been demonstrated as a simple and effective module to regularize the training process of deep neural networks.
We propose contextual dropout with an efficient structural design as a simple and scalable sample-dependent dropout module.
Our experimental results show that the proposed method outperforms baseline methods in terms of both accuracy and quality of uncertainty estimation.
arXiv Detail & Related papers (2021-03-06T19:30:32Z) - Advanced Dropout: A Model-free Methodology for Bayesian Dropout
Optimization [62.8384110757689]
Overfitting ubiquitously exists in real-world applications of deep neural networks (DNNs)
The advanced dropout technique applies a model-free and easily implemented distribution with parametric prior, and adaptively adjusts dropout rate.
We evaluate the effectiveness of the advanced dropout against nine dropout techniques on seven computer vision datasets.
arXiv Detail & Related papers (2020-10-11T13:19:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.