A Survey on Dropout Methods and Experimental Verification in
Recommendation
- URL: http://arxiv.org/abs/2204.02027v1
- Date: Tue, 5 Apr 2022 07:08:21 GMT
- Title: A Survey on Dropout Methods and Experimental Verification in
Recommendation
- Authors: Yangkun Li, Weizhi Ma, Chong Chen, Min Zhang, Yiqun Liu, Shaoping Ma,
Yuekui Yang
- Abstract summary: Overfitting is a common problem in machine learning, which means the model too closely fits the training data while performing poorly in the test data.
Among various methods of coping with overfitting, dropout is one of the representative ways.
From randomly dropping neurons to dropping neural structures, dropout has achieved great success in improving model performances.
- Score: 34.557554809126415
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Overfitting is a common problem in machine learning, which means the model
too closely fits the training data while performing poorly in the test data.
Among various methods of coping with overfitting, dropout is one of the
representative ways. From randomly dropping neurons to dropping neural
structures, dropout has achieved great success in improving model performances.
Although various dropout methods have been designed and widely applied in past
years, their effectiveness, application scenarios, and contributions have not
been comprehensively summarized and empirically compared by far. It is the
right time to make a comprehensive survey.
In this paper, we systematically review previous dropout methods and classify
them into three major categories according to the stage where dropout operation
is performed. Specifically, more than seventy dropout methods published in top
AI conferences or journals (e.g., TKDE, KDD, TheWebConf, SIGIR) are involved.
The designed taxonomy is easy to understand and capable of including new
dropout methods. Then, we further discuss their application scenarios,
connections, and contributions. To verify the effectiveness of distinct dropout
methods, extensive experiments are conducted on recommendation scenarios with
abundant heterogeneous information. Finally, we propose some open problems and
potential research directions about dropout that worth to be further explored.
Related papers
- Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - Zero-shot Retrieval: Augmenting Pre-trained Models with Search Engines [83.65380507372483]
Large pre-trained models can dramatically reduce the amount of task-specific data required to solve a problem, but they often fail to capture domain-specific nuances out of the box.
This paper shows how to leverage recent advances in NLP and multi-modal learning to augment a pre-trained model with search engine retrieval.
arXiv Detail & Related papers (2023-11-29T05:33:28Z) - Systematic Evaluation of Predictive Fairness [60.0947291284978]
Mitigating bias in training on biased datasets is an important open problem.
We examine the performance of various debiasing methods across multiple tasks.
We find that data conditions have a strong influence on relative model performance.
arXiv Detail & Related papers (2022-10-17T05:40:13Z) - On Modality Bias Recognition and Reduction [70.69194431713825]
We study the modality bias problem in the context of multi-modal classification.
We propose a plug-and-play loss function method, whereby the feature space for each label is adaptively learned.
Our method yields remarkable performance improvements compared with the baselines.
arXiv Detail & Related papers (2022-02-25T13:47:09Z) - A Survey on Evidential Deep Learning For Single-Pass Uncertainty
Estimation [0.0]
Evidential Deep Learning: For unfamiliar data, they admit "what they don't know" and fall back onto a prior belief.
This survey aims to familiarize the reader with an alternative class of models based on the concept of Evidential Deep Learning: For unfamiliar data, they admit "what they don't know" and fall back onto a prior belief.
arXiv Detail & Related papers (2021-10-06T20:13:57Z) - CODA: Constructivism Learning for Instance-Dependent Dropout
Architecture Construction [3.2238887070637805]
We propose Constructivism learning for instance-dependent Dropout Architecture (CODA)
Based on the theory we have designed a better drop out technique, Uniform Process Mixture Models.
We have evaluated our proposed method on 5 real-world datasets and compared the performance with other state-of-the-art dropout techniques.
arXiv Detail & Related papers (2021-06-15T21:32:28Z) - Contextual Dropout: An Efficient Sample-Dependent Dropout Module [60.63525456640462]
Dropout has been demonstrated as a simple and effective module to regularize the training process of deep neural networks.
We propose contextual dropout with an efficient structural design as a simple and scalable sample-dependent dropout module.
Our experimental results show that the proposed method outperforms baseline methods in terms of both accuracy and quality of uncertainty estimation.
arXiv Detail & Related papers (2021-03-06T19:30:32Z) - Advanced Dropout: A Model-free Methodology for Bayesian Dropout
Optimization [62.8384110757689]
Overfitting ubiquitously exists in real-world applications of deep neural networks (DNNs)
The advanced dropout technique applies a model-free and easily implemented distribution with parametric prior, and adaptively adjusts dropout rate.
We evaluate the effectiveness of the advanced dropout against nine dropout techniques on seven computer vision datasets.
arXiv Detail & Related papers (2020-10-11T13:19:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.