AutoADR: Automatic Model Design for Ad Relevance
- URL: http://arxiv.org/abs/2010.07075v1
- Date: Wed, 14 Oct 2020 13:24:43 GMT
- Title: AutoADR: Automatic Model Design for Ad Relevance
- Authors: Yiren Chen, Yaming Yang, Hong Sun, Yujing Wang, Yu Xu, Wei Shen, Rong
Zhou, Yunhai Tong, Jing Bai, Ruofei Zhang
- Abstract summary: Large-scale pre-trained models are memory and computation intensive.
How to design an effective yet efficient model architecture is another challenging problem in online Ad Relevance.
We propose AutoADR -- a novel end-to-end framework to address this challenge.
- Score: 26.890941853575253
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Large-scale pre-trained models have attracted extensive attention in the
research community and shown promising results on various tasks of natural
language processing. However, these pre-trained models are memory and
computation intensive, hindering their deployment into industrial online
systems like Ad Relevance. Meanwhile, how to design an effective yet efficient
model architecture is another challenging problem in online Ad Relevance.
Recently, AutoML shed new lights on architecture design, but how to integrate
it with pre-trained language models remains unsettled. In this paper, we
propose AutoADR (Automatic model design for AD Relevance) -- a novel end-to-end
framework to address this challenge, and share our experience to ship these
cutting-edge techniques into online Ad Relevance system at Microsoft Bing.
Specifically, AutoADR leverages a one-shot neural architecture search algorithm
to find a tailored network architecture for Ad Relevance. The search process is
simultaneously guided by knowledge distillation from a large pre-trained
teacher model (e.g. BERT), while taking the online serving constraints (e.g.
memory and latency) into consideration. We add the model designed by AutoADR as
a sub-model into the production Ad Relevance model. This additional sub-model
improves the Precision-Recall AUC (PR AUC) on top of the original Ad Relevance
model by 2.65X of the normalized shipping bar. More importantly, adding this
automatically designed sub-model leads to a statistically significant 4.6%
Bad-Ad ratio reduction in online A/B testing. This model has been shipped into
Microsoft Bing Ad Relevance Production model.
Related papers
- Towards Automated Model Design on Recommender Systems [21.421326082345136]
We introduce a novel paradigm that utilizes weight sharing to explore abundant solution spaces.
From a co-design perspective, we achieve 2x FLOPs efficiency, 1.8x energy efficiency, and 1.5x performance improvements in recommender models.
arXiv Detail & Related papers (2024-11-12T06:03:47Z) - Mitigating Downstream Model Risks via Model Provenance [30.083382916838623]
We propose a machine-readable model specification format to simplify the creation of model records.
Our solution explicitly traces relationships between upstream and downstream models, enhancing transparency and traceability.
This proof of concept aims to set a new standard for managing foundation models, bridging the gap between innovation and responsible model management.
arXiv Detail & Related papers (2024-10-03T05:52:15Z) - AIDE: An Automatic Data Engine for Object Detection in Autonomous Driving [68.73885845181242]
We propose an Automatic Data Engine (AIDE) that automatically identifies issues, efficiently curates data, improves the model through auto-labeling, and verifies the model through generation of diverse scenarios.
We further establish a benchmark for open-world detection on AV datasets to comprehensively evaluate various learning paradigms, demonstrating our method's superior performance at a reduced cost.
arXiv Detail & Related papers (2024-03-26T04:27:56Z) - Auto-Train-Once: Controller Network Guided Automatic Network Pruning from Scratch [72.26822499434446]
Auto-Train-Once (ATO) is an innovative network pruning algorithm designed to automatically reduce the computational and storage costs of DNNs.
We provide a comprehensive convergence analysis as well as extensive experiments, and the results show that our approach achieves state-of-the-art performance across various model architectures.
arXiv Detail & Related papers (2024-03-21T02:33:37Z) - QualEval: Qualitative Evaluation for Model Improvement [82.73561470966658]
We propose QualEval, which augments quantitative scalar metrics with automated qualitative evaluation as a vehicle for model improvement.
QualEval uses a powerful LLM reasoner and our novel flexible linear programming solver to generate human-readable insights.
We demonstrate that leveraging its insights, for example, improves the absolute performance of the Llama 2 model by up to 15% points relative.
arXiv Detail & Related papers (2023-11-06T00:21:44Z) - A Car Model Identification System for Streamlining the Automobile Sales
Process [0.0]
This project presents an automated solution for the efficient identification of car models and makes from images.
We achieved a notable accuracy of 81.97% employing the EfficientNet (V2 b2) architecture.
The trained model offers the potential for automating information extraction, promising enhanced user experiences across car-selling websites.
arXiv Detail & Related papers (2023-10-19T23:36:17Z) - Re-parameterizing Your Optimizers rather than Architectures [119.08740698936633]
We propose a novel paradigm of incorporating model-specific prior knowledge into Structurals and using them to train generic (simple) models.
As an implementation, we propose a novel methodology to add prior knowledge by modifying the gradients according to a set of model-specific hyper- parameters.
For a simple model trained with a Repr, we focus on a VGG-style plain model and showcase that such a simple model trained with a Repr, which is referred to as Rep-VGG, performs on par with the recent well-designed models.
arXiv Detail & Related papers (2022-05-30T16:55:59Z) - Resource-Aware Pareto-Optimal Automated Machine Learning Platform [1.6746303554275583]
novel platform Resource-Aware AutoML (RA-AutoML)
RA-AutoML enables flexible and generalized algorithms to build machine learning models subjected to multiple objectives.
arXiv Detail & Related papers (2020-10-30T19:37:48Z) - Fast, Accurate, and Simple Models for Tabular Data via Augmented
Distillation [97.42894942391575]
We propose FAST-DAD to distill arbitrarily complex ensemble predictors into individual models like boosted trees, random forests, and deep networks.
Our individual distilled models are over 10x faster and more accurate than ensemble predictors produced by AutoML tools like H2O/AutoSklearn.
arXiv Detail & Related papers (2020-06-25T09:57:47Z) - AutoFIS: Automatic Feature Interaction Selection in Factorization Models
for Click-Through Rate Prediction [75.16836697734995]
We propose a two-stage algorithm called Automatic Feature Interaction Selection (AutoFIS)
AutoFIS can automatically identify important feature interactions for factorization models with computational cost just equivalent to training the target model to convergence.
AutoFIS has been deployed onto the training platform of Huawei App Store recommendation service.
arXiv Detail & Related papers (2020-03-25T06:53:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.