DCN V2: Improved Deep & Cross Network and Practical Lessons for
Web-scale Learning to Rank Systems
- URL: http://arxiv.org/abs/2008.13535v2
- Date: Tue, 20 Oct 2020 21:01:21 GMT
- Title: DCN V2: Improved Deep & Cross Network and Practical Lessons for
Web-scale Learning to Rank Systems
- Authors: Ruoxi Wang, Rakesh Shivanna, Derek Z. Cheng, Sagar Jain, Dong Lin,
Lichan Hong, Ed H. Chi
- Abstract summary: Deep & Cross Network (DCN) was proposed to automatically and efficiently learn bounded-degree predictive feature interactions.
We propose an improved framework DCN-V2 to make DCN more practical in large-scale industrial settings.
- Score: 15.398542784403604
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Learning effective feature crosses is the key behind building recommender
systems. However, the sparse and large feature space requires exhaustive search
to identify effective crosses. Deep & Cross Network (DCN) was proposed to
automatically and efficiently learn bounded-degree predictive feature
interactions. Unfortunately, in models that serve web-scale traffic with
billions of training examples, DCN showed limited expressiveness in its cross
network at learning more predictive feature interactions. Despite significant
research progress made, many deep learning models in production still rely on
traditional feed-forward neural networks to learn feature crosses
inefficiently.
In light of the pros/cons of DCN and existing feature interaction learning
approaches, we propose an improved framework DCN-V2 to make DCN more practical
in large-scale industrial settings. In a comprehensive experimental study with
extensive hyper-parameter search and model tuning, we observed that DCN-V2
approaches outperform all the state-of-the-art algorithms on popular benchmark
datasets. The improved DCN-V2 is more expressive yet remains cost efficient at
feature interaction learning, especially when coupled with a mixture of
low-rank architecture. DCN-V2 is simple, can be easily adopted as building
blocks, and has delivered significant offline accuracy and online business
metrics gains across many web-scale learning to rank systems at Google.
Related papers
- DCN^2: Interplay of Implicit Collision Weights and Explicit Cross Layers for Large-Scale Recommendation [1.1027313935007121]
We introduce three significant algorithmic improvements to the DCNv2 architecture, detailing their formulation and behavior at scale.<n>The enhanced architecture we refer to as DCN2 is actively used in a live recommender system, processing over 0.5 billion predictions per second across diverse use cases.<n>These improvements effectively address key limitations observed in the DCNv2, including information loss in Cross layers, implicit management of collisions through learnable lookup-level weights, and explicit modeling of pairwise similarities with a custom layer that emulates FFMs' behavior.
arXiv Detail & Related papers (2025-06-24T06:44:42Z) - Self-cross Feature based Spiking Neural Networks for Efficient Few-shot Learning [16.156610945877986]
We propose a few-shot learning framework based on Spiking Neural Networks (SNNs)<n>We apply the combination of temporal efficient training loss and Info Info loss to optimize the temporal spike dynamics of trains and enhance the discriminative power.
arXiv Detail & Related papers (2025-05-12T16:51:08Z) - Parallel Multi-path Feed Forward Neural Networks (PMFFNN) for Long Columnar Datasets: A Novel Approach to Complexity Reduction [0.0]
We introduce a novel architecture called Parallel Multi-path Feed Forward Neural Networks (PMFFNN)
By doing so, the architecture ensures that each subset of features receives focused attention, which is often neglected in traditional models.
PMFFNN outperforms traditional FFNNs and 1D CNNs, providing an optimized solution for managing large-scale data.
arXiv Detail & Related papers (2024-11-09T00:48:32Z) - DCNv3: Towards Next Generation Deep Cross Network for CTR Prediction [17.19859591493946]
This paper proposes the next generation deep cross network: Deep Cross Network v3 (DCNv3), along with its two sub-networks: Linear Cross Network (LCN) and Exponential Cross Network (ECN) for CTR prediction.
Comprehensive experiments on six datasets demonstrate the effectiveness, efficiency, and interpretability of DCNv3.
arXiv Detail & Related papers (2024-07-18T09:49:13Z) - RLEEGNet: Integrating Brain-Computer Interfaces with Adaptive AI for
Intuitive Responsiveness and High-Accuracy Motor Imagery Classification [0.0]
We introduce a framework that leverages Reinforcement Learning with Deep Q-Networks (DQN) for classification tasks.
We present a preprocessing technique for multiclass motor imagery (MI) classification in a One-Versus-The-Rest (OVR) manner.
The integration of DQN with a 1D-CNN-LSTM architecture optimize the decision-making process in real-time.
arXiv Detail & Related papers (2024-02-09T02:03:13Z) - Efficient Model Adaptation for Continual Learning at the Edge [15.334881190102895]
Most machine learning (ML) systems assume stationary and matching data distributions during training and deployment.
Data distributions often shift over time due to changes in environmental factors, sensor characteristics, and task-of-interest.
This paper presents theAdaptor-Reconfigurator (EAR) framework for efficient continual learning under domain shifts.
arXiv Detail & Related papers (2023-08-03T23:55:17Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Unifying Synergies between Self-supervised Learning and Dynamic
Computation [53.66628188936682]
We present a novel perspective on the interplay between SSL and DC paradigms.
We show that it is feasible to simultaneously learn a dense and gated sub-network from scratch in a SSL setting.
The co-evolution during pre-training of both dense and gated encoder offers a good accuracy-efficiency trade-off.
arXiv Detail & Related papers (2023-01-22T17:12:58Z) - Learning Ability of Interpolating Deep Convolutional Neural Networks [28.437011792990347]
We study the learning ability of an important family of deep neural networks, deep convolutional neural networks (DCNNs)
We show that by adding well-defined layers to a non-interpolating DCNN, we can obtain some interpolating DCNNs that maintain the good learning rates of the non-interpolating DCNN.
Our work provides theoretical verification of how overfitted DCNNs generalize well.
arXiv Detail & Related papers (2022-10-25T17:22:31Z) - Recurrent Bilinear Optimization for Binary Neural Networks [58.972212365275595]
BNNs neglect the intrinsic bilinear relationship of real-valued weights and scale factors.
Our work is the first attempt to optimize BNNs from the bilinear perspective.
We obtain robust RBONNs, which show impressive performance over state-of-the-art BNNs on various models and datasets.
arXiv Detail & Related papers (2022-09-04T06:45:33Z) - Comparison Analysis of Traditional Machine Learning and Deep Learning
Techniques for Data and Image Classification [62.997667081978825]
The purpose of the study is to analyse and compare the most common machine learning and deep learning techniques used for computer vision 2D object classification tasks.
Firstly, we will present the theoretical background of the Bag of Visual words model and Deep Convolutional Neural Networks (DCNN)
Secondly, we will implement a Bag of Visual Words model, the VGG16 CNN Architecture.
arXiv Detail & Related papers (2022-04-11T11:34:43Z) - Contextual HyperNetworks for Novel Feature Adaptation [43.49619456740745]
Contextual HyperNetwork (CHN) generates parameters for extending the base model to a new feature.
At prediction time, the CHN requires only a single forward pass through a neural network, yielding a significant speed-up.
We show that this system obtains improved few-shot learning performance for novel features over existing imputation and meta-learning baselines.
arXiv Detail & Related papers (2021-04-12T23:19:49Z) - Optimization-driven Machine Learning for Intelligent Reflecting Surfaces
Assisted Wireless Networks [82.33619654835348]
Intelligent surface (IRS) has been employed to reshape the wireless channels by controlling individual scattering elements' phase shifts.
Due to the large size of scattering elements, the passive beamforming is typically challenged by the high computational complexity.
In this article, we focus on machine learning (ML) approaches for performance in IRS-assisted wireless networks.
arXiv Detail & Related papers (2020-08-29T08:39:43Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.