Enhancing Personalized Ranking With Differentiable Group AUC
Optimization
- URL: http://arxiv.org/abs/2304.09176v1
- Date: Mon, 17 Apr 2023 09:39:40 GMT
- Title: Enhancing Personalized Ranking With Differentiable Group AUC
Optimization
- Authors: Xiao Sun, Bo Zhang, Chenrui Zhang, Han Ren, Mingchen Cai
- Abstract summary: We propose the PDAOM loss, a personalized and differentiable AUC Optimization method with Maximum violation.
The proposed PDAOM loss not only improves the AUC and GAUC metrics in the offline evaluation, but also reduces the complexity of the training objective.
Online evaluation of the PDAOM loss on the 'Guess What You Like' feed recommendation application in Meituan manifests 1.40% increase in click count and 0.65% increase in order count.
- Score: 10.192514219354651
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: AUC is a common metric for evaluating the performance of a classifier.
However, most classifiers are trained with cross entropy, and it does not
optimize the AUC metric directly, which leaves a gap between the training and
evaluation stage. In this paper, we propose the PDAOM loss, a Personalized and
Differentiable AUC Optimization method with Maximum violation, which can be
directly applied when training a binary classifier and optimized with
gradient-based methods. Specifically, we construct the pairwise exponential
loss with difficult pair of positive and negative samples within sub-batches
grouped by user ID, aiming to guide the classifier to pay attention to the
relation between hard-distinguished pairs of opposite samples from the
perspective of independent users. Compared to the origin form of pairwise
exponential loss, the proposed PDAOM loss not only improves the AUC and GAUC
metrics in the offline evaluation, but also reduces the computation complexity
of the training objective. Furthermore, online evaluation of the PDAOM loss on
the 'Guess What You Like' feed recommendation application in Meituan manifests
1.40% increase in click count and 0.65% increase in order count compared to the
baseline model, which is a significant improvement in this well-developed
online life service recommendation system.
Related papers
- Ordinal Preference Optimization: Aligning Human Preferences via NDCG [28.745322441961438]
We develop an end-to-end preference optimization algorithm by approxing NDCG with a differentiable surrogate loss.
OPO outperforms existing pairwise and listwise approaches on evaluation sets and general benchmarks like AlpacaEval.
arXiv Detail & Related papers (2024-10-06T03:49:28Z) - Bridging and Modeling Correlations in Pairwise Data for Direct Preference Optimization [75.1240295759264]
We propose an effective framework for Bridging and Modeling Correlations in pairwise data, named BMC.
We increase the consistency and informativeness of the pairwise preference signals through targeted modifications.
We identify that DPO alone is insufficient to model these correlations and capture nuanced variations.
arXiv Detail & Related papers (2024-08-14T11:29:47Z) - Prior Constraints-based Reward Model Training for Aligning Large Language Models [58.33118716810208]
This paper proposes a Prior Constraints-based Reward Model (namely PCRM) training method to mitigate this problem.
PCRM incorporates prior constraints, specifically, length ratio and cosine similarity between outputs of each comparison pair, during reward model training to regulate optimization magnitude and control score margins.
Experimental results demonstrate that PCRM significantly improves alignment performance by effectively constraining reward score scaling.
arXiv Detail & Related papers (2024-04-01T07:49:11Z) - Lower-Left Partial AUC: An Effective and Efficient Optimization Metric
for Recommendation [52.45394284415614]
We propose a new optimization metric, Lower-Left Partial AUC (LLPAUC), which is computationally efficient like AUC but strongly correlates with Top-K ranking metrics.
LLPAUC considers only the partial area under the ROC curve in the Lower-Left corner to push the optimization focus on Top-K.
arXiv Detail & Related papers (2024-02-29T13:58:33Z) - Learning by Sorting: Self-supervised Learning with Group Ordering
Constraints [75.89238437237445]
This paper proposes a new variation of the contrastive learning objective, Group Ordering Constraints (GroCo)
It exploits the idea of sorting the distances of positive and negative pairs and computing the respective loss based on how many positive pairs have a larger distance than the negative pairs, and thus are not ordered correctly.
We evaluate the proposed formulation on various self-supervised learning benchmarks and show that it not only leads to improved results compared to vanilla contrastive learning but also shows competitive performance to comparable methods in linear probing and outperforms current methods in k-NN performance.
arXiv Detail & Related papers (2023-01-05T11:17:55Z) - Deep Negative Correlation Classification [82.45045814842595]
Existing deep ensemble methods naively train many different models and then aggregate their predictions.
We propose deep negative correlation classification (DNCC)
DNCC yields a deep classification ensemble where the individual estimator is both accurate and negatively correlated.
arXiv Detail & Related papers (2022-12-14T07:35:20Z) - Adaptive Adversarial Logits Pairing [65.51670200266913]
An adversarial training solution Adversarial Logits Pairing (ALP) tends to rely on fewer high-contribution features compared with vulnerable ones.
Motivated by these observations, we design an Adaptive Adversarial Logits Pairing (AALP) solution by modifying the training process and training target of ALP.
AALP consists of an adaptive feature optimization module with Guided Dropout to systematically pursue fewer high-contribution features.
arXiv Detail & Related papers (2020-05-25T03:12:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.