Sustainable Transparency in Recommender Systems: Bayesian Ranking of
Images for Explainability
- URL: http://arxiv.org/abs/2308.01196v2
- Date: Thu, 21 Dec 2023 11:27:00 GMT
- Title: Sustainable Transparency in Recommender Systems: Bayesian Ranking of
Images for Explainability
- Authors: Jorge Paz-Ruza, Amparo Alonso-Betanzos, Berta Guijarro-Berdi\~nas,
Brais Cancela, Carlos Eiras-Franco
- Abstract summary: Recommender Systems have become crucial in the modern world, commonly guiding users towards relevant content or products.
personalized explanations have emerged as a solution, offering justifications for recommendations.
BRIE is a novel model where we leverage Pairwise Ranking to enhance the training process, allowing us to consistently outperform state-of-the-art models in six real-world datasets.
- Score: 5.499796332553708
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Recommender Systems have become crucial in the modern world, commonly guiding
users towards relevant content or products, and having a large influence over
the decisions of users and citizens. However, ensuring transparency and user
trust in these systems remains a challenge; personalized explanations have
emerged as a solution, offering justifications for recommendations. Among the
existing approaches for generating personalized explanations, using existing
visual content created by users is a promising option to maximize transparency
and user trust. State-of-the-art models that follow this approach, despite
leveraging highly optimized architectures, employ surrogate learning tasks that
do not efficiently model the objective of ranking images as explanations for a
given recommendation; this leads to a suboptimal training process with high
computational costs that may not be reduced without affecting model
performance. This work presents BRIE, a novel model where we leverage Bayesian
Pairwise Ranking to enhance the training process, allowing us to consistently
outperform state-of-the-art models in six real-world datasets while reducing
its model size by up to 64 times and its CO${_2}$ emissions by up to 75% in
training and inference.
Related papers
- Slow Thinking for Sequential Recommendation [88.46598279655575]
We present a novel slow thinking recommendation model, named STREAM-Rec.
Our approach is capable of analyzing historical user behavior, generating a multi-step, deliberative reasoning process, and delivering personalized recommendations.
In particular, we focus on two key challenges: (1) identifying the suitable reasoning patterns in recommender systems, and (2) exploring how to effectively stimulate the reasoning capabilities of traditional recommenders.
arXiv Detail & Related papers (2025-04-13T15:53:30Z) - Dynamic Noise Preference Optimization for LLM Self-Improvement via Synthetic Data [51.62162460809116]
We introduce Dynamic Noise Preference Optimization (DNPO) to ensure consistent improvements across iterations.
In experiments with Zephyr-7B, DNPO consistently outperforms existing methods, showing an average performance boost of 2.6%.
DNPO shows a significant improvement in model-generated data quality, with a 29.4% win-loss rate gap compared to the baseline in GPT-4 evaluations.
arXiv Detail & Related papers (2025-02-08T01:20:09Z) - Recommendation and Temptation [3.734925590025741]
We propose a novel user model that accounts for dual-self behavior and develop an optimal recommendation strategy.
We evaluate our approach through both synthetic simulations and simulations based on real-world data from the MovieLens dataset.
arXiv Detail & Related papers (2024-12-13T22:44:22Z) - VLFeedback: A Large-Scale AI Feedback Dataset for Large Vision-Language Models Alignment [55.7956150385255]
We investigate the efficacy of AI feedback to scale supervision for aligning vision-language models.
We introduce VLFeedback, the first large-scale vision-language feedback dataset.
We train Silkie, an LVLM fine-tuned via direct preference optimization on VLFeedback.
arXiv Detail & Related papers (2024-10-12T07:56:47Z) - Self-Taught Evaluators [77.92610887220594]
We present an approach that aims to im-proves without human annotations, using synthetic training data only.
Our Self-Taught Evaluator can improve a strong LLM from 75.4 to 88.3 on RewardBench.
arXiv Detail & Related papers (2024-08-05T17:57:02Z) - Positive-Unlabelled Learning for Improving Image-based Recommender System Explainability [2.9748898344267785]
This work proposes a new explainer training pipeline by leveraging Positive-Unlabelled (PU) Learning techniques.
Experiments show this PU-based approach outperforms the state-of-the-art non-PU method in six popular real-world datasets.
arXiv Detail & Related papers (2024-07-09T10:40:31Z) - Enhancing Large Vision Language Models with Self-Training on Image Comprehension [131.14381425260706]
We introduce Self-Training on Image (STIC), which emphasizes a self-training approach specifically for image comprehension.
First, the model self-constructs a preference for image descriptions using unlabeled images.
To further self-improve reasoning on the extracted visual information, we let the model reuse a small portion of existing instruction-tuning data.
arXiv Detail & Related papers (2024-05-30T05:53:49Z) - Efficient Model-agnostic Alignment via Bayesian Persuasion [13.42367964190663]
We introduce a model-agnostic and lightweight Bayesian Persuasion Alignment framework.
In the persuasion process, the small model (Advisor) observes the information item (i.e., state) and persuades large models (Receiver) to elicit improved responses.
We show that GPT-2 can significantly improve the performance of various models, achieving an average enhancement of 16.1% in mathematical reasoning ability and 13.7% in code generation.
arXiv Detail & Related papers (2024-05-29T02:57:07Z) - Confidence-aware Reward Optimization for Fine-tuning Text-to-Image Models [85.96013373385057]
Fine-tuning text-to-image models with reward functions trained on human feedback data has proven effective for aligning model behavior with human intent.
However, excessive optimization with such reward models, which serve as mere proxy objectives, can compromise the performance of fine-tuned models.
We propose TextNorm, a method that enhances alignment based on a measure of reward model confidence estimated across a set of semantically contrastive text prompts.
arXiv Detail & Related papers (2024-04-02T11:40:38Z) - COURIER: Contrastive User Intention Reconstruction for Large-Scale Visual Recommendation [33.903096803803706]
We argue that a visual feature pre-training method tailored for recommendation is necessary for further improvements beyond existing modality features.
We propose an effective user intention reconstruction module to mine visual features related to user interests from behavior histories.
arXiv Detail & Related papers (2023-06-08T07:45:24Z) - Occlusion-Aware Self-Supervised Monocular 6D Object Pose Estimation [88.8963330073454]
We propose a novel monocular 6D pose estimation approach by means of self-supervised learning.
We leverage current trends in noisy student training and differentiable rendering to further self-supervise the model.
Our proposed self-supervision outperforms all other methods relying on synthetic data.
arXiv Detail & Related papers (2022-03-19T15:12:06Z) - Fast Multi-Step Critiquing for VAE-based Recommender Systems [27.207067974031805]
We present M&Ms-VAE, a novel variational autoencoder for recommendation and explanation.
We train the model under a weak supervision scheme to simulate both fully and partially observed variables.
We then leverage the generalization ability of a trained M&Ms-VAE model to embed the user preference and the critique separately.
arXiv Detail & Related papers (2021-05-03T12:26:09Z) - Learning User Representations with Hypercuboids for Recommender Systems [26.80987554753327]
Our model explicitly models user interests as a hypercuboid instead of a point in the space.
We present two variants of hypercuboids to enhance the capability in capturing the diversities of user interests.
A neural architecture is also proposed to facilitate user hypercuboid learning by capturing the activity sequences (e.g., buy and rate) of users.
arXiv Detail & Related papers (2020-11-11T12:50:00Z) - Explainable Recommender Systems via Resolving Learning Representations [57.24565012731325]
Explanations could help improve user experience and discover system defects.
We propose a novel explainable recommendation model through improving the transparency of the representation learning process.
arXiv Detail & Related papers (2020-08-21T05:30:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.