Modeling Uncertainty in Personalized Emotion Prediction with Normalizing
Flows
- URL: http://arxiv.org/abs/2312.06034v1
- Date: Sun, 10 Dec 2023 23:21:41 GMT
- Title: Modeling Uncertainty in Personalized Emotion Prediction with Normalizing
Flows
- Authors: Piotr Mi{\l}kowski, Konrad Karanowski, Patryk Wielopolski, Jan
Koco\'n, Przemys{\l}aw Kazienko, Maciej Zi\k{e}ba
- Abstract summary: This work proposes a novel approach to capture the uncertainty of the forecast using conditional Normalizing Flows.
We validated our method on three challenging, subjective NLP tasks, including emotion recognition and hate speech.
The information brought by the developed methods makes it possible to build hybrid models whose effectiveness surpasses classic solutions.
- Score: 6.32047610997385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Designing predictive models for subjective problems in natural language
processing (NLP) remains challenging. This is mainly due to its
non-deterministic nature and different perceptions of the content by different
humans. It may be solved by Personalized Natural Language Processing (PNLP),
where the model exploits additional information about the reader to make more
accurate predictions. However, current approaches require complete information
about the recipients to be straight embedded. Besides, the recent methods focus
on deterministic inference or simple frequency-based estimations of the
probabilities. In this work, we overcome this limitation by proposing a novel
approach to capture the uncertainty of the forecast using conditional
Normalizing Flows. This allows us to model complex multimodal distributions and
to compare various models using negative log-likelihood (NLL). In addition, the
new solution allows for various interpretations of possible reader perception
thanks to the available sampling function. We validated our method on three
challenging, subjective NLP tasks, including emotion recognition and hate
speech. The comparative analysis of generalized and personalized approaches
revealed that our personalized solutions significantly outperform the baseline
and provide more precise uncertainty estimates. The impact on the text
interpretability and uncertainty studies are presented as well. The information
brought by the developed methods makes it possible to build hybrid models whose
effectiveness surpasses classic solutions. In addition, an analysis and
visualization of the probabilities of the given decisions for texts with high
entropy of annotations and annotators with mixed views were carried out.
Related papers
- DISCO: DISCovering Overfittings as Causal Rules for Text Classification Models [6.369258625916601]
Post-hoc interpretability methods fail to capture the models' decision-making process fully.
Our paper introduces DISCO, a novel method for discovering global, rule-based explanations.
DISCO supports interactive explanations, enabling human inspectors to distinguish spurious causes in the rule-based output.
arXiv Detail & Related papers (2024-11-07T12:12:44Z) - On Uncertainty In Natural Language Processing [2.5076643086429993]
This thesis studies how uncertainty in natural language processing can be characterized from a linguistic, statistical and neural perspective.
We propose a method for calibrated sampling in natural language generation based on non-exchangeable conformal prediction.
Lastly, we develop an approach to quantify confidence in large black-box language models using auxiliary predictors.
arXiv Detail & Related papers (2024-10-04T14:08:02Z) - It's All in the Mix: Wasserstein Machine Learning with Mixed Features [5.739657897440173]
We present a practically efficient algorithm to solve mixed-feature problems.
We demonstrate that our approach can significantly outperform existing methods that are to the presence of discrete features.
arXiv Detail & Related papers (2023-12-19T15:15:52Z) - LaPLACE: Probabilistic Local Model-Agnostic Causal Explanations [1.0370398945228227]
We introduce LaPLACE-explainer, designed to provide probabilistic cause-and-effect explanations for machine learning models.
The LaPLACE-Explainer component leverages the concept of a Markov blanket to establish statistical boundaries between relevant and non-relevant features.
Our approach offers causal explanations and outperforms LIME and SHAP in terms of local accuracy and consistency of explained features.
arXiv Detail & Related papers (2023-10-01T04:09:59Z) - Quantification of Predictive Uncertainty via Inference-Time Sampling [57.749601811982096]
We propose a post-hoc sampling strategy for estimating predictive uncertainty accounting for data ambiguity.
The method can generate different plausible outputs for a given input and does not assume parametric forms of predictive distributions.
arXiv Detail & Related papers (2023-08-03T12:43:21Z) - Human Trajectory Forecasting with Explainable Behavioral Uncertainty [63.62824628085961]
Human trajectory forecasting helps to understand and predict human behaviors, enabling applications from social robots to self-driving cars.
Model-free methods offer superior prediction accuracy but lack explainability, while model-based methods provide explainability but cannot predict well.
We show that BNSP-SFM achieves up to a 50% improvement in prediction accuracy, compared with 11 state-of-the-art methods.
arXiv Detail & Related papers (2023-07-04T16:45:21Z) - Explaining Language Models' Predictions with High-Impact Concepts [11.47612457613113]
We propose a complete framework for extending concept-based interpretability methods to NLP.
We optimize for features whose existence causes the output predictions to change substantially.
Our method achieves superior results on predictive impact, usability, and faithfulness compared to the baselines.
arXiv Detail & Related papers (2023-05-03T14:48:27Z) - Tailoring Language Generation Models under Total Variation Distance [55.89964205594829]
The standard paradigm of neural language generation adopts maximum likelihood estimation (MLE) as the optimizing method.
We develop practical bounds to apply it to language generation.
We introduce the TaiLr objective that balances the tradeoff of estimating TVD.
arXiv Detail & Related papers (2023-02-26T16:32:52Z) - NUQ: Nonparametric Uncertainty Quantification for Deterministic Neural
Networks [151.03112356092575]
We show the principled way to measure the uncertainty of predictions for a classifier based on Nadaraya-Watson's nonparametric estimate of the conditional label distribution.
We demonstrate the strong performance of the method in uncertainty estimation tasks on a variety of real-world image datasets.
arXiv Detail & Related papers (2022-02-07T12:30:45Z) - Interpretable Social Anchors for Human Trajectory Forecasting in Crowds [84.20437268671733]
We propose a neural network-based system to predict human trajectory in crowds.
We learn interpretable rule-based intents, and then utilise the expressibility of neural networks to model scene-specific residual.
Our architecture is tested on the interaction-centric benchmark TrajNet++.
arXiv Detail & Related papers (2021-05-07T09:22:34Z) - Accounting for Unobserved Confounding in Domain Generalization [107.0464488046289]
This paper investigates the problem of learning robust, generalizable prediction models from a combination of datasets.
Part of the challenge of learning robust models lies in the influence of unobserved confounders.
We demonstrate the empirical performance of our approach on healthcare data from different modalities.
arXiv Detail & Related papers (2020-07-21T08:18:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.