Weakly-Supervised Hierarchical Models for Predicting Persuasive
Strategies in Good-faith Textual Requests
- URL: http://arxiv.org/abs/2101.06351v1
- Date: Sat, 16 Jan 2021 02:31:04 GMT
- Title: Weakly-Supervised Hierarchical Models for Predicting Persuasive
Strategies in Good-faith Textual Requests
- Authors: Jiaao Chen, Diyi Yang
- Abstract summary: We introduce a large-scale multi-domain text corpus for modeling persuasive strategies in good-faith text requests.
We design a hierarchical weakly-supervised latent variable model that can leverage partially labeled data to predict such associated persuasive strategies for each sentence.
Experimental results showed that our proposed method outperformed existing semi-supervised baselines significantly.
- Score: 22.58861442978803
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Modeling persuasive language has the potential to better facilitate our
decision-making processes. Despite its importance, computational modeling of
persuasion is still in its infancy, largely due to the lack of benchmark
datasets that can provide quantitative labels of persuasive strategies to
expedite this line of research. To this end, we introduce a large-scale
multi-domain text corpus for modeling persuasive strategies in good-faith text
requests. Moreover, we design a hierarchical weakly-supervised latent variable
model that can leverage partially labeled data to predict such associated
persuasive strategies for each sentence, where the supervision comes from both
the overall document-level labels and very limited sentence-level labels.
Experimental results showed that our proposed method outperformed existing
semi-supervised baselines significantly. We have publicly released our code at
https://github.com/GT-SALT/Persuasion_Strategy_WVAE.
Related papers
- Breaking the Manual Annotation Bottleneck: Creating a Comprehensive Legal Case Criticality Dataset through Semi-Automated Labeling [16.529070321280447]
This paper introduces the Criticality Prediction dataset, a new resource for evaluating the potential influence of Swiss Supreme Court decisions on future jurisprudence.
Unlike existing approaches that rely on resource-intensive manual annotations, we semi-automatically derive labels leading to a much larger dataset.
We evaluate several multilingual models, including fine-tuned variants and large language models, and find that fine-tuned models consistently outperform zero-shot baselines.
arXiv Detail & Related papers (2024-10-17T11:43:16Z) - Zero-to-Strong Generalization: Eliciting Strong Capabilities of Large Language Models Iteratively without Gold Labels [75.77877889764073]
Large Language Models (LLMs) have demonstrated remarkable performance through supervised fine-tuning or in-context learning using gold labels.
This study explores whether solely utilizing unlabeled data can elicit strong model capabilities.
We propose a new paradigm termed zero-to-strong generalization.
arXiv Detail & Related papers (2024-09-19T02:59:44Z) - How Well Do Text Embedding Models Understand Syntax? [50.440590035493074]
The ability of text embedding models to generalize across a wide range of syntactic contexts remains under-explored.
Our findings reveal that existing text embedding models have not sufficiently addressed these syntactic understanding challenges.
We propose strategies to augment the generalization ability of text embedding models in diverse syntactic scenarios.
arXiv Detail & Related papers (2023-11-14T08:51:00Z) - SimOAP: Improve Coherence and Consistency in Persona-based Dialogue
Generation via Over-sampling and Post-evaluation [54.66399120084227]
Language models trained on large-scale corpora can generate remarkably fluent results in open-domain dialogue.
For the persona-based dialogue generation task, consistency and coherence are great challenges for language models.
A two-stage SimOAP strategy is proposed, i.e., over-sampling and post-evaluation.
arXiv Detail & Related papers (2023-05-18T17:23:00Z) - POUF: Prompt-oriented unsupervised fine-tuning for large pre-trained
models [62.23255433487586]
We propose an unsupervised fine-tuning framework to fine-tune the model or prompt on the unlabeled target data.
We demonstrate how to apply our method to both language-augmented vision and masked-language models by aligning the discrete distributions extracted from the prompts and target data.
arXiv Detail & Related papers (2023-04-29T22:05:22Z) - Multidimensional Perceptron for Efficient and Explainable Long Text
Classification [31.31206469613901]
We propose a simple but effective model, Segment-aWare multIdimensional PErceptron (SWIPE) to replace attention/RNNs in the framework.
SWIPE can effectively learn the label of the entire text with supervised training, while perceive the labels of the segments and estimate their contributions to the long-text labeling.
arXiv Detail & Related papers (2023-04-04T08:49:39Z) - Long Document Summarization with Top-down and Bottom-up Inference [113.29319668246407]
We propose a principled inference framework to improve summarization models on two aspects.
Our framework assumes a hierarchical latent structure of a document where the top-level captures the long range dependency.
We demonstrate the effectiveness of the proposed framework on a diverse set of summarization datasets.
arXiv Detail & Related papers (2022-03-15T01:24:51Z) - Controllable Abstractive Dialogue Summarization with Sketch Supervision [56.59357883827276]
Our model achieves state-of-the-art performance on the largest dialogue summarization corpus SAMSum, with as high as 50.79 in ROUGE-L score.
arXiv Detail & Related papers (2021-05-28T19:05:36Z) - GPT3Mix: Leveraging Large-scale Language Models for Text Augmentation [9.501648136713694]
Large-scale language models such as GPT-3 are excellent few-shot learners, allowing them to be controlled via natural text prompts.
This paper proposes a novel data augmentation technique that leverages large-scale language models to generate realistic text samples.
arXiv Detail & Related papers (2021-04-18T11:39:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.