Extracting Psychological Indicators Using Question Answering
- URL: http://arxiv.org/abs/2305.14891v1
- Date: Wed, 24 May 2023 08:41:23 GMT
- Title: Extracting Psychological Indicators Using Question Answering
- Authors: Luka Pavlovi\'c
- Abstract summary: We propose a method for extracting text spans that may indicate one of the BIG5 psychological traits using a question-answering task with examples that have no answer for the asked question.
We utilized the RoBERTa model fine-tuned on SQuAD 2.0 dataset.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we propose a method for extracting text spans that may indicate
one of the BIG5 psychological traits using a question-answering task with
examples that have no answer for the asked question. We utilized the RoBERTa
model fine-tuned on SQuAD 2.0 dataset. The model was further fine-tuned
utilizing comments from Reddit. We examined the effect of the percentage of
examples with no answer in the training dataset on the overall performance. The
results obtained in this study are in line with the SQuAD 2.0 benchmark and
present a good baseline for further research.
Related papers
- Answerability in Retrieval-Augmented Open-Domain Question Answering [17.177439885871788]
Open-Domain Question Answering (ODQA) retrieval systems can exhibit sub-optimal behavior, providing text excerpts with varying degrees of irrelevance.
Previous attempts to address this gap have relied on a simplistic approach of pairing questions with random text excerpts.
arXiv Detail & Related papers (2024-03-03T09:55:35Z) - Open vs Closed-ended questions in attitudinal surveys -- comparing,
combining, and interpreting using natural language processing [3.867363075280544]
Topic Modeling could significantly reduce the time to extract information from open-ended responses.
Our research uses Topic Modeling to extract information from open-ended questions and compare its performance with closed-ended responses.
arXiv Detail & Related papers (2022-05-03T06:01:03Z) - Question-Answer Sentence Graph for Joint Modeling Answer Selection [122.29142965960138]
We train and integrate state-of-the-art (SOTA) models for computing scores between question-question, question-answer, and answer-answer pairs.
Online inference is then performed to solve the AS2 task on unseen queries.
arXiv Detail & Related papers (2022-02-16T05:59:53Z) - Reference-based Weak Supervision for Answer Sentence Selection using Web
Data [87.18646699292293]
We introduce Reference-based Weak Supervision (RWS), a fully automatic large-scale data pipeline.
RWS harvests high-quality weakly-supervised answers from abundant Web data.
Our experiments indicate that the produced data consistently bolsters TANDA.
arXiv Detail & Related papers (2021-04-18T19:41:17Z) - Learning Compositional Representation for Few-shot Visual Question
Answering [93.4061107793983]
Current methods of Visual Question Answering perform well on the answers with an amount of training data but have limited accuracy on the novel ones with few examples.
We propose to extract the attributes from the answers with enough data, which are later composed to constrain the learning of the few-shot ones.
Experimental results on the VQA v2.0 validation dataset demonstrate the effectiveness of our proposed attribute network.
arXiv Detail & Related papers (2021-02-21T10:16:24Z) - Overcoming Language Priors with Self-supervised Learning for Visual
Question Answering [62.88124382512111]
Most Visual Question Answering (VQA) models suffer from the language prior problem.
We introduce a self-supervised learning framework to solve this problem.
Our method can significantly outperform the state-of-the-art.
arXiv Detail & Related papers (2020-12-17T12:30:12Z) - Harvesting and Refining Question-Answer Pairs for Unsupervised QA [95.9105154311491]
We introduce two approaches to improve unsupervised Question Answering (QA)
First, we harvest lexically and syntactically divergent questions from Wikipedia to automatically construct a corpus of question-answer pairs (named as RefQA)
Second, we take advantage of the QA model to extract more appropriate answers, which iteratively refines data over RefQA.
arXiv Detail & Related papers (2020-05-06T15:56:06Z) - Improving Multi-Turn Response Selection Models with Complementary
Last-Utterance Selection by Instance Weighting [84.9716460244444]
We consider utilizing the underlying correlation in the data resource itself to derive different kinds of supervision signals.
We conduct extensive experiments in two public datasets and obtain significant improvement in both datasets.
arXiv Detail & Related papers (2020-02-18T06:29:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.