An Enhanced Text Classification to Explore Health based Indian
Government Policy Tweets
- URL: http://arxiv.org/abs/2007.06511v2
- Date: Tue, 18 Aug 2020 12:37:18 GMT
- Title: An Enhanced Text Classification to Explore Health based Indian
Government Policy Tweets
- Authors: Aarzoo Dhiman and Durga Toshniwal
- Abstract summary: We propose an improved text classification framework that classifies the Twitter data of different health-based government schemes.
To handle this, we propose a novel GloVe word embeddings and class-specific sentiments based text augmentation approach (named Mod-EDA)
- Score: 2.2082422928825136
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Government-sponsored policy-making and scheme generations is one of the means
of protecting and promoting the social, economic, and personal development of
the citizens. The evaluation of effectiveness of these schemes done by
government only provide the statistical information in terms of facts and
figures which do not include the in-depth knowledge of public perceptions,
experiences and views on the topic. In this research work, we propose an
improved text classification framework that classifies the Twitter data of
different health-based government schemes. The proposed framework leverages the
language representation models (LR models) BERT, ELMO, and USE. However, these
LR models have less real-time applicability due to the scarcity of the ample
annotated data. To handle this, we propose a novel GloVe word embeddings and
class-specific sentiments based text augmentation approach (named Mod-EDA)
which boosts the performance of text classification task by increasing the size
of labeled data. Furthermore, the trained model is leveraged to identify the
level of engagement of citizens towards these policies in different communities
such as middle-income and low-income groups.
Related papers
- Understanding Public Perception of Crime in Bangladesh: A Transformer-Based Approach with Explainability [0.0]
This study investigates the evolving public perception of crime-related news by classifying user-generated comments into three categories: positive, negative, and neutral.<n>We propose a transformer-based model utilizing the XLM-RoBERTa Base architecture, which achieves a classification accuracy of 97%.
arXiv Detail & Related papers (2025-07-28T18:00:25Z) - LLMTaxo: Leveraging Large Language Models for Constructing Taxonomy of Factual Claims from Social Media [4.808928577524327]
This paper introduces LLMTaxo, a novel framework for the automated construction of taxonomy of factual claims from social media.
We implement this framework with different models across three distinct datasets and introduce specially designed taxonomy evaluation metrics.
With the evaluations from both human evaluators and GPT-4, the results indicate that LLMTaxo effectively categorizes factual claims from social media.
arXiv Detail & Related papers (2025-04-11T18:43:16Z) - Political Actor Agent: Simulating Legislative System for Roll Call Votes Prediction with Large Language Models [9.0463587094323]
Political Actor Agent (PAA) is a novel framework that utilizes Large Language Models to overcome limitations.
By employing role-playing architectures and simulating legislative system, PAA provides a scalable and interpretable paradigm for predicting roll-call votes.
We conducted comprehensive experiments using voting records from the 117-118th U.S. House of Representatives, validating the superior performance and interpretability of PAA.
arXiv Detail & Related papers (2024-12-10T03:06:28Z) - Persuasion with Large Language Models: a Survey [49.86930318312291]
Large Language Models (LLMs) have created new disruptive possibilities for persuasive communication.
In areas such as politics, marketing, public health, e-commerce, and charitable giving, such LLM Systems have already achieved human-level or even super-human persuasiveness.
Our survey suggests that the current and future potential of LLM-based persuasion poses profound ethical and societal risks.
arXiv Detail & Related papers (2024-11-11T10:05:52Z) - Evaluating and Adapting Large Language Models to Represent Folktales in Low-Resource Languages [3.1125246891252125]
Large language models (LLMs) claim to be able to represent low-resource langauges such as Irish and Gaelic.
We present two classification tasks to explore how useful these representations are, and three adaptations to improve the performance of these models.
arXiv Detail & Related papers (2024-11-08T14:26:56Z) - Are Large Language Models Good Classifiers? A Study on Edit Intent Classification in Scientific Document Revisions [62.12545440385489]
Large language models (LLMs) have brought substantial advancements in text generation, but their potential for enhancing classification tasks remains underexplored.
We propose a framework for thoroughly investigating fine-tuning LLMs for classification, including both generation- and encoding-based approaches.
We instantiate this framework in edit intent classification (EIC), a challenging and underexplored classification task.
arXiv Detail & Related papers (2024-10-02T20:48:28Z) - Entity Extraction from High-Level Corruption Schemes via Large Language Models [4.820586736502356]
This article proposes a new micro-benchmark dataset for algorithms and models that identify individuals and organizations in news articles.
Experimental efforts are also reported, using this dataset, to identify individuals and organizations in financial-crime-related articles.
arXiv Detail & Related papers (2024-09-05T10:27:32Z) - A Survey on Natural Language Counterfactual Generation [7.022371235308068]
Natural language counterfactual generation aims to minimally modify a given text such that the modified text will be classified into a different class.
We propose a new taxonomy that systematically categorizes the generation methods into four groups and summarizes the metrics for evaluating the generation quality.
arXiv Detail & Related papers (2024-07-04T15:13:59Z) - Bias and Fairness in Large Language Models: A Survey [73.87651986156006]
We present a comprehensive survey of bias evaluation and mitigation techniques for large language models (LLMs)
We first consolidate, formalize, and expand notions of social bias and fairness in natural language processing.
We then unify the literature by proposing three intuitive, two for bias evaluation, and one for mitigation.
arXiv Detail & Related papers (2023-09-02T00:32:55Z) - DeAR: Debiasing Vision-Language Models with Additive Residuals [5.672132510411465]
Large pre-trained vision-language models (VLMs) provide rich, adaptable image and text representations.
These models suffer from societal biases owing to the skewed distribution of various identity groups in the training data.
We present DeAR, a novel debiasing method that learns additive residual image representations to offset the original representations.
arXiv Detail & Related papers (2023-03-18T14:57:43Z) - Toward Fairness in Text Generation via Mutual Information Minimization
based on Importance Sampling [23.317845744611375]
We propose to minimize the mutual information between the semantics in the generated text sentences and their demographic polarity.
In this way, the mentioning of a demographic group is encouraged to be independent from how it is described in the generated text.
We also propose a distillation mechanism that preserves the language modeling ability of the PLMs after debiasing.
arXiv Detail & Related papers (2023-02-25T18:29:02Z) - Dependency Induction Through the Lens of Visual Perception [81.91502968815746]
We propose an unsupervised grammar induction model that leverages word concreteness and a structural vision-based to jointly learn constituency-structure and dependency-structure grammars.
Our experiments show that the proposed extension outperforms the current state-of-the-art visually grounded models in constituency parsing even with a smaller grammar size.
arXiv Detail & Related papers (2021-09-20T18:40:37Z) - Towards Understanding and Mitigating Social Biases in Language Models [107.82654101403264]
Large-scale pretrained language models (LMs) can be potentially dangerous in manifesting undesirable representational biases.
We propose steps towards mitigating social biases during text generation.
Our empirical results and human evaluation demonstrate effectiveness in mitigating bias while retaining crucial contextual information.
arXiv Detail & Related papers (2021-06-24T17:52:43Z) - Towards Debiasing Sentence Representations [109.70181221796469]
We show that Sent-Debias is effective in removing biases, and at the same time, preserves performance on sentence-level downstream tasks.
We hope that our work will inspire future research on characterizing and removing social biases from widely adopted sentence representations for fairer NLP.
arXiv Detail & Related papers (2020-07-16T04:22:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.