Health Disparities through Generative AI Models: A Comparison Study
Using A Domain Specific large language model
- URL: http://arxiv.org/abs/2310.18355v1
- Date: Mon, 23 Oct 2023 21:24:05 GMT
- Title: Health Disparities through Generative AI Models: A Comparison Study
Using A Domain Specific large language model
- Authors: Yohn Jairo Parra Bautista, Vinicious Lima, Carlos Theran, Richard Alo
- Abstract summary: An artificial intelligence program called large language models (LLMs) can understand and generate human language.
We introduce the comparative investigation of domain-specific large language models such as SciBERT.
We believe clinicians can use generative AI to create a draft response when communicating asynchronously with patients.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Health disparities are differences in health outcomes and access to
healthcare between different groups, including racial and ethnic minorities,
low-income people, and rural residents. An artificial intelligence (AI) program
called large language models (LLMs) can understand and generate human language,
improving health communication and reducing health disparities. There are many
challenges in using LLMs in human-doctor interaction, including the need for
diverse and representative data, privacy concerns, and collaboration between
healthcare providers and technology experts. We introduce the comparative
investigation of domain-specific large language models such as SciBERT with a
multi-purpose LLMs BERT. We used cosine similarity to analyze text queries
about health disparities in exam rooms when factors such as race are used
alone. Using text queries, SciBERT fails when it doesn't differentiate between
queries text: "race" alone and "perpetuates health disparities." We believe
clinicians can use generative AI to create a draft response when communicating
asynchronously with patients. However, careful attention must be paid to ensure
they are developed and implemented ethically and equitably.
Related papers
- Evaluating and Mitigating Bias in AI-Based Medical Text Generation [35.24191727599811]
AI systems may reflect and amplify human bias, and reduce the quality of their performance in historically under-served populations.
In this study, we investigate the fairness problem in text generation within the medical field.
We propose an algorithm that selectively optimize those underperformed groups to reduce bias.
arXiv Detail & Related papers (2025-04-24T06:10:40Z) - BRIGHTER: BRIdging the Gap in Human-Annotated Textual Emotion Recognition Datasets for 28 Languages [93.92804151830744]
We present BRIGHTER -- a collection of multi-labeled datasets in 28 different languages.
We describe the data collection and annotation processes and the challenges of building these datasets.
We show that BRIGHTER datasets are a step towards bridging the gap in text-based emotion recognition.
arXiv Detail & Related papers (2025-02-17T15:39:50Z) - Negation Blindness in Large Language Models: Unveiling the NO Syndrome in Image Generation [63.064204206220936]
Foundational Large Language Models (LLMs) have changed the way we perceive technology.
They have been shown to excel in tasks ranging from poem writing to coding to essay generation and puzzle solving.
With the incorporation of image generation capability, they have become more comprehensive and versatile AI tools.
Currently identified flaws include hallucination, biases, and bypassing restricted commands to generate harmful content.
arXiv Detail & Related papers (2024-08-27T14:40:16Z) - Rel-A.I.: An Interaction-Centered Approach To Measuring Human-LM Reliance [73.19687314438133]
We study how reliance is affected by contextual features of an interaction.
We find that contextual characteristics significantly affect human reliance behavior.
Our results show that calibration and language quality alone are insufficient in evaluating the risks of human-LM interactions.
arXiv Detail & Related papers (2024-07-10T18:00:05Z) - Language Model Alignment in Multilingual Trolley Problems [138.5684081822807]
Building on the Moral Machine experiment, we develop a cross-lingual corpus of moral dilemma vignettes in over 100 languages called MultiTP.
Our analysis explores the alignment of 19 different LLMs with human judgments, capturing preferences across six moral dimensions.
We discover significant variance in alignment across languages, challenging the assumption of uniform moral reasoning in AI systems.
arXiv Detail & Related papers (2024-07-02T14:02:53Z) - RFBES at SemEval-2024 Task 8: Investigating Syntactic and Semantic
Features for Distinguishing AI-Generated and Human-Written Texts [0.8437187555622164]
This article investigates the problem of AI-generated text detection from two different aspects: semantics and syntax.
We present an AI model that can distinguish AI-generated texts from human-written ones with high accuracy on both multilingual and monolingual tasks.
arXiv Detail & Related papers (2024-02-19T00:40:17Z) - Divergences between Language Models and Human Brains [59.100552839650774]
We systematically explore the divergences between human and machine language processing.
We identify two domains that LMs do not capture well: social/emotional intelligence and physical commonsense.
Our results show that fine-tuning LMs on these domains can improve their alignment with human brain responses.
arXiv Detail & Related papers (2023-11-15T19:02:40Z) - LaTeX: Language Pattern-aware Triggering Event Detection for Adverse
Experience during Pandemics [10.292364075312667]
The COVID-19 pandemic has accentuated socioeconomic disparities across various racial ethnic groups in the United States.
This paper explores the role of social media in both addressing scarcity and challenges.
We analyze language patterns related to four types of adverse experiences.
arXiv Detail & Related papers (2023-10-05T23:09:31Z) - Unbiased Pain Assessment through Wearables and EHR Data: Multi-attribute
Fairness Loss-based CNN Approach [3.799109312082668]
We propose a Multi-attribute Fairness Loss (MAFL) based CNN model to account for any sensitive attributes included in the data.
We compare the proposed model with well-known existing mitigation procedures, and studies reveal that the implemented model performs favorably in contrast to state-of-the-art methods.
arXiv Detail & Related papers (2023-07-03T09:21:36Z) - Is This Abstract Generated by AI? A Research for the Gap between
AI-generated Scientific Text and Human-written Scientific Text [13.438933219811188]
We investigate the gap between scientific content generated by AI and written by humans.
We find that there exists a writing style'' gap between AI-generated scientific text and human-written scientific text.
arXiv Detail & Related papers (2023-01-24T04:23:20Z) - CBLUE: A Chinese Biomedical Language Understanding Evaluation Benchmark [51.38557174322772]
We present the first Chinese Biomedical Language Understanding Evaluation benchmark.
It is a collection of natural language understanding tasks including named entity recognition, information extraction, clinical diagnosis normalization, single-sentence/sentence-pair classification.
We report empirical results with the current 11 pre-trained Chinese models, and experimental results show that state-of-the-art neural models perform by far worse than the human ceiling.
arXiv Detail & Related papers (2021-06-15T12:25:30Z) - Crossing the Conversational Chasm: A Primer on Multilingual
Task-Oriented Dialogue Systems [51.328224222640614]
Current state-of-the-art ToD models based on large pretrained neural language models are data hungry.
Data acquisition for ToD use cases is expensive and tedious.
arXiv Detail & Related papers (2021-04-17T15:19:56Z) - Towards Fairness in Classifying Medical Conversations into SOAP Sections [2.1485350418225244]
We identify and understand disparities in a model that classifies doctor-patient conversations into sections of a medical SOAP note.
A deeper analysis of the language in these conversations suggests these differences are related to and often attributable to the type of medical appointment.
Our findings stress the importance of understanding the disparities that may exist in the data itself and how that affects a model's ability to equally distribute benefits.
arXiv Detail & Related papers (2020-12-02T14:55:22Z) - Assessing the Severity of Health States based on Social Media Posts [62.52087340582502]
We propose a multiview learning framework that models both the textual content as well as contextual-information to assess the severity of the user's health state.
The diverse NLU views demonstrate its effectiveness on both the tasks and as well as on the individual disease to assess a user's health.
arXiv Detail & Related papers (2020-09-21T03:45:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.