Toward Improving Health Literacy in Patient Education Materials with
Neural Machine Translation Models
- URL: http://arxiv.org/abs/2209.06723v1
- Date: Wed, 14 Sep 2022 15:30:54 GMT
- Title: Toward Improving Health Literacy in Patient Education Materials with
Neural Machine Translation Models
- Authors: David Oniani, Sreekanth Sreekumar, Renuk DeAlmeida, Dinuk DeAlmeida,
Vivian Hui, Young Ji Lee, Yiye Zhang, Leming Zhou, Yanshan Wang
- Abstract summary: People with low health literacy usually have trouble understanding health information, following post-visit instructions, and using prescriptions.
We propose to leverage natural language processing techniques to improve health literacy in patient education materials by automatically translating illiterate languages in a given sentence.
We trained and tested the state-of-the-art neural machine translation (NMT) models on a silver standard training dataset and a gold standard testing dataset.
- Score: 4.402142901902557
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Health literacy is the central focus of Healthy People 2030, the fifth
iteration of the U.S. national goals and objectives. People with low health
literacy usually have trouble understanding health information, following
post-visit instructions, and using prescriptions, which results in worse health
outcomes and serious health disparities. In this study, we propose to leverage
natural language processing techniques to improve health literacy in patient
education materials by automatically translating illiterate languages in a
given sentence. We scraped patient education materials from four online health
information websites: MedlinePlus.gov, Drugs.com, Mayoclinic.org and
Reddit.com. We trained and tested the state-of-the-art neural machine
translation (NMT) models on a silver standard training dataset and a gold
standard testing dataset, respectively. The experimental results showed that
the Bidirectional Long Short-Term Memory (BiLSTM) NMT model outperformed
Bidirectional Encoder Representations from Transformers (BERT)-based NMT
models. We also verified the effectiveness of NMT models in translating health
illiterate languages by comparing the ratio of health illiterate language in
the sentence. The proposed NMT models were able to identify the correct
complicated words and simplify into layman language while at the same time the
models suffer from sentence completeness, fluency, readability, and have
difficulty in translating certain medical terms.
Related papers
- On Creating an English-Thai Code-switched Machine Translation in Medical Domain [2.0737832185611524]
Machine translation (MT) in the medical domain plays a pivotal role in enhancing healthcare quality and disseminating medical knowledge.
Despite advancements in English-Thai MT technology, common MT approaches often underperform in the medical field due to their inability to precisely translate medical terminologies.
Our research prioritizes not merely improving translation accuracy but also maintaining medical terminology in English.
arXiv Detail & Related papers (2024-10-21T17:25:32Z) - Adapting Mental Health Prediction Tasks for Cross-lingual Learning via Meta-Training and In-context Learning with Large Language Model [3.3590922002216193]
We use model-agnostic meta-learning and leveraging large language models (LLMs) to address this gap.
We first apply a meta-learning model with self-supervision, which results in improved model initialisation for rapid adaptation and cross-lingual transfer.
In parallel, we use LLMs' in-context learning capabilities to assess their performance accuracy across the Swahili mental health prediction tasks.
arXiv Detail & Related papers (2024-04-13T17:11:35Z) - Importance-Aware Data Augmentation for Document-Level Neural Machine
Translation [51.74178767827934]
Document-level neural machine translation (DocNMT) aims to generate translations that are both coherent and cohesive.
Due to its longer input length and limited availability of training data, DocNMT often faces the challenge of data sparsity.
We propose a novel Importance-Aware Data Augmentation (IADA) algorithm for DocNMT that augments the training data based on token importance information estimated by the norm of hidden states and training gradients.
arXiv Detail & Related papers (2024-01-27T09:27:47Z) - Qilin-Med-VL: Towards Chinese Large Vision-Language Model for General
Healthcare [14.646414629627001]
This study introduces Qilin-Med-VL, the first Chinese large vision-language model designed to integrate the analysis of textual and visual data.
We also release ChiMed-VL, a dataset consisting of more than 1M image-text pairs.
arXiv Detail & Related papers (2023-10-27T08:05:21Z) - Code-Switching with Word Senses for Pretraining in Neural Machine
Translation [107.23743153715799]
We introduce Word Sense Pretraining for Neural Machine Translation (WSP-NMT)
WSP-NMT is an end-to-end approach for pretraining multilingual NMT models leveraging word sense-specific information from Knowledge Bases.
Our experiments show significant improvements in overall translation quality.
arXiv Detail & Related papers (2023-10-21T16:13:01Z) - CBLUE: A Chinese Biomedical Language Understanding Evaluation Benchmark [51.38557174322772]
We present the first Chinese Biomedical Language Understanding Evaluation benchmark.
It is a collection of natural language understanding tasks including named entity recognition, information extraction, clinical diagnosis normalization, single-sentence/sentence-pair classification.
We report empirical results with the current 11 pre-trained Chinese models, and experimental results show that state-of-the-art neural models perform by far worse than the human ceiling.
arXiv Detail & Related papers (2021-06-15T12:25:30Z) - Improving the Lexical Ability of Pretrained Language Models for
Unsupervised Neural Machine Translation [127.81351683335143]
Cross-lingual pretraining requires models to align the lexical- and high-level representations of the two languages.
Previous research has shown that this is because the representations are not sufficiently aligned.
In this paper, we enhance the bilingual masked language model pretraining with lexical-level information by using type-level cross-lingual subword embeddings.
arXiv Detail & Related papers (2021-03-18T21:17:58Z) - Leveraging Monolingual Data with Self-Supervision for Multilingual
Neural Machine Translation [54.52971020087777]
Using monolingual data significantly boosts the translation quality of low-resource languages in multilingual models.
Self-supervision improves zero-shot translation quality in multilingual models.
We get up to 33 BLEU on ro-en translation without any parallel data or back-translation.
arXiv Detail & Related papers (2020-05-11T00:20:33Z) - It's Easier to Translate out of English than into it: Measuring Neural
Translation Difficulty by Cross-Mutual Information [90.35685796083563]
Cross-mutual information (XMI) is an asymmetric information-theoretic metric of machine translation difficulty.
XMI exploits the probabilistic nature of most neural machine translation models.
We present the first systematic and controlled study of cross-lingual translation difficulties using modern neural translation systems.
arXiv Detail & Related papers (2020-05-05T17:38:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.