Assessing the Human-Likeness of LLM-Driven Digital Twins in Simulating Health Care System Trust
- URL: http://arxiv.org/abs/2512.08939v1
- Date: Mon, 27 Oct 2025 02:56:22 GMT
- Title: Assessing the Human-Likeness of LLM-Driven Digital Twins in Simulating Health Care System Trust
- Authors: Yuzhou Wu, Mingyang Wu, Di Liu, Rong Yin, Kang Li,
- Abstract summary: Large Language Model (LLM)-driven Human Digital Twins are showing great potential in healthcare system research.<n>However, its actual simulation ability for complex human psychological traits, such as distrust in the healthcare system remains unclear.<n>This study suggests that the current LLM-driven Digital Twins have limitations in modeling complex human attitudes.
- Score: 11.272529608962996
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Serving as an emerging and powerful tool, Large Language Model (LLM)-driven Human Digital Twins are showing great potential in healthcare system research. However, its actual simulation ability for complex human psychological traits, such as distrust in the healthcare system, remains unclear. This research gap particularly impacts health professionals' trust and usage of LLM-based Artificial Intelligence (AI) systems in assisting their routine work. In this study, based on the Twin-2K-500 dataset, we systematically evaluated the simulation results of the LLM-driven human digital twin using the Health Care System Distrust Scale (HCSDS) with an established human-subject sample, analyzing item-level distributions, summary statistics, and demographic subgroup patterns. Results showed that the simulated responses by the digital twin were significantly more centralized with lower variance and had fewer selections of extreme options (all p<0.001). While the digital twin broadly reproduces human results in major demographic patterns, such as age and gender, it exhibits relatively low sensitivity in capturing minor differences in education levels. The LLM-based digital twin simulation has the potential to simulate population trends, but it also presents challenges in making detailed, specific distinctions in subgroups of human beings. This study suggests that the current LLM-driven Digital Twins have limitations in modeling complex human attitudes, which require careful calibration and validation before applying them in inferential analyses or policy simulations in health systems engineering. Future studies are necessary to examine the emotional reasoning mechanism of LLMs before their use, particularly for studies that involve simulations sensitive to social topics, such as human-automation trust.
Related papers
- Psychometric Comparability of LLM-Based Digital Twins [2.7740826124350355]
We benchmark digital twins against human gold standards across models, tasks and testing how person-specific inputs shape performance.<n>Across studies, digital twins achieved high population-level accuracy and strong within-participant profile correlations.<n>Digital twins under-reproduce biases, showing normative prediction, compressed variance and limited sensitivity to temporal information.
arXiv Detail & Related papers (2025-12-22T18:04:27Z) - Large language models replicate and predict human cooperation across experiments in game theory [0.8166364251367626]
How closely large language models mirror actual human decision-making remains poorly understood.<n>We develop a digital twin of game-theoretic experiments and introduce a systematic prompting and probing framework for machine-behavioral evaluation.<n>We find that Llama reproduces human cooperation patterns with high fidelity, capturing human deviations from rational choice theory.
arXiv Detail & Related papers (2025-11-06T16:21:27Z) - Leveraging LLM-based agents for social science research: insights from citation network simulations [132.4334196445918]
We introduce the CiteAgent framework, designed to generate citation networks based on human-behavior simulation.<n>CiteAgent captures predominant phenomena in real-world citation networks, including power-law distribution, citational distortion, and shrinking diameter.<n>We establish two LLM-based research paradigms in social science, allowing us to validate and challenge existing theories.
arXiv Detail & Related papers (2025-11-05T08:47:04Z) - Scaling Law in LLM Simulated Personality: More Detailed and Realistic Persona Profile Is All You Need [17.298070053011802]
This research focuses on using large language models (LLMs) to simulate social experiments, exploring their ability to emulate human personality in virtual persona role-playing.<n>The research develops an end-to-end evaluation framework, including individual-level analysis of stability and identifiability.
arXiv Detail & Related papers (2025-10-10T05:52:07Z) - Population-Aligned Persona Generation for LLM-based Social Simulation [58.84363795421489]
We propose a systematic framework for synthesizing high-quality, population-aligned persona sets for social simulation.<n>Our approach begins by leveraging large language models to generate narrative personas from long-term social media data.<n>To address the needs of specific simulation contexts, we introduce a task-specific module that adapts the globally aligned persona set to targeted subpopulations.
arXiv Detail & Related papers (2025-09-12T10:43:47Z) - Organ-Agents: Virtual Human Physiology Simulator via LLMs [66.40796430669158]
Organ-Agents is a multi-agent framework that simulates human physiology via LLM-driven agents.<n>We curated data from 7,134 sepsis patients and 7,895 controls, generating high-resolution trajectories across 9 systems and 125 variables.<n>Organ-Agents achieved high simulation accuracy on 4,509 held-out patients, with per-system MSEs 0.16 and robustness across SOFA-based severity strata.
arXiv Detail & Related papers (2025-08-20T01:58:45Z) - Enhancing Patient-Centric Communication: Leveraging LLMs to Simulate Patient Perspectives [19.462374723301792]
Large Language Models (LLMs) have demonstrated impressive capabilities in role-playing scenarios.<n>By mimicking human behavior, LLMs can anticipate responses based on concrete demographic or professional profiles.<n>We evaluate the effectiveness of LLMs in simulating individuals with diverse backgrounds and analyze the consistency of these simulated behaviors.
arXiv Detail & Related papers (2025-01-12T22:49:32Z) - Med-Real2Sim: Non-Invasive Medical Digital Twins using Physics-Informed Self-Supervised Learning [15.106435744696013]
A digital twin is a virtual replica of a real-world physical phenomena that uses mathematical modeling to characterize and simulate its defining features.
We propose a method to identify digital twin model parameters using only noninvasive patient health data.
arXiv Detail & Related papers (2024-02-29T23:04:42Z) - Redefining Digital Health Interfaces with Large Language Models [69.02059202720073]
Large Language Models (LLMs) have emerged as general-purpose models with the ability to process complex information.
We show how LLMs can provide a novel interface between clinicians and digital technologies.
We develop a new prognostic tool using automated machine learning.
arXiv Detail & Related papers (2023-10-05T14:18:40Z) - Reducing a complex two-sided smartwatch examination for Parkinson's
Disease to an efficient one-sided examination preserving machine learning
accuracy [63.20765930558542]
We have recorded participants performing technology-based assessments in a prospective study to research Parkinson's Disease (PD)
This study provided the largest PD sample size of two-hand synchronous smartwatch measurements.
arXiv Detail & Related papers (2022-05-11T09:12:59Z) - MIMO: Mutual Integration of Patient Journey and Medical Ontology for
Healthcare Representation Learning [49.57261599776167]
We propose an end-to-end robust Transformer-based solution, Mutual Integration of patient journey and Medical Ontology (MIMO) for healthcare representation learning and predictive analytics.
arXiv Detail & Related papers (2021-07-20T07:04:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.