Validation and Transparency in AI systems for pharmacovigilance: a case
study applied to the medical literature monitoring of adverse events
- URL: http://arxiv.org/abs/2201.00692v1
- Date: Tue, 21 Dec 2021 21:02:24 GMT
- Title: Validation and Transparency in AI systems for pharmacovigilance: a case
study applied to the medical literature monitoring of adverse events
- Authors: Bruno Ohana, Jack Sullivan and Nicole Baker
- Abstract summary: We present a case study on how to operationalize existing guidance for validated AI systems in pharmacovigilance.
We describe an AI system designed with the goal of reducing effort to mitigate activities built in close collaboration with subject matter experts.
- Score: 0.483420384410068
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Recent advances in artificial intelligence applied to biomedical text are
opening exciting opportunities for improving pharmacovigilance activities
currently burdened by the ever growing volumes of real world data. To fully
realize these opportunities, existing regulatory guidance and industry best
practices should be taken into consideration in order to increase the overall
trustworthiness of the system and enable broader adoption. In this paper we
present a case study on how to operationalize existing guidance for validated
AI systems in pharmacovigilance focusing on the specific task of medical
literature monitoring (MLM) of adverse events from the scientific literature.
We describe an AI system designed with the goal of reducing effort in MLM
activities built in close collaboration with subject matter experts and
considering guidance for validated systems in pharmacovigilance and AI
transparency. In particular we make use of public disclosures as a useful risk
control measure to mitigate system misuse and earn user trust. In addition we
present experimental results showing the system can significantly remove
screening effort while maintaining high levels of recall (filtering 55% of
irrelevant articles on average, for a target recall of 0.99 on suspected
adverse articles) and provide a robust method for tuning the desired recall to
suit a particular risk profile.
Related papers
- InferAct: Inferring Safe Actions for LLM-Based Agents Through Preemptive Evaluation and Human Feedback [70.54226917774933]
This paper introduces InferAct, a novel approach to proactively detect potential errors before risky actions are executed.
InferAct acts as a human proxy, detecting unsafe actions and alerting users for intervention.
Experiments on three widely-used tasks demonstrate the effectiveness of InferAct.
arXiv Detail & Related papers (2024-07-16T15:24:44Z) - The Need for Guardrails with Large Language Models in Medical Safety-Critical Settings: An Artificial Intelligence Application in the Pharmacovigilance Ecosystem [0.6965384453064829]
Large language models (LLMs) are useful tools with the capacity for performing specific types of knowledge work at an effective scale.
However, deployments in high-risk and safety-critical domains pose unique challenges, notably the issue of hallucinations.
This is particularly concerning in settings such as drug safety, where inaccuracies could lead to patient harm.
We have developed and demonstrated a proof of concept suite of guardrails specifically designed to mitigate certain types of hallucinations and errors for drug safety.
arXiv Detail & Related papers (2024-07-01T19:52:41Z) - Advancing Healthcare Automation: Multi-Agent System for Medical Necessity Justification [0.0]
This paper explores the application of Multi-Agent System (MAS) that utilize specialized LLM agents to automate Prior Authorization task.
We demonstrate that GPT-4 checklist achieves an accuracy of 86.2% in predicting item-level judgments with evidence, and 95.6% in determining overall checklist judgment.
arXiv Detail & Related papers (2024-04-27T18:40:05Z) - Explainable Machine Learning-Based Security and Privacy Protection Framework for Internet of Medical Things Systems [1.8434042562191815]
The Internet of Medical Things (IoMT) transcends traditional medical boundaries, enabling a transition from reactive treatment to proactive prevention.
Its benefits are countered by significant security challenges that endanger the lives of its users due to the sensitivity and value of the processed data.
A new framework for Intrusion Detection Systems (IDS) is introduced, leveraging Artificial Neural Networks (ANN) for intrusion detection while utilizing Federated Learning (FL) for privacy preservation.
arXiv Detail & Related papers (2024-03-14T11:57:26Z) - Retrieval Augmented Thought Process for Private Data Handling in Healthcare [53.89406286212502]
We introduce the Retrieval-Augmented Thought Process (RATP)
RATP formulates the thought generation of Large Language Models (LLMs)
On a private dataset of electronic medical records, RATP achieves 35% additional accuracy compared to in-context retrieval-augmented generation for the question-answering task.
arXiv Detail & Related papers (2024-02-12T17:17:50Z) - Prioritizing Safeguarding Over Autonomy: Risks of LLM Agents for Science [65.77763092833348]
Intelligent agents powered by large language models (LLMs) have demonstrated substantial promise in autonomously conducting experiments and facilitating scientific discoveries across various disciplines.
While their capabilities are promising, these agents also introduce novel vulnerabilities that demand careful consideration for safety.
This paper conducts a thorough examination of vulnerabilities in LLM-based agents within scientific domains, shedding light on potential risks associated with their misuse and emphasizing the need for safety measures.
arXiv Detail & Related papers (2024-02-06T18:54:07Z) - RAISE -- Radiology AI Safety, an End-to-end lifecycle approach [5.829180249228172]
The integration of AI into radiology introduces opportunities for improved clinical care provision and efficiency.
The focus should be on ensuring models meet the highest standards of safety, effectiveness and efficacy.
The roadmap presented herein aims to expedite the achievement of deployable, reliable, and safe AI in radiology.
arXiv Detail & Related papers (2023-11-24T15:59:14Z) - U-PASS: an Uncertainty-guided deep learning Pipeline for Automated Sleep
Staging [61.6346401960268]
We propose a machine learning pipeline called U-PASS tailored for clinical applications that incorporates uncertainty estimation at every stage of the process.
We apply our uncertainty-guided deep learning pipeline to the challenging problem of sleep staging and demonstrate that it systematically improves performance at every stage.
arXiv Detail & Related papers (2023-06-07T08:27:36Z) - Self-Verification Improves Few-Shot Clinical Information Extraction [73.6905567014859]
Large language models (LLMs) have shown the potential to accelerate clinical curation via few-shot in-context learning.
They still struggle with issues regarding accuracy and interpretability, especially in mission-critical domains such as health.
Here, we explore a general mitigation framework using self-verification, which leverages the LLM to provide provenance for its own extraction and check its own outputs.
arXiv Detail & Related papers (2023-05-30T22:05:11Z) - Overcoming Failures of Imagination in AI Infused System Development and
Deployment [71.9309995623067]
NeurIPS 2020 requested that research paper submissions include impact statements on "potential nefarious uses and the consequences of failure"
We argue that frameworks of harms must be context-aware and consider a wider range of potential stakeholders, system affordances, as well as viable proxies for assessing harms in the widest sense.
arXiv Detail & Related papers (2020-11-26T18:09:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.