Operationalizing Contextual Integrity in Privacy-Conscious Assistants
- URL: http://arxiv.org/abs/2408.02373v2
- Date: Fri, 13 Sep 2024 13:09:41 GMT
- Title: Operationalizing Contextual Integrity in Privacy-Conscious Assistants
- Authors: Sahra Ghalebikesabi, Eugene Bagdasaryan, Ren Yi, Itay Yona, Ilia Shumailov, Aneesh Pappu, Chongyang Shi, Laura Weidinger, Robert Stanforth, Leonard Berrada, Pushmeet Kohli, Po-Sen Huang, Borja Balle,
- Abstract summary: We propose to operationalize contextual integrity (CI) to steer advanced AI assistants to behave in accordance with privacy expectations.
In particular, we design and evaluate a number of strategies to steer assistants' information-sharing actions to be CI compliant.
Our evaluation is based on a novel form filling benchmark composed of human annotations of common webform applications.
- Score: 34.70330533067581
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Advanced AI assistants combine frontier LLMs and tool access to autonomously perform complex tasks on behalf of users. While the helpfulness of such assistants can increase dramatically with access to user information including emails and documents, this raises privacy concerns about assistants sharing inappropriate information with third parties without user supervision. To steer information-sharing assistants to behave in accordance with privacy expectations, we propose to operationalize contextual integrity (CI), a framework that equates privacy with the appropriate flow of information in a given context. In particular, we design and evaluate a number of strategies to steer assistants' information-sharing actions to be CI compliant. Our evaluation is based on a novel form filling benchmark composed of human annotations of common webform applications, and it reveals that prompting frontier LLMs to perform CI-based reasoning yields strong results.
Related papers
- Navigating the Unknown: A Chat-Based Collaborative Interface for Personalized Exploratory Tasks [35.09558253658275]
This paper introduces the Collaborative Assistant for Personalized Exploration (CARE)
CARE is a system designed to enhance personalization in exploratory tasks by combining a multi-agent LLM framework with a structured user interface.
Our findings highlight CARE's potential to transform LLM-based systems from passive information retrievers to proactive partners in personalized problem-solving and exploration.
arXiv Detail & Related papers (2024-10-31T15:30:55Z) - Empowering Users in Digital Privacy Management through Interactive LLM-Based Agents [13.064730964328264]
This paper presents a novel application of large language models (LLMs) to enhance user comprehension of privacy policies through an interactive dialogue agent.
We demonstrate that LLMs significantly outperform traditional models in tasks like Data Practice Identification, Choice Identification, Policy Summarization, and Privacy Question Answering.
We introduce an innovative LLM-based agent that functions as an expert system for processing website privacy policies, guiding users through complex legal language without requiring them to pose specific questions.
arXiv Detail & Related papers (2024-10-15T02:16:59Z) - AI Delegates with a Dual Focus: Ensuring Privacy and Strategic Self-Disclosure [42.96087647326612]
We conduct a pilot study to investigate user preferences for AI delegates across various social relations and task scenarios.
We then propose a novel AI delegate system that enables privacy-conscious self-disclosure.
Our user study demonstrates that the proposed AI delegate strategically protects privacy, pioneering its use in diverse and dynamic social interactions.
arXiv Detail & Related papers (2024-09-26T08:45:15Z) - Constraining Participation: Affordances of Feedback Features in Interfaces to Large Language Models [49.74265453289855]
Large language models (LLMs) are now accessible to anyone with a computer, a web browser, and an internet connection via browser-based interfaces.
This paper examines the affordances of interactive feedback features in ChatGPT's interface, analysing how they shape user input and participation in iteration.
arXiv Detail & Related papers (2024-08-27T13:50:37Z) - Collection, usage and privacy of mobility data in the enterprise and public administrations [55.2480439325792]
Security measures such as anonymization are needed to protect individuals' privacy.
Within our study, we conducted expert interviews to gain insights into practices in the field.
We survey privacy-enhancing methods in use, which generally do not comply with state-of-the-art standards of differential privacy.
arXiv Detail & Related papers (2024-07-04T08:29:27Z) - The Ethics of Advanced AI Assistants [53.89899371095332]
This paper focuses on the opportunities and the ethical and societal risks posed by advanced AI assistants.
We define advanced AI assistants as artificial agents with natural language interfaces, whose function is to plan and execute sequences of actions on behalf of a user.
We consider the deployment of advanced assistants at a societal scale, focusing on cooperation, equity and access, misinformation, economic impact, the environment and how best to evaluate advanced AI assistants.
arXiv Detail & Related papers (2024-04-24T23:18:46Z) - Tell Me More! Towards Implicit User Intention Understanding of Language
Model Driven Agents [110.25679611755962]
Current language model-driven agents often lack mechanisms for effective user participation, which is crucial given the vagueness commonly found in user instructions.
We introduce Intention-in-Interaction (IN3), a novel benchmark designed to inspect users' implicit intentions through explicit queries.
We empirically train Mistral-Interact, a powerful model that proactively assesses task vagueness, inquires user intentions, and refines them into actionable goals.
arXiv Detail & Related papers (2024-02-14T14:36:30Z) - Experts-in-the-Loop: Establishing an Effective Workflow in Crafting
Privacy Q&A [0.0]
We propose a dynamic workflow for transforming privacy policies into privacy question-and-answer (Q&A) pairs.
Thereby, we facilitate interdisciplinary collaboration among legal experts and conversation designers.
Our proposed workflow underscores continuous improvement and monitoring throughout the construction of privacy Q&As.
arXiv Detail & Related papers (2023-11-18T20:32:59Z) - PLUE: Language Understanding Evaluation Benchmark for Privacy Policies
in English [77.79102359580702]
We introduce the Privacy Policy Language Understanding Evaluation benchmark, a multi-task benchmark for evaluating the privacy policy language understanding.
We also collect a large corpus of privacy policies to enable privacy policy domain-specific language model pre-training.
We demonstrate that domain-specific continual pre-training offers performance improvements across all tasks.
arXiv Detail & Related papers (2022-12-20T05:58:32Z) - Privacy Policy Question Answering Assistant: A Query-Guided Extractive
Summarization Approach [18.51811191325837]
We propose an automated privacy policy question answering assistant that extracts a summary in response to the input user query.
This is a challenging task because users articulate their privacy-related questions in a very different language than the legal language of the policy.
Our pipeline is able to find an answer for 89% of the user queries in the privacyQA dataset.
arXiv Detail & Related papers (2021-09-29T18:00:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.