Privacy in Human-AI Romantic Relationships: Concerns, Boundaries, and Agency
- URL: http://arxiv.org/abs/2601.16824v1
- Date: Fri, 23 Jan 2026 15:23:37 GMT
- Title: Privacy in Human-AI Romantic Relationships: Concerns, Boundaries, and Agency
- Authors: Rongjun Ma, Shijing He, Jose Luis Martin-Navarro, Xiao Zhan, Jose Such,
- Abstract summary: This work investigates privacy in human-AI romantic relationships through an interview study (N=17)<n>We found that these relationships took varied forms, from one-to-one to one-to-many, and were shaped by multiple actors, including creators, platforms, and moderators.<n>As intimacy deepened, these boundaries became more permeable, though some participants voiced concerns such as conversation exposure and sought to preserve anonymity.
- Score: 8.378494376906334
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: An increasing number of LLM-based applications are being developed to facilitate romantic relationships with AI partners, yet the safety and privacy risks in these partnerships remain largely underexplored. In this work, we investigate privacy in human-AI romantic relationships through an interview study (N=17), examining participants' experiences and privacy perceptions across stages of exploration, intimacy, and dissolution, alongside platforms they used. We found that these relationships took varied forms, from one-to-one to one-to-many, and were shaped by multiple actors, including creators, platforms, and moderators. AI partners were perceived as having agency, actively negotiating privacy boundaries with participants and sometimes encouraging disclosure of personal details. As intimacy deepened, these boundaries became more permeable, though some participants voiced concerns such as conversation exposure and sought to preserve anonymity. Overall, platform affordances and diverse romantic dynamics expand the privacy landscape, underscoring the need to rethink how privacy is constructed in human-AI intimacy.
Related papers
- Chatting with Confidants or Corporations? Privacy Management with AI Companions [2.1485350418225244]
We conducted in-depth interviews with fifteen users of companion AI platforms such as Replika and Character.AI.<n>Our findings reveal that users blend interpersonal habits with institutional awareness.<n>Many feel uncertain or powerless regarding platform-level data control.
arXiv Detail & Related papers (2026-01-13T15:15:37Z) - Honey Trap or Romantic Utopia: A Case Study of Final Fantasy XIV Players PII Disclosure in Intimate Partner-Seeking Posts [2.7624021966289605]
We conducted a case study on Final Fantasy XIV (FFXIV) players intimate partner seeking posts on social media.<n>Our findings reveal that players disclose sensitive personal information and share vulnerabilities to establish trust.<n>We propose design implications for reducing privacy and safety risks and fostering healthier social interactions in virtual worlds.
arXiv Detail & Related papers (2025-03-12T20:53:06Z) - Multi-P$^2$A: A Multi-perspective Benchmark on Privacy Assessment for Large Vision-Language Models [65.2761254581209]
We evaluate the privacy preservation capabilities of 21 open-source and 2 closed-source Large Vision-Language Models (LVLMs)<n>Based on Multi-P$2$A, we evaluate the privacy preservation capabilities of 21 open-source and 2 closed-source LVLMs.<n>Our results reveal that current LVLMs generally pose a high risk of facilitating privacy breaches.
arXiv Detail & Related papers (2024-12-27T07:33:39Z) - Activity Recognition on Avatar-Anonymized Datasets with Masked Differential Privacy [64.32494202656801]
Privacy-preserving computer vision is an important emerging problem in machine learning and artificial intelligence.<n>We present anonymization pipeline that replaces sensitive human subjects in video datasets with synthetic avatars within context.<n>We also proposeMaskDP to protect non-anonymized but privacy sensitive background information.
arXiv Detail & Related papers (2024-10-22T15:22:53Z) - PrivacyLens: Evaluating Privacy Norm Awareness of Language Models in Action [54.11479432110771]
PrivacyLens is a novel framework designed to extend privacy-sensitive seeds into expressive vignettes and further into agent trajectories.<n>We instantiate PrivacyLens with a collection of privacy norms grounded in privacy literature and crowdsourced seeds.<n>State-of-the-art LMs, like GPT-4 and Llama-3-70B, leak sensitive information in 25.68% and 38.69% of cases, even when prompted with privacy-enhancing instructions.
arXiv Detail & Related papers (2024-08-29T17:58:38Z) - Privacy or Transparency? Negotiated Smartphone Access as a Signifier of Trust in Romantic Relationships [4.963840132735274]
We investigate how individuals think about sharing smartphone access with romantic partners as a function of trust in relationships.
We find that there is little consensus about the level of smartphone access that is desirable in romantic contexts.
We find individuals have crossed these boundaries, violating their partners' privacy and betraying their trust.
arXiv Detail & Related papers (2024-07-06T00:52:34Z) - Linkage on Security, Privacy and Fairness in Federated Learning: New Balances and New Perspectives [48.48294460952039]
This survey offers comprehensive descriptions of the privacy, security, and fairness issues in federated learning.
We contend that there exists a trade-off between privacy and fairness and between security and sharing.
arXiv Detail & Related papers (2024-06-16T10:31:45Z) - Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory [82.7042006247124]
We show that even the most capable AI models reveal private information in contexts that humans would not, 39% and 57% of the time, respectively.
Our work underscores the immediate need to explore novel inference-time privacy-preserving approaches, based on reasoning and theory of mind.
arXiv Detail & Related papers (2023-10-27T04:15:30Z) - A Unified View of Differentially Private Deep Generative Modeling [60.72161965018005]
Data with privacy concerns comes with stringent regulations that frequently prohibited data access and data sharing.
Overcoming these obstacles is key for technological progress in many real-world application scenarios that involve privacy sensitive data.
Differentially private (DP) data publishing provides a compelling solution, where only a sanitized form of the data is publicly released.
arXiv Detail & Related papers (2023-09-27T14:38:16Z) - User Perception of Privacy with Ubiquitous Devices [5.33024001730262]
This study aims to explore and discover various concerns related to perception of privacy in this era of ubiquitous technologies.
Key themes like attitude towards privacy in public and private spaces, privacy awareness, consent seeking, dilemmas/confusions related to various technologies, impact of attitude and beliefs on individuals actions regarding how to protect oneself from invasion of privacy in both public and private spaces.
arXiv Detail & Related papers (2021-07-23T05:01:44Z) - Privacy threats in intimate relationships [0.720851507101878]
This article provides an overview of intimate threats: a class of privacy threats that can arise within our families, romantic partnerships, close friendships, and caregiving relationships.
We survey a range of intimate relationships and describe their common features.
Based on these features, we explore implications for both technical privacy design and policy, and offer design recommendations for ameliorating intimate privacy risks.
arXiv Detail & Related papers (2020-06-06T16:21:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.