Automated detection of dark patterns in cookie banners: how to do it
poorly and why it is hard to do it any other way
- URL: http://arxiv.org/abs/2204.11836v1
- Date: Thu, 21 Apr 2022 12:10:27 GMT
- Title: Automated detection of dark patterns in cookie banners: how to do it
poorly and why it is hard to do it any other way
- Authors: Than Htut Soe, Cristiana Teixeira Santos, and Marija Slavkovik
- Abstract summary: A dataset of cookie banners of 300 news websites was used to train a prediction model that does exactly that.
The accuracy of the trained model is promising, but allows a lot of room for improvement.
We provide an in-depth analysis of the interdisciplinary challenges that automated dark pattern detection poses to artificial intelligence.
- Score: 7.2834950390171205
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cookie banners, the pop ups that appear to collect your consent for data
collection, are a tempting ground for dark patterns. Dark patterns are design
elements that are used to influence the user's choice towards an option that is
not in their interest. The use of dark patterns renders consent elicitation
meaningless and voids the attempts to improve a fair collection and use of
data. Can machine learning be used to automatically detect the presence of dark
patterns in cookie banners? In this work, a dataset of cookie banners of 300
news websites was used to train a prediction model that does exactly that. The
machine learning pipeline we used includes feature engineering, parameter
search, training a Gradient Boosted Tree classifier and evaluation. The
accuracy of the trained model is promising, but allows a lot of room for
improvement. We provide an in-depth analysis of the interdisciplinary
challenges that automated dark pattern detection poses to artificial
intelligence. The dataset and all the code created using machine learning is
available at the url to repository removed for review.
Related papers
- Detecting Deceptive Dark Patterns in E-commerce Platforms [0.0]
Dark patterns are deceptive user interfaces employed by e-commerce websites to manipulate user's behavior in a way that benefits the website, often unethically.
Existing solutions include UIGuard, which uses computer vision and natural language processing, and approaches that categorize dark patterns based on detectability or utilize machine learning models trained on datasets.
We propose combining web scraping techniques with fine-tuned BERT language models and generative capabilities to identify dark patterns, including outliers.
arXiv Detail & Related papers (2024-05-27T16:32:40Z) - Predicting Long-horizon Futures by Conditioning on Geometry and Time [49.86180975196375]
We explore the task of generating future sensor observations conditioned on the past.
We leverage the large-scale pretraining of image diffusion models which can handle multi-modality.
We create a benchmark for video prediction on a diverse set of videos spanning indoor and outdoor scenes.
arXiv Detail & Related papers (2024-04-17T16:56:31Z) - Why is the User Interface a Dark Pattern? : Explainable Auto-Detection
and its Analysis [1.4474137122906163]
Dark patterns are deceptive user interface designs for online services that make users behave in unintended ways.
We study interpretable dark pattern auto-detection, that is, why a particular user interface is detected as having dark patterns.
Our findings may prevent users from being manipulated by dark patterns, and aid in the construction of more equitable internet services.
arXiv Detail & Related papers (2023-12-30T03:53:58Z) - Learn to Unlearn for Deep Neural Networks: Minimizing Unlearning
Interference with Gradient Projection [56.292071534857946]
Recent data-privacy laws have sparked interest in machine unlearning.
Challenge is to discard information about the forget'' data without altering knowledge about remaining dataset.
We adopt a projected-gradient based learning method, named as Projected-Gradient Unlearning (PGU)
We provide empirically evidence to demonstrate that our unlearning method can produce models that behave similar to models retrained from scratch across various metrics even when the training dataset is no longer accessible.
arXiv Detail & Related papers (2023-12-07T07:17:24Z) - Cookiescanner: An Automated Tool for Detecting and Evaluating GDPR
Consent Notices on Websites [1.3416250383686867]
We present emphcookiescanner, an automated scanning tool that detects and extracts consent notices.
We found that manually filter lists have the highest precision but recall fewer consent notices than our keyword-based methods.
Our BERT model achieves high precision for English notices, which is in line with previous work, but suffers from low recall due to insufficient candidate extraction.
arXiv Detail & Related papers (2023-09-12T13:04:00Z) - AidUI: Toward Automated Recognition of Dark Patterns in User Interfaces [6.922187804798161]
UI dark patterns can lead end-users toward (unknowingly) taking actions that they may not have intended.
We introduce AidUI, a novel approach that uses computer vision and natural language processing techniques to recognize ten unique UI dark patterns.
AidUI achieves an overall precision of 0.66, recall of 0.67, F1-score of 0.65 in detecting dark pattern instances, and is able to localize detected patterns with an IoU score of 0.84.
arXiv Detail & Related papers (2023-03-12T23:46:04Z) - Dark patterns in e-commerce: a dataset and its baseline evaluations [0.14680035572775535]
We constructed a dataset for dark pattern detection with state-of-the-art machine learning methods.
As a result of 5-fold cross-validation, we achieved the highest accuracy of 0.975 with RoBERTa.
arXiv Detail & Related papers (2022-11-12T01:53:49Z) - Reasoning-Modulated Representations [85.08205744191078]
We study a common setting where our task is not purely opaque.
Our approach paves the way for a new class of data-efficient representation learning.
arXiv Detail & Related papers (2021-07-19T13:57:13Z) - Visualising Deep Network's Time-Series Representations [93.73198973454944]
Despite the popularisation of machine learning models, more often than not they still operate as black boxes with no insight into what is happening inside the model.
In this paper, a method that addresses that issue is proposed, with a focus on visualising multi-dimensional time-series data.
Experiments on a high-frequency stock market dataset show that the method provides fast and discernible visualisations.
arXiv Detail & Related papers (2021-03-12T09:53:34Z) - Data Augmentation for Object Detection via Differentiable Neural
Rendering [71.00447761415388]
It is challenging to train a robust object detector when annotated data is scarce.
Existing approaches to tackle this problem include semi-supervised learning that interpolates labeled data from unlabeled data.
We introduce an offline data augmentation method for object detection, which semantically interpolates the training data with novel views.
arXiv Detail & Related papers (2021-03-04T06:31:06Z) - What Do Deep Nets Learn? Class-wise Patterns Revealed in the Input Space [88.37185513453758]
We propose a method to visualize and understand the class-wise knowledge learned by deep neural networks (DNNs) under different settings.
Our method searches for a single predictive pattern in the pixel space to represent the knowledge learned by the model for each class.
In the adversarial setting, we show that adversarially trained models tend to learn more simplified shape patterns.
arXiv Detail & Related papers (2021-01-18T06:38:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.