Security for Children in the Digital Society -- A Rights-based and
Research Ethics Approach
- URL: http://arxiv.org/abs/2309.12340v2
- Date: Thu, 5 Oct 2023 10:15:38 GMT
- Title: Security for Children in the Digital Society -- A Rights-based and
Research Ethics Approach
- Authors: Laura Schelenz, Ingrid Stapf, Jessica Heesen
- Abstract summary: The project is situated in a German context with a focus on European frameworks for the development of Artificial Intelligence and the protection of children from security risks arising in the course of algorithm-mediated online communication.
The project develops a children's rights approach to questions of security for children online while also developing a research ethics approach for conducting research with children on online harms such as cybergrooming and sexual violence against children.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this position paper, we present initial perspectives and research results
from the project "SIKID - Security for Children in the Digital World." The
project is situated in a German context with a focus on European frameworks for
the development of Artificial Intelligence and the protection of children from
security risks arising in the course of algorithm-mediated online
communication. The project strengthens networks of relevant stakeholders,
explores regulatory measures and informs policy makers, and develops a
children's rights approach to questions of security for children online while
also developing a research ethics approach for conducting research with
children on online harms such as cybergrooming and sexual violence against
children.
Related papers
- Privacy Risks of General-Purpose AI Systems: A Foundation for Investigating Practitioner Perspectives [47.17703009473386]
Powerful AI models have led to impressive leaps in performance across a wide range of tasks.
Privacy concerns have led to a wealth of literature covering various privacy risks and vulnerabilities of AI models.
We conduct a systematic review of these survey papers to provide a concise and usable overview of privacy risks in GPAIS.
arXiv Detail & Related papers (2024-07-02T07:49:48Z) - A Safe Harbor for AI Evaluation and Red Teaming [124.89885800509505]
Some researchers fear that conducting such research or releasing their findings will result in account suspensions or legal reprisal.
We propose that major AI developers commit to providing a legal and technical safe harbor.
We believe these commitments are a necessary step towards more inclusive and unimpeded community efforts to tackle the risks of generative AI.
arXiv Detail & Related papers (2024-03-07T20:55:08Z) - A safety risk assessment framework for children's online safety based on
a novel safety weakness assessment approach [0.43512163406552007]
This paper addresses the problem of children's online safety in the context of the growing digital landscape.
We propose a safety risk assessment approach that focuses specifically on children's online safety.
arXiv Detail & Related papers (2024-01-26T08:50:15Z) - SoK: Safer Digital-Safety Research Involving At-Risk Users [43.45078079505055]
Pursuing research in computer security and privacy is crucial to understanding how to meet the digital-safety needs of at-risk users.
We offer an analysis of 196 academic works to elicit 14 research risks and 36 safety practices used by a growing community of researchers.
We conclude by suggesting areas for future research regarding the reporting, study, and funding of at-risk user research.
arXiv Detail & Related papers (2023-09-01T21:15:39Z) - On the conformance of Android applications with children's data
protection regulations and safeguarding guidelines [3.8029070240258687]
Even apps designed for children do not always comply with legislation or guidance.
This lack of compliance could contribute creating a path to causing physical or mental harm.
arXiv Detail & Related papers (2023-05-15T09:46:56Z) - Beyond the Safeguards: Exploring the Security Risks of ChatGPT [3.1981440103815717]
Increasing popularity of large language models (LLMs) has led to growing concerns about their safety, security risks, and ethical implications.
This paper aims to provide an overview of the different types of security risks associated with ChatGPT, including malicious text and code generation, private data disclosure, fraudulent services, information gathering, and producing unethical content.
arXiv Detail & Related papers (2023-05-13T21:01:14Z) - Towards Safer Generative Language Models: A Survey on Safety Risks,
Evaluations, and Improvements [76.80453043969209]
This survey presents a framework for safety research pertaining to large models.
We begin by introducing safety issues of wide concern, then delve into safety evaluation methods for large models.
We explore the strategies for enhancing large model safety from training to deployment.
arXiv Detail & Related papers (2023-02-18T09:32:55Z) - Imagining new futures beyond predictive systems in child welfare: A
qualitative study with impacted stakeholders [89.6319385008397]
We conducted a set of seven design workshops with 35 stakeholders who have been impacted by the child welfare system.
We found that participants worried current PRMs perpetuate or exacerbate existing problems in child welfare.
Participants suggested new ways to use data and data-driven tools to better support impacted communities.
arXiv Detail & Related papers (2022-05-18T13:49:55Z) - Understanding parents' perceptions of children's cybersecurity awareness
in Norway [3.590867023672793]
We conducted a study with 25 parents living in Norway with children aged between 10 to 15.
Results include a list of cybersecurity awareness needs for children from a parental perspective, a list of learning resources for children, and a list of challenges for parents to ensure cybersecurity at home.
Future research should focus on defining cybersecurity theories and practices that contribute to children's and parents' awareness about cybersecurity risks, needs, and solutions.
arXiv Detail & Related papers (2021-08-05T10:42:49Z) - Interactive Storytelling for Children: A Case-study of Design and
Development Considerations for Ethical Conversational AI [0.20999222360659603]
Conversational Artificial Intelligence (CAI) systems and Intelligent Personal Assistants (IPA) are becoming ubiquitous in our lives.
This paper reflects on the social context within a specific case of technology development, as substantiated and supported by argumentation from within the literature.
Further research that engages with developers and stakeholders in the ethics of storytelling through CAI is highlighted as a matter of urgency.
arXiv Detail & Related papers (2021-07-20T15:11:45Z) - Adversarial Machine Learning Attacks and Defense Methods in the Cyber
Security Domain [58.30296637276011]
This paper summarizes the latest research on adversarial attacks against security solutions based on machine learning techniques.
It is the first to discuss the unique challenges of implementing end-to-end adversarial attacks in the cyber security domain.
arXiv Detail & Related papers (2020-07-05T18:22:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.