Character comes from practice: longitudinal practice-based ethics
training in data science
- URL: http://arxiv.org/abs/2401.04454v1
- Date: Tue, 9 Jan 2024 09:37:44 GMT
- Title: Character comes from practice: longitudinal practice-based ethics
training in data science
- Authors: Louise Bezuidenhout, Emanuele Ratti
- Abstract summary: We discuss how the goal of RCR training is to foster the cultivation of certain moral abilities.
While the ideal is the cultivation of virtues, the limited space allowed by RCR modules can only facilitate the cultivation of superficial abilities.
Third, we operationalize our approach by stressing that (proto-)virtue acquisition occurs through the technical and social tasks of daily data science activities.
- Score: 0.5439020425818999
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this chapter, we propose a non-traditional RCR training in data science
that is grounded into a virtue theory framework. First, we delineate the
approach in more theoretical detail, by discussing how the goal of RCR training
is to foster the cultivation of certain moral abilities. We specify the nature
of these abilities: while the ideal is the cultivation of virtues, the limited
space allowed by RCR modules can only facilitate the cultivation of superficial
abilities or proto-virtues, which help students to familiarize with moral and
political issues in the data science environment. Third, we operationalize our
approach by stressing that (proto-)virtue acquisition (like skill acquisition)
occurs through the technical and social tasks of daily data science activities,
where these repetitive tasks provide the opportunities to develop
(proto-)virtue capacity and to support the development of ethically robust data
systems. Finally, we discuss a concrete example of how this approach has been
implemented. In particular, we describe how this method is applied to teach
data ethics to students participating in the CODATA-RDA Data Science Summer
Schools.
Related papers
- Exploring the Precise Dynamics of Single-Layer GAN Models: Leveraging Multi-Feature Discriminators for High-Dimensional Subspace Learning [0.0]
We study the training dynamics of a single-layer GAN model from the perspective of subspace learning.
By bridging our analysis to the realm of subspace learning, we systematically compare the efficacy of GAN-based methods against conventional approaches.
arXiv Detail & Related papers (2024-11-01T10:21:12Z) - Teaching Design Science as a Method for Effective Research Development [0.24578723416255752]
Applying Design Science Research (DSR) methodology is becoming a popular working resource for Information Systems (IS) and Software engineering studies.
This chapter includes examples of DSR, a teaching methodology, learning objectives, and recommendations.
We have created a survey artifact intended to gather data on the experiences of design science users.
arXiv Detail & Related papers (2024-07-13T10:43:06Z) - FAIR Enough: How Can We Develop and Assess a FAIR-Compliant Dataset for Large Language Models' Training? [3.0406004578714008]
The rapid evolution of Large Language Models highlights the necessity for ethical considerations and data integrity in AI development.
While FAIR principles are crucial for ethical data stewardship, their specific application in the context of LLM training data remains an under-explored area.
We propose a novel framework designed to integrate FAIR principles into the LLM development lifecycle.
arXiv Detail & Related papers (2024-01-19T21:21:02Z) - Combatting Human Trafficking in the Cyberspace: A Natural Language
Processing-Based Methodology to Analyze the Language in Online Advertisements [55.2480439325792]
This project tackles the pressing issue of human trafficking in online C2C marketplaces through advanced Natural Language Processing (NLP) techniques.
We introduce a novel methodology for generating pseudo-labeled datasets with minimal supervision, serving as a rich resource for training state-of-the-art NLP models.
A key contribution is the implementation of an interpretability framework using Integrated Gradients, providing explainable insights crucial for law enforcement.
arXiv Detail & Related papers (2023-11-22T02:45:01Z) - Towards a General Framework for Continual Learning with Pre-training [55.88910947643436]
We present a general framework for continual learning of sequentially arrived tasks with the use of pre-training.
We decompose its objective into three hierarchical components, including within-task prediction, task-identity inference, and task-adaptive prediction.
We propose an innovative approach to explicitly optimize these components with parameter-efficient fine-tuning (PEFT) techniques and representation statistics.
arXiv Detail & Related papers (2023-10-21T02:03:38Z) - Model-Based Reinforcement Learning with Multi-Task Offline Pretraining [59.82457030180094]
We present a model-based RL method that learns to transfer potentially useful dynamics and action demonstrations from offline data to a novel task.
The main idea is to use the world models not only as simulators for behavior learning but also as tools to measure the task relevance.
We demonstrate the advantages of our approach compared with the state-of-the-art methods in Meta-World and DeepMind Control Suite.
arXiv Detail & Related papers (2023-06-06T02:24:41Z) - Constrained Reinforcement Learning for Robotics via Scenario-Based
Programming [64.07167316957533]
It is crucial to optimize the performance of DRL-based agents while providing guarantees about their behavior.
This paper presents a novel technique for incorporating domain-expert knowledge into a constrained DRL training loop.
Our experiments demonstrate that using our approach to leverage expert knowledge dramatically improves the safety and the performance of the agent.
arXiv Detail & Related papers (2022-06-20T07:19:38Z) - Model Positionality and Computational Reflexivity: Promoting Reflexivity
in Data Science [10.794642538442107]
We describe how the concepts of positionality and reflexivity can be adapted to provide a framework for understanding data science work.
We describe the challenges of adapting these concepts for data science work and offer annotator fingerprinting and position mining as promising solutions.
arXiv Detail & Related papers (2022-03-08T16:02:03Z) - Self-Supervised Representation Learning: Introduction, Advances and
Challenges [125.38214493654534]
Self-supervised representation learning methods aim to provide powerful deep feature learning without the requirement of large annotated datasets.
This article introduces this vibrant area including key concepts, the four main families of approach and associated state of the art, and how self-supervised methods are applied to diverse modalities of data.
arXiv Detail & Related papers (2021-10-18T13:51:22Z) - Opinionated practices for teaching reproducibility: motivation, guided
instruction and practice [0.0]
Predictive modelling is often one of the most interesting topics to novices in data science.
Students are not as intrinsically motivated to learn this topic, and it is not an easy one for them to learn.
Providing extra motivation, guided instruction and lots of practice are key to effectively teaching this topic.
arXiv Detail & Related papers (2021-09-17T19:15:41Z) - AWAC: Accelerating Online Reinforcement Learning with Offline Datasets [84.94748183816547]
We show that our method, advantage weighted actor critic (AWAC), enables rapid learning of skills with a combination of prior demonstration data and online experience.
Our results show that incorporating prior data can reduce the time required to learn a range of robotic skills to practical time-scales.
arXiv Detail & Related papers (2020-06-16T17:54:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.