Developing Open Source Educational Resources for Machine Learning and
Data Science
- URL: http://arxiv.org/abs/2107.14330v1
- Date: Wed, 28 Jul 2021 10:20:20 GMT
- Title: Developing Open Source Educational Resources for Machine Learning and
Data Science
- Authors: Ludwig Bothmann (1), Sven Strickroth (2), Giuseppe Casalicchio (1),
David R\"ugamer (1), Marius Lindauer (3), Fabian Scheipl (1), Bernd Bischl
(1) ((1) Department of Statistics, Ludwig-Maximilians-Universit\"at
M\"unchen, Germany, (2) Institute of Computer Science,
Ludwig-Maximilians-Universit\"at M\"unchen, Germany, (3) Institute of
Information Process, Leibniz University Hannover, Germany)
- Abstract summary: We describe the specific requirements for Open Educational Resources (OER) in Machine Learning (ML) and Data Science (DS)
We argue that it is especially important for these fields to make source files publicly available, leading to Open Source Educational Resources (OSER)
We outline how OSER can be used for blended learning scenarios and share our experiences in university education.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Education should not be a privilege but a common good. It should be openly
accessible to everyone, with as few barriers as possible; even more so for key
technologies such as Machine Learning (ML) and Data Science (DS). Open
Educational Resources (OER) are a crucial factor for greater educational
equity. In this paper, we describe the specific requirements for OER in ML and
DS and argue that it is especially important for these fields to make source
files publicly available, leading to Open Source Educational Resources (OSER).
We present our view on the collaborative development of OSER, the challenges
this poses, and first steps towards their solutions. We outline how OSER can be
used for blended learning scenarios and share our experiences in university
education. Finally, we discuss additional challenges such as credit assignment
or granting certificates.
Related papers
- OpenR: An Open Source Framework for Advanced Reasoning with Large Language Models [61.14336781917986]
We introduce OpenR, an open-source framework for enhancing the reasoning capabilities of large language models (LLMs)
OpenR unifies data acquisition, reinforcement learning training, and non-autoregressive decoding into a cohesive software platform.
Our work is the first to provide an open-source framework that explores the core techniques of OpenAI's o1 model with reinforcement learning.
arXiv Detail & Related papers (2024-10-12T23:42:16Z) - A Survey on RAG Meeting LLMs: Towards Retrieval-Augmented Large Language Models [71.25225058845324]
Large Language Models (LLMs) have demonstrated revolutionary abilities in language understanding and generation.
Retrieval-Augmented Generation (RAG) can offer reliable and up-to-date external knowledge.
RA-LLMs have emerged to harness external and authoritative knowledge bases, rather than relying on the model's internal knowledge.
arXiv Detail & Related papers (2024-05-10T02:48:45Z) - Is open source software culture enough to make AI a common ? [0.0]
Language models (LM) are increasingly deployed in the field of artificial intelligence (AI)
The question arises as to whether they can be a common resource managed and maintained by a community of users.
We highlight the potential benefits of treating the data and resources needed to create LMs as commons.
arXiv Detail & Related papers (2024-03-19T14:43:52Z) - Private Knowledge Sharing in Distributed Learning: A Survey [50.51431815732716]
The rise of Artificial Intelligence has revolutionized numerous industries and transformed the way society operates.
It is crucial to utilize information in learning processes that are either distributed or owned by different entities.
Modern data-driven services have been developed to integrate distributed knowledge entities into their outcomes.
arXiv Detail & Related papers (2024-02-08T07:18:23Z) - Adapting Large Language Models for Education: Foundational Capabilities, Potentials, and Challenges [60.62904929065257]
Large language models (LLMs) offer possibility for resolving this issue by comprehending individual requests.
This paper reviews the recently emerged LLM research related to educational capabilities, including mathematics, writing, programming, reasoning, and knowledge-based question answering.
arXiv Detail & Related papers (2023-12-27T14:37:32Z) - Opening up ChatGPT: Tracking openness, transparency, and accountability
in instruction-tuned text generators [0.11470070927586018]
We evaluate projects in terms of openness of code, training data, model weights, RLHF data, licensing, scientific documentation, and access methods.
We find that while there is a fast-growing list of projects billing themselves as 'open source', many inherit undocumented data of dubious legality.
Degrees of openness are relevant to fairness and accountability at all points.
arXiv Detail & Related papers (2023-07-08T07:08:20Z) - Can We Trust AI-Generated Educational Content? Comparative Analysis of
Human and AI-Generated Learning Resources [4.528957284486784]
Large language models (LLMs) appear to offer a promising solution to the rapid creation of learning materials at scale.
We compare the quality of resources generated by an LLM with those created by students as part of a learnersourcing activity.
Our results show that the quality of AI-generated resources, as perceived by students, is equivalent to the quality of resources generated by their peers.
arXiv Detail & Related papers (2023-06-18T09:49:21Z) - A Survey of Machine Unlearning [56.017968863854186]
Recent regulations now require that, on request, private information about a user must be removed from computer systems.
ML models often remember' the old data.
Recent works on machine unlearning have not been able to completely solve the problem.
arXiv Detail & Related papers (2022-09-06T08:51:53Z) - OpenFed: A Comprehensive and Versatile Open-Source Federated Learning
Framework [5.893286029670115]
We propose OpenFed, an open-source software framework for end-to-end Federated Learning.
For researchers, OpenFed provides a framework wherein new methods can be easily implemented and fairly evaluated.
For downstream users, OpenFed allows Federated Learning to be plugged and play within different subject-matter contexts.
arXiv Detail & Related papers (2021-09-16T10:31:59Z) - Extracting Topics from Open Educational Resources [0.0]
We propose an OER topic extraction approach, applying text mining techniques, to generate high-quality OER metadata about topic distribution.
This is done by: 1) collecting 123 lectures from Coursera and Khan Academy in the area of data science related skills, 2) applying Latent Dirichlet Allocation (LDA) on the collected resources in order to extract existing topics related to these skills, and 3) defining topic distributions covered by a particular OER.
arXiv Detail & Related papers (2020-06-19T12:50:55Z) - Explore, Discover and Learn: Unsupervised Discovery of State-Covering
Skills [155.11646755470582]
'Explore, Discover and Learn' (EDL) is an alternative approach to information-theoretic skill discovery.
We show that EDL offers significant advantages, such as overcoming the coverage problem, reducing the dependence of learned skills on the initial state, and allowing the user to define a prior over which behaviors should be learned.
arXiv Detail & Related papers (2020-02-10T10:49:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.