Democratizing Machine Learning for Interdisciplinary Scholars: Report on
Organizing the NLP+CSS Online Tutorial Series
- URL: http://arxiv.org/abs/2211.15971v1
- Date: Tue, 29 Nov 2022 07:06:45 GMT
- Title: Democratizing Machine Learning for Interdisciplinary Scholars: Report on
Organizing the NLP+CSS Online Tutorial Series
- Authors: Ian Stewart and Katherine Keith
- Abstract summary: Existing tutorials are often costly to participants, presume extensive programming knowledge, and are not tailored to specific application fields.
We organized a year-long, free, online tutorial series targeted at teaching advanced natural language processing (NLP) methods to computational social science (CSS) scholars.
Although live participation was more limited than expected, a comparison of pre- and post-tutorial surveys showed an increase in participants' perceived knowledge of almost one point on a 7-point Likert scale.
- Score: 0.9645196221785691
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many scientific fields -- including biology, health, education, and the
social sciences -- use machine learning (ML) to help them analyze data at an
unprecedented scale. However, ML researchers who develop advanced methods
rarely provide detailed tutorials showing how to apply these methods. Existing
tutorials are often costly to participants, presume extensive programming
knowledge, and are not tailored to specific application fields. In an attempt
to democratize ML methods, we organized a year-long, free, online tutorial
series targeted at teaching advanced natural language processing (NLP) methods
to computational social science (CSS) scholars. Two organizers worked with
fifteen subject matter experts to develop one-hour presentations with hands-on
Python code for a range of ML methods and use cases, from data pre-processing
to analyzing temporal variation of language change. Although live participation
was more limited than expected, a comparison of pre- and post-tutorial surveys
showed an increase in participants' perceived knowledge of almost one point on
a 7-point Likert scale. Furthermore, participants asked thoughtful questions
during tutorials and engaged readily with tutorial content afterwards, as
demonstrated by 10K~total views of posted tutorial recordings. In this report,
we summarize our organizational efforts and distill five principles for
democratizing ML+X tutorials. We hope future organizers improve upon these
principles and continue to lower barriers to developing ML skills for
researchers of all fields.
Related papers
- Hands-On Tutorial: Labeling with LLM and Human-in-the-Loop [7.925650087629884]
This tutorial is designed for NLP practitioners from both research and industry backgrounds.
We will present the basics of each strategy, highlight their benefits and limitations, and discuss in detail real-life case studies.
The tutorial includes a hands-on workshop, where attendees will be guided in implementing a hybrid annotation setup.
arXiv Detail & Related papers (2024-11-07T11:51:14Z) - Tutorly: Turning Programming Videos Into Apprenticeship Learning Environments with LLMs [1.6961276655027102]
Our work transforms programming videos into one-on-one tutoring experiences using the cognitive apprenticeship framework.
Tutorly, developed as a JupyterLab, allows learners to set personalized learning goals.
arXiv Detail & Related papers (2024-05-21T17:17:34Z) - Aligning Large Language Models with Human: A Survey [53.6014921995006]
Large Language Models (LLMs) trained on extensive textual corpora have emerged as leading solutions for a broad array of Natural Language Processing (NLP) tasks.
Despite their notable performance, these models are prone to certain limitations such as misunderstanding human instructions, generating potentially biased content, or factually incorrect information.
This survey presents a comprehensive overview of these alignment technologies, including the following aspects.
arXiv Detail & Related papers (2023-07-24T17:44:58Z) - Towards Understanding Machine Learning Testing in Practise [23.535630175567146]
We propose to study visualisations of Machine Learning pipelines by mining Jupyter notebooks.
First, gather general insights and trends using a qualitative study of a smaller sample of notebooks.
And then use the knowledge gained from the qualitative study to design an empirical study using a larger sample of notebooks.
arXiv Detail & Related papers (2023-05-08T18:52:26Z) - What are the Machine Learning best practices reported by practitioners
on Stack Exchange? [4.882319198853359]
We present a study listing 127 Machine Learning best practices systematically mining 242 posts of 14 different Stack Exchange (STE) websites.
The list of practices is presented in a set of categories related to different stages of the implementation process of an ML-enabled system.
arXiv Detail & Related papers (2023-01-25T10:50:28Z) - A Survey of Knowledge Enhanced Pre-trained Language Models [78.56931125512295]
We present a comprehensive review of Knowledge Enhanced Pre-trained Language Models (KE-PLMs)
For NLU, we divide the types of knowledge into four categories: linguistic knowledge, text knowledge, knowledge graph (KG) and rule knowledge.
The KE-PLMs for NLG are categorized into KG-based and retrieval-based methods.
arXiv Detail & Related papers (2022-11-11T04:29:02Z) - Knowledgeable Salient Span Mask for Enhancing Language Models as
Knowledge Base [51.55027623439027]
We develop two solutions to help the model learn more knowledge from unstructured text in a fully self-supervised manner.
To our best knowledge, we are the first to explore fully self-supervised learning of knowledge in continual pre-training.
arXiv Detail & Related papers (2022-04-17T12:33:34Z) - Tutorial on Deep Learning for Human Activity Recognition [70.94062293989832]
This tutorial was first held at the 2021 ACM International Symposium on Wearable Computers (ISWC'21)
It provides a hands-on and interactive walk-through of the most important steps in the data pipeline for the deep learning of human activities.
arXiv Detail & Related papers (2021-10-13T12:01:02Z) - Panoramic Learning with A Standardized Machine Learning Formalism [116.34627789412102]
This paper presents a standardized equation of the learning objective, that offers a unifying understanding of diverse ML algorithms.
It also provides guidance for mechanic design of new ML solutions, and serves as a promising vehicle towards panoramic learning with all experiences.
arXiv Detail & Related papers (2021-08-17T17:44:38Z) - ProtoTransformer: A Meta-Learning Approach to Providing Student Feedback [54.142719510638614]
In this paper, we frame the problem of providing feedback as few-shot classification.
A meta-learner adapts to give feedback to student code on a new programming question from just a few examples by instructors.
Our approach was successfully deployed to deliver feedback to 16,000 student exam-solutions in a programming course offered by a tier 1 university.
arXiv Detail & Related papers (2021-07-23T22:41:28Z) - Declarative Machine Learning Systems [7.5717114708721045]
Machine learning (ML) has moved from a academic endeavor to a pervasive technology adopted in almost every aspect of computing.
Recent successes in applying ML in natural sciences revealed that ML can be used to tackle some of the hardest real-world problems humanity faces today.
We believe the next wave of ML systems will allow a larger amount of people, potentially without coding skills, to perform the same tasks.
arXiv Detail & Related papers (2021-07-16T23:57:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.