NeurIPS Competition Instructions and Guide: Causal Insights for Learning
Paths in Education
- URL: http://arxiv.org/abs/2208.12610v2
- Date: Wed, 31 Aug 2022 19:55:54 GMT
- Title: NeurIPS Competition Instructions and Guide: Causal Insights for Learning
Paths in Education
- Authors: Wenbo Gong, Digory Smith, Zichao Wang, Craig Barton, Simon Woodhead,
Nick Pawlowski, Joel Jennings, Cheng Zhang
- Abstract summary: Participants will address two fundamental causal challenges in machine learning in the context of education using time-series data.
The first is to identify the causal relationships between different constructs, where a construct is defined as the smallest element of learning.
The second challenge is to predict the impact of learning one construct on the ability to answer questions on other constructs.
- Score: 15.014508164142075
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this competition, participants will address two fundamental causal
challenges in machine learning in the context of education using time-series
data. The first is to identify the causal relationships between different
constructs, where a construct is defined as the smallest element of learning.
The second challenge is to predict the impact of learning one construct on the
ability to answer questions on other constructs. Addressing these challenges
will enable optimisation of students' knowledge acquisition, which can be
deployed in a real edtech solution impacting millions of students. Participants
will run these tasks in an idealised environment with synthetic data and a
real-world scenario with evaluation data collected from a series of A/B tests.
Related papers
- Detecting Unsuccessful Students in Cybersecurity Exercises in Two Different Learning Environments [0.37729165787434493]
This paper develops automated tools to predict when a student is having difficulty.
In a potential application, such models can aid instructors in detecting struggling students and providing targeted help.
arXiv Detail & Related papers (2024-08-16T04:57:54Z) - Investigating the Pre-Training Dynamics of In-Context Learning: Task Recognition vs. Task Learning [99.05401042153214]
In-context learning (ICL) is potentially attributed to two major abilities: task recognition (TR) and task learning (TL)
We take the first step by examining the pre-training dynamics of the emergence of ICL.
We propose a simple yet effective method to better integrate these two abilities for ICL at inference time.
arXiv Detail & Related papers (2024-06-20T06:37:47Z) - Process Mining for Unstructured Data: Challenges and Research Directions [0.0]
The application of process mining for unstructured data might significantly elevate novel insights into disciplines where unstructured data is a common data format.
To efficiently analyze unstructured data by process mining and to convey confidence into the analysis result, requires bridging multiple challenges.
arXiv Detail & Related papers (2023-11-30T12:09:14Z) - Can Foundation Models Watch, Talk and Guide You Step by Step to Make a
Cake? [62.59699229202307]
Despite advances in AI, it remains a significant challenge to develop interactive task guidance systems.
We created a new multimodal benchmark dataset, Watch, Talk and Guide (WTaG) based on natural interaction between a human user and a human instructor.
We leveraged several foundation models to study to what extent these models can be quickly adapted to perceptually enabled task guidance.
arXiv Detail & Related papers (2023-11-01T15:13:49Z) - Enhancing Student Performance Prediction on Learnersourced Questions
with SGNN-LLM Synergy [11.735587384038753]
We introduce an innovative strategy that synergizes the potential of integrating Signed Graph Neural Networks (SGNNs) and Large Language Model (LLM) embeddings.
Our methodology employs a signed bipartite graph to comprehensively model student answers, complemented by a contrastive learning framework that enhances noise resilience.
arXiv Detail & Related papers (2023-09-23T23:37:55Z) - Accelerating exploration and representation learning with offline
pre-training [52.6912479800592]
We show that exploration and representation learning can be improved by separately learning two different models from a single offline dataset.
We show that learning a state representation using noise-contrastive estimation and a model of auxiliary reward can significantly improve the sample efficiency on the challenging NetHack benchmark.
arXiv Detail & Related papers (2023-03-31T18:03:30Z) - 2021 BEETL Competition: Advancing Transfer Learning for Subject
Independence & Heterogenous EEG Data Sets [89.84774119537087]
We design two transfer learning challenges around diagnostics and Brain-Computer-Interfacing (BCI)
Task 1 is centred on medical diagnostics, addressing automatic sleep stage annotation across subjects.
Task 2 is centred on Brain-Computer Interfacing (BCI), addressing motor imagery decoding across both subjects and data sets.
arXiv Detail & Related papers (2022-02-14T12:12:20Z) - Evaluation Methods and Measures for Causal Learning Algorithms [33.07234268724662]
We focus on the two fundamental causal-inference tasks and causality-aware machine learning tasks.
The survey seeks to bring to the forefront the urgency of developing publicly available benchmarks and consensus-building standards for causal learning evaluation with observational data.
arXiv Detail & Related papers (2022-02-07T00:24:34Z) - AR-LSAT: Investigating Analytical Reasoning of Text [57.1542673852013]
We study the challenge of analytical reasoning of text and introduce a new dataset consisting of questions from the Law School Admission Test from 1991 to 2016.
We analyze what knowledge understanding and reasoning abilities are required to do well on this task.
arXiv Detail & Related papers (2021-04-14T02:53:32Z) - CausalWorld: A Robotic Manipulation Benchmark for Causal Structure and
Transfer Learning [138.40338621974954]
CausalWorld is a benchmark for causal structure and transfer learning in a robotic manipulation environment.
Tasks consist of constructing 3D shapes from a given set of blocks - inspired by how children learn to build complex structures.
arXiv Detail & Related papers (2020-10-08T23:01:13Z) - Federated Edge Learning : Design Issues and Challenges [1.916348196696894]
Federated Learning (FL) is a distributed machine learning technique, where each device contributes to the learning model by independently computing the gradient based on its local training data.
implementing FL at the network edge is challenging due to system and data heterogeneity and resources constraints.
This article proposes a general framework for the data-aware scheduling as a guideline for future research directions.
arXiv Detail & Related papers (2020-08-31T19:56:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.