BabyLM Turns 3: Call for papers for the 2025 BabyLM workshop
- URL: http://arxiv.org/abs/2502.10645v1
- Date: Sat, 15 Feb 2025 02:46:43 GMT
- Title: BabyLM Turns 3: Call for papers for the 2025 BabyLM workshop
- Authors: Lucas Charpentier, Leshem Choshen, Ryan Cotterell, Mustafa Omer Gul, Michael Hu, Jaap Jumelet, Tal Linzen, Jing Liu, Aaron Mueller, Candace Ross, Raj Sanjay Shah, Alex Warstadt, Ethan Wilcox, Adina Williams,
- Abstract summary: BabyLM aims to dissolve the boundaries between cognitive modeling and language modeling.
We call for both workshop papers and for researchers to join the 3rd BabyLM competition.
- Score: 77.62533643491747
- License:
- Abstract: BabyLM aims to dissolve the boundaries between cognitive modeling and language modeling. We call for both workshop papers and for researchers to join the 3rd BabyLM competition. As in previous years, we call for participants in the data-efficient pretraining challenge in the general track. This year, we also offer a new track: INTERACTION. This new track encourages interactive behavior, learning from a teacher, and adapting the teaching material to the student. We also call for papers outside the competition in any relevant areas. These include training efficiency, cognitively plausible research, weak model evaluation, and more.
Related papers
- Findings of the Second BabyLM Challenge: Sample-Efficient Pretraining on Developmentally Plausible Corpora [79.03392191805028]
The BabyLM Challenge is a community effort to close the data-efficiency gap between human and computational language learners.
Participants compete to optimize language model training on a fixed language data budget of 100 million words or less.
arXiv Detail & Related papers (2024-12-06T16:06:08Z) - A surprisal oracle for when every layer counts [2.5716627278119444]
Active Curriculum Language Modeling (ACLM) is a learner directed approach to training a language model.
We propose an updated ACLM process for the BabyLM 2024 task.
arXiv Detail & Related papers (2024-12-04T07:53:45Z) - UpStory: the Uppsala Storytelling dataset [2.7895834501191823]
UpStory is a novel dataset of naturalistic dyadic interactions between primary school aged children.
The dataset contains data for 35 pairs, totalling 3h 40m of audio and video recordings.
An anonymized version of the dataset is made publicly available, containing per-frame head pose, body pose, and face features.
arXiv Detail & Related papers (2024-07-05T08:46:16Z) - [Call for Papers] The 2nd BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus [81.34965784440176]
This CfP provides rules for the BabyLM Challenge 2024-2025.
The overarching goals of the challenge remain the same.
We replace the loose track with a paper track.
We relax the rules around pretraining data.
We introduce a multimodal vision-and-language track.
arXiv Detail & Related papers (2024-04-09T11:04:50Z) - ChapGTP, ILLC's Attempt at Raising a BabyLM: Improving Data Efficiency
by Automatic Task Formation [5.472046616411226]
We present the submission of the ILLC at the University of Amsterdam to the BabyLM challenge (Warstadt et al., 2023)
Our final model, ChapGTP, is a masked language model that was trained for 200 epochs, aided by a novel data augmentation technique called Automatic Task Formation.
We discuss in detail the performance of this model on the three evaluation suites: BLiMP, (Super)GLUE, and MSGS.
arXiv Detail & Related papers (2023-10-17T14:06:06Z) - Call for Papers -- The BabyLM Challenge: Sample-efficient pretraining on
a developmentally plausible corpus [32.51325830633226]
We present the call for papers for the BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus.
This shared task is intended for participants with an interest in small scale language modeling, human language acquisition, low-resource NLP, and cognitive modeling.
arXiv Detail & Related papers (2023-01-27T15:52:50Z) - Don't Copy the Teacher: Data and Model Challenges in Embodied Dialogue [92.01165203498299]
Embodied dialogue instruction following requires an agent to complete a complex sequence of tasks from a natural language exchange.
This paper argues that imitation learning (IL) and related low-level metrics are actually misleading and do not align with the goals of embodied dialogue research.
arXiv Detail & Related papers (2022-10-10T05:51:40Z) - Recitation-Augmented Language Models [85.30591349383849]
We show that RECITE is a powerful paradigm for knowledge-intensive NLP tasks.
Specifically, we show that by utilizing recitation as the intermediate step, a recite-and-answer scheme can achieve new state-of-the-art performance.
arXiv Detail & Related papers (2022-10-04T00:49:20Z) - LID 2020: The Learning from Imperfect Data Challenge Results [242.86700551532272]
Learning from Imperfect Data workshop aims to inspire and facilitate the research in developing novel approaches.
We organize three challenges to find the state-of-the-art approaches in weakly supervised learning setting.
This technical report summarizes the highlights from the challenge.
arXiv Detail & Related papers (2020-10-17T13:06:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.