AutoSpeech 2020: The Second Automated Machine Learning Challenge for
Speech Classification
- URL: http://arxiv.org/abs/2010.13130v1
- Date: Sun, 25 Oct 2020 15:01:41 GMT
- Title: AutoSpeech 2020: The Second Automated Machine Learning Challenge for
Speech Classification
- Authors: Jingsong Wang, Tom Ko, Zhen Xu, Xiawei Guo, Souxiang Liu, Wei-Wei Tu,
Lei Xie
- Abstract summary: The AutoSpeech challenge calls for automated machine learning (AutoML) solutions to automate the process of applying machine learning to speech processing tasks.
This paper outlines the challenge protocol, datasets, evaluation metric, starting kit, and baseline systems.
- Score: 31.22181821515342
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The AutoSpeech challenge calls for automated machine learning (AutoML)
solutions to automate the process of applying machine learning to speech
processing tasks. These tasks, which cover a large variety of domains, will be
shown to the automated system in a random order. Each time when the tasks are
switched, the information of the new task will be hinted with its corresponding
training set. Thus, every submitted solution should contain an adaptation
routine which adapts the system to the new task. Compared to the first edition,
the 2020 edition includes advances of 1) more speech tasks, 2) noisier data in
each task, 3) a modified evaluation metric. This paper outlines the challenge
and describe the competition protocol, datasets, evaluation metric, starting
kit, and baseline systems.
Related papers
- Team MTS @ AutoMin 2021: An Overview of Existing Summarization Approaches and Comparison to Unsupervised Summarization Techniques [0.0]
We propose an unsupervised summarization technique based on clustering and provide a pipeline that includes an adapted automatic speech recognition block able to run on real-life recordings.
The proposed technique outperforms pre-trained summarization models on the automatic minuting task with Rouge 1, Rouge 2 and Rouge L values of 0.21, 0.02 and 0.2 on the dev set, with Rouge 1, Rouge 2, Rouge L, Adequacy, Grammatical correctness and Fluency values of 0.180, 0.035, 0.098, 1.857, 2.304, 1.911 on the test set.
arXiv Detail & Related papers (2024-10-04T13:23:50Z) - Semantic Code Classification for Automated Machine Learning [0.0]
We propose a way to control the output via a sequence of simple actions, that are called semantic code classes.
We present a semantic code classification task and discuss methods for solving this problem on the Natural Language to Machine Learning (NL2ML) dataset.
arXiv Detail & Related papers (2022-01-25T10:40:37Z) - Adapting Document-Grounded Dialog Systems to Spoken Conversations using
Data Augmentation and a Noisy Channel Model [46.93744191416991]
This paper summarizes our submission to Task 2 of the 10th Dialog System Technology Challenge (DSTC10) "Knowledge-grounded Task-oriented Dialogue Modeling on Spoken Conversations"
Similar to the previous year's iteration, the task consists of three subtasks: detecting whether a turn is knowledge seeking, selecting the relevant knowledge document and finally generating a grounded response.
Our best system achieved the 1st rank in the automatic and the 3rd rank in the human evaluation of the challenge.
arXiv Detail & Related papers (2021-12-16T12:51:52Z) - Auto-KWS 2021 Challenge: Task, Datasets, and Baselines [63.82759886293636]
Auto-KWS 2021 challenge calls for automated machine learning (AutoML) solutions to automate the process of applying machine learning to a customized keyword spotting task.
The challenge focuses on the problem of customized keyword spotting, where the target device can only be awakened by an enrolled speaker with his specified keyword.
arXiv Detail & Related papers (2021-03-31T14:56:48Z) - Knowledge Distillation for Improved Accuracy in Spoken Question
Answering [63.72278693825945]
We devise a training strategy to perform knowledge distillation from spoken documents and written counterparts.
Our work makes a step towards distilling knowledge from the language model as a supervision signal.
Experiments demonstrate that our approach outperforms several state-of-the-art language models on the Spoken-SQuAD dataset.
arXiv Detail & Related papers (2020-10-21T15:18:01Z) - AutoML to Date and Beyond: Challenges and Opportunities [30.60364966752454]
AutoML tools aim to make machine learning accessible for non-machine learning experts.
We introduce a new classification system for AutoML systems.
We lay out a roadmap for the future, pinpointing the research required to further automate the end-to-end machine learning pipeline.
arXiv Detail & Related papers (2020-10-21T06:08:21Z) - Induction and Exploitation of Subgoal Automata for Reinforcement
Learning [75.55324974788475]
We present ISA, an approach for learning and exploiting subgoals in episodic reinforcement learning (RL) tasks.
ISA interleaves reinforcement learning with the induction of a subgoal automaton, an automaton whose edges are labeled by the task's subgoals.
A subgoal automaton also consists of two special states: a state indicating the successful completion of the task, and a state indicating that the task has finished without succeeding.
arXiv Detail & Related papers (2020-09-08T16:42:55Z) - The NTT DCASE2020 Challenge Task 6 system: Automated Audio Captioning
with Keywords and Sentence Length Estimation [49.41766997393417]
This report describes the system participating to the Detection and Classification of Acoustic Scenes and Events (DCASE) 2020 Challenge, Task 6.
Our submission focuses on solving two indeterminacy problems in automated audio captioning: word selection indeterminacy and sentence length indeterminacy.
We simultaneously solve the main caption generation and sub indeterminacy problems by estimating keywords and sentence length through multi-task learning.
arXiv Detail & Related papers (2020-07-01T04:26:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.