Overview of the TREC 2019 deep learning track
- URL: http://arxiv.org/abs/2003.07820v2
- Date: Wed, 18 Mar 2020 16:56:56 GMT
- Title: Overview of the TREC 2019 deep learning track
- Authors: Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, Ellen M.
Voorhees
- Abstract summary: The Deep Learning Track is a new track for TREC 2019, with the goal of studying ad hoc ranking in a large data regime.
It is the first track with large human-labeled training sets, introducing two sets corresponding to two tasks.
This year 15 groups submitted a total of 75 runs, using various combinations of deep learning, transfer learning and traditional IR ranking methods.
- Score: 36.23357487158591
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Deep Learning Track is a new track for TREC 2019, with the goal of
studying ad hoc ranking in a large data regime. It is the first track with
large human-labeled training sets, introducing two sets corresponding to two
tasks, each with rigorous TREC-style blind evaluation and reusable test sets.
The document retrieval task has a corpus of 3.2 million documents with 367
thousand training queries, for which we generate a reusable test set of 43
queries. The passage retrieval task has a corpus of 8.8 million passages with
503 thousand training queries, for which we generate a reusable test set of 43
queries. This year 15 groups submitted a total of 75 runs, using various
combinations of deep learning, transfer learning and traditional IR ranking
methods. Deep learning runs significantly outperformed traditional IR runs.
Possible explanations for this result are that we introduced large training
data and we included deep models trained on such data in our judging pools,
whereas some past studies did not have such training data or pooling.
Related papers
- Data Efficient Acoustic Scene Classification using Teacher-Informed Confusing Class Instruction [11.15868814062321]
Three systems are introduced to tackle training splits of different sizes.
For small training splits, we explored reducing the complexity of the provided baseline model by reducing the number of base channels.
For the larger training splits, we use FocusNet to provide confusing class information to an ensemble of multiple Patchout faSt Spectrogram Transformer (PaSST) models and baseline models trained on the original sampling rate of 44.1 kHz.
arXiv Detail & Related papers (2024-09-18T13:16:00Z) - Effective Long-Context Scaling of Foundation Models [90.57254298730923]
We present a series of long-context LLMs that support effective context windows of up to 32,768 tokens.
Our models achieve consistent improvements on most regular tasks and significant improvements on long-context tasks over Llama 2.
arXiv Detail & Related papers (2023-09-27T21:41:49Z) - DataComp: In search of the next generation of multimodal datasets [179.79323076587255]
DataComp is a testbed for dataset experiments centered around a new candidate pool of 12.8 billion image-text pairs from Common Crawl.
Our benchmark consists of multiple compute scales spanning four orders of magnitude.
In particular, our best baseline, DataComp-1B, enables training a CLIP ViT-L/14 from scratch to 79.2% zero-shot accuracy on ImageNet.
arXiv Detail & Related papers (2023-04-27T11:37:18Z) - Task-aware Retrieval with Instructions [91.87694020194316]
We study the problem of retrieval with instructions, where users of a retrieval system explicitly describe their intent along with their queries.
We present TART, a multi-task retrieval system trained on the diverse retrieval tasks with instructions.
TART shows strong capabilities to adapt to a new task via instructions and advances the state of the art on two zero-shot retrieval benchmarks.
arXiv Detail & Related papers (2022-11-16T23:13:22Z) - NEVIS'22: A Stream of 100 Tasks Sampled from 30 Years of Computer Vision
Research [96.53307645791179]
We introduce the Never-Ending VIsual-classification Stream (NEVIS'22), a benchmark consisting of a stream of over 100 visual classification tasks.
Despite being limited to classification, the resulting stream has a rich diversity of tasks from OCR, to texture analysis, scene recognition, and so forth.
Overall, NEVIS'22 poses an unprecedented challenge for current sequential learning approaches due to the scale and diversity of tasks.
arXiv Detail & Related papers (2022-11-15T18:57:46Z) - Large Scale Real-World Multi-Person Tracking [68.27438015329807]
This paper presents a new large scale multi-person tracking dataset -- textttPersonPath22.
It is over an order of magnitude larger than currently available high quality multi-object tracking datasets such as MOT17, HiEve, and MOT20.
arXiv Detail & Related papers (2022-11-03T23:03:13Z) - TREC Deep Learning Track: Reusable Test Collections in the Large Data
Regime [33.202007333667375]
This paper supports the reuse of the TREC DL test collections in three ways.
First we describe the data sets in detail, documenting clearly and in one place some details that are otherwise scattered in track guidelines.
Second, because there is some risk of iteration and selection bias when reusing a data set, we describe the best practices for writing a paper using TREC DL data, without overfitting.
arXiv Detail & Related papers (2021-04-19T15:41:28Z) - Efficiently Teaching an Effective Dense Retriever with Balanced Topic
Aware Sampling [37.01593605084575]
TAS-Balanced is an efficient topic-aware query and balanced margin sampling technique.
We show that our TAS-Balanced training method achieves state-of-the-art low-latency (64ms per query) results on two TREC Deep Learning Track query sets.
arXiv Detail & Related papers (2021-04-14T16:49:18Z) - Overview of the TREC 2020 deep learning track [30.531644711518414]
This year we have a document retrieval task and a passage retrieval task, each with hundreds of thousands of human-labeled training queries.
We evaluate using single-shot TREC-style evaluation, to give us a picture of which ranking methods work best when large data is available.
This year we have further evidence that rankers with BERT-style pretraining outperform other rankers in the large data regime.
arXiv Detail & Related papers (2021-02-15T16:47:00Z) - TREC CAsT 2019: The Conversational Assistance Track Overview [34.65827453762031]
The Conversational Assistance Track (CAsT) is a new track for TREC 2019 to facilitate Conversational Information Seeking (CIS) research.
The document corpus is 38,426,252 passages from the TREC Complex Answer Retrieval (CAR) and Microsoft MAchine Reading COmprehension (MARCO) datasets.
This year 21 groups submitted a total of 65 runs using varying methods for conversational query understanding and ranking.
arXiv Detail & Related papers (2020-03-30T16:58:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.