A Study of the Tasks and Models in Machine Reading Comprehension
- URL: http://arxiv.org/abs/2001.08635v1
- Date: Thu, 23 Jan 2020 16:11:44 GMT
- Title: A Study of the Tasks and Models in Machine Reading Comprehension
- Authors: Chao Wang
- Abstract summary: This report reviews some representative simple-reasoning and complex-reasoning MRC tasks.
It also proposes some open problems for the future research.
- Score: 3.6985039575807246
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To provide a survey on the existing tasks and models in Machine Reading
Comprehension (MRC), this report reviews: 1) the dataset collection and
performance evaluation of some representative simple-reasoning and
complex-reasoning MRC tasks; 2) the architecture designs, attention mechanisms,
and performance-boosting approaches for developing neural-network-based MRC
models; 3) some recently proposed transfer learning approaches to incorporating
text-style knowledge contained in external corpora into the neural networks of
MRC models; 4) some recently proposed knowledge base encoding approaches to
incorporating graph-style knowledge contained in external knowledge bases into
the neural networks of MRC models. Besides, according to what has been achieved
and what are still deficient, this report also proposes some open problems for
the future research.
Related papers
- Retrieval-Enhanced Machine Learning: Synthesis and Opportunities [60.34182805429511]
Retrieval-enhancement can be extended to a broader spectrum of machine learning (ML)
This work introduces a formal framework of this paradigm, Retrieval-Enhanced Machine Learning (REML), by synthesizing the literature in various domains in ML with consistent notations which is missing from the current literature.
The goal of this work is to equip researchers across various disciplines with a comprehensive, formally structured framework of retrieval-enhanced models, thereby fostering interdisciplinary future research.
arXiv Detail & Related papers (2024-07-17T20:01:21Z) - Masked Modeling for Self-supervised Representation Learning on Vision
and Beyond [69.64364187449773]
Masked modeling has emerged as a distinctive approach that involves predicting parts of the original data that are proportionally masked during training.
We elaborate on the details of techniques within masked modeling, including diverse masking strategies, recovering targets, network architectures, and more.
We conclude by discussing the limitations of current techniques and point out several potential avenues for advancing masked modeling research.
arXiv Detail & Related papers (2023-12-31T12:03:21Z) - Rule-Extraction Methods From Feedforward Neural Networks: A Systematic
Literature Review [1.1510009152620668]
Rules offer a transparent and intuitive means of explaining neural networks.
The study specifically addresses feedforward networks with supervised learning and crisp rules.
Future work can extend to other network types, machine learning methods, and fuzzy rule extraction.
arXiv Detail & Related papers (2023-12-20T09:40:07Z) - Relational Concept Bottleneck Models [13.311396882130033]
Concept Bottleneck Models (CBMs) are not designed to solve problems.
R-CBMs are capable of both representing standard CBMs and relational GNNs.
In particular, we show that R-CBMs support the generation of concept-based explanations.
arXiv Detail & Related papers (2023-08-23T08:25:33Z) - Information Screening whilst Exploiting! Multimodal Relation Extraction
with Feature Denoising and Multimodal Topic Modeling [96.75821232222201]
Existing research on multimodal relation extraction (MRE) faces two co-existing challenges, internal-information over-utilization and external-information under-exploitation.
We propose a novel framework that simultaneously implements the idea of internal-information screening and external-information exploiting.
arXiv Detail & Related papers (2023-05-19T14:56:57Z) - Retrieval-Enhanced Machine Learning [110.5237983180089]
We describe a generic retrieval-enhanced machine learning framework, which includes a number of existing models as special cases.
REML challenges information retrieval conventions, presenting opportunities for novel advances in core areas, including optimization.
REML research agenda lays a foundation for a new style of information access research and paves a path towards advancing machine learning and artificial intelligence.
arXiv Detail & Related papers (2022-05-02T21:42:45Z) - Entity-Conditioned Question Generation for Robust Attention Distribution
in Neural Information Retrieval [51.53892300802014]
We show that supervised neural information retrieval models are prone to learning sparse attention patterns over passage tokens.
Using a novel targeted synthetic data generation method, we teach neural IR to attend more uniformly and robustly to all entities in a given passage.
arXiv Detail & Related papers (2022-04-24T22:36:48Z) - MOOCRep: A Unified Pre-trained Embedding of MOOC Entities [4.0963355240233446]
We propose to learn pre-trained representations of MOOC entities using abundant unlabeled data from the structure of MOOCs.
Our experiments reveal that MOOCRep's embeddings outperform state-of-the-art representation learning methods on two tasks important for education community.
arXiv Detail & Related papers (2021-07-12T00:11:25Z) - Deep brain state classification of MEG data [2.9048924265579124]
This paper uses Magnetoencephalography (MEG) data, provided by the Human Connectome Project (HCP), in combination with various deep artificial neural network models to perform brain decoding.
arXiv Detail & Related papers (2020-07-02T05:51:57Z) - Closed Loop Neural-Symbolic Learning via Integrating Neural Perception,
Grammar Parsing, and Symbolic Reasoning [134.77207192945053]
Prior methods learn the neural-symbolic models using reinforcement learning approaches.
We introduce the textbfgrammar model as a textitsymbolic prior to bridge neural perception and symbolic reasoning.
We propose a novel textbfback-search algorithm which mimics the top-down human-like learning procedure to propagate the error.
arXiv Detail & Related papers (2020-06-11T17:42:49Z) - Dynamic Knowledge embedding and tracing [18.717482292051788]
We propose a novel approach to knowledge tracing that combines techniques from matrix factorization with recent progress in recurrent neural networks (RNNs)
The proposed emphDynEmb framework enables the tracking of student knowledge even without the concept/skill tag information.
arXiv Detail & Related papers (2020-05-18T21:56:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.