Automated Discovery of Mathematical Definitions in Text with Deep Neural
Networks
- URL: http://arxiv.org/abs/2011.04521v1
- Date: Mon, 9 Nov 2020 15:57:53 GMT
- Title: Automated Discovery of Mathematical Definitions in Text with Deep Neural
Networks
- Authors: Natalia Vanetik, Marina Litvak, Sergey Shevchuk, and Lior Reznik
- Abstract summary: This paper focuses on automatic detection of one-sentence definitions in mathematical texts.
We apply deep learning methods such as the Convolutional Neural Network (CNN) and the Long Short-Term Memory network (LSTM)
We also present a new dataset for definition extraction from mathematical texts.
- Score: 6.172021438837204
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automatic definition extraction from texts is an important task that has
numerous applications in several natural language processing fields such as
summarization, analysis of scientific texts, automatic taxonomy generation,
ontology generation, concept identification, and question answering. For
definitions that are contained within a single sentence, this problem can be
viewed as a binary classification of sentences into definitions and
non-definitions. In this paper, we focus on automatic detection of one-sentence
definitions in mathematical texts, which are difficult to separate from
surrounding text. We experiment with several data representations, which
include sentence syntactic structure and word embeddings, and apply deep
learning methods such as the Convolutional Neural Network (CNN) and the Long
Short-Term Memory network (LSTM), in order to identify mathematical
definitions. Our experiments demonstrate the superiority of CNN and its
combination with LSTM, when applied on the syntactically-enriched input
representation. We also present a new dataset for definition extraction from
mathematical texts. We demonstrate that this dataset is beneficial for training
supervised models aimed at extraction of mathematical definitions. Our
experiments with different domains demonstrate that mathematical definitions
require special treatment, and that using cross-domain learning is inefficient
for that task.
Related papers
- Complex Mathematical Symbol Definition Structures: A Dataset and Model
for Coordination Resolution in Definition Extraction [27.896132821710783]
We present SymDef, an English language dataset of 5,927 sentences from full-text scientific papers.
This dataset focuses specifically on complex coordination structures such as "respectively" constructions.
We introduce a new definition extraction method that masks mathematical symbols, creates a copy of each sentence for each symbol, specifies a target symbol, and predicts its corresponding definition spans using slot filling.
arXiv Detail & Related papers (2023-05-24T02:53:48Z) - Textual Entailment Recognition with Semantic Features from Empirical
Text Representation [60.31047947815282]
A text entails a hypothesis if and only if the true value of the hypothesis follows the text.
In this paper, we propose a novel approach to identifying the textual entailment relationship between text and hypothesis.
We employ an element-wise Manhattan distance vector-based feature that can identify the semantic entailment relationship between the text-hypothesis pair.
arXiv Detail & Related papers (2022-10-18T10:03:51Z) - TeKo: Text-Rich Graph Neural Networks with External Knowledge [75.91477450060808]
We propose a novel text-rich graph neural network with external knowledge (TeKo)
We first present a flexible heterogeneous semantic network that incorporates high-quality entities.
We then introduce two types of external knowledge, that is, structured triplets and unstructured entity description.
arXiv Detail & Related papers (2022-06-15T02:33:10Z) - Contextualized Semantic Distance between Highly Overlapped Texts [85.1541170468617]
Overlapping frequently occurs in paired texts in natural language processing tasks like text editing and semantic similarity evaluation.
This paper aims to address the issue with a mask-and-predict strategy.
We take the words in the longest common sequence as neighboring words and use masked language modeling (MLM) to predict the distributions on their positions.
Experiments on Semantic Textual Similarity show NDD to be more sensitive to various semantic differences, especially on highly overlapped paired texts.
arXiv Detail & Related papers (2021-10-04T03:59:15Z) - EDS-MEMBED: Multi-sense embeddings based on enhanced distributional
semantic structures via a graph walk over word senses [0.0]
We leverage the rich semantic structures in WordNet to enhance the quality of multi-sense embeddings.
We derive new distributional semantic similarity measures for M-SE from prior ones.
We report evaluation results on 11 benchmark datasets involving WSD and Word Similarity tasks.
arXiv Detail & Related papers (2021-02-27T14:36:55Z) - Accelerating Text Mining Using Domain-Specific Stop Word Lists [57.76576681191192]
We present a novel approach for the automatic extraction of domain-specific words called the hyperplane-based approach.
The hyperplane-based approach can significantly reduce text dimensionality by eliminating irrelevant features.
Results indicate that the hyperplane-based approach can reduce the dimensionality of the corpus by 90% and outperforms mutual information.
arXiv Detail & Related papers (2020-11-18T17:42:32Z) - The geometry of integration in text classification RNNs [20.76659136484842]
We study recurrent networks trained on a battery of both natural and synthetic text classification tasks.
We find the dynamics of these trained RNNs to be both interpretable and low-dimensional.
Our observations span multiple architectures and datasets, reflecting a common mechanism RNNs employ to perform text classification.
arXiv Detail & Related papers (2020-10-28T17:58:53Z) - Context based Text-generation using LSTM networks [0.5330240017302621]
The proposed model is trained to generate text for a given set of input words along with a context vector.
The results are evaluated based on the semantic closeness of the generated text to the given context.
arXiv Detail & Related papers (2020-04-30T18:39:25Z) - Towards Accurate Scene Text Recognition with Semantic Reasoning Networks [52.86058031919856]
We propose a novel end-to-end trainable framework named semantic reasoning network (SRN) for accurate scene text recognition.
GSRM is introduced to capture global semantic context through multi-way parallel transmission.
Results on 7 public benchmarks, including regular text, irregular text and non-Latin long text, verify the effectiveness and robustness of the proposed method.
arXiv Detail & Related papers (2020-03-27T09:19:25Z) - Distributional semantic modeling: a revised technique to train term/word
vector space models applying the ontology-related approach [36.248702416150124]
We design a new technique for the distributional semantic modeling with a neural network-based approach to learn distributed term representations (or term embeddings)
Vec2graph is a Python library for visualizing word embeddings (term embeddings in our case) as dynamic and interactive graphs.
arXiv Detail & Related papers (2020-03-06T18:27:39Z) - Learn to Predict Sets Using Feed-Forward Neural Networks [63.91494644881925]
This paper addresses the task of set prediction using deep feed-forward neural networks.
We present a novel approach for learning to predict sets with unknown permutation and cardinality.
We demonstrate the validity of our set formulations on relevant vision problems.
arXiv Detail & Related papers (2020-01-30T01:52:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.