A Boundary Regression Model for Nested Named Entity Recognition
- URL: http://arxiv.org/abs/2011.14330v2
- Date: Sun, 27 Dec 2020 22:09:22 GMT
- Title: A Boundary Regression Model for Nested Named Entity Recognition
- Authors: Yanping Chen, Lefei Wu, Liyuan Deng, Yongbin Qing, Ruizhang Huang,
Qinghua Zheng, Ping Chen
- Abstract summary: Recognizing named entities (NEs) is commonly conducted as a classification problem that predicts a class tag for an NE candidate in a sentence.
Recent developments in neural networks have adopted deep structures that map categorized features into continuous representations.
In this paper, the regression operation is introduced to locate NEs in a sentence.
- Score: 17.968819067122418
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recognizing named entities (NEs) is commonly conducted as a classification
problem that predicts a class tag for an NE candidate in a sentence. In shallow
structures, categorized features are weighted to support the prediction. Recent
developments in neural networks have adopted deep structures that map
categorized features into continuous representations. This approach unfolds a
dense space saturated with high-order abstract semantic information, where the
prediction is based on distributed feature representations. In this paper, the
regression operation is introduced to locate NEs in a sentence. In this
approach, a deep network is first designed to transform an input sentence into
recurrent feature maps. Bounding boxes are generated from the feature maps,
where a box is an abstract representation of an NE candidate. In addition to
the class tag, each bounding box has two parameters denoting the start position
and the length of an NE candidate. In the training process, the location offset
between a bounding box and a true NE are learned to minimize the location loss.
Based on this motivation, a multiobjective learning framework is designed to
simultaneously locate entities and predict the class probability. By sharing
parameters for locating and predicting, the framework can take full advantage
of annotated data and enable more potent nonlinear function approximators to
enhance model discriminability. Experiments demonstrate state-of-the-art
performance for nested named entities\footnote{Our codes will be available at:
\url{https://github.com/wuyuefei3/BR}}.
Related papers
- Clustering Inductive Biases with Unrolled Networks [4.47196217712431]
We propose an autoencoder architecture (WLSC) whose latent representations are implicitly, locally organized for spectral clustering through a Laplacian quadratic form of a bipartite graph.
We show that our regularization can be interpreted as early-stage specialization of receptive fields to certain classes of stimuli.
arXiv Detail & Related papers (2023-11-30T02:02:30Z) - Understanding Imbalanced Semantic Segmentation Through Neural Collapse [81.89121711426951]
We show that semantic segmentation naturally brings contextual correlation and imbalanced distribution among classes.
We introduce a regularizer on feature centers to encourage the network to learn features closer to the appealing structure.
Our method ranks 1st and sets a new record on the ScanNet200 test leaderboard.
arXiv Detail & Related papers (2023-01-03T13:51:51Z) - Refined Edge Usage of Graph Neural Networks for Edge Prediction [51.06557652109059]
We propose a novel edge prediction paradigm named Edge-aware Message PassIng neuRal nEtworks (EMPIRE)
We first introduce an edge splitting technique to specify use of each edge where each edge is solely used as either the topology or the supervision.
In order to emphasize the differences between pairs connected by supervision edges and pairs unconnected, we further weight the messages to highlight the relative ones that can reflect the differences.
arXiv Detail & Related papers (2022-12-25T23:19:56Z) - Box2Seg: Learning Semantics of 3D Point Clouds with Box-Level
Supervision [65.19589997822155]
We introduce a neural architecture, termed Box2Seg, to learn point-level semantics of 3D point clouds with bounding box-level supervision.
We show that the proposed network can be trained with cheap, or even off-the-shelf bounding box-level annotations and subcloud-level tags.
arXiv Detail & Related papers (2022-01-09T09:07:48Z) - A singular Riemannian geometry approach to Deep Neural Networks II.
Reconstruction of 1-D equivalence classes [78.120734120667]
We build the preimage of a point in the output manifold in the input space.
We focus for simplicity on the case of neural networks maps from n-dimensional real spaces to (n - 1)-dimensional real spaces.
arXiv Detail & Related papers (2021-12-17T11:47:45Z) - Index $t$-SNE: Tracking Dynamics of High-Dimensional Datasets with
Coherent Embeddings [1.7188280334580195]
This paper presents a methodology to reuse an embedding to create a new one, where cluster positions are preserved.
The proposed algorithm has the same complexity as the original $t$-SNE to embed new items, and a lower one when considering the embedding of a dataset sliced into sub-pieces.
arXiv Detail & Related papers (2021-09-22T06:45:37Z) - Towards Efficient Scene Understanding via Squeeze Reasoning [71.1139549949694]
We propose a novel framework called Squeeze Reasoning.
Instead of propagating information on the spatial map, we first learn to squeeze the input feature into a channel-wise global vector.
We show that our approach can be modularized as an end-to-end trained block and can be easily plugged into existing networks.
arXiv Detail & Related papers (2020-11-06T12:17:01Z) - Interpreting Graph Neural Networks for NLP With Differentiable Edge
Masking [63.49779304362376]
Graph neural networks (GNNs) have become a popular approach to integrating structural inductive biases into NLP models.
We introduce a post-hoc method for interpreting the predictions of GNNs which identifies unnecessary edges.
We show that we can drop a large proportion of edges without deteriorating the performance of the model.
arXiv Detail & Related papers (2020-10-01T17:51:19Z) - Supervised Topological Maps [0.76146285961466]
Controlling the internal representation space of a neural network is a desirable feature because it allows to generate new data in a supervised manner.
We will show how this can be achieved while building a low-dimensional mapping of the input stream, by deriving a generalized algorithm starting from Self Organizing Maps (SOMs)
arXiv Detail & Related papers (2020-08-14T14:30:16Z) - Explaining Predictions by Approximating the Local Decision Boundary [3.60160227126201]
We present a new procedure for local decision boundary approximation (DBA)
We train a variational autoencoder to learn a Euclidean latent space of encoded data representations.
We exploit attribute annotations to map the latent space to attributes that are meaningful to the user.
arXiv Detail & Related papers (2020-06-14T19:12:42Z) - PushNet: Efficient and Adaptive Neural Message Passing [1.9121961872220468]
Message passing neural networks have recently evolved into a state-of-the-art approach to representation learning on graphs.
Existing methods perform synchronous message passing along all edges in multiple subsequent rounds.
We consider a novel asynchronous message passing approach where information is pushed only along the most relevant edges until convergence.
arXiv Detail & Related papers (2020-03-04T18:15:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.