Semi-supervised URL Segmentation with Recurrent Neural Networks
Pre-trained on Knowledge Graph Entities
- URL: http://arxiv.org/abs/2011.03138v1
- Date: Thu, 5 Nov 2020 23:31:00 GMT
- Title: Semi-supervised URL Segmentation with Recurrent Neural Networks
Pre-trained on Knowledge Graph Entities
- Authors: Hao Zhang and Jae Ro and Richard Sproat
- Abstract summary: We show effectiveness of a tagging model based on Recurrent Neural Networks (RNNs) using characters as input.
To compensate for the lack of training data, we propose a pre-training method on synthesisd entity names in a large knowledge database.
- Score: 8.855143852360328
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Breaking domain names such as openresearch into component words open and
research is important for applications like Text-to-Speech synthesis and web
search. We link this problem to the classic problem of Chinese word
segmentation and show the effectiveness of a tagging model based on Recurrent
Neural Networks (RNNs) using characters as input. To compensate for the lack of
training data, we propose a pre-training method on concatenated entity names in
a large knowledge database. Pre-training improves the model by 33% and brings
the sequence accuracy to 85%.
Related papers
- Disambiguation of Company names via Deep Recurrent Networks [101.90357454833845]
We propose a Siamese LSTM Network approach to extract -- via supervised learning -- an embedding of company name strings.
We analyse how an Active Learning approach to prioritise the samples to be labelled leads to a more efficient overall learning pipeline.
arXiv Detail & Related papers (2023-03-07T15:07:57Z) - Language Semantics Interpretation with an Interaction-based Recurrent
Neural Networks [0.0]
This paper proposes a novel influence score (I-score), a greedy search algorithm called Backward Dropping Algorithm (BDA), and a novel feature engineering technique called the "dagger technique"
The proposed methods are applied to improve prediction performance with an 81% error reduction comparing with other popular peers.
arXiv Detail & Related papers (2021-11-02T00:39:21Z) - On Improving Deep Learning Trace Analysis with System Call Arguments [1.3299507495084417]
Kernel traces are sequences of low-level events comprising a name and multiple arguments.
We introduce a general approach to learning a representation of the event names along with their arguments using both embedding and encoding.
arXiv Detail & Related papers (2021-03-11T19:26:34Z) - A Novel Deep Learning Method for Textual Sentiment Analysis [3.0711362702464675]
This paper proposes a convolutional neural network integrated with a hierarchical attention layer to extract informative words.
The proposed model has higher classification accuracy and can extract informative words.
Applying incremental transfer learning can significantly enhance the classification performance.
arXiv Detail & Related papers (2021-02-23T12:11:36Z) - R$^2$-Net: Relation of Relation Learning Network for Sentence Semantic
Matching [58.72111690643359]
We propose a Relation of Relation Learning Network (R2-Net) for sentence semantic matching.
We first employ BERT to encode the input sentences from a global perspective.
Then a CNN-based encoder is designed to capture keywords and phrase information from a local perspective.
To fully leverage labels for better relation information extraction, we introduce a self-supervised relation of relation classification task.
arXiv Detail & Related papers (2020-12-16T13:11:30Z) - CMT in TREC-COVID Round 2: Mitigating the Generalization Gaps from Web
to Special Domain Search [89.48123965553098]
This paper presents a search system to alleviate the special domain adaption problem.
The system utilizes the domain-adaptive pretraining and few-shot learning technologies to help neural rankers mitigate the domain discrepancy.
Our system performs the best among the non-manual runs in Round 2 of the TREC-COVID task.
arXiv Detail & Related papers (2020-11-03T09:10:48Z) - Named Entity Recognition for Social Media Texts with Semantic
Augmentation [70.44281443975554]
Existing approaches for named entity recognition suffer from data sparsity problems when conducted on short and informal texts.
We propose a neural-based approach to NER for social media texts where both local (from running text) and augmented semantics are taken into account.
arXiv Detail & Related papers (2020-10-29T10:06:46Z) - Learning to Match Jobs with Resumes from Sparse Interaction Data using
Multi-View Co-Teaching Network [83.64416937454801]
Job-resume interaction data is sparse and noisy, which affects the performance of job-resume match algorithms.
We propose a novel multi-view co-teaching network from sparse interaction data for job-resume matching.
Our model is able to outperform state-of-the-art methods for job-resume matching.
arXiv Detail & Related papers (2020-09-25T03:09:54Z) - Improving Joint Layer RNN based Keyphrase Extraction by Using
Syntactical Features [0.6724914680904501]
We propose to modify the input layer of JRNN to extract more than one sequence of keywords.
Our method achieved.9597 in accuracy and.7691 in F1.
arXiv Detail & Related papers (2020-09-15T14:20:04Z) - Wake Word Detection with Alignment-Free Lattice-Free MMI [66.12175350462263]
Always-on spoken language interfaces, e.g. personal digital assistants, rely on a wake word to start processing spoken input.
We present novel methods to train a hybrid DNN/HMM wake word detection system from partially labeled training data.
We evaluate our methods on two real data sets, showing 50%--90% reduction in false rejection rates at pre-specified false alarm rates over the best previously published figures.
arXiv Detail & Related papers (2020-05-17T19:22:25Z) - TNT-KID: Transformer-based Neural Tagger for Keyword Identification [7.91883337742071]
We present a novel algorithm for keyword identification called Transformer-based Neural Tagger for Keyword IDentification (TNT-KID)
By adapting the transformer architecture for a specific task at hand and leveraging language model pretraining on a domain specific corpus, the model is capable of overcoming deficiencies of both supervised and unsupervised state-of-the-art approaches to keyword extraction.
arXiv Detail & Related papers (2020-03-20T09:55:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.