Leveraging Subword Embeddings for Multinational Address Parsing
- URL: http://arxiv.org/abs/2006.16152v3
- Date: Sun, 2 May 2021 14:52:14 GMT
- Title: Leveraging Subword Embeddings for Multinational Address Parsing
- Authors: Marouane Yassine, David Beauchemin, Fran\c{c}ois Laviolette, Luc
Lamontagne
- Abstract summary: We build a single model capable of learning to parse addresses from multiple countries at the same time.
We achieve accuracies around 99 % on the countries used for training with no pre-processing nor post-processing needed.
We explore the possibility of transferring the address parsing knowledge obtained by training on some countries' addresses to others with no further training in a zero-shot transfer learning setting.
- Score: 0.0764671395172401
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Address parsing consists of identifying the segments that make up an address
such as a street name or a postal code. Because of its importance for tasks
like record linkage, address parsing has been approached with many techniques.
Neural network methods defined a new state-of-the-art for address parsing.
While this approach yielded notable results, previous work has only focused on
applying neural networks to achieve address parsing of addresses from one
source country. We propose an approach in which we employ subword embeddings
and a Recurrent Neural Network architecture to build a single model capable of
learning to parse addresses from multiple countries at the same time while
taking into account the difference in languages and address formatting systems.
We achieved accuracies around 99 % on the countries used for training with no
pre-processing nor post-processing needed. We explore the possibility of
transferring the address parsing knowledge obtained by training on some
countries' addresses to others with no further training in a zero-shot transfer
learning setting. We achieve good results for 80 % of the countries (33 out of
41), almost 50 % of which (20 out of 41) is near state-of-the-art performance.
In addition, we propose an open-source Python implementation of our trained
models.
Related papers
- AddressCLIP: Empowering Vision-Language Models for City-wide Image Address Localization [57.34659640776723]
We propose an end-to-end framework named AddressCLIP to solve the problem with more semantics.
We have built three datasets from Pittsburgh and San Francisco on different scales specifically for the IAL problem.
arXiv Detail & Related papers (2024-07-11T03:18:53Z) - A Novel Cartography-Based Curriculum Learning Method Applied on RoNLI: The First Romanian Natural Language Inference Corpus [71.77214818319054]
Natural language inference is a proxy for natural language understanding.
There is no publicly available NLI corpus for the Romanian language.
We introduce the first Romanian NLI corpus (RoNLI) comprising 58K training sentence pairs.
arXiv Detail & Related papers (2024-05-20T08:41:15Z) - Methods for Matching English Language Addresses [1.2930673139458417]
We formalize a framework to generate matching and mismatching pairs of addresses in the English language.
We evaluate various methods to automatically perform address matching.
arXiv Detail & Related papers (2024-03-14T10:39:14Z) - Deepparse : An Extendable, and Fine-Tunable State-Of-The-Art Library for
Parsing Multinational Street Addresses [0.0]
This paper presents Deepparse, a Python open-source, extendable, fine-tunable address parsing solution under LGPL-3.0 licence.
It can parse addresses written in any language and use any address standard.
The library supports fine-tuning with new data to generate a custom address.
arXiv Detail & Related papers (2023-11-20T15:37:33Z) - CDFSL-V: Cross-Domain Few-Shot Learning for Videos [58.37446811360741]
Few-shot video action recognition is an effective approach to recognizing new categories with only a few labeled examples.
Existing methods in video action recognition rely on large labeled datasets from the same domain.
We propose a novel cross-domain few-shot video action recognition method that leverages self-supervised learning and curriculum learning.
arXiv Detail & Related papers (2023-09-07T19:44:27Z) - Meta-Learning a Cross-lingual Manifold for Semantic Parsing [75.26271012018861]
Localizing a semantic to support new languages requires effective cross-lingual generalization.
We introduce a first-order meta-learning algorithm to train a semantic annotated with maximal sample efficiency during cross-lingual transfer.
Results across six languages on ATIS demonstrate that our combination of steps yields accurate semantics sampling $le$10% of source training data in each new language.
arXiv Detail & Related papers (2022-09-26T10:42:17Z) - Multinational Address Parsing: A Zero-Shot Evaluation [0.3211619859724084]
Address parsing consists of identifying the segments that make up an address, such as a street name or a postal code.
Previous work on neural networks has only focused on parsing addresses from a single source country.
This paper explores the possibility of transferring the address parsing knowledge acquired by training deep learning models on some countries' addresses to others.
arXiv Detail & Related papers (2021-12-07T21:40:43Z) - Unsupervised Paraphrasing with Pretrained Language Models [85.03373221588707]
We propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting.
Our recipe consists of task-adaptation, self-supervision, and a novel decoding algorithm named Dynamic Blocking.
We show with automatic and human evaluations that our approach achieves state-of-the-art performance on both the Quora Question Pair and the ParaNMT datasets.
arXiv Detail & Related papers (2020-10-24T11:55:28Z) - Deep Contextual Embeddings for Address Classification in E-commerce [0.03222802562733786]
E-commerce customers in developing nations like India tend to follow no fixed format while entering shipping addresses.
It is imperative to understand the language of addresses, so that shipments can be routed without delays.
We propose a novel approach towards understanding customer addresses by deriving motivation from recent advances in Natural Language Processing (NLP)
arXiv Detail & Related papers (2020-07-06T19:06:34Z) - Exploring the Limits of Transfer Learning with a Unified Text-to-Text
Transformer [64.22926988297685]
Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on a downstream task, has emerged as a powerful technique in natural language processing (NLP)
In this paper, we explore the landscape of introducing transfer learning techniques for NLP by a unified framework that converts all text-based language problems into a text-to-text format.
arXiv Detail & Related papers (2019-10-23T17:37:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.