Multinational Address Parsing: A Zero-Shot Evaluation
- URL: http://arxiv.org/abs/2112.04008v1
- Date: Tue, 7 Dec 2021 21:40:43 GMT
- Title: Multinational Address Parsing: A Zero-Shot Evaluation
- Authors: Marouane Yassine and David Beauchemin and Fran\c{c}ois Laviolette and
Luc Lamontagne
- Abstract summary: Address parsing consists of identifying the segments that make up an address, such as a street name or a postal code.
Previous work on neural networks has only focused on parsing addresses from a single source country.
This paper explores the possibility of transferring the address parsing knowledge acquired by training deep learning models on some countries' addresses to others.
- Score: 0.3211619859724084
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Address parsing consists of identifying the segments that make up an address,
such as a street name or a postal code. Because of its importance for tasks
like record linkage, address parsing has been approached with many techniques,
the latest relying on neural networks. While these models yield notable
results, previous work on neural networks has only focused on parsing addresses
from a single source country. This paper explores the possibility of
transferring the address parsing knowledge acquired by training deep learning
models on some countries' addresses to others with no further training in a
zero-shot transfer learning setting. We also experiment using an attention
mechanism and a domain adversarial training algorithm in the same zero-shot
transfer setting to improve performance. Both methods yield state-of-the-art
performance for most of the tested countries while giving good results to the
remaining countries. We also explore the effect of incomplete addresses on our
best model, and we evaluate the impact of using incomplete addresses during
training. In addition, we propose an open-source Python implementation of some
of our trained models.
Related papers
- AddressCLIP: Empowering Vision-Language Models for City-wide Image Address Localization [57.34659640776723]
We propose an end-to-end framework named AddressCLIP to solve the problem with more semantics.
We have built three datasets from Pittsburgh and San Francisco on different scales specifically for the IAL problem.
arXiv Detail & Related papers (2024-07-11T03:18:53Z) - Methods for Matching English Language Addresses [1.2930673139458417]
We formalize a framework to generate matching and mismatching pairs of addresses in the English language.
We evaluate various methods to automatically perform address matching.
arXiv Detail & Related papers (2024-03-14T10:39:14Z) - Deepparse : An Extendable, and Fine-Tunable State-Of-The-Art Library for
Parsing Multinational Street Addresses [0.0]
This paper presents Deepparse, a Python open-source, extendable, fine-tunable address parsing solution under LGPL-3.0 licence.
It can parse addresses written in any language and use any address standard.
The library supports fine-tuning with new data to generate a custom address.
arXiv Detail & Related papers (2023-11-20T15:37:33Z) - CDFSL-V: Cross-Domain Few-Shot Learning for Videos [58.37446811360741]
Few-shot video action recognition is an effective approach to recognizing new categories with only a few labeled examples.
Existing methods in video action recognition rely on large labeled datasets from the same domain.
We propose a novel cross-domain few-shot video action recognition method that leverages self-supervised learning and curriculum learning.
arXiv Detail & Related papers (2023-09-07T19:44:27Z) - Ensemble Transfer Learning for Multilingual Coreference Resolution [60.409789753164944]
A problem that frequently occurs when working with a non-English language is the scarcity of annotated training data.
We design a simple but effective ensemble-based framework that combines various transfer learning techniques.
We also propose a low-cost TL method that bootstraps coreference resolution models by utilizing Wikipedia anchor texts.
arXiv Detail & Related papers (2023-01-22T18:22:55Z) - On Generalizing Beyond Domains in Cross-Domain Continual Learning [91.56748415975683]
Deep neural networks often suffer from catastrophic forgetting of previously learned knowledge after learning a new task.
Our proposed approach learns new tasks under domain shift with accuracy boosts up to 10% on challenging datasets such as DomainNet and OfficeHome.
arXiv Detail & Related papers (2022-03-08T09:57:48Z) - Learning from Language Description: Low-shot Named Entity Recognition
via Decomposed Framework [23.501276952950366]
We propose a novel NER framework, namely SpanNER, which learns from natural language supervision and enables the identification of never-seen entity classes.
We perform extensive experiments on 5 benchmark datasets and evaluate the proposed method in the few-shot learning, domain transfer and zero-shot learning settings.
The experimental results show that the proposed method can bring 10%, 23% and 26% improvements in average over the best baselines in few-shot learning, domain transfer and zero-shot learning settings respectively.
arXiv Detail & Related papers (2021-09-11T19:52:09Z) - Unsupervised Paraphrasing with Pretrained Language Models [85.03373221588707]
We propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting.
Our recipe consists of task-adaptation, self-supervision, and a novel decoding algorithm named Dynamic Blocking.
We show with automatic and human evaluations that our approach achieves state-of-the-art performance on both the Quora Question Pair and the ParaNMT datasets.
arXiv Detail & Related papers (2020-10-24T11:55:28Z) - Deep Contextual Embeddings for Address Classification in E-commerce [0.03222802562733786]
E-commerce customers in developing nations like India tend to follow no fixed format while entering shipping addresses.
It is imperative to understand the language of addresses, so that shipments can be routed without delays.
We propose a novel approach towards understanding customer addresses by deriving motivation from recent advances in Natural Language Processing (NLP)
arXiv Detail & Related papers (2020-07-06T19:06:34Z) - Leveraging Subword Embeddings for Multinational Address Parsing [0.0764671395172401]
We build a single model capable of learning to parse addresses from multiple countries at the same time.
We achieve accuracies around 99 % on the countries used for training with no pre-processing nor post-processing needed.
We explore the possibility of transferring the address parsing knowledge obtained by training on some countries' addresses to others with no further training in a zero-shot transfer learning setting.
arXiv Detail & Related papers (2020-06-29T16:14:27Z) - Pre-training Text Representations as Meta Learning [113.3361289756749]
We introduce a learning algorithm which directly optimize model's ability to learn text representations for effective learning of downstream tasks.
We show that there is an intrinsic connection between multi-task pre-training and model-agnostic meta-learning with a sequence of meta-train steps.
arXiv Detail & Related papers (2020-04-12T09:05:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.