Record Deduplication for Entity Distribution Modeling in ASR Transcripts
- URL: http://arxiv.org/abs/2306.06246v1
- Date: Fri, 9 Jun 2023 20:42:11 GMT
- Title: Record Deduplication for Entity Distribution Modeling in ASR Transcripts
- Authors: Tianyu Huang, Chung Hoon Hong, Carl Wivagg, Kanna Shimizu
- Abstract summary: We use record deduplication to retrieve 95% of misrecognized entities.
When used for contextual biasing, our method shows an estimated 5% relative word error rate reduction.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Voice digital assistants must keep up with trending search queries. We rely
on a speech recognition model using contextual biasing with a rapidly updated
set of entities, instead of frequent model retraining, to keep up with trends.
There are several challenges with this approach: (1) the entity set must be
frequently reconstructed, (2) the entity set is of limited size due to latency
and accuracy trade-offs, and (3) finding the true entity distribution for
biasing is complicated by ASR misrecognition. We address these challenges and
define an entity set by modeling customers true requested entity distribution
from ASR output in production using record deduplication, a technique from the
field of entity resolution. Record deduplication resolves or deduplicates
coreferences, including misrecognitions, of the same latent entity. Our method
successfully retrieves 95% of misrecognized entities and when used for
contextual biasing shows an estimated 5% relative word error rate reduction.
Related papers
- List-aware Reranking-Truncation Joint Model for Search and
Retrieval-augmented Generation [80.12531449946655]
We propose a Reranking-Truncation joint model (GenRT) that can perform the two tasks concurrently.
GenRT integrates reranking and truncation via generative paradigm based on encoder-decoder architecture.
Our method achieves SOTA performance on both reranking and truncation tasks for web search and retrieval-augmented LLMs.
arXiv Detail & Related papers (2024-02-05T06:52:53Z) - From Chaos to Clarity: Claim Normalization to Empower Fact-Checking [57.024192702939736]
Claim Normalization (aka ClaimNorm) aims to decompose complex and noisy social media posts into more straightforward and understandable forms.
We propose CACN, a pioneering approach that leverages chain-of-thought and claim check-worthiness estimation.
Our experiments demonstrate that CACN outperforms several baselines across various evaluation measures.
arXiv Detail & Related papers (2023-10-22T16:07:06Z) - Single-Stage Visual Relationship Learning using Conditional Queries [60.90880759475021]
TraCQ is a new formulation for scene graph generation that avoids the multi-task learning problem and the entity pair distribution.
We employ a DETR-based encoder-decoder conditional queries to significantly reduce the entity label space as well.
Experimental results show that TraCQ not only outperforms existing single-stage scene graph generation methods, it also beats many state-of-the-art two-stage methods on the Visual Genome dataset.
arXiv Detail & Related papers (2023-06-09T06:02:01Z) - GPT Self-Supervision for a Better Data Annotator [22.598300095822026]
We propose a Generative Pretrained Transformer (GPT) self-supervision annotation method.
The proposed approach comprises a one-shot tuning phase followed by a generation phase.
The alignment score between the recovered and original data serves as a self-supervision navigator to refine the process.
arXiv Detail & Related papers (2023-06-07T11:33:14Z) - WiCE: Real-World Entailment for Claims in Wikipedia [63.234352061821625]
We propose WiCE, a new fine-grained textual entailment dataset built on natural claim and evidence pairs extracted from Wikipedia.
In addition to standard claim-level entailment, WiCE provides entailment judgments over sub-sentence units of the claim.
We show that real claims in our dataset involve challenging verification and retrieval problems that existing models fail to address.
arXiv Detail & Related papers (2023-03-02T17:45:32Z) - Improving Factual Consistency in Summarization with Compression-Based
Post-Editing [146.24839415743358]
We show that a model-agnostic way to address this problem is post-editing the generated summaries.
We propose to use sentence-compression data to train the post-editing model to take a summary with extrinsic entity errors marked with special tokens.
We show that this model improves factual consistency while maintaining ROUGE, improving entity precision by up to 30% on XSum, and that this model can be applied on top of another post-editor.
arXiv Detail & Related papers (2022-11-11T13:35:38Z) - Space-Efficient Representation of Entity-centric Query Language Models [8.712427362992237]
We introduce a deterministic approximation to probabilistic grammars that avoids the explicit expansion of non-terminals at model creation time.
We obtain a 10% relative word error rate improvement on long tail entity queries compared to when a similarly-sized n-gram model is used.
arXiv Detail & Related papers (2022-06-29T19:59:50Z) - Factual Error Correction for Abstractive Summaries Using Entity
Retrieval [57.01193722520597]
We propose an efficient factual error correction system RFEC based on entities retrieval post-editing process.
RFEC retrieves the evidence sentences from the original document by comparing the sentences with the target summary.
Next, RFEC detects the entity-level errors in the summaries by considering the evidence sentences and substitutes the wrong entities with the accurate entities from the evidence sentences.
arXiv Detail & Related papers (2022-04-18T11:35:02Z) - Automatically Generating Counterfactuals for Relation Exaction [18.740447044960796]
relation extraction (RE) is a fundamental task in natural language processing.
Current deep neural models have achieved high accuracy but are easily affected by spurious correlations.
We develop a novel approach to derive contextual counterfactuals for entities.
arXiv Detail & Related papers (2022-02-22T04:46:10Z) - A Discriminative Entity-Aware Language Model for Virtual Assistants [4.2854663014000876]
High-quality automatic speech recognition (ASR) is essential for virtual assistants (VAs) to work well.
In this work, we start from the observation that many ASR errors on named entities are inconsistent with real-world knowledge.
We extend previous discriminative n-gram language modeling approaches to incorporate real-world knowledge from a Knowledge Graph.
arXiv Detail & Related papers (2021-06-21T17:50:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.