Reciprocal Feature Learning via Explicit and Implicit Tasks in Scene
Text Recognition
- URL: http://arxiv.org/abs/2105.06229v1
- Date: Thu, 13 May 2021 12:27:35 GMT
- Title: Reciprocal Feature Learning via Explicit and Implicit Tasks in Scene
Text Recognition
- Authors: Hui Jiang and Yunlu Xu and Zhanzhan Cheng and Shiliang Pu and Yi Niu
and Wenqi Ren and Fei Wu and Wenming Tan
- Abstract summary: In this work, we excavate the implicit task, character counting within the traditional text recognition, without additional labor annotation cost.
We design a two-branch reciprocal feature learning framework in order to adequately utilize the features from both the tasks.
Experiments on 7 benchmarks show the advantages of the proposed methods in both text recognition and the new-built character counting tasks.
- Score: 60.36540008537054
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Text recognition is a popular topic for its broad applications. In this work,
we excavate the implicit task, character counting within the traditional text
recognition, without additional labor annotation cost. The implicit task plays
as an auxiliary branch for complementing the sequential recognition. We design
a two-branch reciprocal feature learning framework in order to adequately
utilize the features from both the tasks. Through exploiting the complementary
effect between explicit and implicit tasks, the feature is reliably enhanced.
Extensive experiments on 7 benchmarks show the advantages of the proposed
methods in both text recognition and the new-built character counting tasks. In
addition, it is convenient yet effective to equip with variable networks and
tasks. We offer abundant ablation studies, generalizing experiments with deeper
understanding on the tasks. Code is available.
Related papers
- Leveraging knowledge distillation for partial multi-task learning from multiple remote sensing datasets [2.1178416840822023]
Partial multi-task learning where training examples are annotated for one of the target tasks is a promising idea in remote sensing.
This paper proposes using knowledge distillation to replace the need of ground truths for the alternate task and enhance the performance of such approach.
arXiv Detail & Related papers (2024-05-24T09:48:50Z) - Task-Driven Exploration: Decoupling and Inter-Task Feedback for Joint Moment Retrieval and Highlight Detection [7.864892339833315]
We propose a novel task-driven top-down framework for joint moment retrieval and highlight detection.
The framework introduces a task-decoupled unit to capture task-specific and common representations.
Comprehensive experiments and in-depth ablation studies on QVHighlights, TVSum, and Charades-STA datasets corroborate the effectiveness and flexibility of the proposed framework.
arXiv Detail & Related papers (2024-04-14T14:06:42Z) - Sharing Knowledge in Multi-Task Deep Reinforcement Learning [57.38874587065694]
We study the benefit of sharing representations among tasks to enable the effective use of deep neural networks in Multi-Task Reinforcement Learning.
We prove this by providing theoretical guarantees that highlight the conditions for which is convenient to share representations among tasks.
arXiv Detail & Related papers (2024-01-17T19:31:21Z) - Pre-training Multi-task Contrastive Learning Models for Scientific
Literature Understanding [52.723297744257536]
Pre-trained language models (LMs) have shown effectiveness in scientific literature understanding tasks.
We propose a multi-task contrastive learning framework, SciMult, to facilitate common knowledge sharing across different literature understanding tasks.
arXiv Detail & Related papers (2023-05-23T16:47:22Z) - Pretext Tasks selection for multitask self-supervised speech
representation learning [23.39079406674442]
This paper introduces a method to select a group of pretext tasks among a set of candidates.
Experiments conducted on speaker recognition and automatic speech recognition validate our approach.
arXiv Detail & Related papers (2021-07-01T16:36:29Z) - Towards All-around Knowledge Transferring: Learning From Task-irrelevant
Labels [44.036667329736225]
Existing efforts mainly focus on transferring task-relevant knowledge from other similar data to tackle the issue.
To date, no large-scale studies have been performed to investigate the impact of task-irrelevant features.
We propose Task-Irrelevant Transfer Learning to exploit taskirrelevant features, which mainly are extracted from task-irrelevant labels.
arXiv Detail & Related papers (2020-11-17T06:43:58Z) - Auxiliary Learning by Implicit Differentiation [54.92146615836611]
Training neural networks with auxiliary tasks is a common practice for improving the performance on a main task of interest.
Here, we propose a novel framework, AuxiLearn, that targets both challenges based on implicit differentiation.
First, when useful auxiliaries are known, we propose learning a network that combines all losses into a single coherent objective function.
Second, when no useful auxiliary task is known, we describe how to learn a network that generates a meaningful, novel auxiliary task.
arXiv Detail & Related papers (2020-06-22T19:35:07Z) - A Cross-Task Analysis of Text Span Representations [52.28565379517174]
We find that the optimal span representation varies by task, and can also vary within different facets of individual tasks.
We also find that the choice of span representation has a bigger impact with a fixed pretrained encoder than with a fine-tuned encoder.
arXiv Detail & Related papers (2020-06-06T13:37:51Z) - TRIE: End-to-End Text Reading and Information Extraction for Document
Understanding [56.1416883796342]
We propose a unified end-to-end text reading and information extraction network.
multimodal visual and textual features of text reading are fused for information extraction.
Our proposed method significantly outperforms the state-of-the-art methods in both efficiency and accuracy.
arXiv Detail & Related papers (2020-05-27T01:47:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.