MixingBoard: a Knowledgeable Stylized Integrated Text Generation
Platform
- URL: http://arxiv.org/abs/2005.08365v2
- Date: Thu, 2 Jul 2020 18:40:26 GMT
- Title: MixingBoard: a Knowledgeable Stylized Integrated Text Generation
Platform
- Authors: Xiang Gao, Michel Galley, Bill Dolan
- Abstract summary: MixingBoard is a platform for building demos with a focus on knowledge grounded stylized text generation.
A user interface for local development, remote access, a webpage API are provided to make it simple for users to build their own demos.
- Score: 32.50773822686633
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present MixingBoard, a platform for quickly building demos with a focus on
knowledge grounded stylized text generation. We unify existing text generation
algorithms in a shared codebase and further adapt earlier algorithms for
constrained generation. To borrow advantages from different models, we
implement strategies for cross-model integration, from the token probability
level to the latent space level. An interface to external knowledge is provided
via a module that retrieves on-the-fly relevant knowledge from passages on the
web or any document collection. A user interface for local development, remote
webpage access, and a RESTful API are provided to make it simple for users to
build their own demos.
Related papers
- Unitxt: Flexible, Shareable and Reusable Data Preparation and Evaluation
for Generative AI [15.220987187105607]
Unitxt is an innovative library for customizable textual data preparation and evaluation tailored to generative language models.
Unitxt integrates with common libraries like HFace and LM-eval-harness, enabling easy customization and sharing between practitioners.
Beyond being a tool, Unitxt is a community-driven platform, empowering users to build, share, and advance their pipelines.
arXiv Detail & Related papers (2024-01-25T08:57:33Z) - REST: Retrieval-Based Speculative Decoding [69.06115086237207]
We introduce Retrieval-Based Speculative Decoding (REST), a novel algorithm designed to speed up language model generation.
Unlike previous methods that rely on a draft language model for speculative decoding, REST harnesses the power of retrieval to generate draft tokens.
When benchmarked on 7B and 13B language models in a single-batch setting, REST achieves a significant speedup of 1.62X to 2.36X on code or text generation.
arXiv Detail & Related papers (2023-11-14T15:43:47Z) - KETM:A Knowledge-Enhanced Text Matching method [0.0]
We introduce a new model for text matching called the Knowledge Enhanced Text Matching model (KETM)
We use Wiktionary to retrieve the text word definitions as our external knowledge.
We fuse text and knowledge using a gating mechanism to learn the ratio of text and knowledge fusion.
arXiv Detail & Related papers (2023-08-11T17:08:14Z) - TextFormer: A Query-based End-to-End Text Spotter with Mixed Supervision [61.186488081379]
We propose TextFormer, a query-based end-to-end text spotter with Transformer architecture.
TextFormer builds upon an image encoder and a text decoder to learn a joint semantic understanding for multi-task modeling.
It allows for mutual training and optimization of classification, segmentation, and recognition branches, resulting in deeper feature sharing.
arXiv Detail & Related papers (2023-06-06T03:37:41Z) - Text-Blueprint: An Interactive Platform for Plan-based Conditional
Generation [84.95981645040281]
Planning can be a useful intermediate step to render conditional generation less opaque and more grounded.
We present a web browser-based demonstration for query-focused summarization that uses a sequence of question-answer pairs.
arXiv Detail & Related papers (2023-04-28T18:14:48Z) - Open Domain Question Answering over Virtual Documents: A Unified
Approach for Data and Text [62.489652395307914]
We use the data-to-text method as a means for encoding structured knowledge for knowledge-intensive applications, i.e. open-domain question answering (QA)
Specifically, we propose a verbalizer-retriever-reader framework for open-domain QA over data and text where verbalized tables from Wikipedia and triples from Wikidata are used as augmented knowledge sources.
We show that our Unified Data and Text QA, UDT-QA, can effectively benefit from the expanded knowledge index, leading to large gains over text-only baselines.
arXiv Detail & Related papers (2021-10-16T00:11:21Z) - DomiKnowS: A Library for Integration of Symbolic Domain Knowledge in
Deep Learning [12.122347427933637]
We demonstrate a library for the integration of domain knowledge in deep learning architectures.
Using this library, the structure of the data is expressed symbolically via graph declarations.
The domain knowledge can be defined explicitly, which improves the models' explainability.
arXiv Detail & Related papers (2021-08-27T16:06:42Z) - VX2TEXT: End-to-End Learning of Video-Based Text Generation From
Multimodal Inputs [103.99315770490163]
We present a framework for text generation from multimodal inputs consisting of video plus text, speech, or audio.
Experiments demonstrate that our approach based on a single architecture outperforms the state-of-the-art on three video-based text-generation tasks.
arXiv Detail & Related papers (2021-01-28T15:22:36Z) - A New Perspective for Flexible Feature Gathering in Scene Text
Recognition Via Character Anchor Pooling [32.82620509088932]
We propose a pair of coupling modules, termed as Character Anchoring Module (CAM) and Anchor Pooling Module (APM)
CAM localizes the text in a shape-insensitive way by design by anchoring characters individually. APM then interpolates and gathers features flexibly along the character anchors which enables sequence learning.
arXiv Detail & Related papers (2020-02-10T03:01:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.