Refined Commonsense Knowledge from Large-Scale Web Contents
- URL: http://arxiv.org/abs/2112.04596v1
- Date: Tue, 30 Nov 2021 20:26:09 GMT
- Title: Refined Commonsense Knowledge from Large-Scale Web Contents
- Authors: Tuan-Phong Nguyen, Simon Razniewski, Julien Romero, Gerhard Weikum
- Abstract summary: Commonsense knowledge (CSK) about concepts and their properties is useful for AI applications.
This paper presents a method, called ASCENT++, to automatically build a large-scale knowledge base (KB) of CSK assertions.
- Score: 24.10708502359049
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Commonsense knowledge (CSK) about concepts and their properties is useful for
AI applications. Prior works like ConceptNet, COMET and others compiled large
CSK collections, but are restricted in their expressiveness to
subject-predicate-object (SPO) triples with simple concepts for S and strings
for P and O. This paper presents a method, called ASCENT++, to automatically
build a large-scale knowledge base (KB) of CSK assertions, with refined
expressiveness and both better precision and recall than prior works. ASCENT++
goes beyond SPO triples by capturing composite concepts with subgroups and
aspects, and by refining assertions with semantic facets. The latter is
important to express the temporal and spatial validity of assertions and
further qualifiers. ASCENT++ combines open information extraction with
judicious cleaning and ranking by typicality and saliency scores. For high
coverage, our method taps into the large-scale crawl C4 with broad web
contents. The evaluation with human judgements shows the superior quality of
the ASCENT++ KB, and an extrinsic evaluation for QA-support tasks underlines
the benefits of ASCENT++. A web interface, data and code can be accessed at
https://www.mpi-inf.mpg.de/ascentpp.
Related papers
- Discover-then-Name: Task-Agnostic Concept Bottlenecks via Automated Concept Discovery [52.498055901649025]
Concept Bottleneck Models (CBMs) have been proposed to address the 'black-box' problem of deep neural networks.
We propose a novel CBM approach -- called Discover-then-Name-CBM (DN-CBM) -- that inverts the typical paradigm.
Our concept extraction strategy is efficient, since it is agnostic to the downstream task, and uses concepts already known to the model.
arXiv Detail & Related papers (2024-07-19T17:50:11Z) - EWEK-QA: Enhanced Web and Efficient Knowledge Graph Retrieval for Citation-based Question Answering Systems [103.91826112815384]
citation-based QA systems are suffering from two shortcomings.
They usually rely only on web as a source of extracted knowledge and adding other external knowledge sources can hamper the efficiency of the system.
We propose our enhanced web and efficient knowledge graph (KG) retrieval solution (EWEK-QA) to enrich the content of the extracted knowledge fed to the system.
arXiv Detail & Related papers (2024-06-14T19:40:38Z) - CSCNET: Class-Specified Cascaded Network for Compositional Zero-Shot
Learning [62.090051975043544]
Attribute and object (A-O) disentanglement is a fundamental and critical problem for Compositional Zero-shot Learning (CZSL)
We propose a novel A-O disentangled framework for CZSL, namely Class-specified Cascaded Network (CSCNet)
arXiv Detail & Related papers (2024-03-09T14:18:41Z) - Semantic Communications for Artificial Intelligence Generated Content
(AIGC) Toward Effective Content Creation [75.73229320559996]
This paper develops a conceptual model for the integration of AIGC and SemCom.
A novel framework that employs AIGC technology is proposed as an encoder and decoder for semantic information.
The framework can adapt to different types of content generated, the required quality, and the semantic information utilized.
arXiv Detail & Related papers (2023-08-09T13:17:21Z) - CAR: Conceptualization-Augmented Reasoner for Zero-Shot Commonsense
Question Answering [56.592385613002584]
We propose Conceptualization-Augmented Reasoner (CAR) to tackle the task of zero-shot commonsense question answering.
CAR abstracts a commonsense knowledge triple to many higher-level instances, which increases the coverage of CommonSense Knowledge Bases.
CAR more robustly generalizes to answering questions about zero-shot commonsense scenarios than existing methods.
arXiv Detail & Related papers (2023-05-24T08:21:31Z) - Extracting Cultural Commonsense Knowledge at Scale [28.856786775318486]
CANDLE is an end-to-end methodology for extracting high-quality cultural commonsense knowledge at scale.
It organizes assertions into coherent clusters for 3 domains of subjects (geography, religion, occupation) and several cultural facets.
It includes judicious techniques for classification-based filtering and scoring of interestingness.
arXiv Detail & Related papers (2022-10-14T12:53:57Z) - Supporting Vision-Language Model Inference with Confounder-pruning Knowledge Prompt [71.77504700496004]
Vision-language models are pre-trained by aligning image-text pairs in a common space to deal with open-set visual concepts.
To boost the transferability of the pre-trained models, recent works adopt fixed or learnable prompts.
However, how and what prompts can improve inference performance remains unclear.
arXiv Detail & Related papers (2022-05-23T07:51:15Z) - Augmenting Pre-trained Language Models with QA-Memory for Open-Domain
Question Answering [38.071375112873675]
We propose a question-answer augmented encoder-decoder model and accompanying pretraining strategy.
This yields an end-to-end system that outperforms prior QA retrieval methods on single-hop QA tasks.
arXiv Detail & Related papers (2022-04-10T02:33:00Z) - Advanced Semantics for Commonsense Knowledge Extraction [32.43213645631101]
Commonsense knowledge (CSK) about concepts and their properties is useful for AI applications such as robust chatbots.
This paper presents a methodology, called Ascent, to automatically build a large-scale knowledge base (KB) of CSK assertions.
Ascent goes beyond triples by capturing composite concepts with subgroups and aspects, and by refining assertions with semantic facets.
arXiv Detail & Related papers (2020-11-02T11:37:17Z) - Joint Reasoning for Multi-Faceted Commonsense Knowledge [28.856786775318486]
Commonsense knowledge (CSK) supports a variety of AI applications, from visual understanding to chatbots.
Prior works on acquiring CSK have compiled statements that associate concepts, like everyday objects or activities, with properties that hold for most or some instances of the concept.
This paper introduces a multi-faceted model of CSK statements and methods for joint reasoning over sets of inter-related statements.
arXiv Detail & Related papers (2020-01-13T11:34:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.