Expedition: A System for the Unsupervised Learning of a Hierarchy of
Concepts
- URL: http://arxiv.org/abs/2112.09348v1
- Date: Fri, 17 Dec 2021 06:49:18 GMT
- Title: Expedition: A System for the Unsupervised Learning of a Hierarchy of
Concepts
- Authors: Omid Madani
- Abstract summary: We present a system for bottom-up cumulative learning of myriad concepts corresponding to meaningful character strings.
The learning is self-supervised in that the concepts discovered are used as predictors as well as targets of prediction.
We devise an objective for segmenting with the learned concepts, derived from comparing to a baseline prediction system.
- Score: 0.522145960878624
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a system for bottom-up cumulative learning of myriad concepts
corresponding to meaningful character strings, and their part-related and
prediction edges. The learning is self-supervised in that the concepts
discovered are used as predictors as well as targets of prediction. We devise
an objective for segmenting with the learned concepts, derived from comparing
to a baseline prediction system, that promotes making and using larger
concepts, which in turn allows for predicting larger spans of text, and we
describe a simple technique to promote exploration, i.e. trying out newly
generated concepts in the segmentation process. We motivate and explain a
layering of the concepts, to help separate the (conditional) distributions
learnt among concepts. The layering of the concepts roughly corresponds to a
part-whole concept hierarchy. With rudimentary segmentation and learning
algorithms, the system is promising in that it acquires many concepts (tens of
thousands in our small-scale experiments), and it learns to segment text well:
when fed with English text with spaces removed, starting at the character
level, much of what is learned respects word or phrase boundaries, and over
time the average number of "bad" splits within segmentations, i.e. splits
inside words, decreases as larger concepts are discovered and the system learns
when to use them during segmentation. We report on promising experiments when
the input text is converted to binary and the system begins with only two
concepts, "0" and "1". The system is transparent, in the sense that it is easy
to tell what the concepts learned correspond to, and which ones are active in a
segmentation, or how the system "sees" its input. We expect this framework to
be extensible and we discuss the current limitations and a number of directions
for enhancing the learning and inference capabilities.
Related papers
- Explaining Explainability: Understanding Concept Activation Vectors [35.37586279472797]
Recent interpretability methods propose using concept-based explanations to translate internal representations of deep learning models into a language that humans are familiar with: concepts.
This requires understanding which concepts are present in the representation space of a neural network.
In this work, we investigate three properties of Concept Activation Vectors (CAVs), which are learnt using a probe dataset of concept exemplars.
We introduce tools designed to detect the presence of these properties, provide insight into how they affect the derived explanations, and provide recommendations to minimise their impact.
arXiv Detail & Related papers (2024-04-04T17:46:20Z) - Rewrite Caption Semantics: Bridging Semantic Gaps for
Language-Supervised Semantic Segmentation [100.81837601210597]
We propose Concept Curation (CoCu) to bridge the gap between visual and textual semantics in pre-training data.
CoCu achieves superb zero-shot transfer performance and greatly boosts language-supervised segmentation baseline by a large margin.
arXiv Detail & Related papers (2023-09-24T00:05:39Z) - Towards Open Vocabulary Learning: A Survey [146.90188069113213]
Deep neural networks have made impressive advancements in various core tasks like segmentation, tracking, and detection.
Recently, open vocabulary settings were proposed due to the rapid progress of vision language pre-training.
This paper provides a thorough review of open vocabulary learning, summarizing and analyzing recent developments in the field.
arXiv Detail & Related papers (2023-06-28T02:33:06Z) - Efficient Induction of Language Models Via Probabilistic Concept
Formation [13.632454840363916]
We present a novel approach to the acquisition of language models from corpora.
The framework builds on Cobweb, an early system for constructing taxonomic hierarchies of probabilistic concepts.
We explore three new extensions to Cobweb -- the Word, Leaf, and Path variants.
arXiv Detail & Related papers (2022-12-22T18:16:58Z) - Visual Superordinate Abstraction for Robust Concept Learning [80.15940996821541]
Concept learning constructs visual representations that are connected to linguistic semantics.
We ascribe the bottleneck to a failure of exploring the intrinsic semantic hierarchy of visual concepts.
We propose a visual superordinate abstraction framework for explicitly modeling semantic-aware visual subspaces.
arXiv Detail & Related papers (2022-05-28T14:27:38Z) - SegDiscover: Visual Concept Discovery via Unsupervised Semantic
Segmentation [29.809900593362844]
SegDiscover is a novel framework that discovers semantically meaningful visual concepts from imagery datasets with complex scenes without supervision.
Our method generates concept primitives from raw images, discovering concepts by clustering in the latent space of a self-supervised pretrained encoder, and concept refinement via neural network smoothing.
arXiv Detail & Related papers (2022-04-22T20:44:42Z) - Discovering Concepts in Learned Representations using Statistical
Inference and Interactive Visualization [0.76146285961466]
Concept discovery is important for bridging the gap between non-deep learning experts and model end-users.
Current approaches include hand-crafting concept datasets and then converting them to latent space directions.
In this study, we offer another two approaches to guide user discovery of meaningful concepts, one based on multiple hypothesis testing, and another on interactive visualization.
arXiv Detail & Related papers (2022-02-09T22:29:48Z) - A Minimalist Dataset for Systematic Generalization of Perception,
Syntax, and Semantics [131.93113552146195]
We present a new dataset, Handwritten arithmetic with INTegers (HINT), to examine machines' capability of learning generalizable concepts.
In HINT, machines are tasked with learning how concepts are perceived from raw signals such as images.
We undertake extensive experiments with various sequence-to-sequence models, including RNNs, Transformers, and GPT-3.
arXiv Detail & Related papers (2021-03-02T01:32:54Z) - Computational linguistic assessment of textbook and online learning
media by means of threshold concepts in business education [59.003956312175795]
From a linguistic perspective, threshold concepts are instances of specialized vocabularies, exhibiting particular linguistic features.
The profiles of 63 threshold concepts from business education have been investigated in textbooks, newspapers, and Wikipedia.
The three kinds of resources can indeed be distinguished in terms of their threshold concepts' profiles.
arXiv Detail & Related papers (2020-08-05T12:56:16Z) - Concept Learners for Few-Shot Learning [76.08585517480807]
We propose COMET, a meta-learning method that improves generalization ability by learning to learn along human-interpretable concept dimensions.
We evaluate our model on few-shot tasks from diverse domains, including fine-grained image classification, document categorization and cell type annotation.
arXiv Detail & Related papers (2020-07-14T22:04:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.