Concept Evolution in Deep Learning Training: A Unified Interpretation
Framework and Discoveries
- URL: http://arxiv.org/abs/2203.16475v4
- Date: Tue, 22 Aug 2023 19:00:49 GMT
- Title: Concept Evolution in Deep Learning Training: A Unified Interpretation
Framework and Discoveries
- Authors: Haekyu Park, Seongmin Lee, Benjamin Hoover, Austin P. Wright, Omar
Shaikh, Rahul Duggal, Nilaksh Das, Kevin Li, Judy Hoffman, Duen Horng Chau
- Abstract summary: ConceptEvo is a unified interpretation framework for deep neural networks (DNNs)
It reveals the inception and evolution of learned concepts during training.
It is applicable to both modern DNN architectures, such as ConvNeXt, and classic DNNs, such as VGGs and InceptionV3.
- Score: 45.88354622464973
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present ConceptEvo, a unified interpretation framework for deep neural
networks (DNNs) that reveals the inception and evolution of learned concepts
during training. Our work addresses a critical gap in DNN interpretation
research, as existing methods primarily focus on post-training interpretation.
ConceptEvo introduces two novel technical contributions: (1) an algorithm that
generates a unified semantic space, enabling side-by-side comparison of
different models during training, and (2) an algorithm that discovers and
quantifies important concept evolutions for class predictions. Through a
large-scale human evaluation and quantitative experiments, we demonstrate that
ConceptEvo successfully identifies concept evolutions across different models,
which are not only comprehensible to humans but also crucial for class
predictions. ConceptEvo is applicable to both modern DNN architectures, such as
ConvNeXt, and classic DNNs, such as VGGs and InceptionV3.
Related papers
- A Self-explaining Neural Architecture for Generalizable Concept Learning [29.932706137805713]
We show that present SOTA concept learning approaches suffer from two major problems - lack of concept fidelity and limited concept interoperability.
We propose a novel self-explaining architecture for concept learning across domains.
We demonstrate the efficacy of our proposed approach over current SOTA concept learning approaches on four widely used real-world datasets.
arXiv Detail & Related papers (2024-05-01T06:50:18Z) - A survey on Concept-based Approaches For Model Improvement [2.1516043775965565]
Concepts are known to be the thinking ground of humans.
We provide a systematic review and taxonomy of various concept representations and their discovery algorithms in Deep Neural Networks (DNNs)
We also provide details on concept-based model improvement literature marking the first comprehensive survey of these methods.
arXiv Detail & Related papers (2024-03-21T17:09:20Z) - ConcEPT: Concept-Enhanced Pre-Training for Language Models [57.778895980999124]
ConcEPT aims to infuse conceptual knowledge into pre-trained language models.
It exploits external entity concept prediction to predict the concepts of entities mentioned in the pre-training contexts.
Results of experiments show that ConcEPT gains improved conceptual knowledge with concept-enhanced pre-training.
arXiv Detail & Related papers (2024-01-11T05:05:01Z) - Interpretable Neural-Symbolic Concept Reasoning [7.1904050674791185]
Concept-based models aim to address this issue by learning tasks based on a set of human-understandable concepts.
We propose the Deep Concept Reasoner (DCR), the first interpretable concept-based model that builds upon concept embeddings.
arXiv Detail & Related papers (2023-04-27T09:58:15Z) - Discovering Latent Concepts Learned in BERT [21.760620298330235]
We study what latent concepts exist in the pre-trained BERT model.
We also release a novel BERT ConceptNet dataset (BCN) consisting of 174 concept labels and 1M annotated instances.
arXiv Detail & Related papers (2022-05-15T09:45:34Z) - Convolutional Cobweb: A Model of Incremental Learning from 2D Images [0.0]
This paper presents a new concept formation approach that supports the ability to incrementally learn and predict labels for visual images.
We experimentally evaluate this new approach by applying it to an incremental variation of the MNIST digit recognition task.
arXiv Detail & Related papers (2022-01-18T04:39:31Z) - A Minimalist Dataset for Systematic Generalization of Perception,
Syntax, and Semantics [131.93113552146195]
We present a new dataset, Handwritten arithmetic with INTegers (HINT), to examine machines' capability of learning generalizable concepts.
In HINT, machines are tasked with learning how concepts are perceived from raw signals such as images.
We undertake extensive experiments with various sequence-to-sequence models, including RNNs, Transformers, and GPT-3.
arXiv Detail & Related papers (2021-03-02T01:32:54Z) - Formalising Concepts as Grounded Abstractions [68.24080871981869]
This report shows how representation learning can be used to induce concepts from raw data.
The main technical goal of this report is to show how techniques from representation learning can be married with a lattice-theoretic formulation of conceptual spaces.
arXiv Detail & Related papers (2021-01-13T15:22:01Z) - Concept Learners for Few-Shot Learning [76.08585517480807]
We propose COMET, a meta-learning method that improves generalization ability by learning to learn along human-interpretable concept dimensions.
We evaluate our model on few-shot tasks from diverse domains, including fine-grained image classification, document categorization and cell type annotation.
arXiv Detail & Related papers (2020-07-14T22:04:17Z) - A Chain Graph Interpretation of Real-World Neural Networks [58.78692706974121]
We propose an alternative interpretation that identifies NNs as chain graphs (CGs) and feed-forward as an approximate inference procedure.
The CG interpretation specifies the nature of each NN component within the rich theoretical framework of probabilistic graphical models.
We demonstrate with concrete examples that the CG interpretation can provide novel theoretical support and insights for various NN techniques.
arXiv Detail & Related papers (2020-06-30T14:46:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.