Image classification network enhancement methods based on knowledge
injection
- URL: http://arxiv.org/abs/2401.04441v1
- Date: Tue, 9 Jan 2024 09:11:41 GMT
- Title: Image classification network enhancement methods based on knowledge
injection
- Authors: Yishuang Tian, Ning Wang, Liang Zhang
- Abstract summary: This paper proposes a multi-level hierarchical deep learning algorithm.
It is composed of multi-level hierarchical deep neural network architecture and multi-level hierarchical deep learning framework.
The experimental results show that the proposed algorithm can effectively explain the hidden information of the neural network.
- Score: 8.885876832491917
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The current deep neural network algorithm still stays in the end-to-end
training supervision method like Image-Label pairs, which makes traditional
algorithm is difficult to explain the reason for the results, and the
prediction logic is difficult to understand and analyze. The current algorithm
does not use the existing human knowledge information, which makes the model
not in line with the human cognition model and makes the model not suitable for
human use. In order to solve the above problems, the present invention provides
a deep neural network training method based on the human knowledge, which uses
the human cognition model to construct the deep neural network training model,
and uses the existing human knowledge information to construct the deep neural
network training model. This paper proposes a multi-level hierarchical deep
learning algorithm, which is composed of multi-level hierarchical deep neural
network architecture and multi-level hierarchical deep learning framework. The
experimental results show that the proposed algorithm can effectively explain
the hidden information of the neural network. The goal of our study is to
improve the interpretability of deep neural networks (DNNs) by providing an
analysis of the impact of knowledge injection on the classification task. We
constructed a knowledge injection dataset with matching knowledge data and
image classification data. The knowledge injection dataset is the benchmark
dataset for the experiments in the paper. Our model expresses the improvement
in interpretability and classification task performance of hidden layers at
different scales.
Related papers
- Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - Graph Neural Operators for Classification of Spatial Transcriptomics
Data [1.408706290287121]
We propose a study incorporating various graph neural network approaches to validate the efficacy of applying neural operators towards prediction of brain regions in mouse brain tissue samples.
We were able to achieve an F1 score of nearly 72% for the graph neural operator approach which outperformed all baseline and other graph network approaches.
arXiv Detail & Related papers (2023-02-01T18:32:06Z) - Neural Routing in Meta Learning [9.070747377130472]
We aim to improve the model performance of the current meta learning algorithms by selectively using only parts of the model conditioned on the input tasks.
In this work, we describe an approach that investigates task-dependent dynamic neuron selection in deep convolutional neural networks (CNNs) by leveraging the scaling factor in the batch normalization layer.
We find that the proposed approach, neural routing in meta learning (NRML), outperforms one of the well-known existing meta learning baselines on few-shot classification tasks.
arXiv Detail & Related papers (2022-10-14T16:31:24Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Multi-Semantic Image Recognition Model and Evaluating Index for
explaining the deep learning models [31.387124252490377]
We first propose a multi-semantic image recognition model, which enables human beings to understand the decision-making process of the neural network.
We then presents a new evaluation index, which can quantitatively assess the model interpretability.
This paper also exhibits the relevant baseline performance with current state-of-the-art deep learning models.
arXiv Detail & Related papers (2021-09-28T07:18:05Z) - Improving a neural network model by explanation-guided training for
glioma classification based on MRI data [0.0]
Interpretability methods have become a popular way to gain insight into the decision-making process of deep learning models.
We propose a method for explanation-guided training that uses a Layer-wise relevance propagation (LRP) technique.
We experimentally verified our method on a convolutional neural network (CNN) model for low-grade and high-grade glioma classification problems.
arXiv Detail & Related papers (2021-07-05T13:27:28Z) - FF-NSL: Feed-Forward Neural-Symbolic Learner [70.978007919101]
This paper introduces a neural-symbolic learning framework, called Feed-Forward Neural-Symbolic Learner (FF-NSL)
FF-NSL integrates state-of-the-art ILP systems based on the Answer Set semantics, with neural networks, in order to learn interpretable hypotheses from labelled unstructured data.
arXiv Detail & Related papers (2021-06-24T15:38:34Z) - Learning Structures for Deep Neural Networks [99.8331363309895]
We propose to adopt the efficient coding principle, rooted in information theory and developed in computational neuroscience.
We show that sparse coding can effectively maximize the entropy of the output signals.
Our experiments on a public image classification dataset demonstrate that using the structure learned from scratch by our proposed algorithm, one can achieve a classification accuracy comparable to the best expert-designed structure.
arXiv Detail & Related papers (2021-05-27T12:27:24Z) - Rectified Linear Postsynaptic Potential Function for Backpropagation in
Deep Spiking Neural Networks [55.0627904986664]
Spiking Neural Networks (SNNs) usetemporal spike patterns to represent and transmit information, which is not only biologically realistic but also suitable for ultra-low-power event-driven neuromorphic implementation.
This paper investigates the contribution of spike timing dynamics to information encoding, synaptic plasticity and decision making, providing a new perspective to design of future DeepSNNs and neuromorphic hardware systems.
arXiv Detail & Related papers (2020-03-26T11:13:07Z) - Neural Architecture Search For Fault Diagnosis [6.226564415963648]
Deep learning is suitable for processing big data, and has a strong feature extraction ability to realize end-to-end fault diagnosis systems.
Neural architecture search (NAS) is developing rapidly, and is becoming one of the next directions for deep learning.
In this paper, we proposed a NAS method for fault diagnosis using reinforcement learning.
arXiv Detail & Related papers (2020-02-19T04:03:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.