Incorporating Expert Rules into Neural Networks in the Framework of
Concept-Based Learning
- URL: http://arxiv.org/abs/2402.14726v1
- Date: Thu, 22 Feb 2024 17:33:49 GMT
- Title: Incorporating Expert Rules into Neural Networks in the Framework of
Concept-Based Learning
- Authors: Andrei V. Konstantinov and Lev V. Utkin
- Abstract summary: It is proposed how to combine logical rules and neural networks predicting the concept probabilities.
We provide several approaches for solving the stated problem and for training neural networks.
The code of proposed algorithms is publicly available.
- Score: 2.9370710299422598
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A problem of incorporating the expert rules into machine learning models for
extending the concept-based learning is formulated in the paper. It is proposed
how to combine logical rules and neural networks predicting the concept
probabilities. The first idea behind the combination is to form constraints for
a joint probability distribution over all combinations of concept values to
satisfy the expert rules. The second idea is to represent a feasible set of
probability distributions in the form of a convex polytope and to use its
vertices or faces. We provide several approaches for solving the stated problem
and for training neural networks which guarantee that the output probabilities
of concepts would not violate the expert rules. The solution of the problem can
be viewed as a way for combining the inductive and deductive learning. Expert
rules are used in a broader sense when any logical function that connects
concepts and class labels or just concepts with each other can be regarded as a
rule. This feature significantly expands the class of the proposed results.
Numerical examples illustrate the approaches. The code of proposed algorithms
is publicly available.
Related papers
- LinSATNet: The Positive Linear Satisfiability Neural Networks [116.65291739666303]
This paper studies how to introduce the popular positive linear satisfiability to neural networks.
We propose the first differentiable satisfiability layer based on an extension of the classic Sinkhorn algorithm for jointly encoding multiple sets of marginal distributions.
arXiv Detail & Related papers (2024-07-18T22:05:21Z) - FI-CBL: A Probabilistic Method for Concept-Based Learning with Expert Rules [2.2120851074630177]
The main idea behind the method is to divide each concept-annotated image into patches, to transform the patches into embeddings by using an autoencoder, and to cluster the embeddings.
To find concepts of a new image, the method implements the frequentist inference by computing prior and posterior probabilities of concepts.
Numerical experiments show that FI-CBL outperforms the concept bottleneck model in cases when the number of training data is small.
arXiv Detail & Related papers (2024-06-28T13:05:17Z) - Deep Concept Removal [29.65899467379793]
We address the problem of concept removal in deep neural networks.
We propose a novel method based on adversarial linear classifiers trained on a concept dataset.
We also introduce an implicit gradient-based technique to tackle the challenges associated with adversarial training.
arXiv Detail & Related papers (2023-10-09T14:31:03Z) - Abstracting Concept-Changing Rules for Solving Raven's Progressive
Matrix Problems [54.26307134687171]
Raven's Progressive Matrix (RPM) is a classic test to realize such ability in machine intelligence by selecting from candidates.
Recent studies suggest that solving RPM in an answer-generation way boosts a more in-depth understanding of rules.
We propose a deep latent variable model for Concept-changing Rule ABstraction (CRAB) by learning interpretable concepts and parsing concept-changing rules in the latent space.
arXiv Detail & Related papers (2023-07-15T07:16:38Z) - Interpretable Neural-Symbolic Concept Reasoning [7.1904050674791185]
Concept-based models aim to address this issue by learning tasks based on a set of human-understandable concepts.
We propose the Deep Concept Reasoner (DCR), the first interpretable concept-based model that builds upon concept embeddings.
arXiv Detail & Related papers (2023-04-27T09:58:15Z) - Bayesian Learning for Neural Networks: an algorithmic survey [95.42181254494287]
This self-contained survey engages and introduces readers to the principles and algorithms of Bayesian Learning for Neural Networks.
It provides an introduction to the topic from an accessible, practical-algorithmic perspective.
arXiv Detail & Related papers (2022-11-21T21:36:58Z) - Robust Training and Verification of Implicit Neural Networks: A
Non-Euclidean Contractive Approach [64.23331120621118]
This paper proposes a theoretical and computational framework for training and robustness verification of implicit neural networks.
We introduce a related embedded network and show that the embedded network can be used to provide an $ell_infty$-norm box over-approximation of the reachable sets of the original network.
We apply our algorithms to train implicit neural networks on the MNIST dataset and compare the robustness of our models with the models trained via existing approaches in the literature.
arXiv Detail & Related papers (2022-08-08T03:13:24Z) - PROTOtypical Logic Tensor Networks (PROTO-LTN) for Zero Shot Learning [2.236663830879273]
Logic Networks (LTNs) are neuro-symbolic systems based on a differentiable, first-order logic grounded into a deep neural network.
We focus here on the subsumption or textttisOfClass predicate, which is fundamental to encode most semantic image interpretation tasks.
We propose a common textttisOfClass predicate, whose level of truth is a function of the distance between an object embedding and the corresponding class prototype.
arXiv Detail & Related papers (2022-06-26T18:34:07Z) - Kernelized Concept Erasure [108.65038124096907]
We propose a kernelization of a linear minimax game for concept erasure.
It is possible to prevent specific non-linear adversaries from predicting the concept.
However, the protection does not transfer to different nonlinear adversaries.
arXiv Detail & Related papers (2022-01-28T15:45:13Z) - Paraconsistent Foundations for Probabilistic Reasoning, Programming and
Concept Formation [0.0]
It is argued that 4-valued paraconsistent truth values (called here "p-bits") can serve as a conceptual, mathematical and practical foundation for highly AI-relevant forms of probabilistic logic and programming and concept formation.
It is shown that appropriate averaging-across-situations and renormalization of 4-valued p-bits operating in accordance with Constructible Duality (CD) logic yields PLN (Probabilistic Logic Networks) strength-and-confidence truth values.
arXiv Detail & Related papers (2020-12-28T20:14:49Z) - An Integer Linear Programming Framework for Mining Constraints from Data [81.60135973848125]
We present a general framework for mining constraints from data.
In particular, we consider the inference in structured output prediction as an integer linear programming (ILP) problem.
We show that our approach can learn to solve 9x9 Sudoku puzzles and minimal spanning tree problems from examples without providing the underlying rules.
arXiv Detail & Related papers (2020-06-18T20:09:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.