Logical Information Cells I
- URL: http://arxiv.org/abs/2108.04751v1
- Date: Tue, 10 Aug 2021 15:31:26 GMT
- Title: Logical Information Cells I
- Authors: Jean-Claude Belfiore, Daniel Bennequin and Xavier Giraud
- Abstract summary: In this study we explore the spontaneous apparition of visible intelligible reasoning in simple artificial networks.
We start with the reproduction of a DNN model of natural neurons in monkeys.
We then study a bit more complex tasks, a priori involving predicate logic.
- Score: 10.411800812671952
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this study we explore the spontaneous apparition of visible intelligible
reasoning in simple artificial networks, and we connect this experimental
observation with a notion of semantic information. We start with the
reproduction of a DNN model of natural neurons in monkeys, studied by
Neromyliotis and Moschovakis in 2017 and 2018, to explain how "motor equivalent
neurons", coding only for the action of pointing, are supplemented by other
neurons for specifying the actor of the action, the eye E, the hand H, or the
eye and the hand together EH. There appear inner neurons performing a logical
work, making intermediary proposition, for instance E V EH. Then, we remarked
that adding a second hidden layer and choosing a symmetric metric for learning,
the activities of the neurons become almost quantized and more informative.
Using the work of Carnap and Bar-Hillel 1952, we define a measure of the
logical value for collections of such cells. The logical score growths with the
depth of the layer, i.e. the information on the output decision increases,
which confirms a kind of bottleneck principle. Then we study a bit more complex
tasks, a priori involving predicate logic. We compare the logic and the
measured weights. This shows, for groups of neurons, a neat correlation between
the logical score and the size of the weights. It exhibits a form of sparsity
between the layers. The most spectacular result concerns the triples which can
conclude for all conditions: when applying their weight matrices to their
logical matrix, we recover the classification. This shows that weights
precisely perform the proofs.
Related papers
- Arithmetic Without Algorithms: Language Models Solve Math With a Bag of Heuristics [43.86518549293703]
We show that large language models (LLMs) perform arithmetic using neither robust algorithms nor memorization.
Overall, our experimental results show that LLMs perform arithmetic using neither robust algorithms nor memorization.
arXiv Detail & Related papers (2024-10-28T17:59:06Z) - Artificial Kuramoto Oscillatory Neurons [65.16453738828672]
It has long been known in both neuroscience and AI that ''binding'' between neurons leads to a form of competitive learning.
We introduce Artificial rethinking together with arbitrary connectivity designs such as fully connected convolutional, or attentive mechanisms.
We show that this idea provides performance improvements across a wide spectrum of tasks such as unsupervised object discovery, adversarial robustness, uncertainty, and reasoning.
arXiv Detail & Related papers (2024-10-17T17:47:54Z) - Confidence Regulation Neurons in Language Models [91.90337752432075]
This study investigates the mechanisms by which large language models represent and regulate uncertainty in next-token predictions.
Entropy neurons are characterized by an unusually high weight norm and influence the final layer normalization (LayerNorm) scale to effectively scale down the logits.
token frequency neurons, which we describe here for the first time, boost or suppress each token's logit proportionally to its log frequency, thereby shifting the output distribution towards or away from the unigram distribution.
arXiv Detail & Related papers (2024-06-24T01:31:03Z) - Interpreting the Second-Order Effects of Neurons in CLIP [73.54377859089801]
We interpret the function of individual neurons in CLIP by automatically describing them using text.
We present the "second-order lens", analyzing the effect flowing from a neuron through the later attention heads, directly to the output.
Our results indicate that an automated interpretation of neurons can be used for model deception and for introducing new model capabilities.
arXiv Detail & Related papers (2024-06-06T17:59:52Z) - Linear Explanations for Individual Neurons [12.231741536057378]
We show that the highest activation range is only responsible for a very small percentage of the neuron's causal effect.
In addition, inputs causing lower activations are often very different and can't be reliably predicted by only looking at high activations.
arXiv Detail & Related papers (2024-05-10T23:48:37Z) - Identifying Interpretable Visual Features in Artificial and Biological
Neural Systems [3.604033202771937]
Single neurons in neural networks are often interpretable in that they represent individual, intuitively meaningful features.
Many neurons exhibit $textitmixed selectivity$, i.e., they represent multiple unrelated features.
We propose an automated method for quantifying visual interpretability and an approach for finding meaningful directions in network activation space.
arXiv Detail & Related papers (2023-10-17T17:41:28Z) - Disentanglement with Biological Constraints: A Theory of Functional Cell
Types [20.929056085868613]
This work provides a mathematical understanding of why single neurons in the brain often represent single human-interpretable factors.
It also steps towards an understanding task structure shapes the structure of brain representation.
arXiv Detail & Related papers (2022-09-30T14:27:28Z) - The Causal Neural Connection: Expressiveness, Learnability, and
Inference [125.57815987218756]
An object called structural causal model (SCM) represents a collection of mechanisms and sources of random variation of the system under investigation.
In this paper, we show that the causal hierarchy theorem (Thm. 1, Bareinboim et al., 2020) still holds for neural models.
We introduce a special type of SCM called a neural causal model (NCM), and formalize a new type of inductive bias to encode structural constraints necessary for performing causal inferences.
arXiv Detail & Related papers (2021-07-02T01:55:18Z) - Compositional Explanations of Neurons [52.71742655312625]
We describe a procedure for explaining neurons in deep representations by identifying compositional logical concepts.
We use this procedure to answer several questions on interpretability in models for vision and natural language processing.
arXiv Detail & Related papers (2020-06-24T20:37:05Z) - Non-linear Neurons with Human-like Apical Dendrite Activations [81.18416067005538]
We show that a standard neuron followed by our novel apical dendrite activation (ADA) can learn the XOR logical function with 100% accuracy.
We conduct experiments on six benchmark data sets from computer vision, signal processing and natural language processing.
arXiv Detail & Related papers (2020-02-02T21:09:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.