Logical Information Cells I
- URL: http://arxiv.org/abs/2108.04751v1
- Date: Tue, 10 Aug 2021 15:31:26 GMT
- Title: Logical Information Cells I
- Authors: Jean-Claude Belfiore, Daniel Bennequin and Xavier Giraud
- Abstract summary: In this study we explore the spontaneous apparition of visible intelligible reasoning in simple artificial networks.
We start with the reproduction of a DNN model of natural neurons in monkeys.
We then study a bit more complex tasks, a priori involving predicate logic.
- Score: 10.411800812671952
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this study we explore the spontaneous apparition of visible intelligible
reasoning in simple artificial networks, and we connect this experimental
observation with a notion of semantic information. We start with the
reproduction of a DNN model of natural neurons in monkeys, studied by
Neromyliotis and Moschovakis in 2017 and 2018, to explain how "motor equivalent
neurons", coding only for the action of pointing, are supplemented by other
neurons for specifying the actor of the action, the eye E, the hand H, or the
eye and the hand together EH. There appear inner neurons performing a logical
work, making intermediary proposition, for instance E V EH. Then, we remarked
that adding a second hidden layer and choosing a symmetric metric for learning,
the activities of the neurons become almost quantized and more informative.
Using the work of Carnap and Bar-Hillel 1952, we define a measure of the
logical value for collections of such cells. The logical score growths with the
depth of the layer, i.e. the information on the output decision increases,
which confirms a kind of bottleneck principle. Then we study a bit more complex
tasks, a priori involving predicate logic. We compare the logic and the
measured weights. This shows, for groups of neurons, a neat correlation between
the logical score and the size of the weights. It exhibits a form of sparsity
between the layers. The most spectacular result concerns the triples which can
conclude for all conditions: when applying their weight matrices to their
logical matrix, we recover the classification. This shows that weights
precisely perform the proofs.
Related papers
- Arithmetic Without Algorithms: Language Models Solve Math With a Bag of Heuristics [43.86518549293703]
We show that large language models (LLMs) perform arithmetic using neither robust algorithms nor memorization.
Overall, our experimental results show that LLMs perform arithmetic using neither robust algorithms nor memorization.
arXiv Detail & Related papers (2024-10-28T17:59:06Z) - Don't Cut Corners: Exact Conditions for Modularity in Biologically Inspired Representations [52.48094670415497]
We develop a theory of when biologically inspired representations modularise with respect to source variables (sources)
We derive necessary and sufficient conditions on a sample of sources that determine whether the neurons in an optimal biologically-inspired linear autoencoder modularise.
Our theory applies to any dataset, extending far beyond the case of statistical independence studied in previous work.
arXiv Detail & Related papers (2024-10-08T17:41:37Z) - Confidence Regulation Neurons in Language Models [91.90337752432075]
This study investigates the mechanisms by which large language models represent and regulate uncertainty in next-token predictions.
Entropy neurons are characterized by an unusually high weight norm and influence the final layer normalization (LayerNorm) scale to effectively scale down the logits.
token frequency neurons, which we describe here for the first time, boost or suppress each token's logit proportionally to its log frequency, thereby shifting the output distribution towards or away from the unigram distribution.
arXiv Detail & Related papers (2024-06-24T01:31:03Z) - Interpreting the Second-Order Effects of Neurons in CLIP [73.54377859089801]
We interpret the function of individual neurons in CLIP by automatically describing them using text.
We present the "second-order lens", analyzing the effect flowing from a neuron through the later attention heads, directly to the output.
Our results indicate that a scalable understanding of neurons can be used for model deception and for introducing new model capabilities.
arXiv Detail & Related papers (2024-06-06T17:59:52Z) - Identifying Interpretable Visual Features in Artificial and Biological
Neural Systems [3.604033202771937]
Single neurons in neural networks are often interpretable in that they represent individual, intuitively meaningful features.
Many neurons exhibit $textitmixed selectivity$, i.e., they represent multiple unrelated features.
We propose an automated method for quantifying visual interpretability and an approach for finding meaningful directions in network activation space.
arXiv Detail & Related papers (2023-10-17T17:41:28Z) - Disentanglement with Biological Constraints: A Theory of Functional Cell
Types [20.929056085868613]
This work provides a mathematical understanding of why single neurons in the brain often represent single human-interpretable factors.
It also steps towards an understanding task structure shapes the structure of brain representation.
arXiv Detail & Related papers (2022-09-30T14:27:28Z) - The Causal Neural Connection: Expressiveness, Learnability, and
Inference [125.57815987218756]
An object called structural causal model (SCM) represents a collection of mechanisms and sources of random variation of the system under investigation.
In this paper, we show that the causal hierarchy theorem (Thm. 1, Bareinboim et al., 2020) still holds for neural models.
We introduce a special type of SCM called a neural causal model (NCM), and formalize a new type of inductive bias to encode structural constraints necessary for performing causal inferences.
arXiv Detail & Related papers (2021-07-02T01:55:18Z) - Compositional Explanations of Neurons [52.71742655312625]
We describe a procedure for explaining neurons in deep representations by identifying compositional logical concepts.
We use this procedure to answer several questions on interpretability in models for vision and natural language processing.
arXiv Detail & Related papers (2020-06-24T20:37:05Z) - Non-linear Neurons with Human-like Apical Dendrite Activations [81.18416067005538]
We show that a standard neuron followed by our novel apical dendrite activation (ADA) can learn the XOR logical function with 100% accuracy.
We conduct experiments on six benchmark data sets from computer vision, signal processing and natural language processing.
arXiv Detail & Related papers (2020-02-02T21:09:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.