Structural Information Learning Machinery: Learning from Observing,
Associating, Optimizing, Decoding, and Abstracting
- URL: http://arxiv.org/abs/2001.09637v1
- Date: Mon, 27 Jan 2020 09:14:46 GMT
- Title: Structural Information Learning Machinery: Learning from Observing,
Associating, Optimizing, Decoding, and Abstracting
- Authors: Angsheng Li
- Abstract summary: We propose the model of it structural information learning machines (SiLeM)
A SiLeM machine learns the laws or rules of nature.
- Score: 0.913755431537592
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the present paper, we propose the model of {\it structural information
learning machines} (SiLeM for short), leading to a mathematical definition of
learning by merging the theories of computation and information. Our model
shows that the essence of learning is {\it to gain information}, that to gain
information is {\it to eliminate uncertainty} embedded in a data space, and
that to eliminate uncertainty of a data space can be reduced to an optimization
problem, that is, an {\it information optimization problem}, which can be
realized by a general {\it encoding tree method}. The principle and criterion
of the structural information learning machines are maximization of {\it
decoding information} from the data points observed together with the
relationships among the data points, and semantical {\it interpretation} of
syntactical {\it essential structure}, respectively. A SiLeM machine learns the
laws or rules of nature. It observes the data points of real world, builds the
{\it connections} among the observed data and constructs a {\it data space},
for which the principle is to choose the way of connections of data points so
that the {\it decoding information} of the data space is maximized, finds the
{\it encoding tree} of the data space that minimizes the dynamical uncertainty
of the data space, in which the encoding tree is hence referred to as a {\it
decoder}, due to the fact that it has already eliminated the maximum amount of
uncertainty embedded in the data space, interprets the {\it semantics} of the
decoder, an encoding tree, to form a {\it knowledge tree}, extracts the {\it
remarkable common features} for both semantical and syntactical features of the
modules decoded by a decoder to construct {\it trees of abstractions},
providing the foundations for {\it intuitive reasoning} in the learning when
new data are observed.
Related papers
- Learning Order Forest for Qualitative-Attribute Data Clustering [52.612779710298526]
This paper discovers a tree-like distance structure to flexibly represent the local order relationship among intra-attribute qualitative values.<n>A joint learning mechanism is proposed to iteratively obtain more appropriate tree structures and clusters.<n>Experiments demonstrate that the joint learning adapts the forest to the clustering task to yield accurate results.
arXiv Detail & Related papers (2026-03-03T07:49:50Z) - AdmTree: Compressing Lengthy Context with Adaptive Semantic Trees [66.39371821756649]
We propose AdmTree, a novel framework for adaptive, hierarchical context compression.<n>AdmTree segments input based on information density, utilizing gist tokens to summarize variable-length segments as the leaves of a semantic binary tree.<n>By preserving fine-grained details alongside global semantic coherence, mitigating positional bias, and dynamically adapting to content, AdmTree robustly retains the semantic information of long contexts.
arXiv Detail & Related papers (2025-12-04T08:04:19Z) - Knowledge prompt chaining for semantic modeling [0.0]
We propose a novel automatic semantic modeling framework: Knowledge Prompt Chaining.
It canserialize the graph-structured knowledge and inject it into the LLMs properly.
Based on experimental results, our method achieves better performance than existing leading techniques.
arXiv Detail & Related papers (2025-01-15T03:00:57Z) - Topology-aware Reinforcement Feature Space Reconstruction for Graph Data [22.5530178427691]
Reconstructing a good feature space is essential to augment the AI power of data, improve model generalization, and increase the availability of downstream ML models.
We use topology-aware reinforcement learning to automate and optimize feature space reconstruction for graph data.
Our approach combines the extraction of core subgraphs to capture essential structural information with a graph neural network (GNN) to encode topological features and reduce computing complexity.
arXiv Detail & Related papers (2024-11-08T18:01:05Z) - Dissecting embedding method: learning higher-order structures from data [0.0]
Geometric deep learning methods for data learning often include set of assumptions on the geometry of the feature space.
These assumptions together with data being discrete and finite can cause some generalisations, which are likely to create wrong interpretations of the data and models outputs.
arXiv Detail & Related papers (2024-10-14T08:19:39Z) - Struct-X: Enhancing Large Language Models Reasoning with Structured Data [38.558614152006975]
Struct-X operates through five key phases: read-model-fill-reflect-reason''
It encodes structured data into a topological space using graph embeddings.
It fills in missing entity information with knowledge retrieval modules.
The final phase involves constructing a topological network with selected tokens.
arXiv Detail & Related papers (2024-07-17T13:06:25Z) - Semantic Loss Functions for Neuro-Symbolic Structured Prediction [74.18322585177832]
We discuss the semantic loss, which injects knowledge about such structure, defined symbolically, into training.
It is agnostic to the arrangement of the symbols, and depends only on the semantics expressed thereby.
It can be combined with both discriminative and generative neural models.
arXiv Detail & Related papers (2024-05-12T22:18:25Z) - Homological Convolutional Neural Networks [4.615338063719135]
We propose a novel deep learning architecture that exploits the data structural organization through topologically constrained network representations.
We test our model on 18 benchmark datasets against 5 classic machine learning and 3 deep learning models.
arXiv Detail & Related papers (2023-08-26T08:48:51Z) - Principled and Efficient Motif Finding for Structure Learning of Lifted
Graphical Models [5.317624228510748]
Structure learning is a core problem in AI central to the fields of neuro-symbolic AI and statistical relational learning.
We present the first principled approach for mining structural motifs in lifted graphical models.
We show that we outperform state-of-the-art structure learning approaches by up to 6% in terms of accuracy and up to 80% in terms of runtime.
arXiv Detail & Related papers (2023-02-09T12:21:55Z) - Deep Equilibrium Assisted Block Sparse Coding of Inter-dependent
Signals: Application to Hyperspectral Imaging [71.57324258813675]
A dataset of inter-dependent signals is defined as a matrix whose columns demonstrate strong dependencies.
A neural network is employed to act as structure prior and reveal the underlying signal interdependencies.
Deep unrolling and Deep equilibrium based algorithms are developed, forming highly interpretable and concise deep-learning-based architectures.
arXiv Detail & Related papers (2022-03-29T21:00:39Z) - Incorporating Constituent Syntax for Coreference Resolution [50.71868417008133]
We propose a graph-based method to incorporate constituent syntactic structures.
We also explore to utilise higher-order neighbourhood information to encode rich structures in constituent trees.
Experiments on the English and Chinese portions of OntoNotes 5.0 benchmark show that our proposed model either beats a strong baseline or achieves new state-of-the-art performance.
arXiv Detail & Related papers (2022-02-22T07:40:42Z) - Representation Learning for Sequence Data with Deep Autoencoding
Predictive Components [96.42805872177067]
We propose a self-supervised representation learning method for sequence data, based on the intuition that useful representations of sequence data should exhibit a simple structure in the latent space.
We encourage this latent structure by maximizing an estimate of predictive information of latent feature sequences, which is the mutual information between past and future windows at each time step.
We demonstrate that our method recovers the latent space of noisy dynamical systems, extracts predictive features for forecasting tasks, and improves automatic speech recognition when used to pretrain the encoder on large amounts of unlabeled data.
arXiv Detail & Related papers (2020-10-07T03:34:01Z) - Auto-Encoding Twin-Bottleneck Hashing [141.5378966676885]
This paper proposes an efficient and adaptive code-driven graph.
It is updated by decoding in the context of an auto-encoder.
Experiments on benchmarked datasets clearly show the superiority of our framework over the state-of-the-art hashing methods.
arXiv Detail & Related papers (2020-02-27T05:58:12Z) - A Theory of Usable Information Under Computational Constraints [103.5901638681034]
We propose a new framework for reasoning about information in complex systems.
Our foundation is based on a variational extension of Shannon's information theory.
We show that by incorporating computational constraints, $mathcalV$-information can be reliably estimated from data.
arXiv Detail & Related papers (2020-02-25T06:09:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.