Representation Synthesis by Probabilistic Many-Valued Logic Operation in Self-Supervised Learning
- URL: http://arxiv.org/abs/2309.04148v3
- Date: Thu, 03 Oct 2024 03:55:04 GMT
- Title: Representation Synthesis by Probabilistic Many-Valued Logic Operation in Self-Supervised Learning
- Authors: Hiroki Nakamura, Masashi Okada, Tadahiro Taniguchi,
- Abstract summary: We propose a new self-supervised learning (SSL) method for representations that enable logic operations.
Our method can generate a representation that has the features of both representations or only those features common to both representations.
Experiments on image retrieval using MNIST and PascalVOC showed that the representations of our method can be operated by OR and AND operations.
- Score: 9.339914898177186
- License:
- Abstract: In this paper, we propose a new self-supervised learning (SSL) method for representations that enable logic operations. Representation learning has been applied to various tasks, such as image generation and retrieval. The logical controllability of representations is important for these tasks. Although some methods have been shown to enable the intuitive control of representations using natural languages as the inputs, representation control via logic operations between representations has not been demonstrated. Some SSL methods using representation synthesis (e.g., elementwise mean and maximum operations) have been proposed, but the operations performed in these methods do not incorporate logic operations. In this work, we propose a logic-operable self-supervised representation learning method by replacing the existing representation synthesis with the OR operation on the probabilistic extension of many-valued logic. The representations comprise a set of feature-possession degrees, which are truth values indicating the presence or absence of each feature in the image, and realize the logic operations (e.g., OR and AND). Our method can generate a representation that has the features of both representations or only those features common to both representations. In addition, the expression of the ambiguous presence of a feature is realized by indicating the feature-possession degree by the probability distribution of truth values of the many-valued logic. We showed that our method performs competitively in single and multi-label classification tasks compared with prior SSL methods using synthetic representations. Moreover, experiments on image retrieval using MNIST and PascalVOC showed that the representations of our method can be operated by OR and AND operations.
Related papers
- A Probabilistic Model Behind Self-Supervised Learning [53.64989127914936]
In self-supervised learning (SSL), representations are learned via an auxiliary task without annotated labels.
We present a generative latent variable model for self-supervised learning.
We show that several families of discriminative SSL, including contrastive methods, induce a comparable distribution over representations.
arXiv Detail & Related papers (2024-02-02T13:31:17Z) - Labeling Neural Representations with Inverse Recognition [25.867702786273586]
Inverse Recognition (INVERT) is a scalable approach for connecting learned representations with human-understandable concepts.
In contrast to prior work, INVERT is capable of handling diverse types of neurons, exhibits less computational complexity, and does not rely on the availability of segmentation masks.
We demonstrate the applicability of INVERT in various scenarios, including the identification of representations affected by spurious correlations.
arXiv Detail & Related papers (2023-11-22T18:55:25Z) - Language Models can be Logical Solvers [99.40649402395725]
We introduce LoGiPT, a novel language model that directly emulates the reasoning processes of logical solvers.
LoGiPT is fine-tuned on a newly constructed instruction-tuning dataset derived from revealing and refining the invisible reasoning process of deductive solvers.
arXiv Detail & Related papers (2023-11-10T16:23:50Z) - LOGICSEG: Parsing Visual Semantics with Neural Logic Learning and
Reasoning [73.98142349171552]
LOGICSEG is a holistic visual semantic that integrates neural inductive learning and logic reasoning with both rich data and symbolic knowledge.
During fuzzy logic-based continuous relaxation, logical formulae are grounded onto data and neural computational graphs, hence enabling logic-induced network training.
These designs together make LOGICSEG a general and compact neural-logic machine that is readily integrated into existing segmentation models.
arXiv Detail & Related papers (2023-09-24T05:43:19Z) - Evaluating Step-by-Step Reasoning through Symbolic Verification [20.156768135017007]
Pre-trained language models (LMs) have shown remarkable reasoning performance for in-context learning.
LMLP enjoys more than $25%$ higher accuracy than chain-of-thoughts (CoT) on length generalization benchmarks even with smaller model sizes.
arXiv Detail & Related papers (2022-12-16T19:30:01Z) - OPERA:Operation-Pivoted Discrete Reasoning over Text [33.36388276371693]
OPERA is an operation-pivoted discrete reasoning framework for machine reading comprehension.
It uses lightweight symbolic operations as neural modules to facilitate the reasoning ability and interpretability.
Experiments on both DROP and RACENum datasets show the reasoning ability of OPERA.
arXiv Detail & Related papers (2022-04-29T15:41:47Z) - SimSR: Simple Distance-based State Representation for Deep Reinforcement
Learning [14.626797887000901]
This work explores how to learn robust and generalizable state representation from image-based observations with deep reinforcement learning methods.
We devise Simple State Representation (SimSR) operator, which achieves equivalent functionality by anapproximation order in comparison with bi metricsimulation.
Our model generally achieves better performance and has better robustness and good generalization.
arXiv Detail & Related papers (2021-12-31T04:39:54Z) - Unifying AI Algorithms with Probabilistic Programming using Implicitly
Defined Representations [0.2580765958706854]
Scruff is a new framework for developing AI systems using probabilistic programming.
It enables a variety of representations to be included, such as code with choices, neural networks, differential equations, and constraint systems.
We show how a relatively small set of operations can serve to unify a variety of AI algorithms.
arXiv Detail & Related papers (2021-10-05T19:49:30Z) - How Fine-Tuning Allows for Effective Meta-Learning [50.17896588738377]
We present a theoretical framework for analyzing representations derived from a MAML-like algorithm.
We provide risk bounds on the best predictor found by fine-tuning via gradient descent, demonstrating that the algorithm can provably leverage the shared structure.
This separation result underscores the benefit of fine-tuning-based methods, such as MAML, over methods with "frozen representation" objectives in few-shot learning.
arXiv Detail & Related papers (2021-05-05T17:56:00Z) - Conditional Meta-Learning of Linear Representations [57.90025697492041]
Standard meta-learning for representation learning aims to find a common representation to be shared across multiple tasks.
In this work we overcome this issue by inferring a conditioning function, mapping the tasks' side information into a representation tailored to the task at hand.
We propose a meta-algorithm capable of leveraging this advantage in practice.
arXiv Detail & Related papers (2021-03-30T12:02:14Z) - Model-free Representation Learning and Exploration in Low-rank MDPs [64.72023662543363]
We present the first model-free representation learning algorithms for low rank MDPs.
Key algorithmic contribution is a new minimax representation learning objective.
Result can accommodate general function approximation to scale to complex environments.
arXiv Detail & Related papers (2021-02-14T00:06:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.