Mathematical conjecture generation using machine intelligence
- URL: http://arxiv.org/abs/2306.07277v1
- Date: Mon, 12 Jun 2023 17:58:38 GMT
- Title: Mathematical conjecture generation using machine intelligence
- Authors: Challenger Mishra, Subhayan Roy Moulik, Rahul Sarkar
- Abstract summary: We focus on strict inequalities of type f g and associate them with a vector space.
We develop a structural understanding of this conjecture space by studying linear automorphisms of this manifold.
We propose an algorithmic pipeline to generate novel conjectures using geometric gradient descent.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Conjectures have historically played an important role in the development of
pure mathematics. We propose a systematic approach to finding abstract patterns
in mathematical data, in order to generate conjectures about mathematical
inequalities, using machine intelligence. We focus on strict inequalities of
type f < g and associate them with a vector space. By geometerising this space,
which we refer to as a conjecture space, we prove that this space is isomorphic
to a Banach manifold. We develop a structural understanding of this conjecture
space by studying linear automorphisms of this manifold and show that this
space admits several free group actions. Based on these insights, we propose an
algorithmic pipeline to generate novel conjectures using geometric gradient
descent, where the metric is informed by the invariances of the conjecture
space. As proof of concept, we give a toy algorithm to generate novel
conjectures about the prime counting function and diameters of Cayley graphs of
non-abelian simple groups. We also report private communications with
colleagues in which some conjectures were proved, and highlight that some
conjectures generated using this procedure are still unproven. Finally, we
propose a pipeline of mathematical discovery in this space and highlight the
importance of domain expertise in this pipeline.
Related papers
- Disentangled Representation Learning with the Gromov-Monge Gap [65.73194652234848]
Learning disentangled representations from unlabelled data is a fundamental challenge in machine learning.
We introduce a novel approach to disentangled representation learning based on quadratic optimal transport.
We demonstrate the effectiveness of our approach for quantifying disentanglement across four standard benchmarks.
arXiv Detail & Related papers (2024-07-10T16:51:32Z) - Learning Formal Mathematics From Intrinsic Motivation [34.986025832497255]
Minimo is an agent that learns to pose problems for itself (conjecturing) and solve them (theorem proving)
We combine methods for constrained decoding and type-directed synthesis to sample valid conjectures from a language model.
Our agent targets generating hard but provable conjectures - a moving target, since its own theorem proving ability also improves as it trains.
arXiv Detail & Related papers (2024-06-30T13:34:54Z) - Machine learning and information theory concepts towards an AI
Mathematician [77.63761356203105]
The current state-of-the-art in artificial intelligence is impressive, especially in terms of mastery of language, but not so much in terms of mathematical reasoning.
This essay builds on the idea that current deep learning mostly succeeds at system 1 abilities.
It takes an information-theoretical posture to ask questions about what constitutes an interesting mathematical statement.
arXiv Detail & Related papers (2024-03-07T15:12:06Z) - Considerations on Approaches and Metrics in Automated Theorem
Generation/Finding in Geometry [0.0]
We present and discuss different approaches for the automatic discovery of geometric theorems (and properties)
We will argue that judging whether a theorem prover is able to produce interesting theorems remains a non deterministic task.
arXiv Detail & Related papers (2024-01-22T12:51:19Z) - Toward a Geometric Theory of Manifold Untangling [12.229735866373073]
We conjecture that there is a more general solution to manifold untangling in the topological space without artificially defining any distance metric.
General strategies of both global manifold embedding and local manifold flattening are presented and connected with existing work on the untangling of image, audio, and language data.
arXiv Detail & Related papers (2023-03-07T19:47:01Z) - Algebraic Machine Learning with an Application to Chemistry [0.0]
We develop a machine learning pipeline that captures fine-grain geometric information without relying on smoothness assumptions.
In particular, we propose a for numerically detecting points lying near the singular locus of the underlying variety.
arXiv Detail & Related papers (2022-05-11T22:41:19Z) - A singular Riemannian geometry approach to Deep Neural Networks I.
Theoretical foundations [77.86290991564829]
Deep Neural Networks are widely used for solving complex problems in several scientific areas, such as speech recognition, machine translation, image analysis.
We study a particular sequence of maps between manifold, with the last manifold of the sequence equipped with a Riemannian metric.
We investigate the theoretical properties of the maps of such sequence, eventually we focus on the case of maps between implementing neural networks of practical interest.
arXiv Detail & Related papers (2021-12-17T11:43:30Z) - Bayesian Quadrature on Riemannian Data Manifolds [79.71142807798284]
A principled way to model nonlinear geometric structure inherent in data is provided.
However, these operations are typically computationally demanding.
In particular, we focus on Bayesian quadrature (BQ) to numerically compute integrals over normal laws.
We show that by leveraging both prior knowledge and an active exploration scheme, BQ significantly reduces the number of required evaluations.
arXiv Detail & Related papers (2021-02-12T17:38:04Z) - Disentangling by Subspace Diffusion [72.1895236605335]
We show that fully unsupervised factorization of a data manifold is possible if the true metric of the manifold is known.
Our work reduces the question of whether unsupervised metric learning is possible, providing a unifying insight into the geometric nature of representation learning.
arXiv Detail & Related papers (2020-06-23T13:33:19Z) - Counterexamples to the Low-Degree Conjecture [80.3668228845075]
A conjecture of Hopkins posits that for certain high-dimensional hypothesis testing problems, no-time algorithm can outperform so-called "simple statistics"
This conjecture formalizes the beliefs surrounding a line of recent work that seeks to understand statistical-versus-computational tradeoffs.
arXiv Detail & Related papers (2020-04-17T21:08:11Z) - Estimate of the Neural Network Dimension using Algebraic Topology and
Lie Theory [0.0]
In this paper we present an approach to determine the smallest possible number of neurons in a layer of a neural network.
We specify the required dimensions precisely, assuming that there is a smooth manifold on or near which the data are located.
We derive the theory and validate it experimentally on toy data sets.
arXiv Detail & Related papers (2020-04-06T14:15:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.