Counting-Based Effective Dimension and Discrete Regularizations
- URL: http://arxiv.org/abs/2205.11520v1
- Date: Mon, 23 May 2022 17:35:39 GMT
- Title: Counting-Based Effective Dimension and Discrete Regularizations
- Authors: Ivan Horv\'ath, Peter Marko\v{s}, Robert Mendris
- Abstract summary: Effective number theory determines all additive ways to assign counts to collections of objects with probabilities or other additive weights.
This effective counting dimension (ECD) specifies how the number of objects in a support scales with their total number.
ECD can be used to characterize targets of discrete regularizations in physics and other quantitative sciences.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Effective number theory determines all additive ways to assign counts to
collections of objects with probabilities or other additive weights. Here we
construct all counting-based schemes to select effective supports on such
collections, and show that it leads to a unique notion of effective dimension.
This effective counting dimension (ECD) specifies how the number of objects in
a support scales with their total number, and its uniqueness means that all
schemes yield the same value. Hence, ECD is well defined and can be used to
characterize targets of discrete regularizations in physics and other
quantitative sciences. Given its generality, ECD may help to connect and
interpret results from widely distinct areas. Our analysis makes recent studies
of effective spatial dimensions in lattice quantum chromodynamics and Anderson
localization models well founded. We address the reliability of regularization
removals in practice and perform the respective numerical analysis in the
context of 3d Anderson criticality. Our arguments suggest that measure-based
dimensions (Minkowski, Hausdorff) of fixed sets have good probabilistic
extensions.
Related papers
- Model-free Estimation of Latent Structure via Multiscale Nonparametric Maximum Likelihood [13.175343048302697]
We propose a model-free approach for estimating such latent structures whenever they are present, without assuming they exist a priori.
As an application, we design a clustering algorithm based on the proposed procedure and demonstrate its effectiveness in capturing a wide range of latent structures.
arXiv Detail & Related papers (2024-10-29T17:11:33Z) - Size-invariance Matters: Rethinking Metrics and Losses for Imbalanced Multi-object Salient Object Detection [133.66006666465447]
Current metrics are size-sensitive, where larger objects are focused, and smaller ones tend to be ignored.
We argue that the evaluation should be size-invariant because bias based on size is unjustified without additional semantic information.
We develop an optimization framework tailored to this goal, achieving considerable improvements in detecting objects of different sizes.
arXiv Detail & Related papers (2024-05-16T03:01:06Z) - Uncertainty quantification in metric spaces [3.637162892228131]
This paper introduces a novel uncertainty quantification framework for regression models where the response takes values in a separable metric space.
The proposed algorithms can efficiently handle large datasets and are agnostic to the predictive base model used.
arXiv Detail & Related papers (2024-05-08T15:06:02Z) - LieDetect: Detection of representation orbits of compact Lie groups from point clouds [0.0]
We suggest a new algorithm to estimate representations of compact Lie groups from finite samples of their orbits.<n>The knowledge of the representation type permits the reconstruction of its orbit, which is useful for identifying the Lie group that generates the action.<n>The algorithm is tested for synthetic data up to dimension 32, as well as real-life applications in image analysis, harmonic analysis, density estimation, equivariant neural networks, chemical conformational spaces, and classical mechanics systems.
arXiv Detail & Related papers (2023-07-26T18:32:43Z) - Enriching Disentanglement: From Logical Definitions to Quantitative Metrics [59.12308034729482]
Disentangling the explanatory factors in complex data is a promising approach for data-efficient representation learning.
We establish relationships between logical definitions and quantitative metrics to derive theoretically grounded disentanglement metrics.
We empirically demonstrate the effectiveness of the proposed metrics by isolating different aspects of disentangled representations.
arXiv Detail & Related papers (2023-05-19T08:22:23Z) - Concrete Score Matching: Generalized Score Matching for Discrete Data [109.12439278055213]
"Concrete score" is a generalization of the (Stein) score for discrete settings.
"Concrete Score Matching" is a framework to learn such scores from samples.
arXiv Detail & Related papers (2022-11-02T00:41:37Z) - Tangential Wasserstein Projections [0.4297070083645048]
We develop a notion of projections between sets of probability measures using the geometric properties of the 2-Wasserstein space.
The idea is to work on regular tangent cones of the Wasserstein space using generalized geodesics.
arXiv Detail & Related papers (2022-07-29T14:59:58Z) - Intrinsic dimension estimation for discrete metrics [65.5438227932088]
In this letter we introduce an algorithm to infer the intrinsic dimension (ID) of datasets embedded in discrete spaces.
We demonstrate its accuracy on benchmark datasets, and we apply it to analyze a metagenomic dataset for species fingerprinting.
This suggests that evolutive pressure acts on a low-dimensional manifold despite the high-dimensionality of sequences' space.
arXiv Detail & Related papers (2022-07-20T06:38:36Z) - DDAC-SpAM: A Distributed Algorithm for Fitting High-dimensional Sparse
Additive Models with Feature Division and Decorrelation [16.232378903482143]
We propose a new distributed statistical learning algorithm, DDAC-SpAM, which divides the features under a high-dimensional sparse additive model.
The effectiveness and efficiency of the proposed algorithm are demonstrated through theoretical analysis and empirical results on both synthetic and real data.
Our approach provides a practical solution for fitting sparse additive models, with promising applications in a wide range of domains.
arXiv Detail & Related papers (2022-05-16T18:31:03Z) - ISDE : Independence Structure Density Estimation [0.0]
Multidimensional density estimation suffers from the curse of dimensionality.
We propose ISDE (Independence Structure Density Estimation), an algorithm designed to estimate a density and an undirected graphical model.
We show how it performs both quantitatively and qualitatively on real-world datasets.
arXiv Detail & Related papers (2022-03-18T08:01:04Z) - Optimal radial basis for density-based atomic representations [58.720142291102135]
We discuss how to build an adaptive, optimal numerical basis that is chosen to represent most efficiently the structural diversity of the dataset at hand.
For each training dataset, this optimal basis is unique, and can be computed at no additional cost with respect to the primitive basis.
We demonstrate that this construction yields representations that are accurate and computationally efficient.
arXiv Detail & Related papers (2021-05-18T17:57:08Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.