General Rough Modeling of Cluster Analysis
- URL: http://arxiv.org/abs/2106.04683v1
- Date: Wed, 2 Jun 2021 20:54:10 GMT
- Title: General Rough Modeling of Cluster Analysis
- Authors: A. Mani
- Abstract summary: The essence of the approach is explained in brief and supported by an example.
A new general rough method of analyzing clusterings is invented, and this opens the subject to clearer conceptions and contamination-free proofs.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this research, a general theoretical framework for clustering is proposed
over specific partial algebraic systems by the present author. Her theory helps
in isolating minimal assumptions necessary for different concepts of clustering
information in any form to be realized in a situation (and therefore in a
semantics). \emph{It is well-known that of the limited number of proofs in the
theory of hard and soft clustering that are known to exist, most involve
statistical assumptions}. Many methods seem to work because they seem to work
in specific empirical practice. A new general rough method of analyzing
clusterings is invented, and this opens the subject to clearer conceptions and
contamination-free theoretical proofs. Numeric ideas of validation are also
proposed to be replaced by those based on general rough approximation. The
essence of the approach is explained in brief and supported by an example.
Related papers
- A Canonicalization Perspective on Invariant and Equivariant Learning [54.44572887716977]
We introduce a canonicalization perspective that provides an essential and complete view of the design of frames.
We show that there exists an inherent connection between frames and canonical forms.
We design novel frames for eigenvectors that are strictly superior to existing methods.
arXiv Detail & Related papers (2024-05-28T17:22:15Z) - Class-wise Generalization Error: an Information-Theoretic Analysis [22.877440350595222]
We study the class-generalization error, which quantifies the generalization performance of each individual class.
We empirically validate our proposed bounds in different neural networks and show that they accurately capture the complex class-generalization error behavior.
arXiv Detail & Related papers (2024-01-05T17:05:14Z) - Synergies between Disentanglement and Sparsity: Generalization and
Identifiability in Multi-Task Learning [79.83792914684985]
We prove a new identifiability result that provides conditions under which maximally sparse base-predictors yield disentangled representations.
Motivated by this theoretical result, we propose a practical approach to learn disentangled representations based on a sparsity-promoting bi-level optimization problem.
arXiv Detail & Related papers (2022-11-26T21:02:09Z) - Explainable Clustering via Exemplars: Complexity and Efficient
Approximation Algorithms [30.369731369945296]
We propose an explainable-by-design clustering approach that finds clusters and exemplars to explain each cluster.
The use of exemplars for understanding is supported by the exemplar-based school of concept definition in psychology.
We show that finding a small set of exemplars to explain even a single cluster is computationally intractable.
arXiv Detail & Related papers (2022-09-20T12:09:51Z) - Granular Directed Rough Sets, Concept Organization and Soft Clustering [0.0]
Up-directed rough sets are introduced and studied by the present author in earlier papers.
This is extended by her in two different granular directions, with a surprising algebraic semantics.
This research is expected to see significant theoretical and practical applications in related domains.
arXiv Detail & Related papers (2022-08-13T11:01:05Z) - The Dynamics of Riemannian Robbins-Monro Algorithms [101.29301565229265]
We propose a family of Riemannian algorithms generalizing and extending the seminal approximation framework of Robbins and Monro.
Compared to their Euclidean counterparts, Riemannian algorithms are much less understood due to lack of a global linear structure on the manifold.
We provide a general template of almost sure convergence results that mirrors and extends the existing theory for Euclidean Robbins-Monro schemes.
arXiv Detail & Related papers (2022-06-14T12:30:11Z) - Granular Generalized Variable Precision Rough Sets and Rational
Approximations [0.24366811507669117]
Granular approximations as per the procedures of VPRS are likely to be more rational than those constructed from a classical perspective under certain conditions.
meta applications to cluster validation, image segmentation and dynamic sorting are invented.
arXiv Detail & Related papers (2022-05-28T08:08:26Z) - Set-valued classification -- overview via a unified framework [15.109906768606644]
Multi-class datasets can be extremely ambiguous and single-output predictions fail to deliver satisfactory performance.
By allowing predictors to predict a set of label candidates, set-valued classification offers a natural way to deal with this ambiguity.
We provide infinite sample optimal set-valued classification strategies and review a general plug-in principle to construct data-driven algorithms.
arXiv Detail & Related papers (2021-02-24T14:54:07Z) - In Search of Robust Measures of Generalization [79.75709926309703]
We develop bounds on generalization error, optimization error, and excess risk.
When evaluated empirically, most of these bounds are numerically vacuous.
We argue that generalization measures should instead be evaluated within the framework of distributional robustness.
arXiv Detail & Related papers (2020-10-22T17:54:25Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z) - Marginal likelihood computation for model selection and hypothesis
testing: an extensive review [66.37504201165159]
This article provides a comprehensive study of the state-of-the-art of the topic.
We highlight limitations, benefits, connections and differences among the different techniques.
Problems and possible solutions with the use of improper priors are also described.
arXiv Detail & Related papers (2020-05-17T18:31:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.