Estimation of Classification Rules from Partially Classified Data
- URL: http://arxiv.org/abs/2004.06237v1
- Date: Mon, 13 Apr 2020 23:35:25 GMT
- Title: Estimation of Classification Rules from Partially Classified Data
- Authors: Geoffrey J. McLachlan, Daniel Ahfock
- Abstract summary: We consider the situation where the observed sample contains some observations whose class of origin is known, and where the remaining observations in the sample are unclassified.
For class-conditional distributions taken to be known up to a vector of unknown parameters, the aim is to estimate the Bayes' rule of allocation for the allocation of subsequent unclassified observations.
- Score: 0.9137554315375919
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the situation where the observed sample contains some
observations whose class of origin is known (that is, they are classified with
respect to the g underlying classes of interest), and where the remaining
observations in the sample are unclassified (that is, their class labels are
unknown). For class-conditional distributions taken to be known up to a vector
of unknown parameters, the aim is to estimate the Bayes' rule of allocation for
the allocation of subsequent unclassified observations. Estimation on the basis
of both the classified and unclassified data can be undertaken in a
straightforward manner by fitting a g-component mixture model by maximum
likelihood (ML) via the EM algorithm in the situation where the observed data
can be assumed to be an observed random sample from the adopted mixture
distribution. This assumption applies if the missing-data mechanism is
ignorable in the terminology pioneered by Rubin (1976). An initial likelihood
approach was to use the so-called classification ML approach whereby the
missing labels are taken to be parameters to be estimated along with the
parameters of the class-conditional distributions. However, as it can lead to
inconsistent estimates, the focus of attention switched to the mixture ML
approach after the appearance of the EM algorithm (Dempster et al., 1977).
Particular attention is given here to the asymptotic relative efficiency (ARE)
of the Bayes' rule estimated from a partially classified sample. Lastly, we
consider briefly some recent results in situations where the missing label
pattern is non-ignorable for the purposes of ML estimation for the mixture
model.
Related papers
- A Tale of Sampling and Estimation in Discounted Reinforcement Learning [50.43256303670011]
We present a minimax lower bound on the discounted mean estimation problem.
We show that estimating the mean by directly sampling from the discounted kernel of the Markov process brings compelling statistical properties.
arXiv Detail & Related papers (2023-04-11T09:13:17Z) - Parametric Classification for Generalized Category Discovery: A Baseline
Study [70.73212959385387]
Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples.
We investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem.
We propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers.
arXiv Detail & Related papers (2022-11-21T18:47:11Z) - On the Semi-supervised Expectation Maximization [5.481082183778667]
We focus on a semi-supervised case to learn the model from labeled and unlabeled samples.
The analysis clearly demonstrates how the labeled samples improve the convergence rate for the exponential family mixture model.
arXiv Detail & Related papers (2022-11-01T15:42:57Z) - Estimating the Contamination Factor's Distribution in Unsupervised
Anomaly Detection [7.174572371800215]
Anomaly detection methods identify examples that do not follow the expected behaviour.
The proportion of examples marked as anomalies equals the expected proportion of anomalies, called contamination factor.
We introduce a method for estimating the posterior distribution of the contamination factor of a given unlabeled dataset.
arXiv Detail & Related papers (2022-10-19T11:51:25Z) - Scalable Marginal Likelihood Estimation for Model Selection in Deep
Learning [78.83598532168256]
Marginal-likelihood based model-selection is rarely used in deep learning due to estimation difficulties.
Our work shows that marginal likelihoods can improve generalization and be useful when validation data is unavailable.
arXiv Detail & Related papers (2021-04-11T09:50:24Z) - Prediction in the presence of response-dependent missing labels [28.932172873182115]
limitations of sensing technologies result in missing labels in wildfire data.
We develop a new methodology and non-labeled algorithm P(ositive) U(ccurrence) M(agnitude) M(ixture) which jointly estimates the occurrence and detection likelihood of positive samples.
arXiv Detail & Related papers (2021-03-25T01:43:33Z) - Distribution-free uncertainty quantification for classification under
label shift [105.27463615756733]
We focus on uncertainty quantification (UQ) for classification problems via two avenues.
We first argue that label shift hurts UQ, by showing degradation in coverage and calibration.
We examine these techniques theoretically in a distribution-free framework and demonstrate their excellent practical performance.
arXiv Detail & Related papers (2021-03-04T20:51:03Z) - Cautious Active Clustering [79.23797234241471]
We consider the problem of classification of points sampled from an unknown probability measure on a Euclidean space.
Our approach is to consider the unknown probability measure as a convex combination of the conditional probabilities for each class.
arXiv Detail & Related papers (2020-08-03T23:47:31Z) - Learning from Aggregate Observations [82.44304647051243]
We study the problem of learning from aggregate observations where supervision signals are given to sets of instances.
We present a general probabilistic framework that accommodates a variety of aggregate observations.
Simple maximum likelihood solutions can be applied to various differentiable models.
arXiv Detail & Related papers (2020-04-14T06:18:50Z) - A Unified View of Label Shift Estimation [45.472049320861856]
There are two dominant approaches for estimating the label marginal.
We present a unified view of the two methods and the first theoretical characterization of MLLS.
Our analysis attributes BBSE's statistical inefficiency to a loss of information due to coarse calibration.
arXiv Detail & Related papers (2020-03-17T06:28:50Z) - Bayesian Semi-supervised Multi-category Classification under Nonparanormality [2.307581190124002]
Semi-supervised learning is a model training method that uses both labeled and unlabeled data.
This paper proposes a fully Bayes semi-supervised learning algorithm that can be applied to any multi-category classification problem.
arXiv Detail & Related papers (2020-01-11T21:31:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.