Knowledge-based Document Classification with Shannon Entropy
- URL: http://arxiv.org/abs/2206.02363v1
- Date: Mon, 6 Jun 2022 05:39:10 GMT
- Title: Knowledge-based Document Classification with Shannon Entropy
- Authors: AtMa P.O. Chan
- Abstract summary: We propose a novel knowledge-based model equipped with Shannon Entropy, which measures the richness of information and favors uniform and diverse keyword matches.
We show that the Shannon Entropy significantly improves the recall at fixed level of false positive rate.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Document classification is the detection specific content of interest in text
documents. In contrast to the data-driven machine learning classifiers,
knowledge-based classifiers can be constructed based on domain specific
knowledge, which usually takes the form of a collection of subject related
keywords. While typical knowledge-based classifiers compute a prediction score
based on the keyword abundance, it generally suffers from noisy detections due
to the lack of guiding principle in gauging the keyword matches. In this paper,
we propose a novel knowledge-based model equipped with Shannon Entropy, which
measures the richness of information and favors uniform and diverse keyword
matches. Without invoking any positive sample, such method provides a simple
and explainable solution for document classification. We show that the Shannon
Entropy significantly improves the recall at fixed level of false positive
rate. Also, we show that the model is more robust against change of data
distribution at inference while compared with traditional machine learning,
particularly when the positive training samples are very limited.
Related papers
- Label-template based Few-Shot Text Classification with Contrastive Learning [7.964862748983985]
We propose a simple and effective few-shot text classification framework.
Label templates are embedded into input sentences to fully utilize the potential value of class labels.
supervised contrastive learning is utilized to model the interaction information between support samples and query samples.
arXiv Detail & Related papers (2024-12-13T12:51:50Z) - Context-Specific Refinements of Bayesian Network Classifiers [1.9136291802656262]
We study the relationship between our novel classes of classifiers and Bayesian networks.
We introduce and implement data-driven learning routines for our models.
The study demonstrates that models embedding asymmetric information can enhance classification accuracy.
arXiv Detail & Related papers (2024-05-28T15:50:50Z) - Learning Context-aware Classifier for Semantic Segmentation [88.88198210948426]
In this paper, contextual hints are exploited via learning a context-aware classifier.
Our method is model-agnostic and can be easily applied to generic segmentation models.
With only negligible additional parameters and +2% inference time, decent performance gain has been achieved on both small and large models.
arXiv Detail & Related papers (2023-03-21T07:00:35Z) - Anomaly Detection using Ensemble Classification and Evidence Theory [62.997667081978825]
We present a novel approach for novel detection using ensemble classification and evidence theory.
A pool selection strategy is presented to build a solid ensemble classifier.
We use uncertainty for the anomaly detection approach.
arXiv Detail & Related papers (2022-12-23T00:50:41Z) - An Upper Bound for the Distribution Overlap Index and Its Applications [22.92968284023414]
This paper proposes an easy-to-compute upper bound for the overlap index between two probability distributions.
The proposed bound shows its value in one-class classification and domain shift analysis.
Our work shows significant promise toward broadening the applications of overlap-based metrics.
arXiv Detail & Related papers (2022-12-16T20:02:03Z) - Gacs-Korner Common Information Variational Autoencoder [102.89011295243334]
We propose a notion of common information that allows one to quantify and separate the information that is shared between two random variables.
We demonstrate that our formulation allows us to learn semantically meaningful common and unique factors of variation even on high-dimensional data such as images and videos.
arXiv Detail & Related papers (2022-05-24T17:47:26Z) - Determination of class-specific variables in nonparametric
multiple-class classification [0.0]
We propose a probability-based nonparametric multiple-class classification method, and integrate it with the ability of identifying high impact variables for individual class.
We report the properties of the proposed method, and use both synthesized and real data sets to illustrate its properties under different classification situations.
arXiv Detail & Related papers (2022-05-07T10:08:58Z) - Resolving label uncertainty with implicit posterior models [71.62113762278963]
We propose a method for jointly inferring labels across a collection of data samples.
By implicitly assuming the existence of a generative model for which a differentiable predictor is the posterior, we derive a training objective that allows learning under weak beliefs.
arXiv Detail & Related papers (2022-02-28T18:09:44Z) - Information Theoretic Meta Learning with Gaussian Processes [74.54485310507336]
We formulate meta learning using information theoretic concepts; namely, mutual information and the information bottleneck.
By making use of variational approximations to the mutual information, we derive a general and tractable framework for meta learning.
arXiv Detail & Related papers (2020-09-07T16:47:30Z) - Self-Attentive Classification-Based Anomaly Detection in Unstructured
Logs [59.04636530383049]
We propose Logsy, a classification-based method to learn log representations.
We show an average improvement of 0.25 in the F1 score, compared to the previous methods.
arXiv Detail & Related papers (2020-08-21T07:26:55Z) - Concept Matching for Low-Resource Classification [36.871182660669746]
We propose a model to tackle classification tasks in the presence of very little training data.
We approximate the notion of exact match with a theoretically sound mechanism that computes a probability of matching in the input space.
arXiv Detail & Related papers (2020-06-01T13:34:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.