Measuring the Novelty of Natural Language Text Using the Conjunctive
Clauses of a Tsetlin Machine Text Classifier
- URL: http://arxiv.org/abs/2011.08755v1
- Date: Tue, 17 Nov 2020 16:35:21 GMT
- Title: Measuring the Novelty of Natural Language Text Using the Conjunctive
Clauses of a Tsetlin Machine Text Classifier
- Authors: Bimal Bhattarai, Ole-Christoffer Granmo, Lei Jiao
- Abstract summary: Most supervised text classification approaches assume a closed world, counting on all classes being present in the data at training time.
This assumption can lead to unpredictable behaviour during operation, whenever novel, previously unseen, classes appear.
We extend the recently introduced Tsetlin machine (TM) with a novelty scoring mechanism.
- Score: 12.087658145293522
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most supervised text classification approaches assume a closed world,
counting on all classes being present in the data at training time. This
assumption can lead to unpredictable behaviour during operation, whenever
novel, previously unseen, classes appear. Although deep learning-based methods
have recently been used for novelty detection, they are challenging to
interpret due to their black-box nature. This paper addresses
\emph{interpretable} open-world text classification, where the trained
classifier must deal with novel classes during operation. To this end, we
extend the recently introduced Tsetlin machine (TM) with a novelty scoring
mechanism. The mechanism uses the conjunctive clauses of the TM to measure to
what degree a text matches the classes covered by the training data. We
demonstrate that the clauses provide a succinct interpretable description of
known topics, and that our scoring mechanism makes it possible to discern novel
topics from the known ones. Empirically, our TM-based approach outperforms
seven other novelty detection schemes on three out of five datasets, and
performs second and third best on the remaining, with the added benefit of an
interpretable propositional logic-based representation.
Related papers
- A Fixed-Point Approach to Unified Prompt-Based Counting [51.20608895374113]
This paper aims to establish a comprehensive prompt-based counting framework capable of generating density maps for objects indicated by various prompt types, such as box, point, and text.
Our model excels in prominent class-agnostic datasets and exhibits superior performance in cross-dataset adaptation tasks.
arXiv Detail & Related papers (2024-03-15T12:05:44Z) - XAI-CLASS: Explanation-Enhanced Text Classification with Extremely Weak
Supervision [6.406111099707549]
XAI-CLASS is a novel explanation-enhanced weakly-supervised text classification method.
It incorporates word saliency prediction as an auxiliary task.
XAI-CLASS outperforms other weakly-supervised text classification methods significantly.
arXiv Detail & Related papers (2023-10-31T23:24:22Z) - WOT-Class: Weakly Supervised Open-world Text Classification [41.77945049159303]
We work on a novel problem of weakly supervised open-world text classification.
We propose a novel framework WOT-Class that lifts strong assumptions.
Experiments on 7 popular text classification datasets demonstrate that WOT-Class outperforms strong baselines.
arXiv Detail & Related papers (2023-05-21T08:51:24Z) - Three ways to improve feature alignment for open vocabulary detection [88.65076922242184]
Key problem in zero-shot open vocabulary detection is how to align visual and text features, so that the detector performs well on unseen classes.
Previous approaches train the feature pyramid and detection head from scratch, which breaks the vision-text feature alignment established during pretraining.
We propose three methods to alleviate these issues. Firstly, a simple scheme is used to augment the text embeddings which prevents overfitting to a small number of classes seen during training.
Secondly, the feature pyramid network and the detection head are modified to include trainable shortcuts.
Finally, a self-training approach is used to leverage a larger corpus of
arXiv Detail & Related papers (2023-03-23T17:59:53Z) - Class Enhancement Losses with Pseudo Labels for Zero-shot Semantic
Segmentation [40.09476732999614]
Mask proposal models have significantly improved the performance of zero-shot semantic segmentation.
The use of a background' embedding during training in these methods is problematic as the resulting model tends to over-learn and assign all unseen classes as the background class instead of their correct labels.
This paper proposes novel class enhancement losses to bypass the use of the background embbedding during training, and simultaneously exploit the semantic relationship between text embeddings and mask proposals by ranking the similarity scores.
arXiv Detail & Related papers (2023-01-18T06:55:02Z) - Task-Specific Embeddings for Ante-Hoc Explainable Text Classification [6.671252951387647]
We propose an alternative training objective in which we learn task-specific embeddings of text.
Our proposed objective learns embeddings such that all texts that share the same target class label should be close together.
We present extensive experiments which show that the benefits of ante-hoc explainability and incremental learning come at no cost in overall classification accuracy.
arXiv Detail & Related papers (2022-11-30T19:56:25Z) - Textual Entailment Recognition with Semantic Features from Empirical
Text Representation [60.31047947815282]
A text entails a hypothesis if and only if the true value of the hypothesis follows the text.
In this paper, we propose a novel approach to identifying the textual entailment relationship between text and hypothesis.
We employ an element-wise Manhattan distance vector-based feature that can identify the semantic entailment relationship between the text-hypothesis pair.
arXiv Detail & Related papers (2022-10-18T10:03:51Z) - Distant finetuning with discourse relations for stance classification [55.131676584455306]
We propose a new method to extract data with silver labels from raw text to finetune a model for stance classification.
We also propose a 3-stage training framework where the noisy level in the data used for finetuning decreases over different stages.
Our approach ranks 1st among 26 competing teams in the stance classification track of the NLPCC 2021 shared task Argumentative Text Understanding for AI Debater.
arXiv Detail & Related papers (2022-04-27T04:24:35Z) - Semi-supervised New Event Type Induction and Description via Contrastive
Loss-Enforced Batch Attention [56.46649994444616]
We present a novel approach to semi-supervised new event type induction using a masked contrastive loss.
We extend our approach to two new tasks: predicting the type name of the discovered clusters and linking them to FrameNet frames.
arXiv Detail & Related papers (2022-02-12T00:32:22Z) - Word-level Human Interpretable Scoring Mechanism for Novel Text
Detection Using Tsetlin Machines [16.457778420360537]
We propose a Tsetlin machine architecture for scoring individual words according to their contribution to novelty.
Our approach encodes a description of the novel documents using the linguistic patterns captured by TM clauses.
We then adopt this description to measure how much a word contributes to making documents novel.
arXiv Detail & Related papers (2021-05-10T23:41:14Z) - Learning Interpretable and Discrete Representations with Adversarial
Training for Unsupervised Text Classification [87.28408260725138]
TIGAN learns to encode texts into two disentangled representations, including a discrete code and a continuous noise.
The extracted topical words for representing latent topics show that TIGAN learns coherent and highly interpretable topics.
arXiv Detail & Related papers (2020-04-28T02:53:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.