Contrastive estimation reveals topic posterior information to linear
models
- URL: http://arxiv.org/abs/2003.02234v1
- Date: Wed, 4 Mar 2020 18:20:55 GMT
- Title: Contrastive estimation reveals topic posterior information to linear
models
- Authors: Christopher Tosh and Akshay Krishnamurthy and Daniel Hsu
- Abstract summary: Contrastive learning is an approach to representation learning that utilizes naturally occurring similar and dissimilar pairs of data points to find useful embeddings of data.
We prove that contrastive learning is capable of recovering a representation of documents that reveals their underlying topic posterior information to linear models.
- Score: 38.80336134485453
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contrastive learning is an approach to representation learning that utilizes
naturally occurring similar and dissimilar pairs of data points to find useful
embeddings of data. In the context of document classification under topic
modeling assumptions, we prove that contrastive learning is capable of
recovering a representation of documents that reveals their underlying topic
posterior information to linear models. We apply this procedure in a
semi-supervised setup and demonstrate empirically that linear classifiers with
these representations perform well in document classification tasks with very
few training examples.
Related papers
- Reinforcing Pre-trained Models Using Counterfactual Images [54.26310919385808]
This paper proposes a novel framework to reinforce classification models using language-guided generated counterfactual images.
We identify model weaknesses by testing the model using the counterfactual image dataset.
We employ the counterfactual images as an augmented dataset to fine-tune and reinforce the classification model.
arXiv Detail & Related papers (2024-06-19T08:07:14Z) - Contextualizing the Limits of Model & Evaluation Dataset Curation on
Semantic Similarity Classification Tasks [1.8130068086063336]
This paper demonstrates how the limitations of pre-trained models and open evaluation datasets factor into assessing the performance of binary semantic similarity classification tasks.
As (1) end-user-facing documentation around the curation of these datasets and pre-trained model training regimes is often not easily accessible and (2) given the lower friction and higher demand to quickly deploy such systems in real-world contexts, our study reinforces prior work showing performance disparities across datasets, embedding techniques and distance metrics.
arXiv Detail & Related papers (2023-11-03T17:12:07Z) - Learning ECG signal features without backpropagation [0.0]
We propose a novel method to generate representations for time series-type data.
This method relies on ideas from theoretical physics to construct a compact representation in a data-driven way.
We demonstrate the effectiveness of our approach on the task of ECG signal classification, achieving state-of-the-art performance.
arXiv Detail & Related papers (2023-07-04T21:35:49Z) - Estimating class separability of text embeddings with persistent homology [1.9956517534421363]
This paper introduces an unsupervised method to estimate the class separability of text datasets from a topological point of view.
We show how this technique can be applied to detect when the training process stops improving the separability of the embeddings.
Our results, validated across binary and multi-class text classification tasks, show that the proposed method's estimates of class separability align with those obtained from supervised methods.
arXiv Detail & Related papers (2023-05-24T10:58:09Z) - Incorporating Relevance Feedback for Information-Seeking Retrieval using
Few-Shot Document Re-Ranking [56.80065604034095]
We introduce a kNN approach that re-ranks documents based on their similarity with the query and the documents the user considers relevant.
To evaluate our different integration strategies, we transform four existing information retrieval datasets into the relevance feedback scenario.
arXiv Detail & Related papers (2022-10-19T16:19:37Z) - Generalization Properties of Retrieval-based Models [50.35325326050263]
Retrieval-based machine learning methods have enjoyed success on a wide range of problems.
Despite growing literature showcasing the promise of these models, the theoretical underpinning for such models remains underexplored.
We present a formal treatment of retrieval-based models to characterize their generalization ability.
arXiv Detail & Related papers (2022-10-06T00:33:01Z) - VLCDoC: Vision-Language Contrastive Pre-Training Model for Cross-Modal
Document Classification [3.7798600249187295]
Multimodal learning from document data has achieved great success lately as it allows to pre-train semantically meaningful features as a prior into a learnable downstream task.
In this paper, we approach the document classification problem by learning cross-modal representations through language and vision cues.
The proposed method exploits high-level interactions and learns relevant semantic information from effective attention flows within and across modalities.
arXiv Detail & Related papers (2022-05-24T12:28:12Z) - Layer-wise Analysis of a Self-supervised Speech Representation Model [26.727775920272205]
Self-supervised learning approaches have been successful for pre-training speech representation models.
Not much has been studied about the type or extent of information encoded in the pre-trained representations themselves.
arXiv Detail & Related papers (2021-07-10T02:13:25Z) - Combining Feature and Instance Attribution to Detect Artifacts [62.63504976810927]
We propose methods to facilitate identification of training data artifacts.
We show that this proposed training-feature attribution approach can be used to uncover artifacts in training data.
We execute a small user study to evaluate whether these methods are useful to NLP researchers in practice.
arXiv Detail & Related papers (2021-07-01T09:26:13Z) - Relation-Guided Representation Learning [53.60351496449232]
We propose a new representation learning method that explicitly models and leverages sample relations.
Our framework well preserves the relations between samples.
By seeking to embed samples into subspace, we show that our method can address the large-scale and out-of-sample problem.
arXiv Detail & Related papers (2020-07-11T10:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.