Persistence Homology of TEDtalk: Do Sentence Embeddings Have a
Topological Shape?
- URL: http://arxiv.org/abs/2103.14131v1
- Date: Thu, 25 Mar 2021 20:52:17 GMT
- Title: Persistence Homology of TEDtalk: Do Sentence Embeddings Have a
Topological Shape?
- Authors: Shouman Das, Syed A. Haque, Md. Iftekhar Tanveer
- Abstract summary: We investigate the possibility of applying TDA to improve the classification accuracy of public speaking rating.
We calculated emphpersistence image vectors for the sentence embeddings of TEDtalk data and feed this vectors as additional inputs to our machine learning models.
From our results, we could not conclude that the topological shapes of the sentence embeddings can help us train a better model for public speaking rating.
- Score: 3.1675545188012078
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: \emph{Topological data analysis} (TDA) has recently emerged as a new
technique to extract meaningful discriminitve features from high dimensional
data. In this paper, we investigate the possibility of applying TDA to improve
the classification accuracy of public speaking rating. We calculated
\emph{persistence image vectors} for the sentence embeddings of TEDtalk data
and feed this vectors as additional inputs to our machine learning models. We
have found a negative result that this topological information does not improve
the model accuracy significantly. In some cases, it makes the accuracy slightly
worse than the original one. From our results, we could not conclude that the
topological shapes of the sentence embeddings can help us train a better model
for public speaking rating.
Related papers
- Towards Model-Agnostic Posterior Approximation for Fast and Accurate Variational Autoencoders [22.77397537980102]
We show that we can compute a deterministic, model-agnostic posterior approximation (MAPA) of the true model's posterior.
We present preliminary results on low-dimensional synthetic data that (1) MAPA captures the trend of the true posterior, and (2) our MAPA-based inference performs better density estimation with less computation than baselines.
arXiv Detail & Related papers (2024-03-13T20:16:21Z) - Towards Fine-Grained Information: Identifying the Type and Location of
Translation Errors [80.22825549235556]
Existing approaches can not synchronously consider error position and type.
We build an FG-TED model to predict the textbf addition and textbfomission errors.
Experiments show that our model can identify both error type and position concurrently, and gives state-of-the-art results.
arXiv Detail & Related papers (2023-02-17T16:20:33Z) - Topological Data Analysis for Speech Processing [10.00176964652466]
We show that a simple linear classifier built on top of such features outperforms a fine-tuned classification head.
We also show that topological features are able to reveal functional roles of speech Transformer heads.
arXiv Detail & Related papers (2022-11-30T18:22:37Z) - Stochastic gradient descent with gradient estimator for categorical
features [3.597778914286147]
Categorical data are present in key areas such as health or supply chain, and this data require specific treatment.
In order to apply recent machine learning models on such data, encoding is needed.
In order to build interpretable models, one-hot encoding is still a very good, but such encoding sparse data.
We show what this estimator minimizes in theory and show its efficiency on different datasets with multiple model architectures.
arXiv Detail & Related papers (2022-09-08T12:24:27Z) - Evaluating State-of-the-Art Classification Models Against Bayes
Optimality [106.50867011164584]
We show that we can compute the exact Bayes error of generative models learned using normalizing flows.
We use our approach to conduct a thorough investigation of state-of-the-art classification models.
arXiv Detail & Related papers (2021-06-07T06:21:20Z) - Generative Text Modeling through Short Run Inference [47.73892773331617]
The present work proposes a short run dynamics for inference. It is variation from the prior distribution of the latent variable and then runs a small number of Langevin dynamics steps guided by its posterior distribution.
We show that the models trained with short run dynamics more accurately model the data, compared to strong language model and VAE baselines, and exhibit no sign of posterior collapse.
arXiv Detail & Related papers (2021-05-27T09:14:35Z) - Learnable Boundary Guided Adversarial Training [66.57846365425598]
We use the model logits from one clean model to guide learning of another one robust model.
We achieve new state-of-the-art robustness on CIFAR-100 without additional real or synthetic data.
arXiv Detail & Related papers (2020-11-23T01:36:05Z) - Improving Robustness by Augmenting Training Sentences with
Predicate-Argument Structures [62.562760228942054]
Existing approaches to improve robustness against dataset biases mostly focus on changing the training objective.
We propose to augment the input sentences in the training data with their corresponding predicate-argument structures.
We show that without targeting a specific bias, our sentence augmentation improves the robustness of transformer models against multiple biases.
arXiv Detail & Related papers (2020-10-23T16:22:05Z) - Evaluating Prediction-Time Batch Normalization for Robustness under
Covariate Shift [81.74795324629712]
We call prediction-time batch normalization, which significantly improves model accuracy and calibration under covariate shift.
We show that prediction-time batch normalization provides complementary benefits to existing state-of-the-art approaches for improving robustness.
The method has mixed results when used alongside pre-training, and does not seem to perform as well under more natural types of dataset shift.
arXiv Detail & Related papers (2020-06-19T05:08:43Z) - Rethinking the Role of Gradient-Based Attribution Methods for Model
Interpretability [8.122270502556374]
Current methods for interpretability of discriminative deep neural networks rely on the model's input-gradients.
We show that these input-gradients can be arbitrarily manipulated without changing the discriminative function.
arXiv Detail & Related papers (2020-06-16T13:17:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.