"Normalized Stress" is Not Normalized: How to Interpret Stress Correctly
- URL: http://arxiv.org/abs/2408.07724v1
- Date: Wed, 14 Aug 2024 13:42:47 GMT
- Title: "Normalized Stress" is Not Normalized: How to Interpret Stress Correctly
- Authors: Kiran Smelser, Jacob Miller, Stephen Kobourov,
- Abstract summary: Stress is among the most commonly employed quality metrics and optimization criteria for dimension reduction projections of high dimensional data.
One of the most commonly employed metrics, normalized stress, is sensitive to uniform scaling of the projection, despite this act not meaningfully changing anything about the projection.
We introduce a simple technique to make normalized stress scale invariant and show that it accurately captures expected behavior on a small benchmark.
- Score: 0.4915744683251151
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Stress is among the most commonly employed quality metrics and optimization criteria for dimension reduction projections of high dimensional data. Complex, high dimensional data is ubiquitous across many scientific disciplines, including machine learning, biology, and the social sciences. One of the primary methods of visualizing these datasets is with two dimensional scatter plots that visually capture some properties of the data. Because visually determining the accuracy of these plots is challenging, researchers often use quality metrics to measure projection accuracy or faithfulness to the full data. One of the most commonly employed metrics, normalized stress, is sensitive to uniform scaling of the projection, despite this act not meaningfully changing anything about the projection. We investigate the effect of scaling on stress and other distance based quality metrics analytically and empirically by showing just how much the values change and how this affects dimension reduction technique evaluations. We introduce a simple technique to make normalized stress scale invariant and show that it accurately captures expected behavior on a small benchmark.
Related papers
- Generalization Below the Edge of Stability: The Role of Data Geometry [60.147710896851045]
We show how data geometry controls generalization in ReLU networks trained below the edge of stability.<n>For data distributions supported on a mixture of low-dimensional balls, we derive generalization bounds that provably adapt to the intrinsic dimension.<n>Our results consolidate disparate empirical findings that have appeared in the literature.
arXiv Detail & Related papers (2025-10-20T21:40:36Z) - How Scale Breaks "Normalized Stress" and KL Divergence: Rethinking Quality Metrics [0.20999222360659606]
Researchers often use quality metrics to measure the accuracy of two-dimensional scatter plots.<n>One of the most commonly employed metrics, normalized stress, is sensitive to uniform scaling (stretching, shrinking) of the projection.<n>We show just how much the values change and how this affects dimension reduction technique evaluations.
arXiv Detail & Related papers (2025-10-09T13:11:31Z) - A statistical theory of overfitting for imbalanced classification [0.6144680854063939]
We develop a statistical theory for high-dimensional imbalanced classification.
We find that dimensionality induces truncation or skewing effects on the logit distribution.
This phenomenon explains why the minority class is more severely affected by overfitting.
arXiv Detail & Related papers (2025-02-17T00:21:33Z) - The Star Geometry of Critic-Based Regularizer Learning [2.2530496464901106]
Variational regularization is a technique to solve statistical inference tasks and inverse problems.
Recent works learn task-dependent regularizers by integrating information about the measurements and ground-truth data.
There is little theory about the structure of regularizers learned via this process and how it relates to the two data distributions.
arXiv Detail & Related papers (2024-08-29T18:34:59Z) - Webcam-based Pupil Diameter Prediction Benefits from Upscaling [6.298516754485939]
This study evaluates the impact of various upscaling methods, ranging from bicubic to advanced super-resolution, on pupil diameter predictions.
Our findings suggest that pupil diameter prediction models trained on upscaled datasets are highly sensitive to the selected upscaling method and scale.
arXiv Detail & Related papers (2024-08-19T20:28:39Z) - Leveraging Gradients for Unsupervised Accuracy Estimation under
Distribution Shift [25.951051758560702]
Estimating test accuracy without access to the ground-truth test labels under varying test environments is a challenging, yet extremely important problem.
We use the norm of classification-layer gradients, backpropagated from the cross-entropy loss after only one step over test data.
Our key idea is that the model should be adjusted with a higher magnitude of gradients when it does not generalize to the test dataset with a distribution shift.
arXiv Detail & Related papers (2024-01-17T01:33:23Z) - NeuralGF: Unsupervised Point Normal Estimation by Learning Neural
Gradient Function [55.86697795177619]
Normal estimation for 3D point clouds is a fundamental task in 3D geometry processing.
We introduce a new paradigm for learning neural gradient functions, which encourages the neural network to fit the input point clouds.
Our excellent results on widely used benchmarks demonstrate that our method can learn more accurate normals for both unoriented and oriented normal estimation tasks.
arXiv Detail & Related papers (2023-11-01T09:25:29Z) - Graph Out-of-Distribution Generalization with Controllable Data
Augmentation [51.17476258673232]
Graph Neural Network (GNN) has demonstrated extraordinary performance in classifying graph properties.
Due to the selection bias of training and testing data, distribution deviation is widespread.
We propose OOD calibration to measure the distribution deviation of virtual samples.
arXiv Detail & Related papers (2023-08-16T13:10:27Z) - An evaluation framework for dimensionality reduction through sectional
curvature [59.40521061783166]
In this work, we aim to introduce the first highly non-supervised dimensionality reduction performance metric.
To test its feasibility, this metric has been used to evaluate the performance of the most commonly used dimension reduction algorithms.
A new parameterized problem instance generator has been constructed in the form of a function generator.
arXiv Detail & Related papers (2023-03-17T11:59:33Z) - Revisiting lp-constrained Softmax Loss: A Comprehensive Study [2.570570340104555]
We investigate the performance of lp-constrained softmax loss classifiers across different norm orders, magnitudes, and data dimensions.
Experimental results suggest collectively that lp-constrained softmax loss classifiers can achieve more accurate classification results.
We suggest that lp normalization is a recommended data representation practice for image classification in terms of performance and convergence.
arXiv Detail & Related papers (2022-06-20T08:03:12Z) - Deep DIC: Deep Learning-Based Digital Image Correlation for End-to-End
Displacement and Strain Measurement [4.999506391054041]
Digital image correlation (DIC) has become an industry standard to retrieve accurate displacement and strain measurement.
Two convolutional neural networks, DisplacementNet and StrainNet, are designed to work together for end-to-end prediction of displacements and strains.
Deep DIC gives highly consistent and comparable predictions of displacement and strain with those obtained from commercial DIC software.
arXiv Detail & Related papers (2021-10-26T14:13:57Z) - Virtual Normal: Enforcing Geometric Constraints for Accurate and Robust
Depth Prediction [87.08227378010874]
We show the importance of the high-order 3D geometric constraints for depth prediction.
By designing a loss term that enforces a simple geometric constraint, we significantly improve the accuracy and robustness of monocular depth estimation.
We show state-of-the-art results of learning metric depth on NYU Depth-V2 and KITTI.
arXiv Detail & Related papers (2021-03-07T00:08:21Z) - REVISE: A Tool for Measuring and Mitigating Bias in Visual Datasets [64.76453161039973]
REVISE (REvealing VIsual biaSEs) is a tool that assists in the investigation of a visual dataset.
It surfacing potential biases along three dimensions: (1) object-based, (2) person-based, and (3) geography-based.
arXiv Detail & Related papers (2020-04-16T23:54:37Z) - Post-Estimation Smoothing: A Simple Baseline for Learning with Side
Information [102.18616819054368]
We propose a post-estimation smoothing operator as a fast and effective method for incorporating structural index data into prediction.
Because the smoothing step is separate from the original predictor, it applies to a broad class of machine learning tasks.
Our experiments on large scale spatial and temporal datasets highlight the speed and accuracy of post-estimation smoothing in practice.
arXiv Detail & Related papers (2020-03-12T18:04:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.