$\sigma^2$R Loss: a Weighted Loss by Multiplicative Factors using
Sigmoidal Functions
- URL: http://arxiv.org/abs/2009.08796v1
- Date: Fri, 18 Sep 2020 12:34:40 GMT
- Title: $\sigma^2$R Loss: a Weighted Loss by Multiplicative Factors using
Sigmoidal Functions
- Authors: Riccardo La Grassa, Ignazio Gallo, Nicola Landro
- Abstract summary: We introduce a new loss function called squared reduction loss ($sigma2$R loss), which is regulated by a sigmoid function to inflate/deflate the error per instance.
Our loss has clear intuition and geometric interpretation, we demonstrate by experiments the effectiveness of our proposal.
- Score: 0.9569316316728905
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In neural networks, the loss function represents the core of the learning
process that leads the optimizer to an approximation of the optimal convergence
error. Convolutional neural networks (CNN) use the loss function as a
supervisory signal to train a deep model and contribute significantly to
achieving the state of the art in some fields of artificial vision.
Cross-entropy and Center loss functions are commonly used to increase the
discriminating power of learned functions and increase the generalization
performance of the model. Center loss minimizes the class intra-class variance
and at the same time penalizes the long distance between the deep features
inside each class. However, the total error of the center loss will be heavily
influenced by the majority of the instances and can lead to a freezing state in
terms of intra-class variance. To address this, we introduce a new loss
function called sigma squared reduction loss ($\sigma^2$R loss), which is
regulated by a sigmoid function to inflate/deflate the error per instance and
then continue to reduce the intra-class variance. Our loss has clear intuition
and geometric interpretation, furthermore, we demonstrate by experiments the
effectiveness of our proposal on several benchmark datasets showing the
intra-class variance reduction and overcoming the results obtained with center
loss and soft nearest neighbour functions.
Related papers
- Large Margin Discriminative Loss for Classification [3.3975558777609915]
We introduce a novel discriminative loss function with large margin in the context of Deep Learning.
This loss boosts the discriminative power of neural nets, represented by intra-class compactness and inter-class separability.
arXiv Detail & Related papers (2024-05-28T18:10:45Z) - On the Dynamics Under the Unhinged Loss and Beyond [104.49565602940699]
We introduce the unhinged loss, a concise loss function, that offers more mathematical opportunities to analyze closed-form dynamics.
The unhinged loss allows for considering more practical techniques, such as time-vary learning rates and feature normalization.
arXiv Detail & Related papers (2023-12-13T02:11:07Z) - Alternate Loss Functions for Classification and Robust Regression Can Improve the Accuracy of Artificial Neural Networks [6.452225158891343]
This paper shows that training speed and final accuracy of neural networks can significantly depend on the loss function used to train neural networks.
Two new classification loss functions that significantly improve performance on a wide variety of benchmark tasks are proposed.
arXiv Detail & Related papers (2023-03-17T12:52:06Z) - Implicit Stochastic Gradient Descent for Training Physics-informed
Neural Networks [51.92362217307946]
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and inverse differential equation problems.
PINNs are trapped in training failures when the target functions to be approximated exhibit high-frequency or multi-scale features.
In this paper, we propose to employ implicit gradient descent (ISGD) method to train PINNs for improving the stability of training process.
arXiv Detail & Related papers (2023-03-03T08:17:47Z) - Xtreme Margin: A Tunable Loss Function for Binary Classification
Problems [0.0]
We provide an overview of a novel loss function, the Xtreme Margin loss function.
Unlike the binary cross-entropy and the hinge loss functions, this loss function provides researchers and practitioners flexibility with their training process.
arXiv Detail & Related papers (2022-10-31T22:39:32Z) - Are All Losses Created Equal: A Neural Collapse Perspective [36.0354919583995]
Cross entropy (CE) is the most commonly used loss to train deep neural networks for classification tasks.
We show through global solution and landscape analyses that a broad family of loss functions including commonly used label smoothing (LS) and focal loss (FL) exhibits Neural Collapse.
arXiv Detail & Related papers (2022-10-04T00:36:45Z) - Adaptive Self-supervision Algorithms for Physics-informed Neural
Networks [59.822151945132525]
Physics-informed neural networks (PINNs) incorporate physical knowledge from the problem domain as a soft constraint on the loss function.
We study the impact of the location of the collocation points on the trainability of these models.
We propose a novel adaptive collocation scheme which progressively allocates more collocation points to areas where the model is making higher errors.
arXiv Detail & Related papers (2022-07-08T18:17:06Z) - Towards an Understanding of Benign Overfitting in Neural Networks [104.2956323934544]
Modern machine learning models often employ a huge number of parameters and are typically optimized to have zero training loss.
We examine how these benign overfitting phenomena occur in a two-layer neural network setting.
We show that it is possible for the two-layer ReLU network interpolator to achieve a near minimax-optimal learning rate.
arXiv Detail & Related papers (2021-06-06T19:08:53Z) - Topological obstructions in neural networks learning [67.8848058842671]
We study global properties of the loss gradient function flow.
We use topological data analysis of the loss function and its Morse complex to relate local behavior along gradient trajectories with global properties of the loss surface.
arXiv Detail & Related papers (2020-12-31T18:53:25Z) - Shaping Deep Feature Space towards Gaussian Mixture for Visual
Classification [74.48695037007306]
We propose a Gaussian mixture (GM) loss function for deep neural networks for visual classification.
With a classification margin and a likelihood regularization, the GM loss facilitates both high classification performance and accurate modeling of the feature distribution.
The proposed model can be implemented easily and efficiently without using extra trainable parameters.
arXiv Detail & Related papers (2020-11-18T03:32:27Z) - AMC-Loss: Angular Margin Contrastive Loss for Improved Explainability in
Image Classification [8.756814963313804]
Angular Margin Contrastive Loss (AMC-Loss) is a new loss function to be used along with the traditional cross-entropy loss.
AMC-Loss employs the discriminative angular distance metric that is equivalent to geodesic distance on a hypersphere manifold.
We find that although the proposed geometrically constrained loss-function improves quantitative results modestly, it has a qualitatively surprisingly beneficial effect on increasing the interpretability of deep-net decisions.
arXiv Detail & Related papers (2020-04-21T08:03:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.