Learning Positional Embeddings for Coordinate-MLPs
- URL: http://arxiv.org/abs/2112.11577v1
- Date: Tue, 21 Dec 2021 23:23:33 GMT
- Title: Learning Positional Embeddings for Coordinate-MLPs
- Authors: Sameera Ramasinghe, Simon Lucey
- Abstract summary: We develop a generic framework to learn the positional embedding based on the classic graph-Laplacian regularization.
We show that the proposed embedding achieves better performance with higher stability compared to the well-established random Fourier features.
- Score: 37.56813817513575
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel method to enhance the performance of coordinate-MLPs by
learning instance-specific positional embeddings. End-to-end optimization of
positional embedding parameters along with network weights leads to poor
generalization performance. Instead, we develop a generic framework to learn
the positional embedding based on the classic graph-Laplacian regularization,
which can implicitly balance the trade-off between memorization and
generalization. This framework is then used to propose a novel positional
embedding scheme, where the hyperparameters are learned per coordinate (i.e,
instance) to deliver optimal performance. We show that the proposed embedding
achieves better performance with higher stability compared to the
well-established random Fourier features (RFF). Further, we demonstrate that
the proposed embedding scheme yields stable gradients, enabling seamless
integration into deep architectures as intermediate layers.
Related papers
- LPGD: A General Framework for Backpropagation through Embedded Optimization Layers [26.251442093195376]
We propose a flexible framework for training architectures with embedded optimization layers.
Lagrangian Proximal Gradient Descent efficiently computes meaningful replacements of degenerate layer derivatives.
We theoretically analyze our method and demonstrate on historical and synthetic data that LPGD converges faster than gradient descent even in a differentiable setup.
arXiv Detail & Related papers (2024-07-08T13:27:41Z) - SPARE: Symmetrized Point-to-Plane Distance for Robust Non-Rigid Registration [76.40993825836222]
We propose SPARE, a novel formulation that utilizes a symmetrized point-to-plane distance for robust non-rigid registration.
The proposed method greatly improves the accuracy of non-rigid registration problems and maintains relatively high solution efficiency.
arXiv Detail & Related papers (2024-05-30T15:55:04Z) - Beyond Single-Model Views for Deep Learning: Optimization versus
Generalizability of Stochastic Optimization Algorithms [13.134564730161983]
This paper adopts a novel approach to deep learning optimization, focusing on gradient descent (SGD) and its variants.
We show that SGD and its variants demonstrate performance on par with flat-minimas like SAM, albeit with half the gradient evaluations.
Our study uncovers several key findings regarding the relationship between training loss and hold-out accuracy, as well as the comparable performance of SGD and noise-enabled variants.
arXiv Detail & Related papers (2024-03-01T14:55:22Z) - PARL: A Unified Framework for Policy Alignment in Reinforcement Learning from Human Feedback [106.63518036538163]
We present a novel unified bilevel optimization-based framework, textsfPARL, formulated to address the recently highlighted critical issue of policy alignment in reinforcement learning.
Our framework addressed these concerns by explicitly parameterizing the distribution of the upper alignment objective (reward design) by the lower optimal variable.
Our empirical results substantiate that the proposed textsfPARL can address the alignment concerns in RL by showing significant improvements.
arXiv Detail & Related papers (2023-08-03T18:03:44Z) - Accelerated Federated Learning with Decoupled Adaptive Optimization [53.230515878096426]
federated learning (FL) framework enables clients to collaboratively learn a shared model while keeping privacy of training data on clients.
Recently, many iterations efforts have been made to generalize centralized adaptive optimization methods, such as SGDM, Adam, AdaGrad, etc., to federated settings.
This work aims to develop novel adaptive optimization methods for FL from the perspective of dynamics of ordinary differential equations (ODEs)
arXiv Detail & Related papers (2022-07-14T22:46:43Z) - FEDNEST: Federated Bilevel, Minimax, and Compositional Optimization [53.78643974257301]
Many contemporary ML problems fall under nested bilevel programming that subsumes minimax and compositional optimization.
We propose FedNest: A federated alternating gradient method to address general nested problems.
arXiv Detail & Related papers (2022-05-04T17:48:55Z) - A Unified Framework of Bundle Adjustment and Feature Matching for
High-Resolution Satellite Images [4.835738511987696]
This article incorpo-rates Bundle adjustment (BA) and feature matching in a unified framework.
Experiments on multi-view high-resolution satellite images show that our proposed method outperforms state-of-the-art orientation techniques.
arXiv Detail & Related papers (2021-07-01T16:40:25Z) - StochasticRank: Global Optimization of Scale-Free Discrete Functions [28.224889996383396]
In this paper, we introduce a powerful and efficient framework for direct optimization of ranking metrics.
We show that classic smoothing approaches may introduce bias and present a universal solution for a proper debiasing.
Our framework applies to any scale-free discrete loss function.
arXiv Detail & Related papers (2020-03-04T15:27:11Z) - Target-Embedding Autoencoders for Supervised Representation Learning [111.07204912245841]
This paper analyzes a framework for improving generalization in a purely supervised setting, where the target space is high-dimensional.
We motivate and formalize the general framework of target-embedding autoencoders (TEA) for supervised prediction, learning intermediate latent representations jointly optimized to be both predictable from features as well as predictive of targets.
arXiv Detail & Related papers (2020-01-23T02:37:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.