Neural Inverse Kinematics
- URL: http://arxiv.org/abs/2205.10837v1
- Date: Sun, 22 May 2022 14:44:07 GMT
- Title: Neural Inverse Kinematics
- Authors: Raphael Bensadoun, Shir Gur, Nitsan Blau, Tom Shenkar, Lior Wolf
- Abstract summary: Inverse kinematic (IK) methods recover the parameters of the joints, given the desired position of selected elements in the kinematic chain.
We propose a neural IK method that employs the hierarchical structure of the problem to sequentially sample valid joint angles conditioned on the desired position.
- Score: 72.85330210991508
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Inverse kinematic (IK) methods recover the parameters of the joints, given
the desired position of selected elements in the kinematic chain. While the
problem is well-defined and low-dimensional, it has to be solved rapidly,
accounting for multiple possible solutions. In this work, we propose a neural
IK method that employs the hierarchical structure of the problem to
sequentially sample valid joint angles conditioned on the desired position and
on the preceding joints along the chain. In our solution, a hypernetwork $f$
recovers the parameters of multiple primary networks {$g_1,g_2,\dots,g_N$,
where $N$ is the number of joints}, such that each $g_i$ outputs a distribution
of possible joint angles, and is conditioned on the sampled values obtained
from the previous primary networks $g_j, j<i$. The hypernetwork can be trained
on readily available pairs of matching joint angles and positions, without
observing multiple solutions. At test time, a high-variance joint distribution
is presented, by sampling sequentially from the primary networks. We
demonstrate the advantage of the proposed method both in comparison to other IK
methods for isolated instances of IK and with regard to following the path of
the end effector in Cartesian space.
Related papers
- Revisiting Multi-Permutation Equivariance through the Lens of Irreducible Representations [3.0222726571099665]
We show that non-Siamese layers can improve performance in tasks like graph anomaly detection, weight space alignment, and learning Wasserstein distances.
We also show empirically that these additional non-Siamese layers can improve performance in tasks like graph anomaly detection, weight space alignment, and learning Wasserstein distances.
arXiv Detail & Related papers (2024-10-09T08:19:31Z) - Equivariant Deep Weight Space Alignment [54.65847470115314]
We propose a novel framework aimed at learning to solve the weight alignment problem.
We first prove that weight alignment adheres to two fundamental symmetries and then, propose a deep architecture that respects these symmetries.
arXiv Detail & Related papers (2023-10-20T10:12:06Z) - Joint Bayesian Inference of Graphical Structure and Parameters with a
Single Generative Flow Network [59.79008107609297]
We propose in this paper to approximate the joint posterior over the structure of a Bayesian Network.
We use a single GFlowNet whose sampling policy follows a two-phase process.
Since the parameters are included in the posterior distribution, this leaves more flexibility for the local probability models.
arXiv Detail & Related papers (2023-05-30T19:16:44Z) - On the Effective Number of Linear Regions in Shallow Univariate ReLU
Networks: Convergence Guarantees and Implicit Bias [50.84569563188485]
We show that gradient flow converges in direction when labels are determined by the sign of a target network with $r$ neurons.
Our result may already hold for mild over- parameterization, where the width is $tildemathcalO(r)$ and independent of the sample size.
arXiv Detail & Related papers (2022-05-18T16:57:10Z) - An application of the splitting-up method for the computation of a
neural network representation for the solution for the filtering equations [68.8204255655161]
Filtering equations play a central role in many real-life applications, including numerical weather prediction, finance and engineering.
One of the classical approaches to approximate the solution of the filtering equations is to use a PDE inspired method, called the splitting-up method.
We combine this method with a neural network representation to produce an approximation of the unnormalised conditional distribution of the signal process.
arXiv Detail & Related papers (2022-01-10T11:01:36Z) - PCAM: Product of Cross-Attention Matrices for Rigid Registration of
Point Clouds [79.99653758293277]
PCAM is a neural network whose key element is a pointwise product of cross-attention matrices.
We show that PCAM achieves state-of-the-art results among methods which, like us, solve steps (a) and (b) jointly via deepnets.
arXiv Detail & Related papers (2021-10-04T09:23:27Z) - A hybrid variance-reduced method for decentralized stochastic non-convex
optimization [15.447966950703947]
textttGTHSGD algorithm specialized local hybrid gradient implements the network to track the global gradient.
textttGTHSGD achieves a network complexity of$O(n-1)$ when the required error tolerance$epsilon$ is small enough.
arXiv Detail & Related papers (2021-02-12T20:13:05Z) - Simultaneous Decision Making for Stochastic Multi-echelon Inventory
Optimization with Deep Neural Networks as Decision Makers [0.7614628596146599]
We propose a framework that uses deep neural networks (DNN) to optimize inventory decisions in complex multi-echelon supply chains.
Our method is suitable for a wide variety of supply chain networks, including general topologies that may contain both assembly and distribution nodes.
arXiv Detail & Related papers (2020-06-10T02:02:52Z) - Efficient Clustering for Stretched Mixtures: Landscape and Optimality [4.2111286819721485]
This paper considers a canonical clustering problem where one receives unlabeled samples drawn from a balanced mixture of two elliptical distributions.
We show that the non-optimal clustering function exhibits desirable geometric properties when the sample size exceeds some constant statistical objectives.
arXiv Detail & Related papers (2020-03-22T17:57:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.