The role of feature space in atomistic learning
- URL: http://arxiv.org/abs/2009.02741v4
- Date: Thu, 10 Dec 2020 10:17:59 GMT
- Title: The role of feature space in atomistic learning
- Authors: Alexander Goscinski and Guillaume Fraux and Giulio Imbalzano and
Michele Ceriotti
- Abstract summary: Physically-inspired descriptors play a key role in the application of machine-learning techniques to atomistic simulations.
We introduce a framework to compare different sets of descriptors, and different ways of transforming them by means of metrics and kernels.
We compare representations built in terms of n-body correlations of the atom density, quantitatively assessing the information loss associated with the use of low-order features.
- Score: 62.997667081978825
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Eficient, physically-inspired descriptors of the structure and composition of
molecules and materials play a key role in the application of machine-learning
techniques to atomistic simulations. The proliferation of approaches, as well
as the fact that each choice of features can lead to very different behavior
depending on how they are used, e.g. by introducing non-linear kernels and
non-Euclidean metrics to manipulate them, makes it difficult to objectively
compare different methods, and to address fundamental questions on how one
feature space is related to another. In this work we introduce a framework to
compare different sets of descriptors, and different ways of transforming them
by means of metrics and kernels, in terms of the structure of the feature space
that they induce. We define diagnostic tools to determine whether alternative
feature spaces contain equivalent amounts of information, and whether the
common information is substantially distorted when going from one feature space
to another. We compare, in particular, representations that are built in terms
of n-body correlations of the atom density, quantitatively assessing the
information loss associated with the use of low-order features. We also
investigate the impact of different choices of basis functions and
hyperparameters of the widely used SOAP and Behler-Parrinello features, and
investigate how the use of non-linear kernels, and of a Wasserstein-type
metric, change the structure of the feature space in comparison to a simpler
linear feature space.
Related papers
- Automatic feature selection and weighting using Differentiable Information Imbalance [41.452380773977154]
We introduce the Differentiable Information Imbalance (DII), an automatic data analysis method to rank information content between sets of features.
Based on the nearest neighbors according to distances in the ground truth feature space, the method finds a low-dimensional subset of the input features.
By employing the Differentiable Information Imbalance as a loss function, the relative feature weights of the inputs are optimized, simultaneously performing unit alignment and relative importance scaling.
arXiv Detail & Related papers (2024-10-30T11:19:10Z) - Normalization in Proportional Feature Spaces [49.48516314472825]
normalization plays an important central role in data representation, characterization, visualization, analysis, comparison, classification, and modeling.
The selection of an appropriate normalization method needs to take into account the type and characteristics of the involved features.
arXiv Detail & Related papers (2024-09-17T17:46:27Z) - Going Beyond Neural Network Feature Similarity: The Network Feature
Complexity and Its Interpretation Using Category Theory [64.06519549649495]
We provide the definition of what we call functionally equivalent features.
These features produce equivalent output under certain transformations.
We propose an efficient algorithm named Iterative Feature Merging.
arXiv Detail & Related papers (2023-10-10T16:27:12Z) - Measuring dissimilarity with diffeomorphism invariance [94.02751799024684]
We introduce DID, a pairwise dissimilarity measure applicable to a wide range of data spaces.
We prove that DID enjoys properties which make it relevant for theoretical study and practical use.
arXiv Detail & Related papers (2022-02-11T13:51:30Z) - Feature Weighted Non-negative Matrix Factorization [92.45013716097753]
We propose the Feature weighted Non-negative Matrix Factorization (FNMF) in this paper.
FNMF learns the weights of features adaptively according to their importances.
It can be solved efficiently with the suggested optimization algorithm.
arXiv Detail & Related papers (2021-03-24T21:17:17Z) - Product Manifold Learning [7.394643746798322]
We present a new paradigm for non-linear independent component analysis called manifold factorization.
We demonstrate the potential use of our method for an important and challenging problem in structural biology.
arXiv Detail & Related papers (2020-10-19T22:51:06Z) - Linear Classifier Combination via Multiple Potential Functions [0.6091702876917279]
We propose a novel concept of calculating a scoring function based on the distance of the object from the decision boundary and its distance to the class centroid.
An important property is that the proposed score function has the same nature for all linear base classifiers.
arXiv Detail & Related papers (2020-10-02T08:11:51Z) - A Functional Perspective on Learning Symmetric Functions with Neural
Networks [48.80300074254758]
We study the learning and representation of neural networks defined on measures.
We establish approximation and generalization bounds under different choices of regularization.
The resulting models can be learned efficiently and enjoy generalization guarantees that extend across input sizes.
arXiv Detail & Related papers (2020-08-16T16:34:33Z) - A Deep Joint Sparse Non-negative Matrix Factorization Framework for
Identifying the Common and Subject-specific Functional Units of Tongue Motion
During Speech [7.870139900799612]
We develop a new deep learning framework to identify common and subject-specific functional units of tongue motion during speech.
We transform NMF with sparse and graph regularizations into modular architectures akin to deep neural networks.
arXiv Detail & Related papers (2020-07-09T15:05:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.