Local Model Feature Transformations
- URL: http://arxiv.org/abs/2004.06149v1
- Date: Mon, 13 Apr 2020 18:41:03 GMT
- Title: Local Model Feature Transformations
- Authors: CScott Brown
- Abstract summary: Local learning methods are a popular class of machine learning algorithms.
Research on locally-learned models has largely been restricted to simple model families.
We extend the local modeling paradigm to Gaussian processes, quadric models and word embedding models.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Local learning methods are a popular class of machine learning algorithms.
The basic idea for the entire cadre is to choose some non-local model family,
to train many of them on small sections of neighboring data, and then to
`stitch' the resulting models together in some way. Due to the limits of
constraining a training dataset to a small neighborhood, research on
locally-learned models has largely been restricted to simple model families.
Also, since simple model families have no complex structure by design, this has
limited use of the individual local models to predictive tasks. We hypothesize
that, using a sufficiently complex local model family, various properties of
the individual local models, such as their learned parameters, can be used as
features for further learning. This dissertation improves upon the current
state of research and works toward establishing this hypothesis by
investigating algorithms for localization of more complex model families and by
studying their applications beyond predictions as a feature extraction
mechanism. We summarize this generic technique of using local models as a
feature extraction step with the term ``local model feature transformations.''
In this document, we extend the local modeling paradigm to Gaussian processes,
orthogonal quadric models and word embedding models, and extend the existing
theory for localized linear classifiers. We then demonstrate applications of
local model feature transformations to epileptic event classification from EEG
readings, activity monitoring via chest accelerometry, 3D surface
reconstruction, 3D point cloud segmentation, handwritten digit classification
and event detection from Twitter feeds.
Related papers
- MASALA: Model-Agnostic Surrogate Explanations by Locality Adaptation [3.587367153279351]
Existing local Explainable AI (XAI) methods select a region of the input space in the vicinity of a given input instance, for which they approximate the behaviour of a model using a simpler and more interpretable surrogate model.
We propose a novel method, MASALA, for generating explanations, which automatically determines the appropriate local region of impactful model behaviour for each individual instance being explained.
arXiv Detail & Related papers (2024-08-19T15:26:45Z) - A Two-Phase Recall-and-Select Framework for Fast Model Selection [13.385915962994806]
We propose a two-phase (coarse-recall and fine-selection) model selection framework.
It aims to enhance the efficiency of selecting a robust model by leveraging the models' training performances on benchmark datasets.
It has been demonstrated that the proposed methodology facilitates the selection of a high-performing model at a rate about 3x times faster than conventional baseline methods.
arXiv Detail & Related papers (2024-03-28T14:44:44Z) - Towards Model-Agnostic Federated Learning over Networks [0.0]
We present a model-agnostic federated learning method for networks of heterogeneous data and models.
Our method is an instance of empirical risk minimization, with the regularization term derived from the network structure of data.
arXiv Detail & Related papers (2023-02-08T22:55:57Z) - Generalization Properties of Retrieval-based Models [50.35325326050263]
Retrieval-based machine learning methods have enjoyed success on a wide range of problems.
Despite growing literature showcasing the promise of these models, the theoretical underpinning for such models remains underexplored.
We present a formal treatment of retrieval-based models to characterize their generalization ability.
arXiv Detail & Related papers (2022-10-06T00:33:01Z) - Low-Rank Constraints for Fast Inference in Structured Models [110.38427965904266]
This work demonstrates a simple approach to reduce the computational and memory complexity of a large class of structured models.
Experiments with neural parameterized structured models for language modeling, polyphonic music modeling, unsupervised grammar induction, and video modeling show that our approach matches the accuracy of standard models at large state spaces.
arXiv Detail & Related papers (2022-01-08T00:47:50Z) - Contrastive Neighborhood Alignment [81.65103777329874]
We present Contrastive Neighborhood Alignment (CNA), a manifold learning approach to maintain the topology of learned features.
The target model aims to mimic the local structure of the source representation space using a contrastive loss.
CNA is illustrated in three scenarios: manifold learning, where the model maintains the local topology of the original data in a dimension-reduced space; model distillation, where a small student model is trained to mimic a larger teacher; and legacy model update, where an older model is replaced by a more powerful one.
arXiv Detail & Related papers (2022-01-06T04:58:31Z) - Multi-Branch Deep Radial Basis Function Networks for Facial Emotion
Recognition [80.35852245488043]
We propose a CNN based architecture enhanced with multiple branches formed by radial basis function (RBF) units.
RBF units capture local patterns shared by similar instances using an intermediate representation.
We show it is the incorporation of local information what makes the proposed model competitive.
arXiv Detail & Related papers (2021-09-07T21:05:56Z) - Clustered Federated Learning via Generalized Total Variation
Minimization [83.26141667853057]
We study optimization methods to train local (or personalized) models for local datasets with a decentralized network structure.
Our main conceptual contribution is to formulate federated learning as total variation minimization (GTV)
Our main algorithmic contribution is a fully decentralized federated learning algorithm.
arXiv Detail & Related papers (2021-05-26T18:07:19Z) - Goal-directed Generation of Discrete Structures with Conditional
Generative Models [85.51463588099556]
We introduce a novel approach to directly optimize a reinforcement learning objective, maximizing an expected reward.
We test our methodology on two tasks: generating molecules with user-defined properties and identifying short python expressions which evaluate to a given target value.
arXiv Detail & Related papers (2020-10-05T20:03:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.