Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach
- URL: http://arxiv.org/abs/2110.04514v2
- Date: Tue, 13 Jun 2023 17:34:36 GMT
- Title: Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach
- Authors: Qitian Wu, Chenxiao Yang, Junchi Yan
- Abstract summary: We propose a new learning paradigm with graph representation and learning.
Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data.
- Score: 80.8446673089281
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We target open-world feature extrapolation problem where the feature space of
input data goes through expansion and a model trained on partially observed
features needs to handle new features in test data without further retraining.
The problem is of much significance for dealing with features incrementally
collected from different fields. To this end, we propose a new learning
paradigm with graph representation and learning. Our framework contains two
modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model
takes features as input and outputs predicted labels; 2) a graph neural network
as an upper model learns to extrapolate embeddings for new features via message
passing over a feature-data graph built from observed data. Based on our
framework, we design two training strategies, a self-supervised approach and an
inductive learning approach, to endow the model with extrapolation ability and
alleviate feature-level over-fitting. We also provide theoretical analysis on
the generalization error on test data with new features, which dissects the
impact of training features and algorithms on generalization performance. Our
experiments over several classification datasets and large-scale advertisement
click prediction datasets demonstrate that our model can produce effective
embeddings for unseen features and significantly outperforms baseline methods
that adopt KNN and local aggregation.
Related papers
- Imbalanced Graph-Level Anomaly Detection via Counterfactual Augmentation and Feature Learning [1.3756846638796]
We propose an imbalanced GLAD method via counterfactual augmentation and feature learning.
We apply the model to brain disease datasets, which can prove the capability of our work.
arXiv Detail & Related papers (2024-07-13T13:40:06Z) - In-Context Learning of Physical Properties: Few-Shot Adaptation to Out-of-Distribution Molecular Graphs [1.8635507597668244]
In-context learning allows for performing nontrivial machine learning tasks during inference only.
In this work, we address the question: can we leverage in-context learning to predict out-of-distribution materials properties?
We employ a compound model in which GPT-2 acts on the output of geometry-aware graph neural networks to adapt in-context information.
arXiv Detail & Related papers (2024-06-03T21:59:21Z) - Generalization Properties of Retrieval-based Models [50.35325326050263]
Retrieval-based machine learning methods have enjoyed success on a wide range of problems.
Despite growing literature showcasing the promise of these models, the theoretical underpinning for such models remains underexplored.
We present a formal treatment of retrieval-based models to characterize their generalization ability.
arXiv Detail & Related papers (2022-10-06T00:33:01Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - Self-Supervised Graph Representation Learning for Neuronal Morphologies [75.38832711445421]
We present GraphDINO, a data-driven approach to learn low-dimensional representations of 3D neuronal morphologies from unlabeled datasets.
We show, in two different species and across multiple brain areas, that this method yields morphological cell type clusterings on par with manual feature-based classification by experts.
Our method could potentially enable data-driven discovery of novel morphological features and cell types in large-scale datasets.
arXiv Detail & Related papers (2021-12-23T12:17:47Z) - From Canonical Correlation Analysis to Self-supervised Graph Neural
Networks [99.44881722969046]
We introduce a conceptually simple yet effective model for self-supervised representation learning with graph data.
We optimize an innovative feature-level objective inspired by classical Canonical Correlation Analysis.
Our method performs competitively on seven public graph datasets.
arXiv Detail & Related papers (2021-06-23T15:55:47Z) - A Framework to Learn with Interpretation [2.3741312212138896]
We present a novel framework to jointly learn a predictive model and its associated interpretation model.
We seek for a small-size dictionary of high level attribute functions that take as inputs the outputs of selected hidden layers.
A detailed pipeline to visualize the learnt features is also developed.
arXiv Detail & Related papers (2020-10-19T09:26:28Z) - Quantifying Challenges in the Application of Graph Representation
Learning [0.0]
We provide an application oriented perspective to a set of popular embedding approaches.
We evaluate their representational power with respect to real-world graph properties.
Our results suggest that "one-to-fit-all" GRL approaches are hard to define in real-world scenarios.
arXiv Detail & Related papers (2020-06-18T03:19:43Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.