Graph-LDA: Graph Structure Priors to Improve the Accuracy in Few-Shot
Classification
- URL: http://arxiv.org/abs/2108.10427v1
- Date: Mon, 23 Aug 2021 21:55:45 GMT
- Title: Graph-LDA: Graph Structure Priors to Improve the Accuracy in Few-Shot
Classification
- Authors: Myriam Bontonou, Nicolas Farrugia, Vincent Gripon
- Abstract summary: We introduce a generic model where observed class signals are supposed to be deteriorated with two sources of noise.
We derive an optimal methodology to classify such signals.
This methodology includes a single parameter, making it particularly suitable for cases where available data is scarce.
- Score: 6.037383467521294
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is very common to face classification problems where the number of
available labeled samples is small compared to their dimension. These
conditions are likely to cause underdetermined settings, with high risk of
overfitting. To improve the generalization ability of trained classifiers,
common solutions include using priors about the data distribution. Among many
options, data structure priors, often represented through graphs, are
increasingly popular in the field. In this paper, we introduce a generic model
where observed class signals are supposed to be deteriorated with two sources
of noise, one independent of the underlying graph structure and isotropic, and
the other colored by a known graph operator. Under this model, we derive an
optimal methodology to classify such signals. Interestingly, this methodology
includes a single parameter, making it particularly suitable for cases where
available data is scarce. Using various real datasets, we showcase the ability
of the proposed model to be implemented in real world scenarios, resulting in
increased generalization accuracy compared to popular alternatives.
Related papers
- A Complete Decomposition of KL Error using Refined Information and Mode Interaction Selection [11.994525728378603]
We revisit the classical formulation of the log-linear model with a focus on higher-order mode interactions.
We find that our learned distributions are able to more efficiently use the finite amount of data which is available in practice.
arXiv Detail & Related papers (2024-10-15T18:08:32Z) - Generation is better than Modification: Combating High Class Homophily Variance in Graph Anomaly Detection [51.11833609431406]
Homophily distribution differences between different classes are significantly greater than those in homophilic and heterophilic graphs.
We introduce a new metric called Class Homophily Variance, which quantitatively describes this phenomenon.
To mitigate its impact, we propose a novel GNN model named Homophily Edge Generation Graph Neural Network (HedGe)
arXiv Detail & Related papers (2024-03-15T14:26:53Z) - Generated Graph Detection [27.591612297045817]
Graph generative models become increasingly effective for data distribution approximation and data augmentation.
We propose the first framework to investigate a set of sophisticated models and their performance in four classification scenarios.
Our solution can sustain for a decent while to curb generated graph misuses.
arXiv Detail & Related papers (2023-06-13T13:18:04Z) - Optimality of Message-Passing Architectures for Sparse Graphs [13.96547777184641]
We study the node classification problem on feature-decorated graphs in the sparse setting, i.e., when the expected degree of a node is $O(1)$ in the number of nodes.
We introduce a notion of Bayes optimality for node classification tasks, called local Bayes optimality.
We show that the optimal message-passing architecture interpolates between a standard in the regime of low graph signal and a typical in the regime of high graph signal.
arXiv Detail & Related papers (2023-05-17T17:31:20Z) - Optimal Propagation for Graph Neural Networks [51.08426265813481]
We propose a bi-level optimization approach for learning the optimal graph structure.
We also explore a low-rank approximation model for further reducing the time complexity.
arXiv Detail & Related papers (2022-05-06T03:37:00Z) - Selecting the suitable resampling strategy for imbalanced data
classification regarding dataset properties [62.997667081978825]
In many application domains such as medicine, information retrieval, cybersecurity, social media, etc., datasets used for inducing classification models often have an unequal distribution of the instances of each class.
This situation, known as imbalanced data classification, causes low predictive performance for the minority class examples.
Oversampling and undersampling techniques are well-known strategies to deal with this problem by balancing the number of examples of each class.
arXiv Detail & Related papers (2021-12-15T18:56:39Z) - A Robust and Generalized Framework for Adversarial Graph Embedding [73.37228022428663]
We propose a robust framework for adversarial graph embedding, named AGE.
AGE generates the fake neighbor nodes as the enhanced negative samples from the implicit distribution.
Based on this framework, we propose three models to handle three types of graph data.
arXiv Detail & Related papers (2021-05-22T07:05:48Z) - Graph Classification by Mixture of Diverse Experts [67.33716357951235]
We present GraphDIVE, a framework leveraging mixture of diverse experts for imbalanced graph classification.
With a divide-and-conquer principle, GraphDIVE employs a gating network to partition an imbalanced graph dataset into several subsets.
Experiments on real-world imbalanced graph datasets demonstrate the effectiveness of GraphDIVE.
arXiv Detail & Related papers (2021-03-29T14:03:03Z) - Imputation of Missing Data with Class Imbalance using Conditional
Generative Adversarial Networks [24.075691766743702]
We propose a new method for imputing missing data based on its class-specific characteristics.
Our Conditional Generative Adversarial Imputation Network (CGAIN) imputes the missing data using class-specific distributions.
We tested our approach on benchmark datasets and achieved superior performance compared with the state-of-the-art and popular imputation approaches.
arXiv Detail & Related papers (2020-12-01T02:26:54Z) - Geometric graphs from data to aid classification tasks with graph
convolutional networks [0.0]
We show that, even if additional relational information is not available in the data set, one can improve classification by constructing geometric graphs from the features themselves.
The improvement in classification accuracy is maximized by graphs that capture sample similarity with relatively low edge density.
arXiv Detail & Related papers (2020-05-08T15:00:45Z) - Block-Approximated Exponential Random Graphs [77.4792558024487]
An important challenge in the field of exponential random graphs (ERGs) is the fitting of non-trivial ERGs on large graphs.
We propose an approximative framework to such non-trivial ERGs that result in dyadic independence (i.e., edge independent) distributions.
Our methods are scalable to sparse graphs consisting of millions of nodes.
arXiv Detail & Related papers (2020-02-14T11:42:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.