Word2VecGD: Neural Graph Drawing with Cosine-Stress Optimization
- URL: http://arxiv.org/abs/2509.17333v1
- Date: Mon, 22 Sep 2025 03:09:55 GMT
- Title: Word2VecGD: Neural Graph Drawing with Cosine-Stress Optimization
- Authors: Minglai Yang, Reyan Ahmed,
- Abstract summary: Instead of relying on exact shortest-path distances, we optimize layouts using cosine dissimilarities.<n>Our framework integrates differentiable stress optimization with gradient descent.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a novel graph visualization method leveraging random walk-based embeddings to replace costly graph-theoretical distance computations. Using word2vec-inspired embeddings, our approach captures both structural and semantic relationships efficiently. Instead of relying on exact shortest-path distances, we optimize layouts using cosine dissimilarities, significantly reducing computational overhead. Our framework integrates differentiable stress optimization with stochastic gradient descent (SGD), supporting multi-criteria layout objectives. Experimental results demonstrate that our method produces high-quality, semantically meaningful layouts while efficiently scaling to large graphs. Code available at: https://github.com/mlyann/graphv_nn
Related papers
- DREAM: Dual-Standard Semantic Homogeneity with Dynamic Optimization for Graph Learning with Label Noise [53.55187452152358]
This paper proposes a novel method, Dual-Standard Semantic Homogeneity with Dynamic Optimization (DREAM) for reliable, relation-informed optimization on graphs with label noise.<n>Specifically, we design a relation-informed dynamic optimization framework that iteratively reevaluates the reliability of each labeled node in the graph.
arXiv Detail & Related papers (2026-01-24T12:54:18Z) - LayoutGKN: Graph Similarity Learning of Floor Plans [11.388505978767627]
Most successful methods to compare graphs rely on costly intermediate cross-graph node-level interactions.<n>We introduce textbfGKN, a more efficient approach that postpones the cross-graph node-level interactions.<n>We show that LayoutGKN computes similarity comparably or better than graph matching networks.
arXiv Detail & Related papers (2025-09-03T21:56:16Z) - CoRe-GD: A Hierarchical Framework for Scalable Graph Visualization with
GNNs [20.706469085872516]
We introduce a scalable Graph Neural Network (GNN) based Graph Drawing framework with sub-quadratic that can learn to optimize stress.
Inspired by classical stress optimization techniques and force-directed layout algorithms, we create a coarsening hierarchy for the input graph.
To enhance information propagation within the network, we propose a novel positional rewiring technique.
arXiv Detail & Related papers (2024-02-09T10:50:45Z) - Robust Attributed Graph Alignment via Joint Structure Learning and
Optimal Transport [26.58964162799207]
We propose SLOTAlign, an unsupervised graph alignment framework that jointly performs Structure Learning and Optimal Transport Alignment.
We incorporate multi-view structure learning to enhance graph representation power and reduce the effect of structure and feature inconsistency inherited across graphs.
The proposed SLOTAlign shows superior performance and strongest robustness over seven unsupervised graph alignment methods and five specialized KG alignment methods.
arXiv Detail & Related papers (2023-01-30T08:41:36Z) - Condensing Graphs via One-Step Gradient Matching [50.07587238142548]
We propose a one-step gradient matching scheme, which performs gradient matching for only one single step without training the network weights.
Our theoretical analysis shows this strategy can generate synthetic graphs that lead to lower classification loss on real graphs.
In particular, we are able to reduce the dataset size by 90% while approximating up to 98% of the original performance.
arXiv Detail & Related papers (2022-06-15T18:20:01Z) - Optimal Propagation for Graph Neural Networks [51.08426265813481]
We propose a bi-level optimization approach for learning the optimal graph structure.
We also explore a low-rank approximation model for further reducing the time complexity.
arXiv Detail & Related papers (2022-05-06T03:37:00Z) - A Regularized Wasserstein Framework for Graph Kernels [32.558913310384476]
We propose a learning framework for graph kernels grounded on regularizing optimal transport.
This framework provides a novel optimal transport distance metric, namely Regularized Wasserstein (RW) discrepancy.
We have empirically validated our method using 12 datasets against 16 state-of-the-art baselines.
arXiv Detail & Related papers (2021-10-06T07:54:04Z) - FGOT: Graph Distances based on Filters and Optimal Transport [62.779521543654134]
Graph comparison deals with identifying similarities and dissimilarities between graphs.
A major obstacle is the unknown alignment of graphs, as well as the lack of accurate and inexpensive comparison metrics.
In this work we introduce the filter graph distance approximation.
arXiv Detail & Related papers (2021-09-09T17:43:07Z) - Learnable Graph Matching: Incorporating Graph Partitioning with Deep
Feature Learning for Multiple Object Tracking [58.30147362745852]
Data association across frames is at the core of Multiple Object Tracking (MOT) task.
Existing methods mostly ignore the context information among tracklets and intra-frame detections.
We propose a novel learnable graph matching method to address these issues.
arXiv Detail & Related papers (2021-03-30T08:58:45Z) - Wasserstein-based Graph Alignment [56.84964475441094]
We cast a new formulation for the one-to-many graph alignment problem, which aims at matching a node in the smaller graph with one or more nodes in the larger graph.
We show that our method leads to significant improvements with respect to the state-of-the-art algorithms for each of these tasks.
arXiv Detail & Related papers (2020-03-12T22:31:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.