GitEvolve: Predicting the Evolution of GitHub Repositories
- URL: http://arxiv.org/abs/2010.04366v1
- Date: Fri, 9 Oct 2020 04:32:15 GMT
- Title: GitEvolve: Predicting the Evolution of GitHub Repositories
- Authors: Honglu Zhou, Hareesh Ravi, Carlos M. Muniz, Vahid Azizi, Linda Ness,
Gerard de Melo, Mubbasir Kapadia
- Abstract summary: We propose GitEvolve, a system to predict the evolution of GitHub repositories.
We map users to groups by modelling common interests to better predict popularity.
The proposed multi-task architecture is generic and can be extended to model information diffusion in other social networks.
- Score: 31.814226661858694
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Software development is becoming increasingly open and collaborative with the
advent of platforms such as GitHub. Given its crucial role, there is a need to
better understand and model the dynamics of GitHub as a social platform.
Previous work has mostly considered the dynamics of traditional social
networking sites like Twitter and Facebook. We propose GitEvolve, a system to
predict the evolution of GitHub repositories and the different ways by which
users interact with them. To this end, we develop an end-to-end multi-task
sequential deep neural network that given some seed events, simultaneously
predicts which user-group is next going to interact with a given repository,
what the type of the interaction is, and when it happens. To facilitate
learning, we use graph based representation learning to encode relationship
between repositories. We map users to groups by modelling common interests to
better predict popularity and to generalize to unseen users during inference.
We introduce an artificial event type to better model varying levels of
activity of repositories in the dataset. The proposed multi-task architecture
is generic and can be extended to model information diffusion in other social
networks. In a series of experiments, we demonstrate the effectiveness of the
proposed model, using multiple metrics and baselines. Qualitative analysis of
the model's ability to predict popularity and forecast trends proves its
applicability.
Related papers
- Deep Generative Models in Robotics: A Survey on Learning from Multimodal Demonstrations [52.11801730860999]
In recent years, the robot learning community has shown increasing interest in using deep generative models to capture the complexity of large datasets.
We present the different types of models that the community has explored, such as energy-based models, diffusion models, action value maps, or generative adversarial networks.
We also present the different types of applications in which deep generative models have been used, from grasp generation to trajectory generation or cost learning.
arXiv Detail & Related papers (2024-08-08T11:34:31Z) - Premonition: Using Generative Models to Preempt Future Data Changes in
Continual Learning [63.850451635362425]
Continual learning requires a model to adapt to ongoing changes in the data distribution.
We show that the combination of a large language model and an image generation model can similarly provide useful premonitions.
We find that the backbone of our pre-trained networks can learn representations useful for the downstream continual learning problem.
arXiv Detail & Related papers (2024-03-12T06:29:54Z) - DARLEI: Deep Accelerated Reinforcement Learning with Evolutionary
Intelligence [77.78795329701367]
We present DARLEI, a framework that combines evolutionary algorithms with parallelized reinforcement learning.
We characterize DARLEI's performance under various conditions, revealing factors impacting diversity of evolved morphologies.
We hope to extend DARLEI in future work to include interactions between diverse morphologies in richer environments.
arXiv Detail & Related papers (2023-12-08T16:51:10Z) - Synergistic Signals: Exploiting Co-Engagement and Semantic Links via
Graph Neural Networks [4.261438296177923]
We study the problem in the context recommender systems at Netflix.
We propose a novel graph-based approach called SemanticGNN.
Our key technical contributions are twofold: (1) we develop a novel relation-aware attention graph neural network (GNN) to handle the imbalanced distribution of relation types in our graph; (2) to handle web-scale graph data that has millions of nodes and billions of edges, we develop a novel distributed graph training paradigm.
arXiv Detail & Related papers (2023-12-07T06:29:26Z) - Modeling Random Networks with Heterogeneous Reciprocity [9.630755176298056]
We develop methodology to model the diverse reciprocal behavior in growing social networks.
We present a preferential attachment model with heterogeneous reciprocity that imitates the attraction users have for popular users.
We apply the presented methods to the analysis of a Facebook wallpost network where users have non-uniform reciprocal behavior patterns.
arXiv Detail & Related papers (2023-08-19T21:21:25Z) - DIGMN: Dynamic Intent Guided Meta Network for Differentiated User
Engagement Forecasting in Online Professional Social Platforms [32.70471436337077]
A major reason for the differences in user engagement patterns is that users have different intents.
We propose a Dynamic Guided Meta Network (DIGMN) which can explicitly model user intent varying with time.
Our method outperforms state-of-the-art baselines significantly.
arXiv Detail & Related papers (2022-10-22T09:57:27Z) - OpenMixup: Open Mixup Toolbox and Benchmark for Visual Representation Learning [53.57075147367114]
We introduce OpenMixup, the first mixup augmentation and benchmark for visual representation learning.
We train 18 representative mixup baselines from scratch and rigorously evaluate them across 11 image datasets.
We also open-source our modular backbones, including a collection of popular vision backbones, optimization strategies, and analysis toolkits.
arXiv Detail & Related papers (2022-09-11T12:46:01Z) - Graph Representation Learning for Popularity Prediction Problem: A
Survey [4.795530213347874]
We present a comprehensive review for existing works using graph representation learning (GRL) methods for popularity prediction problem.
Deep learning method is further classified into six small classes: convolutional neural networks, graph convolutional networks, graph attention networks, graph neural networks, recurrent neural networks, and reinforcement learning.
arXiv Detail & Related papers (2022-03-15T04:11:46Z) - An Expectation-Maximization Perspective on Federated Learning [75.67515842938299]
Federated learning describes the distributed training of models across multiple clients while keeping the data private on-device.
In this work, we view the server-orchestrated federated learning process as a hierarchical latent variable model where the server provides the parameters of a prior distribution over the client-specific model parameters.
We show that with simple Gaussian priors and a hard version of the well known Expectation-Maximization (EM) algorithm, learning in such a model corresponds to FedAvg, the most popular algorithm for the federated learning setting.
arXiv Detail & Related papers (2021-11-19T12:58:59Z) - A Variational Graph Autoencoder for Manipulation Action Recognition and
Prediction [1.1816942730023883]
We introduce a deep graph autoencoder to jointly learn recognition and prediction of manipulation tasks from symbolic scene graphs.
Our network has a variational autoencoder structure with two branches: one for identifying the input graph type and one for predicting the future graphs.
We benchmark our new model against different state-of-the-art methods on two different datasets, MANIAC and MSRC-9, and show that our proposed model can achieve better performance.
arXiv Detail & Related papers (2021-10-25T21:40:42Z) - Connecting the Dots: Multivariate Time Series Forecasting with Graph
Neural Networks [91.65637773358347]
We propose a general graph neural network framework designed specifically for multivariate time series data.
Our approach automatically extracts the uni-directed relations among variables through a graph learning module.
Our proposed model outperforms the state-of-the-art baseline methods on 3 of 4 benchmark datasets.
arXiv Detail & Related papers (2020-05-24T04:02:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.