Public-private funding models in open source software development: A case study on scikit-learn
- URL: http://arxiv.org/abs/2404.06484v5
- Date: Fri, 3 May 2024 15:57:04 GMT
- Title: Public-private funding models in open source software development: A case study on scikit-learn
- Authors: Cailean Osborne,
- Abstract summary: This study is a case study on scikit-learn, a Python library for machine learning funded by public research grants, commercial sponsorship, micro-donations, and a 32 euro million grant announced in France's artificial intelligence strategy.
Through 25 interviews with scikit-learn's maintainers and funders, this study makes two key contributions.
It contributes empirical findings about the benefits and drawbacks of public and private funding in an impactful OSS project, and the governance protocols employed by the maintainers to balance the diverse interests of their community and funders.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Governments are increasingly funding open source software (OSS) development to support software security, digital sovereignty, and national competitiveness in science and innovation, amongst others. However, little is known about how OSS developers evaluate the relative benefits and drawbacks of governmental funding for OSS. This study explores this question through a case study on scikit-learn, a Python library for machine learning, funded by public research grants, commercial sponsorship, micro-donations, and a 32 euro million grant announced in France's artificial intelligence strategy. Through 25 interviews with scikit-learn's maintainers and funders, this study makes two key contributions. First, it contributes empirical findings about the benefits and drawbacks of public and private funding in an impactful OSS project, and the governance protocols employed by the maintainers to balance the diverse interests of their community and funders. Second, it offers practical lessons on funding for OSS developers, governments, and companies based on the experience of scikit-learn. The paper concludes with key recommendations for practitioners and future research directions.
Related papers
- Analyzing the Usage of Donation Platforms for PyPI Libraries [91.97201077607862]
This study analyzes the adoption of donation platforms in the PyPI ecosystem.
GitHub Sponsors is the dominant platform, though many PyPI-listed links are outdated.
arXiv Detail & Related papers (2025-03-11T10:27:31Z) - Ten Challenging Problems in Federated Foundation Models [55.343738234307544]
Federated Foundation Models (FedFMs) represent a distributed learning paradigm that fuses general competences of foundation models as well as privacy-preserving capabilities of federated learning.
This paper provides a comprehensive summary of the ten challenging problems inherent in FedFMs, encompassing foundational theory, utilization of private data, continual learning, unlearning, Non-IID and graph data, bidirectional knowledge transfer, incentive mechanism design, game mechanism design, model watermarking, and efficiency.
arXiv Detail & Related papers (2025-02-14T04:01:15Z) - An Overview of Cyber Security Funding for Open Source Software [3.5880059456896842]
The paper examines two such funding bodies for OSS and the projects they have funded.
The focus of both funding bodies is on software security and cyber security in general.
arXiv Detail & Related papers (2024-12-08T10:48:30Z) - A Toolkit for Measuring the Impacts of Public Funding on Open Source Software Development [0.0]
Impacts of public funding on open source software development remain poorly understood.
We present a taxonomy of potential social, economic, and technological impacts that can be both positive and negative.
With this toolkit, we contribute to the multi-stakeholder conversation about the value and impacts of funding on OSS developers and society at large.
arXiv Detail & Related papers (2024-11-09T01:13:45Z) - OpenR: An Open Source Framework for Advanced Reasoning with Large Language Models [61.14336781917986]
We introduce OpenR, an open-source framework for enhancing the reasoning capabilities of large language models (LLMs)
OpenR unifies data acquisition, reinforcement learning training, and non-autoregressive decoding into a cohesive software platform.
Our work is the first to provide an open-source framework that explores the core techniques of OpenAI's o1 model with reinforcement learning.
arXiv Detail & Related papers (2024-10-12T23:42:16Z) - Systematic Literature Review of Commercial Participation in Open Source Software [20.727062223873986]
We collected 92 papers and organized them based on their research topics.
We found the explored motivations of companies are mainly from economic, technological, and social aspects.
Researchers also explored how commercial participation affects OSS development.
arXiv Detail & Related papers (2024-05-27T06:50:01Z) - On the Societal Impact of Open Foundation Models [93.67389739906561]
We focus on open foundation models, defined here as those with broadly available model weights.
We identify five distinctive properties of open foundation models that lead to both their benefits and risks.
arXiv Detail & Related papers (2024-02-27T16:49:53Z) - Private Knowledge Sharing in Distributed Learning: A Survey [50.51431815732716]
The rise of Artificial Intelligence has revolutionized numerous industries and transformed the way society operates.
It is crucial to utilize information in learning processes that are either distributed or owned by different entities.
Modern data-driven services have been developed to integrate distributed knowledge entities into their outcomes.
arXiv Detail & Related papers (2024-02-08T07:18:23Z) - Investing in the Quantum Future : State of Play and Way Forward for
Quantum Venture Capital [0.0]
The paper introduces concepts to support the emergence of new startups and advocates for funding of scale-up quantum companies.
The paper concludes with proposals to improve the impact of the industry by taking steps to better involve society-at-large.
arXiv Detail & Related papers (2023-11-28T19:31:51Z) - From Distributed Machine Learning to Federated Learning: A Survey [49.7569746460225]
Federated learning emerges as an efficient approach to exploit distributed data and computing resources.
We propose a functional architecture of federated learning systems and a taxonomy of related techniques.
We present the distributed training, data communication, and security of FL systems.
arXiv Detail & Related papers (2021-04-29T14:15:11Z) - Leaving My Fingerprints: Motivations and Challenges of Contributing to
OSS for Social Good [17.145094780239564]
We conducted 21 semi-structured interviews with OSS for Social Good contributors.
We find that the majority of contributors demonstrate a distinction between OSS4SG and OSS.
OSS4SG contributors evaluate the owners of the project significantly more than OSS contributors.
arXiv Detail & Related papers (2021-04-26T21:50:11Z) - The CROSS Incubator: A Case Study for funding and training RSEs [0.8472211480603307]
The Center for Research in Open Source Software (CROSS) at UC Santa Cruz has been very effective at promoting the professional development of research software engineers.
Carlos Maltzahn founded CROSS in 2015 with a generous gift of $2,000,000 from UC Santa Cruz alumnus Dr. Sage Weil and founding memberships of Toshiba America Electronic Components, SK Hynix Memory Solutions, and Micron Technology.
This position paper will present CROSS fellowships as case studies for how university-led open source projects can create a real-world, reproducible model for effectively training, funding and supporting research software engineers.
arXiv Detail & Related papers (2020-11-30T20:16:43Z) - Learnings from Frontier Development Lab and SpaceML -- AI Accelerators
for NASA and ESA [57.06643156253045]
Research with AI and ML technologies lives in a variety of settings with often asynchronous goals and timelines.
We perform a case study of the Frontier Development Lab (FDL), an AI accelerator under a public-private partnership from NASA and ESA.
FDL research follows principled practices that are grounded in responsible development, conduct, and dissemination of AI research.
arXiv Detail & Related papers (2020-11-09T21:23:03Z) - Insights into Fairness through Trust: Multi-scale Trust Quantification
for Financial Deep Learning [94.65749466106664]
A fundamental aspect of fairness that has not been explored in financial deep learning is the concept of trust.
We conduct multi-scale trust quantification on a deep neural network for the purpose of credit card default prediction.
arXiv Detail & Related papers (2020-11-03T19:05:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.