Towards Decentralized and Sustainable Foundation Model Training with the Edge
- URL: http://arxiv.org/abs/2507.01803v1
- Date: Wed, 02 Jul 2025 15:21:40 GMT
- Title: Towards Decentralized and Sustainable Foundation Model Training with the Edge
- Authors: Leyang Xue, Meghana Madhyastha, Randal Burns, Myungjin Lee, Mahesh K. Marina,
- Abstract summary: Foundation models are at the forefront of AI research, appealing for their ability to learn from vast datasets and cater to diverse tasks.<n>We put forward a vision towards decentralized and sustainable foundation model training that leverages the collective compute of sparingly used connected edge AI devices.
- Score: 2.2815302415385297
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Foundation models are at the forefront of AI research, appealing for their ability to learn from vast datasets and cater to diverse tasks. Yet, their significant computational demands raise issues of environmental impact and the risk of centralized control in their development. We put forward a vision towards decentralized and sustainable foundation model training that leverages the collective compute of sparingly used connected edge AI devices. We present the rationale behind our vision, particularly in support of its sustainability benefit. We further outline a set of challenges that need to be addressed to turn this vision into reality.
Related papers
- Towards Depth Foundation Model: Recent Trends in Vision-Based Depth Estimation [75.30238170051291]
Depth estimation is a fundamental task in 3D computer vision, crucial for applications such as 3D reconstruction, free-viewpoint rendering, robotics, autonomous driving, and AR/VR technologies.<n>Traditional methods relying on hardware sensors like LiDAR are often limited by high costs, low resolution, and environmental sensitivity, limiting their applicability in real-world scenarios.<n>Recent advances in vision-based methods offer a promising alternative, yet they face challenges in generalization and stability due to either the low-capacity model architectures or the reliance on domain-specific and small-scale datasets.
arXiv Detail & Related papers (2025-07-15T17:59:59Z) - An Overview of Large Language Models for Statisticians [109.38601458831545]
Large Language Models (LLMs) have emerged as transformative tools in artificial intelligence (AI)<n>This paper explores potential areas where statisticians can make important contributions to the development of LLMs.<n>We focus on issues such as uncertainty quantification, interpretability, fairness, privacy, watermarking and model adaptation.
arXiv Detail & Related papers (2025-02-25T03:40:36Z) - Spatio-Temporal Foundation Models: Vision, Challenges, and Opportunities [48.45951497996322]
Foundation models (STFMs) have revolutionized artificial intelligence, setting new benchmarks in performance and enabling transformative capabilities across a wide range of vision and language tasks.<n>In this paper, we articulate a vision for the future of STFMs, outlining their essential characteristics and generalization capabilities necessary for broad applicability.<n>We explore potential opportunities and directions to advance research towards the aim of effective and broadly applicable STFMs.
arXiv Detail & Related papers (2025-01-15T08:52:28Z) - Addressing the sustainable AI trilemma: a case study on LLM agents and RAG [7.6212949300713015]
Large language models (LLMs) have demonstrated significant capabilities, but their widespread deployment and more advanced applications raise critical sustainability challenges.<n>We propose the concept of the Sustainable AI Trilemma, highlighting the tensions between AI capability, digital equity, and environmental sustainability.
arXiv Detail & Related papers (2025-01-14T17:21:16Z) - Unleashing the Power of Continual Learning on Non-Centralized Devices: A Survey [37.07938402225207]
Non- Continual Learning (NCCL) has become an emerging paradigm for enabling distributed devices to handle streaming data from a joint non-stationary environment.<n>This survey focuses on the development of the non-centralized continual learning algorithms and the real-world deployment across distributed devices.
arXiv Detail & Related papers (2024-12-18T13:33:28Z) - Implementation of Big AI Models for Wireless Networks with Collaborative Edge Computing [10.524645516703643]
Training big AI models poses significant challenges to edge devices.
Traditional approaches usually resort to aggregating training data and sending it to a remote cloud for centralized training.
We propose collaborative edge training, a novel training mechanism that orchestrates a group of trusted edge devices as a resource pool.
arXiv Detail & Related papers (2024-04-27T03:09:39Z) - A Survey on Robotics with Foundation Models: toward Embodied AI [30.999414445286757]
Recent advances in computer vision, natural language processing, and multi-modality learning have shown that the foundation models have superhuman capabilities for specific tasks.
This survey aims to provide a comprehensive and up-to-date overview of foundation models in robotics, focusing on autonomous manipulation and encompassing high-level planning and low-level control.
arXiv Detail & Related papers (2024-02-04T07:55:01Z) - A Survey of Reasoning with Foundation Models [235.7288855108172]
Reasoning plays a pivotal role in various real-world settings such as negotiation, medical diagnosis, and criminal investigation.
We introduce seminal foundation models proposed or adaptable for reasoning.
We then delve into the potential future directions behind the emergence of reasoning abilities within foundation models.
arXiv Detail & Related papers (2023-12-17T15:16:13Z) - Exploring the Robustness of Decentralized Training for Large Language
Models [51.41850749014054]
Decentralized training of large language models has emerged as an effective way to democratize this technology.
This paper explores the robustness of decentralized training from three main perspectives.
arXiv Detail & Related papers (2023-12-01T04:04:03Z) - Causal Reasoning: Charting a Revolutionary Course for Next-Generation
AI-Native Wireless Networks [63.246437631458356]
Next-generation wireless networks (e.g., 6G) will be artificial intelligence (AI)-native.
This article introduces a novel framework for building AI-native wireless networks; grounded in the emerging field of causal reasoning.
We highlight several wireless networking challenges that can be addressed by causal discovery and representation.
arXiv Detail & Related papers (2023-09-23T00:05:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.