Foundation Models Meet Visualizations: Challenges and Opportunities
- URL: http://arxiv.org/abs/2310.05771v1
- Date: Mon, 9 Oct 2023 14:57:05 GMT
- Title: Foundation Models Meet Visualizations: Challenges and Opportunities
- Authors: Weikai Yang, Mengchen Liu, Zheng Wang, and Shixia Liu
- Abstract summary: This paper divides visualizations for foundation models (VIS4FM) and foundation models for visualizations (FM4VIS)
In VIS4FM, we explore the primary role of visualizations in understanding, refining, and evaluating these intricate models.
In FM4VIS, we highlight how foundation models can be utilized to advance the visualization field itself.
- Score: 23.01218856618978
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent studies have indicated that foundation models, such as BERT and GPT,
excel in adapting to a variety of downstream tasks. This adaptability has
established them as the dominant force in building artificial intelligence (AI)
systems. As visualization techniques intersect with these models, a new
research paradigm emerges. This paper divides these intersections into two main
areas: visualizations for foundation models (VIS4FM) and foundation models for
visualizations (FM4VIS). In VIS4FM, we explore the primary role of
visualizations in understanding, refining, and evaluating these intricate
models. This addresses the pressing need for transparency, explainability,
fairness, and robustness. Conversely, within FM4VIS, we highlight how
foundation models can be utilized to advance the visualization field itself.
The confluence of foundation models and visualizations holds great promise, but
it also comes with its own set of challenges. By highlighting these challenges
and the growing opportunities, this paper seeks to provide a starting point for
continued exploration in this promising avenue.
Related papers
- Explainability for Vision Foundation Models: A Survey [3.570403495760109]
Foundation models occupy an ambiguous position in the explainability domain.
Foundation models are characterized by their extensive generalization capabilities and emergent uses.
We discuss the challenges faced by current research in integrating XAI within foundation models.
arXiv Detail & Related papers (2025-01-21T15:18:55Z) - Towards Robust and Fair Vision Learning in Open-World Environments [5.520041242906903]
The dissertation presents four key contributions toward fairness and robustness in vision learning.
First, to address the problem of large-scale data requirements, the dissertation presents a novel Fairness Domain Adaptation approach.
Second, to enable the capability of open-world modeling of vision learning, this dissertation presents a novel Open-world Fairness Continual Learning Framework.
arXiv Detail & Related papers (2024-12-12T16:50:52Z) - ReVLA: Reverting Visual Domain Limitation of Robotic Foundation Models [55.07988373824348]
We study the visual generalization capabilities of three existing robotic foundation models.
Our study shows that the existing models do not exhibit robustness to visual out-of-domain scenarios.
We propose a gradual backbone reversal approach founded on model merging.
arXiv Detail & Related papers (2024-09-23T17:47:59Z) - Forging Vision Foundation Models for Autonomous Driving: Challenges,
Methodologies, and Opportunities [59.02391344178202]
Vision foundation models (VFMs) serve as potent building blocks for a wide range of AI applications.
The scarcity of comprehensive training data, the need for multi-sensor integration, and the diverse task-specific architectures pose significant obstacles to the development of VFMs.
This paper delves into the critical challenge of forging VFMs tailored specifically for autonomous driving, while also outlining future directions.
arXiv Detail & Related papers (2024-01-16T01:57:24Z) - Masked Modeling for Self-supervised Representation Learning on Vision
and Beyond [69.64364187449773]
Masked modeling has emerged as a distinctive approach that involves predicting parts of the original data that are proportionally masked during training.
We elaborate on the details of techniques within masked modeling, including diverse masking strategies, recovering targets, network architectures, and more.
We conclude by discussing the limitations of current techniques and point out several potential avenues for advancing masked modeling research.
arXiv Detail & Related papers (2023-12-31T12:03:21Z) - Towards the Unification of Generative and Discriminative Visual
Foundation Model: A Survey [30.528346074194925]
Visual foundation models (VFMs) have become a catalyst for groundbreaking developments in computer vision.
This review paper delineates the pivotal trajectories of VFMs, emphasizing their scalability and proficiency in generative tasks.
A crucial direction for forthcoming innovation is the amalgamation of generative and discriminative paradigms.
arXiv Detail & Related papers (2023-12-15T19:17:15Z) - Towards Graph Foundation Models: A Survey and Beyond [66.37994863159861]
Foundation models have emerged as critical components in a variety of artificial intelligence applications.
The capabilities of foundation models to generalize and adapt motivate graph machine learning researchers to discuss the potential of developing a new graph learning paradigm.
This article introduces the concept of Graph Foundation Models (GFMs), and offers an exhaustive explanation of their key characteristics and underlying technologies.
arXiv Detail & Related papers (2023-10-18T09:31:21Z) - Learning from models beyond fine-tuning [78.20895343699658]
Learn From Model (LFM) focuses on the research, modification, and design of foundation models (FM) based on the model interface.
The study of LFM techniques can be broadly categorized into five major areas: model tuning, model distillation, model reuse, meta learning and model editing.
This paper gives a comprehensive review of the current methods based on FM from the perspective of LFM.
arXiv Detail & Related papers (2023-10-12T10:20:36Z) - UViM: A Unified Modeling Approach for Vision with Learned Guiding Codes [91.24112204588353]
We introduce UViM, a unified approach capable of modeling a wide range of computer vision tasks.
In contrast to previous models, UViM has the same functional form for all tasks.
We demonstrate the effectiveness of UViM on three diverse and challenging vision tasks.
arXiv Detail & Related papers (2022-05-20T17:47:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.