A Case for Business Process-Specific Foundation Models
- URL: http://arxiv.org/abs/2210.14739v1
- Date: Wed, 26 Oct 2022 14:17:47 GMT
- Title: A Case for Business Process-Specific Foundation Models
- Authors: Yara Rizk, Praveen Venkateswaran, Vatche Isahagian, Vinod Muthusamy
- Abstract summary: We argue that business process data representations have unique characteristics that warrant the development of a new class of foundation models.
These models should tackle the unique challenges of applying AI to business processes which include data scarcity, multi-modal representations, domain specific terminology, and privacy concerns.
- Score: 6.25118865553438
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The inception of large language models has helped advance state-of-the-art
performance on numerous natural language tasks. This has also opened the door
for the development of foundation models for other domains and data modalities
such as images, code, and music. In this paper, we argue that business process
data representations have unique characteristics that warrant the development
of a new class of foundation models to handle tasks like process mining,
optimization, and decision making. These models should also tackle the unique
challenges of applying AI to business processes which include data scarcity,
multi-modal representations, domain specific terminology, and privacy concerns.
Related papers
- An overview of domain-specific foundation model: key technologies, applications and challenges [25.146374189455784]
This article provides a timely and thorough overview of the methodology for customizing domain-specific foundation models.
It introduces basic concepts, outlines the general architecture, and surveys key methods for constructing domain-specific models.
arXiv Detail & Related papers (2024-09-06T13:24:22Z) - HEMM: Holistic Evaluation of Multimodal Foundation Models [91.60364024897653]
Multimodal foundation models can holistically process text alongside images, video, audio, and other sensory modalities.
It is challenging to characterize and study progress in multimodal foundation models, given the range of possible modeling decisions, tasks, and domains.
arXiv Detail & Related papers (2024-07-03T18:00:48Z) - Adapting Large Language Models for Content Moderation: Pitfalls in Data
Engineering and Supervised Fine-tuning [79.53130089003986]
Large Language Models (LLMs) have become a feasible solution for handling tasks in various domains.
In this paper, we introduce how to fine-tune a LLM model that can be privately deployed for content moderation.
arXiv Detail & Related papers (2023-10-05T09:09:44Z) - Generative AI for Business Strategy: Using Foundation Models to Create
Business Strategy Tools [0.7784248206747153]
We propose the use of foundation models for business decision making.
We derive IT artifacts in the form of asequence of signed business networks.
Such artifacts can inform business stakeholders about the state of the market and their own positioning.
arXiv Detail & Related papers (2023-08-27T19:03:12Z) - UniDiff: Advancing Vision-Language Models with Generative and
Discriminative Learning [86.91893533388628]
This paper presents UniDiff, a unified multi-modal model that integrates image-text contrastive learning (ITC), text-conditioned image synthesis learning (IS), and reciprocal semantic consistency modeling (RSC)
UniDiff demonstrates versatility in both multi-modal understanding and generative tasks.
arXiv Detail & Related papers (2023-06-01T15:39:38Z) - Beyond Rule-based Named Entity Recognition and Relation Extraction for
Process Model Generation from Natural Language Text [0.0]
We present an extension to an existing pipeline to make it entirely data driven.
We demonstrate the competitiveness of our improved pipeline, which not only eliminates the substantial overhead associated with feature engineering and rule definition.
We propose an extension to the PET dataset that incorporates information about linguistic references and a corresponding method for resolving them.
arXiv Detail & Related papers (2023-05-06T07:06:47Z) - Scaling Vision-Language Models with Sparse Mixture of Experts [128.0882767889029]
We show that mixture-of-experts (MoE) techniques can achieve state-of-the-art performance on a range of benchmarks over dense models of equivalent computational cost.
Our research offers valuable insights into stabilizing the training of MoE models, understanding the impact of MoE on model interpretability, and balancing the trade-offs between compute performance when scaling vision-language models.
arXiv Detail & Related papers (2023-03-13T16:00:31Z) - Foundation Models for Decision Making: Problems, Methods, and
Opportunities [124.79381732197649]
Foundation models pretrained on diverse data at scale have demonstrated extraordinary capabilities in a wide range of vision and language tasks.
New paradigms are emerging for training foundation models to interact with other agents and perform long-term reasoning.
Research at the intersection of foundation models and decision making holds tremendous promise for creating powerful new systems.
arXiv Detail & Related papers (2023-03-07T18:44:07Z) - Foundation models in brief: A historical, socio-technical focus [2.5991265608180396]
Foundation models can be disruptive for future AI development by scaling up deep learning.
Models achieve state-of-the-art performance on a variety of tasks in domains such as natural language processing and computer vision.
arXiv Detail & Related papers (2022-12-17T22:11:33Z) - Model Reprogramming: Resource-Efficient Cross-Domain Machine Learning [65.268245109828]
In data-rich domains such as vision, language, and speech, deep learning prevails to deliver high-performance task-specific models.
Deep learning in resource-limited domains still faces multiple challenges including (i) limited data, (ii) constrained model development cost, and (iii) lack of adequate pre-trained models for effective finetuning.
Model reprogramming enables resource-efficient cross-domain machine learning by repurposing a well-developed pre-trained model from a source domain to solve tasks in a target domain without model finetuning.
arXiv Detail & Related papers (2022-02-22T02:33:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.