Learning Service Selection Decision Making Behaviors During Scientific Workflow Development
- URL: http://arxiv.org/abs/2404.00420v1
- Date: Sat, 30 Mar 2024 16:58:42 GMT
- Title: Learning Service Selection Decision Making Behaviors During Scientific Workflow Development
- Authors: Xihao Xie, Jia Zhang, Rahul Ramachandran, Tsengdar J. Lee, Seungwon Lee,
- Abstract summary: In this paper, a novel context-aware approach is proposed to recommending next services in a workflow development process.
The problem of next service recommendation is mapped to next word prediction.
Experiments on a real-word repository have demonstrated the effectiveness of this approach.
- Score: 3.341965553962658
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Increasingly, more software services have been published onto the Internet, making it a big challenge to recommend services in the process of a scientific workflow composition. In this paper, a novel context-aware approach is proposed to recommending next services in a workflow development process, through learning service representation and service selection decision making behaviors from workflow provenance. Inspired by natural language sentence generation, the composition process of a scientific workflow is formalized as a step-wise procedure within the context of the goal of workflow, and the problem of next service recommendation is mapped to next word prediction. Historical service dependencies are first extracted from scientific workflow provenance to build a knowledge graph. Service sequences are then generated based on diverse composition path generation strategies. Afterwards, the generated corpus of composition paths are leveraged to study previous decision making strategies. Such a trained goal-oriented next service prediction model will be used to recommend top K candidate services during workflow composition process. Extensive experiments on a real-word repository have demonstrated the effectiveness of this approach.
Related papers
- FlowBench: Revisiting and Benchmarking Workflow-Guided Planning for LLM-based Agents [64.1759086221016]
We present FlowBench, the first benchmark for workflow-guided planning.
FlowBench covers 51 different scenarios from 6 domains, with knowledge presented in diverse formats.
Results indicate that current LLM agents need considerable improvements for satisfactory planning.
arXiv Detail & Related papers (2024-06-21T06:13:00Z) - Systematic Mapping Protocol -- UX Design role in software development
process [55.2480439325792]
We present a systematic mapping protocol for investigating the role of the UX designer in the software development process.
We define the research questions, scope, sources, search strategy, selection criteria, data extraction, and analysis methods that we will use to conduct the mapping study.
arXiv Detail & Related papers (2024-02-20T16:56:46Z) - Trace Encoding in Process Mining: a survey and benchmarking [0.34410212782758054]
Methods are employed across several process mining tasks, including predictive process monitoring, anomalous case detection, clustering trace, etc.
Most papers choose existing encoding methods arbitrarily or employ a strategy based on a specific expert knowledge domain.
This work aims at providing a comprehensive survey on event log encoding by comparing 27 methods.
arXiv Detail & Related papers (2023-01-05T17:25:30Z) - Goal-Driven Context-Aware Next Service Recommendation for Mashup
Composition [6.17189383632496]
Service discovery and recommendation has attracted significant momentum in both academia and industry.
This paper proposes a novel incremental recommend-as-you-go approach to recommending next potential service based on the context of a mashup under construction.
arXiv Detail & Related papers (2022-10-25T16:24:21Z) - Learning Context-Aware Service Representation for Service Recommendation
in Workflow Composition [6.17189383632496]
This paper proposes a novel NLP-inspired approach to recommending services throughout a workflow development process.
A workflow composition process is formalized as a step-wise, context-aware service generation procedure.
Service embeddings are then learned by applying deep learning model from the NLP field.
arXiv Detail & Related papers (2022-05-24T04:18:01Z) - Nemo: Guiding and Contextualizing Weak Supervision for Interactive Data
Programming [77.38174112525168]
We present Nemo, an end-to-end interactive Supervision system that improves overall productivity of WS learning pipeline by an average 20% (and up to 47% in one task) compared to the prevailing WS supervision approach.
arXiv Detail & Related papers (2022-03-02T19:57:32Z) - Video Summarization Using Deep Neural Networks: A Survey [72.98424352264904]
Video summarization technologies aim to create a concise and complete synopsis by selecting the most informative parts of the video content.
This work focuses on the recent advances in the area and provides a comprehensive survey of the existing deep-learning-based methods for generic video summarization.
arXiv Detail & Related papers (2021-01-15T11:41:29Z) - Explainable Artificial Intelligence for Process Mining: A General
Overview and Application of a Novel Local Explanation Approach for Predictive
Process Monitoring [0.0]
This study proposes a conceptual framework sought to establish and promote understanding of decision-making environment.
This study defines the local regions from the validation dataset by using the intermediate latent space representations.
The adopted deep learning classifier achieves a good performance with the Area Under the ROC Curve of 0.94.
arXiv Detail & Related papers (2020-09-04T10:28:56Z) - Pre-training Text Representations as Meta Learning [113.3361289756749]
We introduce a learning algorithm which directly optimize model's ability to learn text representations for effective learning of downstream tasks.
We show that there is an intrinsic connection between multi-task pre-training and model-agnostic meta-learning with a sequence of meta-train steps.
arXiv Detail & Related papers (2020-04-12T09:05:47Z) - Exploring the Limits of Transfer Learning with a Unified Text-to-Text
Transformer [64.22926988297685]
Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on a downstream task, has emerged as a powerful technique in natural language processing (NLP)
In this paper, we explore the landscape of introducing transfer learning techniques for NLP by a unified framework that converts all text-based language problems into a text-to-text format.
arXiv Detail & Related papers (2019-10-23T17:37:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.