dFlow: A Domain Specific Language for the Rapid Development of
open-source Virtual Assistants
- URL: http://arxiv.org/abs/2310.02102v1
- Date: Tue, 3 Oct 2023 14:46:33 GMT
- Title: dFlow: A Domain Specific Language for the Rapid Development of
open-source Virtual Assistants
- Authors: Nikolaos Malamas, Konstantinos Panayiotou, Andreas L. Symeonidis
- Abstract summary: We present textitdFlow, a framework for creating task-specific VAs in a low-code manner.
We describe a system-agnostic VA meta-model, the developed grammar, and all essential processes for developing and deploying smart VAs.
For further convenience, we create a cloud-native architecture and expose it through the Discord platform.
- Score: 1.2277343096128712
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: An increasing number of models and frameworks for Virtual Assistant (VA)
development exist nowadays, following the progress in the Natural Language
Processing (NLP) and Natural Language Understanding (NLU) fields. Regardless of
their performance, popularity, and ease of use, these frameworks require at
least basic expertise in NLP and software engineering, even for simple and
repetitive processes, limiting their use only to the domain and programming
experts. However, since the current state of practice of VA development is a
straightforward process, Model-Driven Engineering approaches can be utilized to
achieve automation and rapid development in a more convenient manner. To this
end, we present \textit{dFlow}, a textual Domain-Specific Language (DSL) that
offers a simplified, reusable, and framework-agnostic language for creating
task-specific VAs in a low-code manner. We describe a system-agnostic VA
meta-model, the developed grammar, and all essential processes for developing
and deploying smart VAs. For further convenience, we create a cloud-native
architecture and expose it through the Discord platform. We conducted a
large-scale empirical evaluation with more than 200 junior software developers
and collected positive feedback, indicating that dFlow can accelerate the
entire VA development process, while also enabling citizen and software
developers with minimum experience to participate.
Related papers
- CLSP: High-Fidelity Contrastive Language-State Pre-training for Agent State Representation [49.19402798479942]
multimodal learning has become an important research area for artificial intelligence.
For intelligent agents, the state is a crucial modality to convey precise information alongside common modalities like images, videos, and language.
We propose a High-Fidelity Contrastive Language-State Pre-training method, which can accurately encode state information into general representations.
arXiv Detail & Related papers (2024-09-24T07:08:00Z) - A Framework to Model ML Engineering Processes [1.9744907811058787]
Development of Machine Learning (ML) based systems is complex and requires multidisciplinary teams with diverse skill sets.
Current process modeling languages are not suitable for describing the development of such systems.
We introduce a framework for modeling ML-based software development processes, built around a domain-specific language.
arXiv Detail & Related papers (2024-04-29T09:17:36Z) - CMULAB: An Open-Source Framework for Training and Deployment of Natural Language Processing Models [59.91221728187576]
This paper introduces the CMU Linguistic Linguistic Backend, an open-source framework that simplifies model deployment and continuous human-in-the-loop fine-tuning of NLP models.
CMULAB enables users to leverage the power of multilingual models to quickly adapt and extend existing tools for speech recognition, OCR, translation, and syntactic analysis to new languages.
arXiv Detail & Related papers (2024-04-03T02:21:46Z) - adaptNMT: an open-source, language-agnostic development environment for
Neural Machine Translation [2.648836772989769]
adaptNMT is designed for both technical and non-technical users who work in the field of machine translation.
The application is built upon the widely-adopted OpenNMT ecosystem.
To support eco-friendly research in the NLP space, a green report also flags the power consumption and kgCO$_2$ emissions generated during model development.
arXiv Detail & Related papers (2024-03-04T12:10:17Z) - LAMM: Language-Assisted Multi-Modal Instruction-Tuning Dataset,
Framework, and Benchmark [81.42376626294812]
We present Language-Assisted Multi-Modal instruction tuning dataset, framework, and benchmark.
Our aim is to establish LAMM as a growing ecosystem for training and evaluating MLLMs.
We present a comprehensive dataset and benchmark, which cover a wide range of vision tasks for 2D and 3D vision.
arXiv Detail & Related papers (2023-06-11T14:01:17Z) - HuSpaCy: an industrial-strength Hungarian natural language processing
toolkit [0.0]
A language processing pipeline should consist of close to state-of-the-art lemmatization, morphosyntactic analysis, entity recognition and word embeddings.
This paper introduces HuSpaCy, an industryready Hungarian language processing pipeline.
arXiv Detail & Related papers (2022-01-06T07:49:45Z) - Reinforced Iterative Knowledge Distillation for Cross-Lingual Named
Entity Recognition [54.92161571089808]
Cross-lingual NER transfers knowledge from rich-resource language to languages with low resources.
Existing cross-lingual NER methods do not make good use of rich unlabeled data in target languages.
We develop a novel approach based on the ideas of semi-supervised learning and reinforcement learning.
arXiv Detail & Related papers (2021-06-01T05:46:22Z) - A Data-Centric Framework for Composable NLP Workflows [109.51144493023533]
Empirical natural language processing systems in application domains (e.g., healthcare, finance, education) involve interoperation among multiple components.
We establish a unified open-source framework to support fast development of such sophisticated NLP in a composable manner.
arXiv Detail & Related papers (2021-03-02T16:19:44Z) - Technology Readiness Levels for Machine Learning Systems [107.56979560568232]
Development and deployment of machine learning systems can be executed easily with modern tools, but the process is typically rushed and means-to-an-end.
We have developed a proven systems engineering approach for machine learning development and deployment.
Our "Machine Learning Technology Readiness Levels" framework defines a principled process to ensure robust, reliable, and responsible systems.
arXiv Detail & Related papers (2021-01-11T15:54:48Z) - Sequence Model Design for Code Completion in the Modern IDE [3.4824234779710452]
We propose a novel design for predicting top-k next tokens that combines static analysis' ability to enumerate all valid keywords and in-scope identifiers with the ability of a language model to place a probability distribution over them.
Our model mixes character-level input representation with token output to represent out-of-vocabulary (OOV) tokens meaningfully and minimize prediction latency.
arXiv Detail & Related papers (2020-04-10T22:40:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.