STAMP 4 NLP -- An Agile Framework for Rapid Quality-Driven NLP
Applications Development
- URL: http://arxiv.org/abs/2111.08408v1
- Date: Tue, 16 Nov 2021 12:20:47 GMT
- Title: STAMP 4 NLP -- An Agile Framework for Rapid Quality-Driven NLP
Applications Development
- Authors: Philipp Kohl and Oliver Schmidts and Lars Kl\"oser and Henri Werth and
Bodo Kraft and Albert Z\"undorf
- Abstract summary: We introduce STAMP 4 NLP as an iterative and incremental process model for developing NLP applications.
With STAMP 4 NLP, we merge software engineering principles with best practices from data science.
Due to our iterative-incremental approach, businesses can deploy an enhanced version of the prototype to their software environment after every iteration.
- Score: 3.86574270083089
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The progress in natural language processing (NLP) research over the last
years, offers novel business opportunities for companies, as automated user
interaction or improved data analysis. Building sophisticated NLP applications
requires dealing with modern machine learning (ML) technologies, which impedes
enterprises from establishing successful NLP projects. Our experience in
applied NLP research projects shows that the continuous integration of research
prototypes in production-like environments with quality assurance builds trust
in the software and shows convenience and usefulness regarding the business
goal. We introduce STAMP 4 NLP as an iterative and incremental process model
for developing NLP applications. With STAMP 4 NLP, we merge software
engineering principles with best practices from data science. Instantiating our
process model allows efficiently creating prototypes by utilizing templates,
conventions, and implementations, enabling developers and data scientists to
focus on the business goals. Due to our iterative-incremental approach,
businesses can deploy an enhanced version of the prototype to their software
environment after every iteration, maximizing potential business value and
trust early and avoiding the cost of successful yet never deployed experiments.
Related papers
- Automatic Evaluation for Text-to-image Generation: Task-decomposed Framework, Distilled Training, and Meta-evaluation Benchmark [62.58869921806019]
We propose a task decomposition evaluation framework based on GPT-4o to automatically construct a new training dataset.
We design innovative training strategies to effectively distill GPT-4o's evaluation capabilities into a 7B open-source MLLM, MiniCPM-V-2.6.
Experimental results demonstrate that our distilled open-source MLLM significantly outperforms the current state-of-the-art GPT-4o-base baseline.
arXiv Detail & Related papers (2024-11-23T08:06:06Z) - Human-In-the-Loop Software Development Agents [12.830816751625829]
Large Language Models (LLMs) are introduced to automatically resolve software development tasks.
We introduce a Human-in-the-loop LLM-based Agents framework (HULA) for software development.
We design, implement, and deploy the HULA framework into Atlassian for internal uses.
arXiv Detail & Related papers (2024-11-19T23:22:33Z) - Think-on-Process: Dynamic Process Generation for Collaborative Development of Multi-Agent System [13.65717444483291]
ToP (Think-on-Process) is a dynamic process generation framework for software development.
Our framework significantly enhances the dynamic process generation capability of the GPT-3.5 and GPT-4.
arXiv Detail & Related papers (2024-09-10T15:02:34Z) - Agent-Driven Automatic Software Improvement [55.2480439325792]
This research proposal aims to explore innovative solutions by focusing on the deployment of agents powered by Large Language Models (LLMs)
The iterative nature of agents, which allows for continuous learning and adaptation, can help surpass common challenges in code generation.
We aim to use the iterative feedback in these systems to further fine-tune the LLMs underlying the agents, becoming better aligned to the task of automated software improvement.
arXiv Detail & Related papers (2024-06-24T15:45:22Z) - Zero-shot LLM-guided Counterfactual Generation: A Case Study on NLP Model Evaluation [15.254775341371364]
We explore the possibility of leveraging large language models for zero-shot counterfactual generation.
We propose a structured pipeline to facilitate this generation, and we hypothesize that the instruction-following and textual understanding capabilities of recent LLMs can be effectively leveraged.
arXiv Detail & Related papers (2024-05-08T03:57:45Z) - Selene: Pioneering Automated Proof in Software Verification [62.09555413263788]
We introduce Selene, which is the first project-level automated proof benchmark constructed based on the real-world industrial-level operating system microkernel, seL4.
Our experimental results with advanced large language models (LLMs), such as GPT-3.5-turbo and GPT-4, highlight the capabilities of LLMs in the domain of automated proof generation.
arXiv Detail & Related papers (2024-01-15T13:08:38Z) - Generative User-Experience Research for Developing Domain-specific Natural Language Processing Applications [4.139846693958609]
This paper proposes a new methodology for integrating generative UX research into developing domain NLP applications.
Generative UX research employs domain users at the initial stages of prototype development, i.e., ideation and concept evaluation, and the last stage for evaluating system usefulness and user utility.
arXiv Detail & Related papers (2023-06-28T12:17:45Z) - EasyNLP: A Comprehensive and Easy-to-use Toolkit for Natural Language
Processing [38.9428437204642]
EasyNLP is designed to make it easy to build NLP applications.
It features knowledge-enhanced pre-training, knowledge distillation and few-shot learning.
EasyNLP has powered over ten business units within Alibaba Group.
arXiv Detail & Related papers (2022-04-30T13:03:53Z) - FedNLP: A Research Platform for Federated Learning in Natural Language
Processing [55.01246123092445]
We present the FedNLP, a research platform for federated learning in NLP.
FedNLP supports various popular task formulations in NLP such as text classification, sequence tagging, question answering, seq2seq generation, and language modeling.
Preliminary experiments with FedNLP reveal that there exists a large performance gap between learning on decentralized and centralized datasets.
arXiv Detail & Related papers (2021-04-18T11:04:49Z) - Technology Readiness Levels for AI & ML [79.22051549519989]
Development of machine learning systems can be executed easily with modern tools, but the process is typically rushed and means-to-an-end.
Engineering systems follow well-defined processes and testing standards to streamline development for high-quality, reliable results.
We propose a proven systems engineering approach for machine learning development and deployment.
arXiv Detail & Related papers (2020-06-21T17:14:34Z) - Towards CRISP-ML(Q): A Machine Learning Process Model with Quality
Assurance Methodology [53.063411515511056]
We propose a process model for the development of machine learning applications.
The first phase combines business and data understanding as data availability oftentimes affects the feasibility of the project.
The sixth phase covers state-of-the-art approaches for monitoring and maintenance of a machine learning applications.
arXiv Detail & Related papers (2020-03-11T08:25:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.