Costs to Consider in Adopting NLP for Your Business
- URL: http://arxiv.org/abs/2012.08958v2
- Date: Thu, 15 Apr 2021 01:38:24 GMT
- Title: Costs to Consider in Adopting NLP for Your Business
- Authors: Made Nindyatama Nityasya, Haryo Akbarianto Wibowo, Radityo Eko
Prasojo, Alham Fikri Aji
- Abstract summary: We show the trade-off between performance gain and the cost across the models to give more insights for AI-pivoting business.
We call for more research into low-cost models, especially for under-resourced languages.
- Score: 3.608765813727773
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advances in Natural Language Processing (NLP) have largely pushed deep
transformer-based models as the go-to state-of-the-art technique without much
regard to the production and utilization cost. Companies planning to adopt
these methods into their business face difficulties because of the lack of
machine, data, and human resources to build them. We compare both the
performance and the cost of classical learning algorithms to the latest ones in
common sequence and text labeling tasks. In our industrial datasets, we find
that classical models often perform on par with deep neural ones despite the
lower cost. We show the trade-off between performance gain and the cost across
the models to give more insights for AI-pivoting business. Further, we call for
more research into low-cost models, especially for under-resourced languages.
Related papers
- Pricing and Competition for Generative AI [3.8677478583601776]
We explore the problem of how developers of new generative AI software can release and price their technology.
We first develop a comparison of two different models for a specific task with respect to user cost-effectiveness.
We then model the pricing problem of generative AI software as a game between two different companies.
arXiv Detail & Related papers (2024-11-04T22:52:45Z) - Prompting Encoder Models for Zero-Shot Classification: A Cross-Domain Study in Italian [75.94354349994576]
This paper explores the feasibility of employing smaller, domain-specific encoder LMs alongside prompting techniques to enhance performance in specialized contexts.
Our study concentrates on the Italian bureaucratic and legal language, experimenting with both general-purpose and further pre-trained encoder-only models.
The results indicate that while further pre-trained models may show diminished robustness in general knowledge, they exhibit superior adaptability for domain-specific tasks, even in a zero-shot setting.
arXiv Detail & Related papers (2024-07-30T08:50:16Z) - Evolve Cost-aware Acquisition Functions Using Large Language Models [11.209139558885035]
This paper introduces EvolCAF, a novel framework that integrates large language models (LLMs) with evolutionary computation (EC) to automatically design cost-aware AFs.
The designed cost-aware AF maximizes the utilization of available information from historical data, surrogate models and budget details.
In comparison to the well-known EIpu and EI-cool methods designed by human experts, our approach showcases remarkable efficiency and generalization across various tasks.
arXiv Detail & Related papers (2024-04-25T12:19:18Z) - Leveraging Data Augmentation for Process Information Extraction [0.0]
We investigate the application of data augmentation for natural language text data.
Data augmentation is an important component in enabling machine learning methods for the task of business process model generation from natural language text.
arXiv Detail & Related papers (2024-04-11T06:32:03Z) - Transferable and Principled Efficiency for Open-Vocabulary Segmentation [82.66423763561697]
Recent success of pre-trained foundation vision-language computation models makes Open-Vocabulary (OVS) possible.
This approach introduces heavy computational overheads for two challenges: 1) large model sizes of the backbone; 2) expensive costs during the fine-tuning.
In this paper, we aim to achieve performance that is comparable to or even better than prior OVS works based on large vision-language foundation models.
arXiv Detail & Related papers (2024-04-11T03:08:53Z) - PanGu-$\pi$: Enhancing Language Model Architectures via Nonlinearity
Compensation [97.78045712375047]
We present a new efficient model architecture for large language models (LLMs)
We show that PanGu-$pi$-7B can achieve a comparable performance to that of benchmarks with about 10% inference speed-up.
In addition, we have deployed PanGu-$pi$-7B in the high-value domains of finance and law, developing an LLM named YunShan for practical application.
arXiv Detail & Related papers (2023-12-27T11:49:24Z) - Power Hungry Processing: Watts Driving the Cost of AI Deployment? [74.19749699665216]
generative, multi-purpose AI systems promise a unified approach to building machine learning (ML) models into technology.
This ambition of generality'' comes at a steep cost to the environment, given the amount of energy these systems require and the amount of carbon that they emit.
We measure deployment cost as the amount of energy and carbon required to perform 1,000 inferences on representative benchmark dataset using these models.
We conclude with a discussion around the current trend of deploying multi-purpose generative ML systems, and caution that their utility should be more intentionally weighed against increased costs in terms of energy and emissions
arXiv Detail & Related papers (2023-11-28T15:09:36Z) - ZooPFL: Exploring Black-box Foundation Models for Personalized Federated
Learning [95.64041188351393]
This paper endeavors to solve both the challenges of limited resources and personalization.
We propose a method named ZOOPFL that uses Zeroth-Order Optimization for Personalized Federated Learning.
To reduce the computation costs and enhance personalization, we propose input surgery to incorporate an auto-encoder with low-dimensional and client-specific embeddings.
arXiv Detail & Related papers (2023-10-08T12:26:13Z) - Cheaply Evaluating Inference Efficiency Metrics for Autoregressive
Transformer APIs [66.30706841821123]
Large language models (LLMs) power many state-of-the-art systems in natural language processing.
LLMs are extremely computationally expensive, even at inference time.
We propose a new metric for comparing inference efficiency across models.
arXiv Detail & Related papers (2023-05-03T21:51:42Z) - Data Augmentation for Neural NLP [0.0]
Data augmentation is a low-cost approach for tackling data scarcity.
This paper gives an overview of current state-of-the-art data augmentation methods used for natural language processing.
arXiv Detail & Related papers (2023-02-22T14:47:15Z) - The NLP Cookbook: Modern Recipes for Transformer based Deep Learning
Architectures [0.0]
Natural Language Processing models have achieved phenomenal success in linguistic and semantic tasks.
Recent NLP architectures have utilized concepts of transfer learning, pruning, quantization, and knowledge distillation to achieve moderate model sizes.
Knowledge Retrievers have been built to extricate explicit data documents from a large corpus of databases with greater efficiency and accuracy.
arXiv Detail & Related papers (2021-03-23T22:38:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.