AI2: The next leap toward native language based and explainable machine
learning framework
- URL: http://arxiv.org/abs/2301.03391v1
- Date: Mon, 9 Jan 2023 14:48:35 GMT
- Title: AI2: The next leap toward native language based and explainable machine
learning framework
- Authors: Jean-S\'ebastien Dessureault, Daniel Massicotte
- Abstract summary: The proposed framework, named AI$2$, uses a natural language interface that allows a non-specialist to benefit from machine learning algorithms.
The primary contribution of the AI$2$ framework allows a user to call the machine learning algorithms in English, making its interface usage easier.
Another contribution is a preprocessing module that helps to describe and to load data properly.
- Score: 1.827510863075184
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The machine learning frameworks flourished in the last decades, allowing
artificial intelligence to get out of academic circles to be applied to
enterprise domains. This field has significantly advanced, but there is still
some meaningful improvement to reach the subsequent expectations. The proposed
framework, named AI$^{2}$, uses a natural language interface that allows a
non-specialist to benefit from machine learning algorithms without necessarily
knowing how to program with a programming language. The primary contribution of
the AI$^{2}$ framework allows a user to call the machine learning algorithms in
English, making its interface usage easier. The second contribution is
greenhouse gas (GHG) awareness. It has some strategies to evaluate the GHG
generated by the algorithm to be called and to propose alternatives to find a
solution without executing the energy-intensive algorithm. Another contribution
is a preprocessing module that helps to describe and to load data properly.
Using an English text-based chatbot, this module guides the user to define
every dataset so that it can be described, normalized, loaded and divided
appropriately. The last contribution of this paper is about explainability. For
decades, the scientific community has known that machine learning algorithms
imply the famous black-box problem. Traditional machine learning methods
convert an input into an output without being able to justify this result. The
proposed framework explains the algorithm's process with the proper texts,
graphics and tables. The results, declined in five cases, present usage
applications from the user's English command to the explained output.
Ultimately, the AI$^{2}$ framework represents the next leap toward native
language-based, human-oriented concerns about machine learning framework.
Related papers
- Autoregressive Large Language Models are Computationally Universal [59.34397993748194]
We show that autoregressive decoding of a transformer-based language model can realize universal computation.
We first show that a universal Turing machine can be simulated by a Lag system with 2027 production rules.
We conclude that, by the Church-Turing thesis, prompted gemini-1.5-pro-001 with extended autoregressive (greedy) decoding is a general purpose computer.
arXiv Detail & Related papers (2024-10-04T06:05:17Z) - From Decoding to Meta-Generation: Inference-time Algorithms for Large Language Models [63.188607839223046]
This survey focuses on the benefits of scaling compute during inference.
We explore three areas under a unified mathematical formalism: token-level generation algorithms, meta-generation algorithms, and efficient generation.
arXiv Detail & Related papers (2024-06-24T17:45:59Z) - CodeGRAG: Bridging the Gap between Natural Language and Programming Language via Graphical Retrieval Augmented Generation [58.84212778960507]
We propose CodeGRAG, a Graphical Retrieval Augmented Code Generation framework to enhance the performance of LLMs.
CodeGRAG builds the graphical view of code blocks based on the control flow and data flow of them to fill the gap between programming languages and natural language.
Various experiments and ablations are done on four datasets including both the C++ and python languages to validate the hard meta-graph prompt, the soft prompting technique, and the effectiveness of the objectives for pretrained GNN expert.
arXiv Detail & Related papers (2024-05-03T02:48:55Z) - Linguacodus: A Synergistic Framework for Transformative Code Generation in Machine Learning Pipelines [0.0]
We introduce a dynamic pipeline that transforms natural language task descriptions into code through high-level data-shaping instructions.
This paper details the fine-tuning process, and sheds light on how natural language descriptions can be translated into functional code.
We propose an algorithm capable of transforming a natural description of an ML task into code with minimal human interaction.
arXiv Detail & Related papers (2024-03-18T08:58:47Z) - Machine Learning Augmented Branch and Bound for Mixed Integer Linear
Programming [11.293025183996832]
Mixed Linear Programming (MILP) offers a powerful modeling language for a wide range of applications.
In recent years, there has been an explosive development in the use of machine learning algorithms for enhancing all main tasks involved in the branch-and-bound algorithm.
In particular, we give detailed attention to machine learning algorithms that automatically optimize some metric of branch-and-bound efficiency.
arXiv Detail & Related papers (2024-02-08T09:19:26Z) - Performance Prediction of Data-Driven Knowledge summarization of High
Entropy Alloys (HEAs) literature implementing Natural Language Processing
algorithms [0.0]
The goal of natural language processing (NLP) is to get a machine intelligence to process words the same way a human brain does.
Five NLP algorithms, namely, Geneism, Sumy, Luhn, Latent Semantic Analysis (LSA), and Kull-back-Liebler (KL) al-gorithm, are implemented.
Luhn algorithm has the highest accuracy score for the knowledge summarization tasks compared to the other used algorithms.
arXiv Detail & Related papers (2023-11-06T16:22:32Z) - Evolution of Natural Language Processing Technology: Not Just Language
Processing Towards General Purpose AI [0.0]
This report provides a technological explanation of how cutting-edge NLP has made it possible to realize the "practice makes perfect" principle.
Achievements exceeding the initial predictions have been reported from the results of learning vast amounts of textual data using deep learning.
It is an accurate example of the learner embodying the concept of "practice makes perfect" by using vast amounts of textual data.
arXiv Detail & Related papers (2023-10-10T00:41:38Z) - Interactive Code Generation via Test-Driven User-Intent Formalization [60.90035204567797]
Large language models (LLMs) produce code from informal natural language (NL) intent.
It is hard to define a notion of correctness since natural language can be ambiguous and lacks a formal semantics.
We describe a language-agnostic abstract algorithm and a concrete implementation TiCoder.
arXiv Detail & Related papers (2022-08-11T17:41:08Z) - BigIssue: A Realistic Bug Localization Benchmark [89.8240118116093]
BigIssue is a benchmark for realistic bug localization.
We provide a general benchmark with a diversity of real and synthetic Java bugs.
We hope to advance the state of the art in bug localization, in turn improving APR performance and increasing its applicability to the modern development cycle.
arXiv Detail & Related papers (2022-07-21T20:17:53Z) - Pre-Trained Language Models for Interactive Decision-Making [72.77825666035203]
We describe a framework for imitation learning in which goals and observations are represented as a sequence of embeddings.
We demonstrate that this framework enables effective generalization across different environments.
For test tasks involving novel goals or novel scenes, initializing policies with language models improves task completion rates by 43.6%.
arXiv Detail & Related papers (2022-02-03T18:55:52Z) - Julia Language in Machine Learning: Algorithms, Applications, and Open
Issues [5.666843255747851]
Machine learning is driving development across many fields in science and engineering.
Currently, the programming languages most commonly used to develop machine learning algorithms include Python, and C/C ++.
This paper summarizes the related research work and developments in the application of the Julia language in machine learning.
arXiv Detail & Related papers (2020-03-23T09:31:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.