Explaining Genetic Programming Trees using Large Language Models
- URL: http://arxiv.org/abs/2403.03397v1
- Date: Wed, 6 Mar 2024 01:38:42 GMT
- Title: Explaining Genetic Programming Trees using Large Language Models
- Authors: Paula Maddigan, Andrew Lensen, Bing Xue
- Abstract summary: Genetic programming (GP) has the potential to generate explainable results, especially when used for dimensionality reduction.
In this research, we investigate the potential of leveraging eXplainable AI (XAI) and large language models (LLMs) to improve the interpretability of GP-based non-linear dimensionality reduction.
- Score: 2.909922147268382
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Genetic programming (GP) has the potential to generate explainable results,
especially when used for dimensionality reduction. In this research, we
investigate the potential of leveraging eXplainable AI (XAI) and large language
models (LLMs) like ChatGPT to improve the interpretability of GP-based
non-linear dimensionality reduction. Our study introduces a novel XAI dashboard
named GP4NLDR, the first approach to combine state-of-the-art GP with an
LLM-powered chatbot to provide comprehensive, user-centred explanations. We
showcase the system's ability to provide intuitive and insightful narratives on
high-dimensional data reduction processes through case studies. Our study
highlights the importance of prompt engineering in eliciting accurate and
pertinent responses from LLMs. We also address important considerations around
data privacy, hallucinatory outputs, and the rapid advancements in generative
AI. Our findings demonstrate its potential in advancing the explainability of
GP algorithms. This opens the door for future research into explaining GP
models with LLMs.
Related papers
- Interpretable Language Modeling via Induction-head Ngram Models [74.26720927767398]
We propose Induction-head ngram models (Induction-Gram) to bolster modern ngram models with a hand-engineered "induction head"
This induction head uses a custom neural similarity metric to efficiently search the model's input context for potential next-word completions.
Experiments show that this simple method significantly improves next-word prediction over baseline interpretable models.
arXiv Detail & Related papers (2024-10-31T12:33:26Z) - ExaRanker-Open: Synthetic Explanation for IR using Open-Source LLMs [60.81649785463651]
We introduce ExaRanker-Open, where we adapt and explore the use of open-source language models to generate explanations.
Our findings reveal that incorporating explanations consistently enhances neural rankers, with benefits escalating as the LLM size increases.
arXiv Detail & Related papers (2024-02-09T11:23:14Z) - Domain Invariant Learning for Gaussian Processes and Bayesian
Exploration [39.83530605880014]
We propose a domain invariant learning algorithm for Gaussian processes (DIL-GP) with a min-max optimization on the likelihood.
Numerical experiments demonstrate the superiority of DIL-GP for predictions on several synthetic and real-world datasets.
arXiv Detail & Related papers (2023-12-18T16:13:34Z) - Graph Neural Prompting with Large Language Models [32.97391910476073]
Graph Neural Prompting (GNP) is a novel plug-and-play method to assist pre-trained language models in learning beneficial knowledge from knowledge graphs.
Extensive experiments on multiple datasets demonstrate the superiority of GNP on both commonsense and biomedical reasoning tasks.
arXiv Detail & Related papers (2023-09-27T06:33:29Z) - Exploring the Potential of Large Language Models (LLMs) in Learning on
Graphs [59.74814230246034]
Large Language Models (LLMs) have been proven to possess extensive common knowledge and powerful semantic comprehension abilities.
We investigate two possible pipelines: LLMs-as-Enhancers and LLMs-as-Predictors.
arXiv Detail & Related papers (2023-07-07T05:31:31Z) - Explaining Patterns in Data with Language Models via Interpretable
Autoprompting [143.4162028260874]
We introduce interpretable autoprompting (iPrompt), an algorithm that generates a natural-language string explaining the data.
iPrompt can yield meaningful insights by accurately finding groundtruth dataset descriptions.
Experiments with an fMRI dataset show the potential for iPrompt to aid in scientific discovery.
arXiv Detail & Related papers (2022-10-04T18:32:14Z) - Gradient-based explanations for Gaussian Process regression and
classification models [0.0]
Gaussian Processes (GPs) have proven themselves as a reliable and effective method in probabilistic Machine Learning.
Thanks to recent and current advances, modeling complex data with GPs is becoming more and more feasible.
We see an increasing interest in so-called explainable approaches - methods that aim to make a Machine Learning model's decision process transparent to humans.
arXiv Detail & Related papers (2022-05-25T14:11:00Z) - Less is More: A Call to Focus on Simpler Models in Genetic Programming
for Interpretable Machine Learning [1.0323063834827415]
Interpretability can be critical for the safe and responsible use of machine learning models in high-stakes applications.
We argue that research in GP for IML needs to focus on searching in the space of low-complexity models.
arXiv Detail & Related papers (2022-04-05T08:28:07Z) - SumGNN: Multi-typed Drug Interaction Prediction via Efficient Knowledge
Graph Summarization [64.56399911605286]
We propose SumGNN: knowledge summarization graph neural network, which is enabled by a subgraph extraction module.
SumGNN outperforms the best baseline by up to 5.54%, and the performance gain is particularly significant in low data relation types.
arXiv Detail & Related papers (2020-10-04T00:14:57Z) - Machine Learning in Nano-Scale Biomedical Engineering [77.75587007080894]
We review the existing research regarding the use of machine learning in nano-scale biomedical engineering.
The main challenges that can be formulated as ML problems are classified into the three main categories.
For each of the presented methodologies, special emphasis is given to its principles, applications, and limitations.
arXiv Detail & Related papers (2020-08-05T15:45:54Z) - Applying Genetic Programming to Improve Interpretability in Machine
Learning Models [0.3908287552267639]
We propose a Genetic Programming (GP) based approach, named Genetic Programming Explainer (GPX)
The method generates a noise set located in the neighborhood of the point of interest, whose prediction should be explained, and fits a local explanation model for the analyzed sample.
Our results indicate that the GPX is able to produce more accurate understanding of complex models than the state of the art.
arXiv Detail & Related papers (2020-05-18T16:09:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.