Genetic Programming for Evolving a Front of Interpretable Models for
Data Visualisation
- URL: http://arxiv.org/abs/2001.09578v1
- Date: Mon, 27 Jan 2020 04:03:19 GMT
- Title: Genetic Programming for Evolving a Front of Interpretable Models for
Data Visualisation
- Authors: Andrew Lensen, Bing Xue, Mengjie Zhang
- Abstract summary: We propose a genetic programming approach named GPtSNE for evolving interpretable mappings from a dataset to high-quality visualisations.
A multi-objective approach is designed that produces a variety of visualisations in a single run which give different trade-offs between visual quality and model complexity.
- Score: 4.4181317696554325
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data visualisation is a key tool in data mining for understanding big
datasets. Many visualisation methods have been proposed, including the
well-regarded state-of-the-art method t-Distributed Stochastic Neighbour
Embedding. However, the most powerful visualisation methods have a significant
limitation: the manner in which they create their visualisation from the
original features of the dataset is completely opaque. Many domains require an
understanding of the data in terms of the original features; there is hence a
need for powerful visualisation methods which use understandable models. In
this work, we propose a genetic programming approach named GPtSNE for evolving
interpretable mappings from a dataset to highquality visualisations. A
multi-objective approach is designed that produces a variety of visualisations
in a single run which give different trade-offs between visual quality and
model complexity. Testing against baseline methods on a variety of datasets
shows the clear potential of GP-tSNE to allow deeper insight into data than
that provided by existing visualisation methods. We further highlight the
benefits of a multi-objective approach through an in-depth analysis of a
candidate front, which shows how multiple models can
Related papers
- Multi-modal Auto-regressive Modeling via Visual Words [96.25078866446053]
We propose the concept of visual words, which maps the visual features to probability distributions over Large Multi-modal Models' vocabulary.
We further explore the distribution of visual features in the semantic space within LMM and the possibility of using text embeddings to represent visual information.
arXiv Detail & Related papers (2024-03-12T14:58:52Z) - StableLLaVA: Enhanced Visual Instruction Tuning with Synthesized
Image-Dialogue Data [129.92449761766025]
We propose a novel data collection methodology that synchronously synthesizes images and dialogues for visual instruction tuning.
This approach harnesses the power of generative models, marrying the abilities of ChatGPT and text-to-image generative models.
Our research includes comprehensive experiments conducted on various datasets.
arXiv Detail & Related papers (2023-08-20T12:43:52Z) - Multi-Objective Genetic Algorithm for Multi-View Feature Selection [0.23343923880060582]
We propose a novel genetic algorithm strategy to overcome limitations of traditional feature selection methods for multi-view data.
Our proposed approach, called the multi-view multi-objective feature selection genetic algorithm (MMFS-GA), simultaneously selects the optimal subset of features within a view and between views.
The results of our evaluations on three benchmark datasets, including synthetic and real data, show improvement over the best baseline methods.
arXiv Detail & Related papers (2023-05-26T13:25:20Z) - T1: Scaling Diffusion Probabilistic Fields to High-Resolution on Unified
Visual Modalities [69.16656086708291]
Diffusion Probabilistic Field (DPF) models the distribution of continuous functions defined over metric spaces.
We propose a new model comprising of a view-wise sampling algorithm to focus on local structure learning.
The model can be scaled to generate high-resolution data while unifying multiple modalities.
arXiv Detail & Related papers (2023-05-24T03:32:03Z) - A Spectral Method for Assessing and Combining Multiple Data
Visualizations [13.193958370464683]
We propose an efficient spectral method for assessing and combining multiple visualizations of a given dataset.
The proposed method provides a quantitative measure -- the visualization eigenscore -- of the relative performance of the visualizations for preserving the structure around each data point.
We analyze multiple simulated and real-world datasets to demonstrate the effectiveness of the eigenscores for evaluating visualizations and the superiority of the proposed consensus visualization.
arXiv Detail & Related papers (2022-10-25T02:13:19Z) - Auto-weighted Multi-view Feature Selection with Graph Optimization [90.26124046530319]
We propose a novel unsupervised multi-view feature selection model based on graph learning.
The contributions are threefold: (1) during the feature selection procedure, the consensus similarity graph shared by different views is learned.
Experiments on various datasets demonstrate the superiority of the proposed method compared with the state-of-the-art methods.
arXiv Detail & Related papers (2021-04-11T03:25:25Z) - Multi-view Data Visualisation via Manifold Learning [0.03222802562733786]
This manuscript proposes extensions of Student's t-distributed SNE, LLE and ISOMAP, to allow for dimensionality reduction and visualisation of multi-view data.
We show that by incorporating the low-dimensional embeddings obtained via the multi-view manifold learning approaches into the K-means algorithm, clusters of the samples are accurately identified.
arXiv Detail & Related papers (2021-01-17T19:54:36Z) - Saliency-driven Class Impressions for Feature Visualization of Deep
Neural Networks [55.11806035788036]
It is advantageous to visualize the features considered to be essential for classification.
Existing visualization methods develop high confidence images consisting of both background and foreground features.
In this work, we propose a saliency-driven approach to visualize discriminative features that are considered most important for a given task.
arXiv Detail & Related papers (2020-07-31T06:11:06Z) - Generative Partial Multi-View Clustering [133.36721417531734]
We propose a generative partial multi-view clustering model, named as GP-MVC, to address the incomplete multi-view problem.
First, multi-view encoder networks are trained to learn common low-dimensional representations, followed by a clustering layer to capture the consistent cluster structure across multiple views.
Second, view-specific generative adversarial networks are developed to generate the missing data of one view conditioning on the shared representation given by other views.
arXiv Detail & Related papers (2020-03-29T17:48:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.