Clustering-Based Subset Selection in Evolutionary Multiobjective
Optimization
- URL: http://arxiv.org/abs/2108.08453v2
- Date: Sun, 29 Aug 2021 13:48:03 GMT
- Title: Clustering-Based Subset Selection in Evolutionary Multiobjective
Optimization
- Authors: Weiyu Chen, Hisao Ishibuchi, and Ke Shang
- Abstract summary: Subset selection is an important component in evolutionary multiobjective optimization (EMO) algorithms.
Clustering-based methods have not been evaluated in the context of subset selection from solution sets obtained by EMO algorithms.
- Score: 11.110675371854988
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Subset selection is an important component in evolutionary multiobjective
optimization (EMO) algorithms. Clustering, as a classic method to group similar
data points together, has been used for subset selection in some fields.
However, clustering-based methods have not been evaluated in the context of
subset selection from solution sets obtained by EMO algorithms. In this paper,
we first review some classic clustering algorithms. We also point out that
another popular subset selection method, i.e., inverted generational distance
(IGD)-based subset selection, can be viewed as clustering. Then, we perform a
comprehensive experimental study to evaluate the performance of various
clustering algorithms in different scenarios. Experimental results are analyzed
in detail, and some suggestions about the use of clustering algorithms for
subset selection are derived. Additionally, we demonstrate that decision
maker's preference can be introduced to clustering-based subset selection.
Related papers
- An incremental preference elicitation-based approach to learning potentially non-monotonic preferences in multi-criteria sorting [53.36437745983783]
We first construct a max-margin optimization-based model to model potentially non-monotonic preferences.
We devise information amount measurement methods and question selection strategies to pinpoint the most informative alternative in each iteration.
Two incremental preference elicitation-based algorithms are developed to learn potentially non-monotonic preferences.
arXiv Detail & Related papers (2024-09-04T14:36:20Z) - Comparison of Clustering Algorithms for Statistical Features of
Vibration Data Sets [0.4806505912512235]
We present an extensive comparison of the clustering algorithms K-means clustering, OPTICS, and Gaussian mixture model clustering (GMM) applied to statistical features extracted from the time and frequency domains of vibration data sets.
Our work showed that averaging (Mean, Median) and variance-based features (Standard Deviation, Interquartile Range) performed significantly better than shape-based features (Skewness, Kurtosis)
With an increase in the specified number of clusters, clustering algorithms performed better, although there were some specific algorithmic restrictions.
arXiv Detail & Related papers (2023-05-11T12:19:30Z) - Sparse and geometry-aware generalisation of the mutual information for joint discriminative clustering and feature selection [19.066989850964756]
We introduce a discriminative clustering model trying to maximise a geometry-aware generalisation of the mutual information called GEMINI.
This algorithm avoids the burden of feature exploration and is easily scalable to high-dimensional data and large amounts of samples while only designing a discriminative clustering model.
Our results show that Sparse GEMINI is a competitive algorithm and has the ability to select relevant subsets of variables with respect to the clustering without using relevance criteria or prior hypotheses.
arXiv Detail & Related papers (2023-02-07T10:52:04Z) - Gradient Based Clustering [72.15857783681658]
We propose a general approach for distance based clustering, using the gradient of the cost function that measures clustering quality.
The approach is an iterative two step procedure (alternating between cluster assignment and cluster center updates) and is applicable to a wide range of functions.
arXiv Detail & Related papers (2022-02-01T19:31:15Z) - A review of systematic selection of clustering algorithms and their
evaluation [0.0]
This paper aims to identify a systematic selection logic for clustering algorithms and corresponding validation concepts.
The goal is to enable potential users to choose an algorithm that fits best to their needs and the properties of their underlying data clustering problem.
arXiv Detail & Related papers (2021-06-24T07:01:46Z) - A Multi-disciplinary Ensemble Algorithm for Clustering Heterogeneous
Datasets [0.76146285961466]
We propose a new evolutionary clustering algorithm (ECAStar) based on social class ranking and meta-heuristic algorithms.
ECAStar is integrated with recombinational evolutionary operators, Levy flight optimisation, and some statistical techniques.
Experiments are conducted to evaluate the ECAStar against five conventional approaches.
arXiv Detail & Related papers (2021-01-01T07:20:50Z) - Cost-sensitive Hierarchical Clustering for Dynamic Classifier Selection [3.42658286826597]
We investigate if a method developed for general algorithm selection named cost-sensitive hierarchical clustering is suited for dynamic classifier selection.
Our experimental results show that our modified CSHC algorithm compares favorably.
arXiv Detail & Related papers (2020-12-14T18:44:51Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z) - Multi-View Spectral Clustering with High-Order Optimal Neighborhood
Laplacian Matrix [57.11971786407279]
Multi-view spectral clustering can effectively reveal the intrinsic cluster structure among data.
This paper proposes a multi-view spectral clustering algorithm that learns a high-order optimal neighborhood Laplacian matrix.
Our proposed algorithm generates the optimal Laplacian matrix by searching the neighborhood of the linear combination of both the first-order and high-order base.
arXiv Detail & Related papers (2020-08-31T12:28:40Z) - Learning to Select Base Classes for Few-shot Classification [96.92372639495551]
We use the Similarity Ratio as an indicator for the generalization performance of a few-shot model.
We then formulate the base class selection problem as a submodular optimization problem over Similarity Ratio.
arXiv Detail & Related papers (2020-04-01T09:55:18Z) - Clustering Binary Data by Application of Combinatorial Optimization
Heuristics [52.77024349608834]
We study clustering methods for binary data, first defining aggregation criteria that measure the compactness of clusters.
Five new and original methods are introduced, using neighborhoods and population behavior optimization metaheuristics.
From a set of 16 data tables generated by a quasi-Monte Carlo experiment, a comparison is performed for one of the aggregations using L1 dissimilarity, with hierarchical clustering, and a version of k-means: partitioning around medoids or PAM.
arXiv Detail & Related papers (2020-01-06T23:33:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.