Group-wise Reinforcement Feature Generation for Optimal and Explainable
Representation Space Reconstruction
- URL: http://arxiv.org/abs/2205.14526v1
- Date: Sat, 28 May 2022 21:34:14 GMT
- Title: Group-wise Reinforcement Feature Generation for Optimal and Explainable
Representation Space Reconstruction
- Authors: Dongjie Wang, Yanjie Fu, Kunpeng Liu, Xiaolin Li, Yan Solihin
- Abstract summary: We reformulate representation space reconstruction into an interactive process of nested feature generation and selection.
We design a group-wise generation strategy to cross a feature group, an operation, and another feature group to generate new features.
We present extensive experiments to demonstrate the effectiveness, efficiency, traceability, and explicitness of our system.
- Score: 25.604176830832586
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Representation (feature) space is an environment where data points are
vectorized, distances are computed, patterns are characterized, and geometric
structures are embedded. Extracting a good representation space is critical to
address the curse of dimensionality, improve model generalization, overcome
data sparsity, and increase the availability of classic models. Existing
literature, such as feature engineering and representation learning, is limited
in achieving full automation (e.g., over heavy reliance on intensive labor and
empirical experiences), explainable explicitness (e.g., traceable
reconstruction process and explainable new features), and flexible optimal
(e.g., optimal feature space reconstruction is not embedded into downstream
tasks). Can we simultaneously address the automation, explicitness, and optimal
challenges in representation space reconstruction for a machine learning task?
To answer this question, we propose a group-wise reinforcement generation
perspective. We reformulate representation space reconstruction into an
interactive process of nested feature generation and selection, where feature
generation is to generate new meaningful and explicit features, and feature
selection is to eliminate redundant features to control feature sizes. We
develop a cascading reinforcement learning method that leverages three
cascading Markov Decision Processes to learn optimal generation policies to
automate the selection of features and operations and the feature crossing. We
design a group-wise generation strategy to cross a feature group, an operation,
and another feature group to generate new features and find the strategy that
can enhance exploration efficiency and augment reward signals of cascading
agents. Finally, we present extensive experiments to demonstrate the
effectiveness, efficiency, traceability, and explicitness of our system.
Related papers
- Topology-aware Reinforcement Feature Space Reconstruction for Graph Data [22.5530178427691]
Reconstructing a good feature space is essential to augment the AI power of data, improve model generalization, and increase the availability of downstream ML models.
We use topology-aware reinforcement learning to automate and optimize feature space reconstruction for graph data.
Our approach combines the extraction of core subgraphs to capture essential structural information with a graph neural network (GNN) to encode topological features and reduce computing complexity.
arXiv Detail & Related papers (2024-11-08T18:01:05Z) - Reinforcement Feature Transformation for Polymer Property Performance Prediction [22.87577374767465]
Existing machine learning models face challenges in effectively learning polymer representations due to low-quality polymer datasets.
This study focuses on improving polymer property performance prediction tasks by reconstructing an optimal and explainable descriptor representation space.
arXiv Detail & Related papers (2024-09-23T23:42:18Z) - GrootVL: Tree Topology is All You Need in State Space Model [66.36757400689281]
GrootVL is a versatile multimodal framework that can be applied to both visual and textual tasks.
Our method significantly outperforms existing structured state space models on image classification, object detection and segmentation.
By fine-tuning large language models, our approach achieves consistent improvements in multiple textual tasks at minor training cost.
arXiv Detail & Related papers (2024-06-04T15:09:29Z) - Feature Selection as Deep Sequential Generative Learning [50.00973409680637]
We develop a deep variational transformer model over a joint of sequential reconstruction, variational, and performance evaluator losses.
Our model can distill feature selection knowledge and learn a continuous embedding space to map feature selection decision sequences into embedding vectors associated with utility scores.
arXiv Detail & Related papers (2024-03-06T16:31:56Z) - Human as Points: Explicit Point-based 3D Human Reconstruction from
Single-view RGB Images [78.56114271538061]
We introduce an explicit point-based human reconstruction framework called HaP.
Our approach is featured by fully-explicit point cloud estimation, manipulation, generation, and refinement in the 3D geometric space.
Our results may indicate a paradigm rollback to the fully-explicit and geometry-centric algorithm design.
arXiv Detail & Related papers (2023-11-06T05:52:29Z) - Feature Interaction Aware Automated Data Representation Transformation [27.26916497306978]
We develop a hierarchical reinforcement learning structure with cascading Markov Decision Processes to automate feature and operation selection.
We reward agents based on the interaction strength between selected features, resulting in intelligent and efficient exploration of the feature space that emulates human decision-making.
arXiv Detail & Related papers (2023-09-29T06:48:16Z) - Self-optimizing Feature Generation via Categorical Hashing
Representation and Hierarchical Reinforcement Crossing [37.73656271138515]
We propose a principled and generic representation-crossing framework to solve self-optimizing feature generation.
We present extensive experimental results to demonstrate the effectiveness and efficiency of the proposed method.
arXiv Detail & Related papers (2023-09-08T22:05:27Z) - Traceable Group-Wise Self-Optimizing Feature Transformation Learning: A
Dual Optimization Perspective [33.45878576396101]
Feature transformation aims to reconstruct an effective representation space by mathematically refining the existing features.
Existing research predominantly focuses on domain knowledge-based feature engineering or learning latent representations.
Our initial work took a pioneering step towards this challenge by introducing a novel self-optimizing framework.
arXiv Detail & Related papers (2023-06-29T12:29:21Z) - Structure-Aware Feature Generation for Zero-Shot Learning [108.76968151682621]
We introduce a novel structure-aware feature generation scheme, termed as SA-GAN, to account for the topological structure in learning both the latent space and the generative networks.
Our method significantly enhances the generalization capability on unseen-classes and consequently improve the classification performance.
arXiv Detail & Related papers (2021-08-16T11:52:08Z) - A Trainable Optimal Transport Embedding for Feature Aggregation and its
Relationship to Attention [96.77554122595578]
We introduce a parametrized representation of fixed size, which embeds and then aggregates elements from a given input set according to the optimal transport plan between the set and a trainable reference.
Our approach scales to large datasets and allows end-to-end training of the reference, while also providing a simple unsupervised learning mechanism with small computational cost.
arXiv Detail & Related papers (2020-06-22T08:35:58Z) - Mutual Information Maximization for Robust Plannable Representations [82.83676853746742]
We present MIRO, an information theoretic representational learning algorithm for model-based reinforcement learning.
We show that our approach is more robust than reconstruction objectives in the presence of distractors and cluttered scenes.
arXiv Detail & Related papers (2020-05-16T21:58:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.