Learning Choice Functions via Pareto-Embeddings
- URL: http://arxiv.org/abs/2007.06927v1
- Date: Tue, 14 Jul 2020 09:34:44 GMT
- Title: Learning Choice Functions via Pareto-Embeddings
- Authors: Karlson Pfannschmidt, Eyke H\"ullermeier
- Abstract summary: We consider the problem of learning to choose from a given set of objects, where each object is represented by a feature vector.
We propose a learning algorithm that minimizes a differentiable loss function suitable for this task.
- Score: 3.1410342959104725
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of learning to choose from a given set of objects,
where each object is represented by a feature vector. Traditional approaches in
choice modelling are mainly based on learning a latent, real-valued utility
function, thereby inducing a linear order on choice alternatives. While this
approach is suitable for discrete (top-1) choices, it is not straightforward
how to use it for subset choices. Instead of mapping choice alternatives to the
real number line, we propose to embed them into a higher-dimensional utility
space, in which we identify choice sets with Pareto-optimal points. To this
end, we propose a learning algorithm that minimizes a differentiable loss
function suitable for this task. We demonstrate the feasibility of learning a
Pareto-embedding on a suite of benchmark datasets.
Related papers
- Approximation of a Pareto Set Segment Using a Linear Model with Sharing Variables [14.161627541155775]
We develop a performance metric that considers both optimality and variable sharing.
We then design an algorithm for finding the model that minimizes the metric to meet the user's requirements.
Experimental results illustrate that we can obtain a linear model that approximates the mapping from the preference vectors to solutions in a local area well.
arXiv Detail & Related papers (2024-03-30T05:42:07Z) - Feature Selection as Deep Sequential Generative Learning [50.00973409680637]
We develop a deep variational transformer model over a joint of sequential reconstruction, variational, and performance evaluator losses.
Our model can distill feature selection knowledge and learn a continuous embedding space to map feature selection decision sequences into embedding vectors associated with utility scores.
arXiv Detail & Related papers (2024-03-06T16:31:56Z) - Feature Selection Based on Orthogonal Constraints and Polygon Area [10.587608254638667]
The goal of feature selection is to choose the optimal subset of features for a recognition task by evaluating the importance of each feature.
This paper introduces a non-monotone linear search between dependencies enhancing features labels.
Experimental results demonstrate that our approach not only effectively captures discriminative dependency but also surpasses traditional methods in reducing dimensions classification performance.
arXiv Detail & Related papers (2024-02-25T08:20:05Z) - Predict-Then-Optimize by Proxy: Learning Joint Models of Prediction and
Optimization [59.386153202037086]
Predict-Then- framework uses machine learning models to predict unknown parameters of an optimization problem from features before solving.
This approach can be inefficient and requires handcrafted, problem-specific rules for backpropagation through the optimization step.
This paper proposes an alternative method, in which optimal solutions are learned directly from the observable features by predictive models.
arXiv Detail & Related papers (2023-11-22T01:32:06Z) - A Performance-Driven Benchmark for Feature Selection in Tabular Deep
Learning [131.2910403490434]
Data scientists typically collect as many features as possible into their datasets, and even engineer new features from existing ones.
Existing benchmarks for tabular feature selection consider classical downstream models, toy synthetic datasets, or do not evaluate feature selectors on the basis of downstream performance.
We construct a challenging feature selection benchmark evaluated on downstream neural networks including transformers.
We also propose an input-gradient-based analogue of Lasso for neural networks that outperforms classical feature selection methods on challenging problems.
arXiv Detail & Related papers (2023-11-10T05:26:10Z) - Learning Choice Functions with Gaussian Processes [0.225596179391365]
In consumer theory, ranking available objects by means of preference relations yields the most common description of individual choices.
We propose a choice-model which allows an individual to express a set-valued choice.
arXiv Detail & Related papers (2023-02-01T12:46:43Z) - A Lagrangian Duality Approach to Active Learning [119.36233726867992]
We consider the batch active learning problem, where only a subset of the training data is labeled.
We formulate the learning problem using constrained optimization, where each constraint bounds the performance of the model on labeled samples.
We show, via numerical experiments, that our proposed approach performs similarly to or better than state-of-the-art active learning methods.
arXiv Detail & Related papers (2022-02-08T19:18:49Z) - Auto-weighted Multi-view Feature Selection with Graph Optimization [90.26124046530319]
We propose a novel unsupervised multi-view feature selection model based on graph learning.
The contributions are threefold: (1) during the feature selection procedure, the consensus similarity graph shared by different views is learned.
Experiments on various datasets demonstrate the superiority of the proposed method compared with the state-of-the-art methods.
arXiv Detail & Related papers (2021-04-11T03:25:25Z) - Feature Selection Using Reinforcement Learning [0.0]
The space of variables or features that can be used to characterize a particular predictor of interest continues to grow exponentially.
Identifying the most characterizing features that minimizes the variance without jeopardizing the bias of our models is critical to successfully training a machine learning model.
arXiv Detail & Related papers (2021-01-23T09:24:37Z) - Joint Adaptive Graph and Structured Sparsity Regularization for
Unsupervised Feature Selection [6.41804410246642]
We propose a joint adaptive graph and structured sparsity regularization unsupervised feature selection (JASFS) method.
A subset of optimal features will be selected in group, and the number of selected features will be determined automatically.
Experimental results on eight benchmarks demonstrate the effectiveness and efficiency of the proposed method.
arXiv Detail & Related papers (2020-10-09T08:17:04Z) - Learning with Differentiable Perturbed Optimizers [54.351317101356614]
We propose a systematic method to transform operations into operations that are differentiable and never locally constant.
Our approach relies on perturbeds, and can be used readily together with existing solvers.
We show how this framework can be connected to a family of losses developed in structured prediction, and give theoretical guarantees for their use in learning tasks.
arXiv Detail & Related papers (2020-02-20T11:11:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.