To tree or not to tree? Assessing the impact of smoothing the decision
boundaries
- URL: http://arxiv.org/abs/2210.03672v1
- Date: Fri, 7 Oct 2022 16:27:13 GMT
- Title: To tree or not to tree? Assessing the impact of smoothing the decision
boundaries
- Authors: Anthea M\'erida, Argyris Kalogeratos and Mathilde Mougeot
- Abstract summary: We quantify how much should the 'rigid' decision boundaries, produced by an algorithm that naturally finds such solutions, be relaxed to obtain a performance improvement.
We show how these two measures can help the user in figuring out how expressive his model should be, before exploring it further via model selection.
- Score: 4.286327408435937
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When analyzing a dataset, it can be useful to assess how smooth the decision
boundaries need to be for a model to better fit the data. This paper addresses
this question by proposing the quantification of how much should the 'rigid'
decision boundaries, produced by an algorithm that naturally finds such
solutions, be relaxed to obtain a performance improvement. The approach we
propose starts with the rigid decision boundaries of a seed Decision Tree (seed
DT), which is used to initialize a Neural DT (NDT). The initial boundaries are
challenged by relaxing them progressively through training the NDT. During this
process, we measure the NDT's performance and decision agreement to its seed
DT. We show how these two measures can help the user in figuring out how
expressive his model should be, before exploring it further via model
selection. The validity of our approach is demonstrated with experiments on
simulated and benchmark datasets.
Related papers
- Uncertainty Measurement of Deep Learning System based on the Convex Hull of Training Sets [0.13265175299265505]
We propose To-hull Uncertainty and Closure Ratio, which measures an uncertainty of trained model based on the convex hull of training data.
It can observe the positional relation between the convex hull of the learned data and an unseen sample and infer how extrapolate the sample is from the convex hull.
arXiv Detail & Related papers (2024-05-25T06:25:24Z) - Online Learning of Decision Trees with Thompson Sampling [12.403737756721467]
Decision Trees are prominent prediction models for interpretable Machine Learning.
We devise a new Monte Carlo Tree Search algorithm, able to produce optimal Decision Trees in an online setting.
arXiv Detail & Related papers (2024-04-09T15:53:02Z) - Reinforcement Learning for Node Selection in Branch-and-Bound [52.2648997215667]
Current state-of-the-art selectors utilize either hand-crafted ensembles that automatically switch between naive sub-node selectors, or learned node selectors that rely on individual node data.
We propose a novel simulation technique that uses reinforcement learning (RL) while considering the entire tree state, rather than just isolated nodes.
arXiv Detail & Related papers (2023-09-29T19:55:56Z) - Limits of Actor-Critic Algorithms for Decision Tree Policies Learning in
IBMDPs [9.587070290189507]
Interpretability of AI models allows for user safety checks to build trust in such AIs.
Decision Trees (DTs) provide a global look at the learned model and transparently reveal which features of the input are critical for making a decision.
Recent Reinforcement Learning framework has been proposed to explore the space of DTs using deep RL.
arXiv Detail & Related papers (2023-09-23T13:06:20Z) - STEERING: Stein Information Directed Exploration for Model-Based
Reinforcement Learning [111.75423966239092]
We propose an exploration incentive in terms of the integral probability metric (IPM) between a current estimate of the transition model and the unknown optimal.
Based on KSD, we develop a novel algorithm algo: textbfSTEin information dirtextbfEcted exploration for model-based textbfReinforcement LearntextbfING.
arXiv Detail & Related papers (2023-01-28T00:49:28Z) - Understanding Deep Learning via Decision Boundary [81.49114762506287]
We show that the neural network with lower decision boundary (DB) variability has better generalizability.
Two new notions, algorithm DB variability and $(epsilon, eta)$-data DB variability, are proposed to measure the decision boundary variability.
arXiv Detail & Related papers (2022-06-03T11:34:12Z) - R(Det)^2: Randomized Decision Routing for Object Detection [64.48369663018376]
We propose a novel approach to combine decision trees and deep neural networks in an end-to-end learning manner for object detection.
To facilitate effective learning, we propose randomized decision routing with node selective and associative losses.
We name this approach as the randomized decision routing for object detection, abbreviated as R(Det)$2$.
arXiv Detail & Related papers (2022-04-02T07:54:58Z) - Rectified Decision Trees: Exploring the Landscape of Interpretable and
Effective Machine Learning [66.01622034708319]
We propose a knowledge distillation based decision trees extension, dubbed rectified decision trees (ReDT)
We extend the splitting criteria and the ending condition of the standard decision trees, which allows training with soft labels.
We then train the ReDT based on the soft label distilled from a well-trained teacher model through a novel jackknife-based method.
arXiv Detail & Related papers (2020-08-21T10:45:25Z) - SODEN: A Scalable Continuous-Time Survival Model through Ordinary
Differential Equation Networks [14.564168076456822]
We propose a flexible model for survival analysis using neural networks along with scalable optimization algorithms.
We demonstrate the effectiveness of the proposed method in comparison to existing state-of-the-art deep learning survival analysis models.
arXiv Detail & Related papers (2020-08-19T19:11:25Z) - Model family selection for classification using Neural Decision Trees [4.286327408435937]
In this paper we propose a method to reduce the scope of exploration needed for the task.
The idea is to quantify how much it would be necessary to depart from trained instances of a given family, reference models (RMs) carrying rigid' decision boundaries.
arXiv Detail & Related papers (2020-06-20T01:27:01Z) - Uncertainty Estimation Using a Single Deep Deterministic Neural Network [66.26231423824089]
We propose a method for training a deterministic deep model that can find and reject out of distribution data points at test time with a single forward pass.
We scale training in these with a novel loss function and centroid updating scheme and match the accuracy of softmax models.
arXiv Detail & Related papers (2020-03-04T12:27:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.