On the Learning with Augmented Class via Forests
- URL: http://arxiv.org/abs/2505.09294v2
- Date: Mon, 14 Jul 2025 08:29:44 GMT
- Title: On the Learning with Augmented Class via Forests
- Authors: Fan Xu, Wuyang Chen, Wei Gao,
- Abstract summary: We focus on learning with augmented class via forests, where an augmented class may appear in testing data yet not in training data.<n>We develop the Learning with Augmented Class via Forests approach, which constructs shallow forests according to the augmented Gini impurity.<n>We also develop deep neural forests via an optimization objective based on our augmented Gini impurity.
- Score: 17.606415934443554
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Decision trees and forests have achieved successes in various real applications, most working with all testing classes known in training data. In this work, we focus on learning with augmented class via forests, where an augmented class may appear in testing data yet not in training data. We incorporate information of augmented class into trees' splitting, that is, augmented Gini impurity, a new splitting criterion is introduced to exploit some unlabeled data from testing distribution. We then develop the Learning with Augmented Class via Forests (short for LACForest) approach, which constructs shallow forests according to the augmented Gini impurity and then splits forests with pseudo-labeled augmented instances for better performance. We also develop deep neural forests via an optimization objective based on our augmented Gini impurity, which essentially utilizes the representation power of neural networks for forests. Theoretically, we present the convergence analysis for our augmented Gini impurity, and we finally conduct experiments to evaluate our approaches. The code is available at https://github.com/nju-xuf/LACForest.
Related papers
- TreeRPO: Tree Relative Policy Optimization [55.97385410074841]
name is a novel method that estimates the mathematical expectations of rewards at various reasoning steps using tree sampling.<n>Building on the group-relative reward training mechanism of GRPO, name innovatively computes rewards based on step-level groups generated during tree sampling.
arXiv Detail & Related papers (2025-06-05T15:56:38Z) - Binary Classification: Is Boosting stronger than Bagging? [5.877778007271621]
We introduce Enhanced Random Forests, an extension of vanilla Random Forests with extra functionalities and adaptive sample and model weighting.
We develop an iterative algorithm for adapting the training sample weights, by favoring the hardest examples, and an approach for finding personalized tree weighting schemes for each new sample.
Our method significantly improves upon regular Random Forests across 15 different binary classification datasets and considerably outperforms other tree methods, including XGBoost.
arXiv Detail & Related papers (2024-10-24T23:22:33Z) - A Coefficient Makes SVRG Effective [51.36251650664215]
Variance Reduced Gradient (SVRG) is a theoretically compelling optimization method.<n>In this work, we demonstrate the potential of SVRG in optimizing real-world neural networks.
arXiv Detail & Related papers (2023-11-09T18:47:44Z) - TreeLearn: A deep learning method for segmenting individual trees from ground-based LiDAR forest point clouds [40.46280139210502]
TreeLearn is a deep learning approach for tree instance segmentation of forest point clouds.<n>TreeLearn is trained on already segmented point clouds in a data-driven manner.<n>We trained TreeLearn on forest point clouds of 6665 trees, labeled using the Lidar360 software.
arXiv Detail & Related papers (2023-09-15T15:20:16Z) - ForensicsForest Family: A Series of Multi-scale Hierarchical Cascade Forests for Detecting GAN-generated Faces [53.739014757621376]
We describe a simple and effective forest-based method set called em ForensicsForest Family to detect GAN-generate faces.
ForenscisForest is a newly proposed Multi-scale Hierarchical Cascade Forest.
Hybrid ForensicsForest integrates the CNN layers into models.
Divide-and-Conquer ForensicsForest can construct a forest model using only a portion of training samplings.
arXiv Detail & Related papers (2023-08-02T06:41:19Z) - Tree Detection and Diameter Estimation Based on Deep Learning [0.0]
Tree perception is an essential building block toward autonomous forestry operations.
Deep neural network models trained on our datasets achieve a precision of 90.4% for tree detection.
Results offer promising avenues toward autonomous tree felling operations.
arXiv Detail & Related papers (2022-10-31T15:51:32Z) - Training Deep Learning Algorithms on Synthetic Forest Images for Tree
Detection [0.0]
We propose to use simulated forest environments to automatically generate 43 k realistic synthetic images with pixel-level annotations.
We also report the promising transfer learning capability of features learned on our synthetic dataset by directly predicting bounding box, segmentation masks and keypoints on real images.
arXiv Detail & Related papers (2022-10-08T20:49:40Z) - Q-learning with online random forests [0.913755431537592]
We provide online random forests as $Q$-function approximators and propose a novel method wherein the random forest is grown as learning proceeds.
We show that expanding forests improve performance, suggesting that expanding forests are viable for other applications of online random forests beyond the reinforcement learning setting.
arXiv Detail & Related papers (2022-04-07T23:00:39Z) - Making CNNs Interpretable by Building Dynamic Sequential Decision
Forests with Top-down Hierarchy Learning [62.82046926149371]
We propose a generic model transfer scheme to make Convlutional Neural Networks (CNNs) interpretable.
We achieve this by building a differentiable decision forest on top of CNNs.
We name the transferred model deep Dynamic Sequential Decision Forest (dDSDF)
arXiv Detail & Related papers (2021-06-05T07:41:18Z) - Growing Deep Forests Efficiently with Soft Routing and Learned
Connectivity [79.83903179393164]
This paper further extends the deep forest idea in several important aspects.
We employ a probabilistic tree whose nodes make probabilistic routing decisions, a.k.a., soft routing, rather than hard binary decisions.
Experiments on the MNIST dataset demonstrate that our empowered deep forests can achieve better or comparable performance than [1],[3].
arXiv Detail & Related papers (2020-12-29T18:05:05Z) - Residual Likelihood Forests [19.97069303172077]
This paper presents a novel ensemble learning approach called Residual Likelihood Forests (RLF)
Our weak learners produce conditional likelihoods that are sequentially optimized using global loss in the context of previous learners.
When compared against several ensemble approaches including Random Forests and Gradient Boosted Trees, RLFs offer a significant improvement in performance.
arXiv Detail & Related papers (2020-11-04T00:59:41Z) - Forest R-CNN: Large-Vocabulary Long-Tailed Object Detection and Instance
Segmentation [75.93960390191262]
We exploit prior knowledge of the relations among object categories to cluster fine-grained classes into coarser parent classes.
We propose a simple yet effective resampling method, NMS Resampling, to re-balance the data distribution.
Our method, termed as Forest R-CNN, can serve as a plug-and-play module being applied to most object recognition models.
arXiv Detail & Related papers (2020-08-13T03:52:37Z) - Neural Random Forest Imitation [24.02961053662835]
We introduce an imitation learning approach by generating training data from a random forest and learning a neural network that imitates its behavior.
This implicit transformation creates very efficient neural networks that learn the decision boundaries of a random forest.
Experiments on several real-world benchmark datasets demonstrate superior performance.
arXiv Detail & Related papers (2019-11-25T11:04:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.