LCE: An Augmented Combination of Bagging and Boosting in Python
- URL: http://arxiv.org/abs/2308.07250v2
- Date: Tue, 15 Aug 2023 19:08:36 GMT
- Title: LCE: An Augmented Combination of Bagging and Boosting in Python
- Authors: Kevin Fauvel, \'Elisa Fromont, V\'eronique Masson, Philippe Faverdin
and Alexandre Termier
- Abstract summary: lcensemble is a high-performing, scalable and user-friendly Python package for the general tasks of classification and regression.
Local Cascade Ensemble (LCE) is a machine learning method that further enhances the prediction performance of the current state-of-the-art methods Random Forest and XGBoost.
- Score: 45.65284933207566
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: lcensemble is a high-performing, scalable and user-friendly Python package
for the general tasks of classification and regression. The package implements
Local Cascade Ensemble (LCE), a machine learning method that further enhances
the prediction performance of the current state-of-the-art methods Random
Forest and XGBoost. LCE combines their strengths and adopts a complementary
diversification approach to obtain a better generalizing predictor. The package
is compatible with scikit-learn, therefore it can interact with scikit-learn
pipelines and model selection tools. It is distributed under the Apache 2.0
license, and its source code is available at
https://github.com/LocalCascadeEnsemble/LCE.
Related papers
- eipy: An Open-Source Python Package for Multi-modal Data Integration
using Heterogeneous Ensembles [3.465746303617158]
eipy is an open-source Python package for developing effective, multi-modal heterogeneous ensembles for classification.
eipy provides both a rigorous, and user-friendly framework for comparing and selecting the best-performing data integration and predictive modeling methods.
arXiv Detail & Related papers (2024-01-17T20:07:47Z) - Stochastic Gradient Descent without Full Data Shuffle [65.97105896033815]
CorgiPile is a hierarchical data shuffling strategy that avoids a full data shuffle while maintaining comparable convergence rate of SGD as if a full shuffle were performed.
Our results show that CorgiPile can achieve comparable convergence rate with the full shuffle based SGD for both deep learning and generalized linear models.
arXiv Detail & Related papers (2022-06-12T20:04:31Z) - DADApy: Distance-based Analysis of DAta-manifolds in Python [51.37841707191944]
DADApy is a python software package for analysing and characterising high-dimensional data.
It provides methods for estimating the intrinsic dimension and the probability density, for performing density-based clustering and for comparing different distance metrics.
arXiv Detail & Related papers (2022-05-04T08:41:59Z) - L0Learn: A Scalable Package for Sparse Learning using L0 Regularization [6.037383467521294]
L0Learn is an open-source package for sparse linear regression classification.
It implements scalable, approximate algorithms, based on coordinate descent and local optimization.
arXiv Detail & Related papers (2022-02-10T03:51:25Z) - PyHHMM: A Python Library for Heterogeneous Hidden Markov Models [63.01207205641885]
PyHHMM is an object-oriented Python implementation of Heterogeneous-Hidden Markov Models (HHMMs)
PyHHMM emphasizes features not supported in similar available frameworks: a heterogeneous observation model, missing data inference, different model order selection criterias, and semi-supervised training.
PyHHMM relies on the numpy, scipy, scikit-learn, and seaborn Python packages, and is distributed under the Apache-2.0 License.
arXiv Detail & Related papers (2022-01-12T07:32:36Z) - IMBENS: Ensemble Class-imbalanced Learning in Python [26.007498723608155]
imbens is an open-source Python toolbox for implementing and deploying ensemble learning algorithms on class-imbalanced data.
imbens is released under the MIT open-source license and can be installed from Python Package Index (PyPI)
arXiv Detail & Related papers (2021-11-24T20:14:20Z) - COMBO: State-of-the-Art Morphosyntactic Analysis [0.0]
COMBO is a fully neural NLP system for accurate part-of-speech tagging, morphological analysis, lemmatisation, and (enhanced) dependency parsing.
It predicts categorical morphosyntactic features whilst also exposing their vector representations, extracted from hidden layers.
It is an easy to install Python package with automatically downloadable pre-trained models for over 40 languages.
arXiv Detail & Related papers (2021-09-11T20:00:20Z) - Picasso: A Sparse Learning Library for High Dimensional Data Analysis in
R and Python [77.33905890197269]
We describe a new library which implements a unified pathwise coordinate optimization for a variety of sparse learning problems.
The library is coded in R++ and has user-friendly sparse experiments.
arXiv Detail & Related papers (2020-06-27T02:39:24Z) - OPFython: A Python-Inspired Optimum-Path Forest Classifier [68.8204255655161]
This paper proposes a Python-based Optimum-Path Forest framework, denoted as OPFython.
As OPFython is a Python-based library, it provides a more friendly environment and a faster prototyping workspace than the C language.
arXiv Detail & Related papers (2020-01-28T15:46:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.