Based on the "Machine Learning" category.
Alternatively, view xgboost alternatives based on common mentions on social networks and blogs.
* Code Quality Rankings and insights are calculated and provided by Lumnify.
They vary from L1 to L5 with "L5" being the highest.
Do you think we are missing an alternative of xgboost or a related project?
Build Status Build Status XGBoost-CI Documentation Status [GitHub license](./LICENSE) CRAN Status Badge PyPI version Conda version Optuna Twitter OpenSSF Scorecard
Community | Documentation | [Resources](demo/README.md) | [Contributors](CONTRIBUTORS.md) | [Release Notes](NEWS.md)
XGBoost is an optimized distributed gradient boosting library designed to be highly efficient, flexible and portable. It implements machine learning algorithms under the Gradient Boosting framework. XGBoost provides a parallel tree boosting (also known as GBDT, GBM) that solve many data science problems in a fast and accurate way. The same code runs on major distributed environment (Kubernetes, Hadoop, SGE, MPI, Dask) and can solve problems beyond billions of examples.
© Contributors, 2021. Licensed under an Apache-2 license.
XGBoost has been developed and used by a group of active community members. Your help is very valuable to make the package better for everyone. Checkout the Community Page.
Become a sponsor and get a logo here. See details at Sponsoring the XGBoost Project. The funds are used to defray the cost of continuous integration and testing infrastructure (https://xgboost-ci.net).
Backers on Open Collective Sponsors on Open Collective
*Note that all licence references and agreements mentioned in the xgboost README section above
are relevant to that project's source code only.
Do not miss the trending, packages, news and articles with our weekly report.