We have seen previously that LightGBM was extremely fast, much faster than xgboost with default settings in R. Recently, to fix a prediction bug in LightGBM, a switch from float to double (for prediction-related functions) was made to fix that issue. Now that…
This post is about benchmarking LightGBM and xgboost (exact method) on a customized Bosch data set. I have seen xgboost being 10 times slower than LightGBM during the Bosch competition, but now we got back with some numbers to…
Laurae: This post is about the new feature of xgboost: the histogram tree grow method. Currently, it provides error in R but works in Python (?). You can find the pull request #1940 here. I’ll get benchmarks on my customized Bosch data set when…
Laurae: the topic post can be found on Kaggle.
Laurae wrote:
(until we all read: “(not available until the second stage of the competition)”)
Laurae: This post is about xgboost’s gblinear and its parameters. Elastic Net? Generalized Linear Model? Gradient Descent? Coordinate Descent?… The post was originally at Kaggle.
Laurae: This post is about what Intel CPU to look for if you want a powerful server/workstation. This is an insight from my “IT background” I do not even have (I hold over 40+ IT certifications, duh). If you do not know what to look for, ask on Reddit on…
Non-Kaggle post about the impact of Virtualized CPU cores / Sockets on Machine Learning / Optimization problems, specifically on xgboost and VMware (Linux host, Windows client). I found this…
Laurae: This post is about plotting data to maximize readability so you can read fast multivariate data vs a single label. Obviously, if there are interactions, it will be harder to notice them and you would go with regression…