Laurae: This post is about xgboost’s gblinear and its parameters. Elastic Net? Generalized Linear Model? Gradient Descent? Coordinate Descent?… The post was originally at Kaggle.
Laurae: This post is about transforming data appropriately when you are provided a multivariate distribution (preferable a low amount of variables) to find an explanation of internal interactions of variables. It takes an example on Allstate…
Laurae: This post is about the rationale between over-predicting/under-predicting and the performance metric you are optimizing. It takes the example of the Matthew’s Correlation Coefficient (MCC). The post was originally…
Laurae: This post is about my team scoring 8th for the Bosch competition, post Leaderboard results (Private Leaderboard open). I also earned two master tiers from the Bosch competition…
Laurae: This post is about the downfall of thinking the generalization of “using Public LB = same for Private LB”. This is a particularly good way to show why you should not only follow your Public LB advancements if…
Laurae: This post is about exploring series of images in order to be able to reconstruct how they are supposed to be ordered. Color theory knowledge is a little bit necessary (how colors are additive)…
Laurae: This post talks about hierarchical classification, but applies also to hierarchical regression. When a label has a known hierarchy, it is tempting to split the label. But…
Laurae: This post is about a row ID leak in a competition that stroke (openly) a competition 3 days before the end. The post was originally at Kaggle. The context of the leak can be found there, by fakeplastictrees. Without the available…
Laurae: This post is about how someone excelling at machine learning should transition to a business environment. It includes an extra post by inversion which summarizes an essential…