Construct your individual book recommender with CatBoost Ranker

In today’s digital world, where information overload and wide product offer is the norm, with the ability to help customers find what they need and like could be a crucial factor to make our company stand out and get ahead of the competition.
Recommender systems can enhance digital experiences facilitating the seek for relevant information or products. At their core, these systems leverage data-driven algorithms to investigate user preferences, behaviors, and interactions, transforming raw data into meaningful recommendations tailored to individual tastes and preferences.
In this text, I provide an in depth explanation of how Gradient Tree Boosting works for classification, regression and recommender systems. I also introduce CatBoost, a state-of-art library for Gradient Tree Boosting, and the way it handles categorical features. Finally, I explain how YetiRank (a rating loss function) works and methods to implement it using CatBoost Ranker in a book recommender dataset.
As at all times, the code is obtainable on Github.
The thought of boosting relies on the hypothesis that a mix of sequential weak learners could be pretty much as good and even higher than a powerful learner [1]. A weak learner is an algorithm whose performance is at the very least barely higher than a random selection and, in case of Gradient Tree Boosting, the weak learner is a Decision Tree. These weak learners in a boosting arrange are trained to handle more complex observations that the previous one couldn’t solve. In this manner, the brand new weak learners can give attention to developing themselves on more complex patterns.
AdaBoost
The primary boosting algorithm with great success for binary classification was AdaBoost [2]. The weak learner in AdaBoost is a choice tree with a single split and, it really works by putting more weight on observations which might be more complex to categorise. The brand new weak learner is added sequentially to focus its training on more complex patterns. The ultimate prediction is made by majority vote…