top of page
Exploring Opportunities in AI & Machine Learning
All Posts


XGBoost Optimizations
XGBoost is one of the fastest gradient boosting algorithms, designed for high-dimensional and large-scale datasets. This guide explains its core optimizations—including approximate split finding, quantile sketches, and weighted quantile sketches—that reduce computation time while maintaining high accuracy.

Aryan
Sep 12, 2025


XGBoost Regularization
XGBoost is a powerful boosting algorithm, but it can overfit if not controlled. Regularization helps by simplifying trees, pruning unnecessary splits, and balancing bias–variance. This guide explains overfitting, how XGBoost improves on Gradient Boosting, and key parameters like gamma, lambda, max_depth, min_child_weight, learning rate, subsample, and early stopping to build robust models.

Aryan
Sep 5, 2025


The Core Math Behind XGBoost
XGBoost isn’t just another boosting algorithm — its strength lies in the mathematics that power its objective function, optimization, and tree-building strategy. In this post, we break down the core math behind XGBoost: from gradients and Hessians to Taylor series approximation, leaf weight derivation, and similarity scores. By the end, you’ll understand how XGBoost balances accuracy with regularization to build powerful predictive models.

Aryan
Aug 26, 2025


XGBoost for Classification
Master classification with XGBoost using a practical, beginner-friendly example. Understand how the algorithm builds decision trees, calculates log loss, optimizes splits, and uses probabilities to make accurate class predictions. A must-read for aspiring machine learning engineers.

Aryan
Aug 16, 2025


XGBoost For Regression
Dive into a step-by-step explanation of how XGBoost handles regression problems using a CGPA vs. salary dataset. Understand residual learning, tree construction, similarity scores, gain calculations, and how each stage progressively refines model accuracy. Ideal for beginners and intermediates mastering XGBoost.

Aryan
Aug 11, 2025


Introduction to XGBoost
XGBoost is one of the most powerful tools for structured/tabular data — known for its speed, scalability, and high performance. In this post, I’ve shared a detailed explanation of what makes XGBoost so effective, along with its history, features, and real-world use. A great resource for anyone learning ML!

Aryan
Jul 26, 2025


Gradient Boosting For Classification - 2
Gradient boosting shines in classification, combining weak learners like decision trees into a powerful model. By iteratively minimizing log loss, it corrects errors, excelling with imbalanced data and complex patterns. Tools like XGBoost and LightGBM offer flexibility via hyperparameters, making gradient boosting a top choice for data scientists tackling real-world classification tasks.

Aryan
Jun 25, 2025


Gradient Boosting For Classification - 1
Discover how Gradient Boosting builds powerful classifiers by turning weak learners into strong ones, step by step. From boosting logic to practical implementation, this blog walks you through an intuitive, beginner-friendly path using real-world data.

Aryan
Jun 20, 2025


Gradient Boosting For Regression - 2
Gradient Boosting is a powerful machine learning technique that builds strong models by combining weak learners. It minimizes errors using gradient descent and is widely used for accurate predictions in classification and regression tasks.

Aryan
May 31, 2025


Gradient Boosting For Regression - 1
Gradient Boosting is a powerful machine learning technique that builds strong models by combining many weak learners. It works by training each model to correct the errors of the previous one using gradient descent. Fast, accurate, and widely used in real-world applications, it’s a must-know for any data science enthusiast.

Aryan
May 29, 2025
bottom of page