top of page


How LSTMs Work: A Deep Dive into Gates and Information Flow
Long Short-Term Memory (LSTM) networks solve the limitations of traditional RNNs through a powerful gating mechanism. This article explains how the Forget, Input, and Output gates work internally, breaking down the math, vector dimensions, and intuition behind cell states and hidden states. A deep, implementation-level guide for serious deep learning practitioners.

Aryan
Feb 4


Backpropagation Through Time (BPTT) Explained Step-by-Step with a Simple RNN Example
Backpropagation in RNNs is often confusing because a single weight affects the loss through multiple time-dependent paths. In this post, I break down Backpropagation Through Time step by step using a small toy dataset, clearly showing how gradients flow across timesteps and why unfolding the network is necessary.

Aryan
Jan 28
bottom of page