4. Gradient Descent: Discovering Adam Optimizer
Mar 10, 2025 · 10 min read · 🚀 Implementation of ADAM In earlier posts, we implemented a momentum-based optimizer, which offered several advantages: Momentum helps avoid getting stuck in local minima, much like in physics 🏃♂️. It smooths updates by acting as a moving averag...
Join discussion


