Inventor of AdamW and SGDR | Generated by AI

Home 2025.12

Question: Who is Ilya Loshchilov in machine learning?

Answer:

Ilya Loshchilov is a prominent researcher in machine learning, currently working as a research scientist at NVIDIA. His expertise spans stochastic optimization, evolutionary computation, deep learning, and reinforcement learning.

He is best known for co-authoring (with Frank Hutter) the influential 2017 paper “Decoupled Weight Decay Regularization,” which introduced AdamW—a modified version of the Adam optimizer that properly decouples weight decay from the adaptive learning rate. AdamW has become the standard optimizer in many deep learning frameworks (e.g., PyTorch, TensorFlow) and is widely used for improving generalization in training neural networks.

Another key contribution is the 2016 paper on SGDR (Stochastic Gradient Descent with Warm Restarts), a learning rate scheduling technique involving cosine annealing with restarts, which enhances training performance and convergence.

Earlier in his career, Loshchilov focused on evolutionary algorithms, including improvements to Covariance Matrix Adaptation Evolution Strategy (CMA-ES), surrogate-assisted optimization, and limited-memory variants for large-scale problems. He conducted much of this work during his PhD and postdoctoral research at the University of Freiburg, Germany.

His publications have garnered over 55,000 citations (as per Google Scholar), reflecting significant impact in optimization methods for machine learning.

References:


Back

x-ai/grok-4.1-fast

Donate