Data Science 365

Bring data into actionable insights.

Member-only story

Learning Rate Schedules and Decay in Keras Optimizers

5 min readOct 5, 2022

--

Photo by Emma Louise Comerford on Unsplash

Importance of dynamic learning rates

The learning rate is one of the most important neural network hyperparameters.

In a previous post, I discussed a systematic approach to choosing the optimal learning rate for neural networks. There, we focused on fixed learning rates.

When using a fixed learning rate, we change the learning rate value only after training!

The most effective way of using the learning rate is to decrease its value during training. This type of learning rate is known as a dynamic learning rate whose value is decreased over time!

A dynamic learning rate will allow the model to properly converge at the optimal point where the error is minimum. Sometimes, a static learning rate will miss that point even if with its best value! A dynamic learning rate will also reduce the time taken to converge for the model.

Main approaches to decrease the learning rate during training

There are mainly two approaches to decrease the learning rate during training.

  • Using optimizers that have adaptive

--

--

Data Science 365
Data Science 365
Rukshan Pramoditha
Rukshan Pramoditha

Written by Rukshan Pramoditha

3,000,000+ Views | BSc in Stats (University of Colombo, Sri Lanka) | Top 50 Data Science, AI/ML Technical Writer on Medium

No responses yet