Member-only story
Back-propagation Demystified [Part 1]
If you have been reading about deep learning, you must have definitely heard the term back-propagation at least once. In this article, I explain back-propagation and computational graphs.
I would recommend you read about gradient descent optimization here, before proceeding with this article. Now, let’s dive right into the explanation.
Introduction
We will first look at a few key steps involved in network training and terms like forward propagation.
Deep learning network training steps
The overall process for a deep learning network training may be summarized in the following steps.
- Data exploration and analysis [This is a huge step involving a lot of sub-steps]
- Choice of an architecture based on information obtained in step 1 and building it using a framework such as Chainer, PyTorch, TensorFlow etc.
- Choice of an appropriate cost function such as binary cross-entropy or mean squared error etc. depend on the task at hand.
- Choice of an optimizer such as Stochastic Gradient Descent, Adam, Adadelta etc.
- Training the model to minimize the selected loss function w.r.t. the network parameters using the training examples.