<aside> 💡

This was not in the lecture, I just investigated more by myself to get a better understanding of Normalization. Hope this material helps.

</aside>

What is Normalization in the literature of Deep Learning?

In Deep Learning, Normalization(정규화) is a technique used to standardize the data being fed into and through a neural network.

The main idea is to transform the features to be on a similar scale. This is done by adjusting the range of the data, often to have a mean of zero and a standard deviation of one.

This process is important because it helps to stabilize and speed up the training of deep neural networks. By ensuring that the inputs to each layer have a consistent distribution, normalization can prevent issues like vanishing or exploding gradients, which can hinder the learning process. Ultimately, this leads to more reliable and faster convergence to a good solution.

When do we Normalize

What are we exactly normalizing?

It depends on the type. But generally, we are trying to normalize the input of each layer before implementing activation function to it during the forward pass.

image.png

Types of Normalization Layer

image.png

Batch Norm