Concept
Batch Normalization 0
Batch Normalization is a technique to improve the training of deep neural networks by normalizing the inputs to each layer, which helps in reducing internal covariate shift and accelerates convergence. It allows for higher learning rates, reduces sensitivity to initialization, and can act as a form of regularization to reduce overfitting.
Relevant Degrees