Regularization_for_Neural_Networks_1718966083
Regularization_for_Neural_Networks_1718966083
Shivang Kainthola
Regularization in Neural Networks
2) Co-Adaptation
3) Large Weights
2) Dropout
3) L1 Regularization
4) L2 Regularization
1) BATCH NORMALIZATION
Internal covariate shift can slow down training and lead to instability.
https://kwokanthony.medium.com/batch-normalization-in-neural-network-simply-explained-115fe281f4cd
Solution : Batch Normalization
https://medium.com/@abheerchrome/batch-normalization-explained-1e78f7eb1e8a
⟶ This dropped out fraction of neurons does not take part in forward
pass (activation and gradient computation) or backpropagation (gradient
updates) during training.
||w||^2 represents the squared L2 norm of the weights (sum of squares of all elements in the weight vector)
||w||_1 represents the L1 norm of the weights, which is the sum of the absolute values of all elements in the weight vector
(w)