Various Methods in Regularization

Regularization refers to controlling the capacity of neural network and prevent it from overfitting. For a better training of RNN, we seperate some part of training dataset to validation dataset. The validation set is used to watch the training process and prevent the network from underfitting or overfitting. Overfitting refers to the difference between training loss and validation loss.

  • The L1 and L2 regularization method add a regularization term to a loss function to penalize certain parameter configuration and prevent the coefficients from fitting so perfectly to the training data which leads to overfitting

  • Dropout - In general, dropout randomly omits a fraction of the connections between two layers of the network during training.Based on the drop out ratio, we randomly choose a set of neurons to deactivate and this helps in regularizing and it is analogous to feature selection in Machine learning paradigm

  • A dropout specifically tailored to RNNs is called RNNDrop

  • For additional regularization techniques please refer here

Last updated