L2 (Ridge) regularization is a popular form of regularizer which computes the squared sum of the model parameters, promoting the optimization procedure to find models with smaller absolute parameters. It is typically used as an extra term in the loss function, to reduce the magnitude of the weights of the model. This can work to prevent overfitting.
\( \theta \) | This is the symbol we use for model weights/parameters. |
\( \mathbf{W} \) | This symbol represents the matrix containing the weights and biases of a layer in a neural network. |
\( \textup{reg} \) | This is the symbol used for representing a regularization function. |