Appropriate Regularization/ Representation
The no free lunch theorem is proof that no single regularization is strategy is certifiably the best at every task, however that is not to say that every strategy is created equal. There may be no best general strategy, but there may be strategies that are often times more useful than others. That being the case, here is a list of some regularization assumptions/ strategies that have been shown to be generally useful:
- Smoothness
- Linearity
- Multiple explanatory factors
- Causal factors
- Depth/ hierarchical factors
- Shared factors across tasks
- Manifolds
- Natural clustering
- Temporal and spatial coherence
- Sparsity
- Simplicity of factor dependencies.
0
2
Tags
Data Science
Related
Data Augmentation in Deep Learning
Early Stopping in Deep Learning
Dropout Regularization in Deep Learning
Which of these techniques are useful for reducing variance (reducing overfitting)?
ElasticNet Regression
If your Neural Network model seems to have high variance, what of the following would be promising things to try?
Regularization in ML and DL
Bagging in Deep Learning
Dropout in Deep Learning
Normalization of Data
Tangent Distance Algorithm
Tangent Propagation Algorithm
Manifold Tangent Classifier
Boosting in Deep Learning
Appropriate Regularization/ Representation
Weight Decay
L1 Regularization