Concept
Tangent Propagation Algorithm
- Similar to tangent distance algorithm
- Closely related to dataset augmentation, both require that the model be invariant to certain specified directions of change in the input. Dataset augmentation is the non-infinitesimal version of tangent propagation
- Trains a neural net classifier with extra penalty to make each output of the neural net locally invariant to known factors of variation
- Factors correspond to movement along the manifold near which examples of the same class concentrate.
- Local invariance achieved by requiring to be orthogonal to known manifold tangent vectors at
- Equivalently, the directional derivative of at in the directions be small by adding a regularization penalty , defined as: , which can be scaled by a hyperparameter, and for most neural networks, we would need to sum over many outputs
- Tangent vectors are derived a priori, usually from knowledge of the effect of transformations
- Has been used for supervised learning and reinforcement learning
- User encodes prior knowledge of task by specifying a set of transformations that should not alter the output, and analytically regularizes the model to resist perturbation in the directions corresponding to the specified transformation
- Only regularizes the model to resist infinitesimal perturbation, and poses difficulties for models based on rectified linear units
- Related to double backprop and adversarial training, both of which require that the model should be invariant to all directions of change in the input as long as the change is small
- Double backprop regularizes the Jacobian to be small
- Adversarial training finds inputs near the original inputs and trains the model to produce the same output on these as on the original inputs
- Adversarial training is the non-infinitesimal version of double backprop
0
1
Updated 2021-06-24
Tags
Data Science
Related
Data Augmentation in Deep Learning
Early Stopping in Deep Learning
Dropout Regularization in Deep Learning
Which of these techniques are useful for reducing variance (reducing overfitting)?
ElasticNet Regression
If your Neural Network model seems to have high variance, what of the following would be promising things to try?
Regularization in ML and DL
Bagging in Deep Learning
Dropout in Deep Learning
Normalization of Data
Tangent Distance Algorithm
Tangent Propagation Algorithm
Manifold Tangent Classifier
Boosting in Deep Learning
Appropriate Regularization/ Representation
Weight Decay
L1 Regularization