Relationship Between Dataset Size and Model Complexity
The amount of available training data dictates the appropriate level of model complexity. For a fixed task and data distribution, model complexity should not increase more rapidly than the dataset size. With fewer training samples, a model is highly susceptible to overfitting, making simpler models difficult to beat. However, as dataset size increases, generalization error typically decreases, allowing for the successful training of more complex architectures like deep neural networks.
0
1
Tags
D2L
Dive into Deep Learning @ D2L
Related
Overfitting a supervised statistical model
Training Error and Test Error
Generalizability of a supervised statistical model
Underfitting a supervised statistical model
Measuring Model Complexity: Rademacher complexity
Bias of Supervised Models in Statistical Learning
Variance of Supervised Models in Statistical Learning
Falsifiability of Machine Learning Models
Notions of Model Complexity
Relationship Between Dataset Size and Model Complexity