Mark As Completed Discussion

Regularization & Generalization

  • Overfitting: model learns noise; low training loss, high validation loss.
  • Underfitting: model too simple; high training and validation loss.
  • Regularization: techniques to improve generalization:
Regularization & Generalization
  • L2 (weight decay): penalize large weights.
  • Early stopping: stop when validation loss worsens.
  • Dropout: randomly drop units during training (simulated in code by masking).
  • Data augmentation: alter inputs (flips/crops/noise) to create variety.