Interface administrators, Administrators (Semantic MediaWiki), Curators (Semantic MediaWiki), Editors (Semantic MediaWiki), Suppressors, Administrators
7,785
edits
No edit summary |
No edit summary |
||
Line 18: | Line 18: | ||
==Overfitting and Underfitting== | ==Overfitting and Underfitting== | ||
Overfitting occurs when a model is too complex and fits the training data | [[Overfitting]] occurs when a model is too complex and fits the [[training data]] too well yet fails to generalize well on unseen data. This leads to low training losses but high [[validation loss]]es. [[Regularization]] techniques like [[L1 regularization|L1]] or [[L2 regularization]] can be used as penalties for complex models in order to minimize overfitting. | ||
Underfitting occurs when a model is too simple and does not fit the training data well, leading to high training loss and validation loss. To avoid this issue, more complex models can be utilized or more data can be collected for training the model. | [[Underfitting]] occurs when a model is too simple and does not fit the training data well, leading to high training loss and validation loss. To avoid this issue, more complex models can be utilized or more data can be collected for training the model. | ||
==Explain Like I'm 5 (ELI5)== | ==Explain Like I'm 5 (ELI5)== | ||
Training loss is like a | Training loss is like a score that tells us how well our model is doing at guessing what we're teaching it. When teaching a model, we give it examples and ask it to guess the answer. The training loss is an indicator of how close these guesses are to reality - the lower the loss, the better! Just like when playing video games and getting high scores means you're doing well! | ||
The training loss is | |||
[[Category:Terms]] [[Category:Machine learning terms]] | [[Category:Terms]] [[Category:Machine learning terms]] |