Interface administrators, Administrators (Semantic MediaWiki), Curators (Semantic MediaWiki), Editors (Semantic MediaWiki), Suppressors, Administrators
7,785
edits
No edit summary |
No edit summary |
||
Line 1: | Line 1: | ||
{{see also|Machine learning terms}} | {{see also|Machine learning terms}} | ||
==Introduction== | ==Introduction== | ||
[[ | In [[machine learning]], [[training loss]] refers to a [[metric]] of how well a [[model]] fits [[training data]]. It is the model's [[loss]] during a particular [[training]] [[iteration]]. The aim of [[training]] a [[machine learning model]] is to find [[parameters]] that produce the optimal fit with given information. To evaluate how well it does this, we use something called a [[loss function]] - also known as [[cost function]] or [[error function]]. | ||
The training loss is the outcome of evaluating a loss function on [[training set]]. It measures the difference between predicted values and actual values ([[labels]]) for that set. The aim of training is to minimize this loss so that the model can make accurate predictions on new, unseen data with confidence. | |||
==Types of Loss Functions== | ==Types of Loss Functions== | ||
Machine learning employs a variety of loss | Machine learning employs a variety of [[loss function]]s, depending on the problem being solved and the model being employed. Some commonly employed loss functions include: | ||
#Mean Squared Error (MSE) Loss: This loss function is the most commonly employed for regression problems. It measures the average squared difference between predicted output and actual output. | |||
#Binary Cross-Entropy Loss: Used in binary classification problems where the objective is to accurately predict one of two possible classes, this statistic measures the difference between predicted probability of a positive class and actual binary label. | |||
#Categorical Cross-Entropy Loss: Used in multiclass classification problems to predict one of several classes, this statistic measures the difference between a predicted probability distribution and an actual one-hot encoded class label. | |||
#Softmax Cross-Entropy Loss: This approach is used for multiclass classification problems with mutually exclusive classes. It calculates the categorical cross-entropy loss for each class and then takes its average across all classes. | |||
==How Training Loss is Used== | ==How Training Loss is Used== |