Jump to content

Accuracy: Difference between revisions

m
No edit summary
 
Line 32: Line 32:
Accuracy is a measure of how good a computer program is at distinguishing things. For instance, if we want it to distinguish between pictures of cats and dogs, accuracy would measure how many pictures it gets right out of all those it looks at. The higher the accuracy, the better equipped your program will be at distinguishing between them.
Accuracy is a measure of how good a computer program is at distinguishing things. For instance, if we want it to distinguish between pictures of cats and dogs, accuracy would measure how many pictures it gets right out of all those it looks at. The higher the accuracy, the better equipped your program will be at distinguishing between them.


[[Category:Terms]] [[Category:Machine learning terms]]
[[Category:Terms]] [[Category:Machine learning terms]] [[Category:not updated]]