Perplexity: Difference between revisions

From AI Wiki
No edit summary
No edit summary
Line 9: Line 9:
==Perplexity in language models==
==Perplexity in language models==
{{see also|language models}}
{{see also|language models}}
Perplexity is a metric used to evaluate the performance of a [[language model]]. It is defined as the inverse probability of the test set normalized by the number of words, or as the average number of bits needed to encode one word using [[cross-entropy]]. Perplexity can be interpreted as the weighted branching factor, with a higher perplexity indicating more confusion in the model's prediction of the next word.
Perplexity is an important measurement for determining how good a [[language model]] is. Essentially, it quantifies the quality of the model's predictions by evaluating the inverse probability of the [[test set]], normalized by the number of words, or by calculating the average number of bits required to encode a single word through [[cross-entropy]]. Perplexity can be perceived as the [[weighted branching factor]], and a high perplexity score represents a higher degree of confusion in the model's next-word predictions, while a low perplexity score implies greater confidence in the model's output.

Revision as of 22:49, 31 January 2023

In order to not have articles being flagged as AI generated content. The text should have high perplexity.

Perplexity measures the amount of information contained in a text. Despite the high redundancy of natural language, which allows us to communicate effectively even in noisy environments, some information can still be lost without compromising overall understanding. However, external factors such as alliteration, which adds an additional layer of meaning, can contribute to high perplexity that AI may not be able to recognize.

AI, based on probabilistic models, struggles with text that contains high variability in terms of word specificity, such as legal documents that require precise use of specific keywords and phrases.

Generative AI for text operates by predicting the most likely word to follow a given word. In a text with perfect perplexity, the AI would always make incorrect predictions.

Perplexity in language models

See also: language models

Perplexity is an important measurement for determining how good a language model is. Essentially, it quantifies the quality of the model's predictions by evaluating the inverse probability of the test set, normalized by the number of words, or by calculating the average number of bits required to encode a single word through cross-entropy. Perplexity can be perceived as the weighted branching factor, and a high perplexity score represents a higher degree of confusion in the model's next-word predictions, while a low perplexity score implies greater confidence in the model's output.