BLEU (Bilingual Evaluation Understudy): Revision history

Diff selection: Mark the radio buttons of the revisions to compare and hit enter or the button at the bottom.
Legend: (cur) = difference with latest revision, (prev) = difference with preceding revision, m = minor edit.

18 March 2023

  • curprev 13:1113:11, 18 March 2023Walle talk contribs 3,877 bytes +3,877 Created page with "{{see also|Machine learning terms}} ==Introduction== The '''Bilingual Evaluation Understudy''' ('''BLEU''') is an automatic evaluation metric used in the field of Natural Language Processing (NLP) to measure the quality of machine-generated translations. Developed by IBM Research in 2002, it compares translations generated by a machine with a set of human-generated reference translations. BLEU scores are widely used in the evaluation of machine translation system..."