lvwerra HF staff commited on
Commit
c23e03b
1 Parent(s): 8103015

Update Space (evaluate main: 05209ece)

Browse files
Files changed (1) hide show
  1. README.md +8 -0
README.md CHANGED
@@ -10,6 +10,14 @@ pinned: false
10
  tags:
11
  - evaluate
12
  - metric
 
 
 
 
 
 
 
 
13
  ---
14
 
15
  # Metric Card for BLEU
10
  tags:
11
  - evaluate
12
  - metric
13
+ description: >-
14
+ BLEU (Bilingual Evaluation Understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another.
15
+ Quality is considered to be the correspondence between a machine's output and that of a human: "the closer a machine translation is to a professional human translation, the better it is"
16
+ – this is the central idea behind BLEU. BLEU was one of the first metrics to claim a high correlation with human judgements of quality, and remains one of the most popular automated and inexpensive metrics.
17
+
18
+ Scores are calculated for individual translated segments—generally sentences—by comparing them with a set of good quality reference translations.
19
+ Those scores are then averaged over the whole corpus to reach an estimate of the translation's overall quality.
20
+ Neither intelligibility nor grammatical correctness are not taken into account.
21
  ---
22
 
23
  # Metric Card for BLEU