Spaces:
Running
Running
Update Space (evaluate main: 05209ece)
Browse files
README.md
CHANGED
@@ -10,6 +10,14 @@ pinned: false
|
|
10 |
tags:
|
11 |
- evaluate
|
12 |
- metric
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
---
|
14 |
|
15 |
# Metric Card for BLEU
|
|
|
10 |
tags:
|
11 |
- evaluate
|
12 |
- metric
|
13 |
+
description: >-
|
14 |
+
BLEU (Bilingual Evaluation Understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another.
|
15 |
+
Quality is considered to be the correspondence between a machine's output and that of a human: "the closer a machine translation is to a professional human translation, the better it is"
|
16 |
+
– this is the central idea behind BLEU. BLEU was one of the first metrics to claim a high correlation with human judgements of quality, and remains one of the most popular automated and inexpensive metrics.
|
17 |
+
|
18 |
+
Scores are calculated for individual translated segments—generally sentences—by comparing them with a set of good quality reference translations.
|
19 |
+
Those scores are then averaged over the whole corpus to reach an estimate of the translation's overall quality.
|
20 |
+
Neither intelligibility nor grammatical correctness are not taken into account.
|
21 |
---
|
22 |
|
23 |
# Metric Card for BLEU
|