Spaces:
Running
Running
Update Space (evaluate main: 05209ece)
Browse files
README.md
CHANGED
|
@@ -10,6 +10,14 @@ pinned: false
|
|
| 10 |
tags:
|
| 11 |
- evaluate
|
| 12 |
- metric
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
---
|
| 14 |
|
| 15 |
# Metric Card for BLEU
|
|
|
|
| 10 |
tags:
|
| 11 |
- evaluate
|
| 12 |
- metric
|
| 13 |
+
description: >-
|
| 14 |
+
BLEU (Bilingual Evaluation Understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another.
|
| 15 |
+
Quality is considered to be the correspondence between a machine's output and that of a human: "the closer a machine translation is to a professional human translation, the better it is"
|
| 16 |
+
– this is the central idea behind BLEU. BLEU was one of the first metrics to claim a high correlation with human judgements of quality, and remains one of the most popular automated and inexpensive metrics.
|
| 17 |
+
|
| 18 |
+
Scores are calculated for individual translated segments—generally sentences—by comparing them with a set of good quality reference translations.
|
| 19 |
+
Those scores are then averaged over the whole corpus to reach an estimate of the translation's overall quality.
|
| 20 |
+
Neither intelligibility nor grammatical correctness are not taken into account.
|
| 21 |
---
|
| 22 |
|
| 23 |
# Metric Card for BLEU
|