xixianliao
commited on
Commit
•
a77e850
1
Parent(s):
5af834f
Update files
Browse files
README.md
CHANGED
@@ -35,7 +35,7 @@ were parallel synthetic data created using the
|
|
35 |
|
36 |
Following the fine-tuning phase, Contrastive Preference Optimization (CPO) was applied to further refine the model's outputs. CPO training involved pairs of "chosen" and "rejected" translations for a total of 4,006 sentences. These sentences were sourced from the Flores development set (997 sentences), the Flores devtest set (1,012 sentences), and the NTREX set (1,997 sentences).
|
37 |
|
38 |
-
The model was evaluated on the Projecte Aina's Catalan-Chinese evaluation dataset, achieving results comparable to those of Google Translate.
|
39 |
|
40 |
## Intended uses and limitations
|
41 |
|
@@ -157,7 +157,7 @@ Following fine-tuning on the M2M100 1.2B model, Contrastive Preference Optimizat
|
|
157 |
|
158 |
### Variable and metrics
|
159 |
|
160 |
-
Below are the evaluation results on the Projecte Aina's Catalan-Chinese test set, compared to Google Translate for the CA-ZH direction. The evaluation was conducted using [`tower-eval`](https://github.com/deep-spin/tower-eval) following the standard setting (beam search with beam size 5, limiting the translation length to 200 tokens). We report the following metrics:
|
161 |
|
162 |
- BLEU: Sacrebleu implementation, version:2.4.0
|
163 |
- ChrF: Sacrebleu implementation.
|
|
|
35 |
|
36 |
Following the fine-tuning phase, Contrastive Preference Optimization (CPO) was applied to further refine the model's outputs. CPO training involved pairs of "chosen" and "rejected" translations for a total of 4,006 sentences. These sentences were sourced from the Flores development set (997 sentences), the Flores devtest set (1,012 sentences), and the NTREX set (1,997 sentences).
|
37 |
|
38 |
+
The model was evaluated on the Projecte Aina's Catalan-Chinese evaluation dataset (unpublished), achieving results comparable to those of Google Translate.
|
39 |
|
40 |
## Intended uses and limitations
|
41 |
|
|
|
157 |
|
158 |
### Variable and metrics
|
159 |
|
160 |
+
Below are the evaluation results on the Projecte Aina's Catalan-Chinese test set (unpublished), compared to Google Translate for the CA-ZH direction. The evaluation was conducted using [`tower-eval`](https://github.com/deep-spin/tower-eval) following the standard setting (beam search with beam size 5, limiting the translation length to 200 tokens). We report the following metrics:
|
161 |
|
162 |
- BLEU: Sacrebleu implementation, version:2.4.0
|
163 |
- ChrF: Sacrebleu implementation.
|