Safetensors
Chinese
Catalan
m2m_100
xixianliao commited on
Commit
a77e850
1 Parent(s): 5af834f

Update files

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -35,7 +35,7 @@ were parallel synthetic data created using the
35
 
36
  Following the fine-tuning phase, Contrastive Preference Optimization (CPO) was applied to further refine the model's outputs. CPO training involved pairs of "chosen" and "rejected" translations for a total of 4,006 sentences. These sentences were sourced from the Flores development set (997 sentences), the Flores devtest set (1,012 sentences), and the NTREX set (1,997 sentences).
37
 
38
- The model was evaluated on the Projecte Aina's Catalan-Chinese evaluation dataset, achieving results comparable to those of Google Translate.
39
 
40
  ## Intended uses and limitations
41
 
@@ -157,7 +157,7 @@ Following fine-tuning on the M2M100 1.2B model, Contrastive Preference Optimizat
157
 
158
  ### Variable and metrics
159
 
160
- Below are the evaluation results on the Projecte Aina's Catalan-Chinese test set, compared to Google Translate for the CA-ZH direction. The evaluation was conducted using [`tower-eval`](https://github.com/deep-spin/tower-eval) following the standard setting (beam search with beam size 5, limiting the translation length to 200 tokens). We report the following metrics:
161
 
162
  - BLEU: Sacrebleu implementation, version:2.4.0
163
  - ChrF: Sacrebleu implementation.
 
35
 
36
  Following the fine-tuning phase, Contrastive Preference Optimization (CPO) was applied to further refine the model's outputs. CPO training involved pairs of "chosen" and "rejected" translations for a total of 4,006 sentences. These sentences were sourced from the Flores development set (997 sentences), the Flores devtest set (1,012 sentences), and the NTREX set (1,997 sentences).
37
 
38
+ The model was evaluated on the Projecte Aina's Catalan-Chinese evaluation dataset (unpublished), achieving results comparable to those of Google Translate.
39
 
40
  ## Intended uses and limitations
41
 
 
157
 
158
  ### Variable and metrics
159
 
160
+ Below are the evaluation results on the Projecte Aina's Catalan-Chinese test set (unpublished), compared to Google Translate for the CA-ZH direction. The evaluation was conducted using [`tower-eval`](https://github.com/deep-spin/tower-eval) following the standard setting (beam search with beam size 5, limiting the translation length to 200 tokens). We report the following metrics:
161
 
162
  - BLEU: Sacrebleu implementation, version:2.4.0
163
  - ChrF: Sacrebleu implementation.