BounharAbdelaziz commited on
Commit
711e129
1 Parent(s): 1ca7c48

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -18
README.md CHANGED
@@ -23,23 +23,9 @@ It achieves the following results on the evaluation set:
23
  - Bleu: 8.3292
24
  - Gen Len: 34.4959
25
 
26
- The finetuning was conducted using a A**100-40GB** and took **23 hours**.
27
-
28
-
29
- ## Training hyperparameters
30
-
31
- The following hyperparameters were used during training:
32
- - learning_rate: 3e-05
33
- - train_batch_size: 22
34
- - eval_batch_size: 22
35
- - seed: 42
36
- - gradient_accumulation_steps: 4
37
- - total_train_batch_size: 88
38
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
39
- - lr_scheduler_type: linear
40
- - lr_scheduler_warmup_ratio: 0.03
41
- - num_epochs: 40
42
 
 
43
 
44
  ## Usage
45
 
@@ -73,9 +59,9 @@ print("Translation:", output_text)
73
 
74
  Let's see an example of transliterating Moroccan Darija Arabizi to Arabic:
75
 
76
- **Input**: "Hello my friend, how's life in Morocco"
77
 
78
- **Output**: "مرحبا يا صاحبي, كيفاش الحياة فالمغرب"
79
 
80
  ## Limiations
81
 
@@ -88,6 +74,20 @@ We're continuously striving to improve our model's performance and usability and
88
  If you have any feedback, suggestions, or encounter any issues, please don't hesitate to reach out to us.
89
 
90
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
91
  ## Training results
92
 
93
  | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
 
23
  - Bleu: 8.3292
24
  - Gen Len: 34.4959
25
 
26
+ The finetuning was conducted using a **A100-40GB** and took **23 hours**.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
 
28
+ Try it out on our dedicated [Terjman-Large Space](https://huggingface.co/spaces/atlasia/Terjman-Large) 🤗
29
 
30
  ## Usage
31
 
 
59
 
60
  Let's see an example of transliterating Moroccan Darija Arabizi to Arabic:
61
 
62
+ **Input**: "Hi my friend, can you tell me a joke in moroccan darija? I'd be happy to hear that from you!"
63
 
64
+ **Output**: "مرحبا صديقي، يمكن لك تقول لي نكتة في داريجا المغربية؟ سأكون سعيدا بسماعها منك!"
65
 
66
  ## Limiations
67
 
 
74
  If you have any feedback, suggestions, or encounter any issues, please don't hesitate to reach out to us.
75
 
76
 
77
+ ## Training hyperparameters
78
+
79
+ The following hyperparameters were used during training:
80
+ - learning_rate: 3e-05
81
+ - train_batch_size: 22
82
+ - eval_batch_size: 22
83
+ - seed: 42
84
+ - gradient_accumulation_steps: 4
85
+ - total_train_batch_size: 88
86
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
87
+ - lr_scheduler_type: linear
88
+ - lr_scheduler_warmup_ratio: 0.03
89
+ - num_epochs: 40
90
+
91
  ## Training results
92
 
93
  | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |