vera-8 commited on
Commit
0e6ee2e
1 Parent(s): f4d80d8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -9
README.md CHANGED
@@ -1,13 +1,47 @@
1
  ---
2
- {}
 
 
 
 
 
 
 
 
 
 
3
  ---
4
- # TRAINING ARGS:
5
- {'per_device_train_batch_size': 16,
6
- 'warmup_ratio': 0.1,
7
- 'num_train_epochs': 1}
8
 
9
- # OPTIMIZER ARGS:
10
- {'lr': 0.001, 'eps': (1e-30, 0.001), 'clip_threshold': 1.0, 'decay_rate': 0.0, 'beta1': None, 'weight_decay': 0.0, 'scale_parameter': False, 'relative_step': False, 'warmup_init': False, 'differentiable': False}
11
 
12
- # DEVICE:
13
- NVIDIA A100 80GB PCIe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ datasets:
3
+ - DEplain/DEplain-APA-sent
4
+ language:
5
+ - de
6
+ metrics:
7
+ - sari
8
+ - bleu
9
+ - bertscore
10
+ library_name: transformers
11
+ base_model: google/mT5-large
12
+ pipeline_tag: text2text-generation
13
  ---
14
+ # Model Card for mT5-large-trimmed_deplain-apa
 
 
 
15
 
16
+ Finetuned mT5-Model for German sentence-level text-simplification.
 
17
 
18
+ ## Model Details
19
+
20
+ ### Model Description
21
+
22
+ - **Model type:** Encoder-Decoder-Transformer
23
+ - **Language(s) (NLP):** German
24
+ - **Finetuned from model:** google/mT5-small
25
+ - **Task**: Text-Simplification
26
+
27
+ ## Training Details
28
+
29
+ ### Training Data
30
+
31
+ [DEplain/DEplain-APA-sent](https://huggingface.co/datasets/DEplain/DEplain-APA-sent)
32
+
33
+ ### Training Procedure
34
+
35
+ Parameter-efficient Fine-Tuning with LoRA
36
+
37
+ #### Training Hyperparameters
38
+ * Batch Size: 16
39
+ * Epochs: 1
40
+ * Learning Rate: 0,001
41
+ * Optimizer: Adafactor
42
+
43
+ #### LoRA Hyperparameters
44
+ * R: 32
45
+ * Alpha: 64
46
+ * Dropout:
47
+ * Target modules: all linear layers