lIlBrother commited on
Commit
59898c9
1 Parent(s): d526a62

Update: 최종 완료 모델에 대한 README 확정

Browse files
Files changed (1) hide show
  1. README.md +12 -11
README.md CHANGED
@@ -20,25 +20,25 @@ model-index:
20
  name: text2text-generation # Optional. Example: Speech Recognition
21
  metrics:
22
  - type: bleu # Required. Example: wer. Use metric id from https://hf.co/metrics
23
- value: 0.9529006548919251 # Required. Example: 20.90
24
  name: eval_bleu # Optional. Example: Test WER
25
- verified: true # Optional. If true, indicates that evaluation was generated by Hugging Face (vs. self-reported).
26
  - type: rouge1 # Required. Example: wer. Use metric id from https://hf.co/metrics
27
- value: 0.9693520563208838 # Required. Example: 20.90
28
  name: eval_rouge1 # Optional. Example: Test WER
29
- verified: true # Optional. If true, indicates that evaluation was generated by Hugging Face (vs. self-reported).
30
  - type: rouge2 # Required. Example: wer. Use metric id from https://hf.co/metrics
31
- value: 0.9444220599246154 # Required. Example: 20.90
32
  name: eval_rouge2 # Optional. Example: Test WER
33
- verified: true # Optional. If true, indicates that evaluation was generated by Hugging Face (vs. self-reported).
34
  - type: rougeL # Required. Example: wer. Use metric id from https://hf.co/metrics
35
- value: 0.9692485601662657 # Required. Example: 20.90
36
  name: eval_rougeL # Optional. Example: Test WER
37
- verified: true # Optional. If true, indicates that evaluation was generated by Hugging Face (vs. self-reported).
38
  - type: rougeLsum # Required. Example: wer. Use metric id from https://hf.co/metrics
39
- value: 0.9692422603343052 # Required. Example: 20.90
40
  name: eval_rougeLsum # Optional. Example: Test WER
41
- verified: true # Optional. If true, indicates that evaluation was generated by Hugging Face (vs. self-reported).
42
  ---
43
 
44
  # ko-TextNumbarT(TNT Model🧨): Try Korean Reading To Number(한글을 숫자로 바꾸는 모델)
@@ -81,7 +81,8 @@ Want see more detail follow this URL [KoGPT_num_converter](https://github.com/dd
81
 
82
  ## Evaluation
83
  Just using `evaluate-metric/bleu` and `evaluate-metric/rouge` in huggingface `evaluate` library <br />
84
- [Training wanDB URL](https://wandb.ai/bart_tadev/BartForConditionalGeneration/runs/1chrc03q?workspace=user-bart_tadev)
 
85
  ## How to Get Started With the Model
86
  ```python
87
  from transformers.pipelines import Text2TextGenerationPipeline
 
20
  name: text2text-generation # Optional. Example: Speech Recognition
21
  metrics:
22
  - type: bleu # Required. Example: wer. Use metric id from https://hf.co/metrics
23
+ value: 0.958234790096092 # Required. Example: 20.90
24
  name: eval_bleu # Optional. Example: Test WER
25
+ verified: false # Optional. If true, indicates that evaluation was generated by Hugging Face (vs. self-reported).
26
  - type: rouge1 # Required. Example: wer. Use metric id from https://hf.co/metrics
27
+ value: 0.9735361877162854 # Required. Example: 20.90
28
  name: eval_rouge1 # Optional. Example: Test WER
29
+ verified: false # Optional. If true, indicates that evaluation was generated by Hugging Face (vs. self-reported).
30
  - type: rouge2 # Required. Example: wer. Use metric id from https://hf.co/metrics
31
+ value: 0.9493975212378124 # Required. Example: 20.90
32
  name: eval_rouge2 # Optional. Example: Test WER
33
+ verified: false # Optional. If true, indicates that evaluation was generated by Hugging Face (vs. self-reported).
34
  - type: rougeL # Required. Example: wer. Use metric id from https://hf.co/metrics
35
+ value: 0.9734558938864928 # Required. Example: 20.90
36
  name: eval_rougeL # Optional. Example: Test WER
37
+ verified: false # Optional. If true, indicates that evaluation was generated by Hugging Face (vs. self-reported).
38
  - type: rougeLsum # Required. Example: wer. Use metric id from https://hf.co/metrics
39
+ value: 0.9734350757552404 # Required. Example: 20.90
40
  name: eval_rougeLsum # Optional. Example: Test WER
41
+ verified: false # Optional. If true, indicates that evaluation was generated by Hugging Face (vs. self-reported).
42
  ---
43
 
44
  # ko-TextNumbarT(TNT Model🧨): Try Korean Reading To Number(한글을 숫자로 바꾸는 모델)
 
81
 
82
  ## Evaluation
83
  Just using `evaluate-metric/bleu` and `evaluate-metric/rouge` in huggingface `evaluate` library <br />
84
+ [Training wanDB URL](https://wandb.ai/bart_tadev/BartForConditionalGeneration/runs/14hyusvf?workspace=user-bart_tadev)
85
+
86
  ## How to Get Started With the Model
87
  ```python
88
  from transformers.pipelines import Text2TextGenerationPipeline