pochunhsu commited on
Commit
1b0edc1
1 Parent(s): f5564b6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -12,7 +12,7 @@ pipeline_tag: text-generation
12
 
13
  Version 1.0 / 20.Feb.2023
14
 
15
- This model is a joint collaboration between CKIP lab at Acedemia Sinica, MediaTek Research, and National Academy for Educational Research.
16
 
17
  ## Table of Contents
18
  1. [Model Details](#model-details)
@@ -36,7 +36,7 @@ BLOOM-zh is trained extendedly on larger amounts of Traditional Chinese text dat
36
  <details>
37
  <summary>Click to expand</summary> <br/>
38
 
39
- **Developed by:** MediaTek Research ([website](https://www.mtkresearch.com/))
40
 
41
  **Model Type:** Transformer-based Language Model
42
 
@@ -44,9 +44,9 @@ BLOOM-zh is trained extendedly on larger amounts of Traditional Chinese text dat
44
 
45
  **Languages:** Multiple; see [training data](#training-data)
46
 
47
- **License:** MEDIATEK RESEARCH License ([link](https://huggingface.co/MediaTek-Research/bloom-1b1-zh/blob/main/LICENSE_MR.md)) and RAIL License v1.0 ([link](https://huggingface.co/spaces/bigscience/license))
48
 
49
- **Release Date Estimate:** Tuesday, 14.February.2023
50
 
51
  **Send Questions to:** info@mtkresearch.com
52
 
@@ -157,7 +157,7 @@ Please refer to [Model card](https://huggingface.co/bigscience/bloom-1b1#uses).
157
  <details>
158
  <summary>Click to expand</summary><br/>
159
 
160
- We trained the 1B1 parameter model on a total of 6 Billion tokens mainly crawled from the internet and provided from National Academy for Educational Research, 75% of the training data is Traditional Chinese, 25% is English.
161
  Details are provided in the [paper](https://arxiv.org/).
162
 
163
  </details>
 
12
 
13
  Version 1.0 / 20.Feb.2023
14
 
15
+ This model is a joint collaboration between CKIP lab at Acedemia Sinica ([website](https://ckip.iis.sinica.edu.tw/)), MediaTek Research ([website](https://www.mtkresearch.com/)), and National Academy for Educational Research ([website](https://www.naer.edu.tw/)).
16
 
17
  ## Table of Contents
18
  1. [Model Details](#model-details)
 
36
  <details>
37
  <summary>Click to expand</summary> <br/>
38
 
39
+ **Developed by:** MediaTek Research
40
 
41
  **Model Type:** Transformer-based Language Model
42
 
 
44
 
45
  **Languages:** Multiple; see [training data](#training-data)
46
 
47
+ **License:** MEDIATEK RESEARCH License ([link](https://huggingface.co/ckip-joint/bloom-1b1-zh/blob/main/LICENSE_MR.md)) and RAIL License v1.0 ([link](https://huggingface.co/spaces/bigscience/license))
48
 
49
+ **Release Date Estimate:** Wednesday, 22.February.2023
50
 
51
  **Send Questions to:** info@mtkresearch.com
52
 
 
157
  <details>
158
  <summary>Click to expand</summary><br/>
159
 
160
+ We trained the 1B1 parameter model on a total of 6 Billion tokens mainly crawled from the internet and provided from National Academy for Educational Research. 75% of the training data is Traditional Chinese, 25% is English.
161
  Details are provided in the [paper](https://arxiv.org/).
162
 
163
  </details>