pellto commited on
Commit
744ecfd
โ€ข
1 Parent(s): a63658b

upload files about models

Browse files
Files changed (6) hide show
  1. README.md +4 -0
  2. config.json +38 -0
  3. merges.txt +0 -0
  4. pytorch_model.bin +3 -0
  5. tokenizer.json +0 -0
  6. vocab.json +0 -0
README.md ADDED
@@ -0,0 +1,4 @@
 
 
 
 
1
+ ํ•ด๋‹น ๋ชจ๋ธ์€ [ํ•ด๋‹น ์‚ฌ์ดํŠธ](https://huggingface.co/gpt2-medium)์—์„œ ๊ฐ€์ ธ์˜จ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.
2
+
3
+
4
+ ํ•ด๋‹น ๋ชจ๋ธ์€ [Teachable NLP](https://ainize.ai/teachable-nlp) ์„œ๋น„์Šค์—์„œ ์‚ฌ์šฉ๋ฉ๋‹ˆ๋‹ค.
config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2Model"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "layer_norm_epsilon": 1e-05,
12
+ "model_type": "gpt2",
13
+ "n_ctx": 1024,
14
+ "n_embd": 1280,
15
+ "n_head": 20,
16
+ "n_inner": null,
17
+ "n_layer": 36,
18
+ "n_positions": 1024,
19
+ "reorder_and_upcast_attn": false,
20
+ "resid_pdrop": 0.1,
21
+ "scale_attn_by_inverse_layer_idx": false,
22
+ "scale_attn_weights": true,
23
+ "summary_activation": null,
24
+ "summary_first_dropout": 0.1,
25
+ "summary_proj_to_labels": true,
26
+ "summary_type": "cls_index",
27
+ "summary_use_proj": true,
28
+ "task_specific_params": {
29
+ "text-generation": {
30
+ "do_sample": true,
31
+ "max_length": 50
32
+ }
33
+ },
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.12.3",
36
+ "use_cache": true,
37
+ "vocab_size": 50257
38
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b5474d9a239e3025f710322e4ea01b690b4a75ccdb7de3d962ac586d2ad62ec
3
+ size 3134034377
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
vocab.json ADDED
The diff for this file is too large to render. See raw diff