miguelvictor commited on
Commit
45b5d97
1 Parent(s): 66948ac

added tokenizer config

Browse files
Files changed (4) hide show
  1. config.json +3 -9
  2. merges.txt +0 -0
  3. tokenizer.json +0 -0
  4. vocab.json +0 -0
config.json CHANGED
@@ -1,15 +1,11 @@
1
  {
2
- "_name_or_path": "gpt2-large",
3
  "_num_labels": 1,
4
  "activation_function": "gelu_new",
5
- "architectures": [
6
- "GPT2LMHeadModel"
7
- ],
8
  "attn_pdrop": 0.1,
9
  "bos_token_id": 50256,
10
  "embd_pdrop": 0.1,
11
  "eos_token_id": 50256,
12
- "gradient_checkpointing": true,
13
  "id2label": {
14
  "0": "LABEL_0"
15
  },
@@ -17,12 +13,11 @@
17
  "label2id": {
18
  "LABEL_0": 0
19
  },
20
- "layer_norm_epsilon": 1e-05,
21
  "model_type": "gpt2",
22
  "n_ctx": 1024,
23
  "n_embd": 1280,
24
  "n_head": 20,
25
- "n_inner": null,
26
  "n_layer": 36,
27
  "n_positions": 1024,
28
  "resid_pdrop": 0.1,
@@ -33,11 +28,10 @@
33
  "summary_use_proj": true,
34
  "task_specific_params": {
35
  "text-generation": {
36
- "do_sample": true,
37
  "max_length": 50
38
  }
39
  },
40
  "transformers_version": "4.5.1",
41
- "use_cache": false,
42
  "vocab_size": 50257
43
  }
1
  {
 
2
  "_num_labels": 1,
3
  "activation_function": "gelu_new",
4
+ "architectures": ["GPT2LMHeadModel"],
 
 
5
  "attn_pdrop": 0.1,
6
  "bos_token_id": 50256,
7
  "embd_pdrop": 0.1,
8
  "eos_token_id": 50256,
 
9
  "id2label": {
10
  "0": "LABEL_0"
11
  },
13
  "label2id": {
14
  "LABEL_0": 0
15
  },
16
+ "layer_norm_epsilon": 1e-5,
17
  "model_type": "gpt2",
18
  "n_ctx": 1024,
19
  "n_embd": 1280,
20
  "n_head": 20,
 
21
  "n_layer": 36,
22
  "n_positions": 1024,
23
  "resid_pdrop": 0.1,
28
  "summary_use_proj": true,
29
  "task_specific_params": {
30
  "text-generation": {
31
+ "do_sample": false,
32
  "max_length": 50
33
  }
34
  },
35
  "transformers_version": "4.5.1",
 
36
  "vocab_size": 50257
37
  }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
vocab.json ADDED
The diff for this file is too large to render. See raw diff