benlipkin commited on
Commit
534537b
1 Parent(s): 28db0d7

args and setup.

Browse files
Files changed (6) hide show
  1. args.json +18 -0
  2. merges.txt +0 -0
  3. special_tokens_map.json +5 -0
  4. tokenizer.json +0 -0
  5. tokenizer_config.json +10 -0
  6. vocab.json +0 -0
args.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "gpt2",
3
+ "dataset": "wikitext",
4
+ "subset": "wikitext-103-v1",
5
+ "output_dir": "output",
6
+ "num_epochs": 20,
7
+ "num_tokens": 100000000,
8
+ "window_size": 256,
9
+ "batch_size": 64,
10
+ "learning_rate": 1e-05,
11
+ "warmup_steps": 3000,
12
+ "scheduler": "cosine",
13
+ "weight_decay": 0.1,
14
+ "random_seed": 42,
15
+ "eval_steps": 1000,
16
+ "patience": 5,
17
+ "id": "26e50955232e9b5c"
18
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|endoftext|>",
4
+ "unk_token": "<|endoftext|>"
5
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<|endoftext|>",
4
+ "eos_token": "<|endoftext|>",
5
+ "model_max_length": 1024,
6
+ "name_or_path": "gpt2",
7
+ "special_tokens_map_file": null,
8
+ "tokenizer_class": "GPT2Tokenizer",
9
+ "unk_token": "<|endoftext|>"
10
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff