soni69 commited on
Commit
f1d63b3
1 Parent(s): fd7b2fb

Trained on Season 1 - 4 all Episodes

Browse files
config.json CHANGED
@@ -8,6 +8,7 @@
8
  "bos_token_id": 50256,
9
  "embd_pdrop": 0.1,
10
  "eos_token_id": 50256,
 
11
  "initializer_range": 0.02,
12
  "layer_norm_epsilon": 1e-05,
13
  "model_type": "gpt2",
@@ -17,9 +18,7 @@
17
  "n_inner": null,
18
  "n_layer": 24,
19
  "n_positions": 1024,
20
- "reorder_and_upcast_attn": false,
21
  "resid_pdrop": 0.1,
22
- "scale_attn_by_inverse_layer_idx": false,
23
  "scale_attn_weights": true,
24
  "summary_activation": null,
25
  "summary_first_dropout": 0.1,
@@ -31,8 +30,7 @@
31
  "max_length": 1000
32
  }
33
  },
34
- "torch_dtype": "float32",
35
- "transformers_version": "4.18.0",
36
  "use_cache": true,
37
  "vocab_size": 50257
38
  }
8
  "bos_token_id": 50256,
9
  "embd_pdrop": 0.1,
10
  "eos_token_id": 50256,
11
+ "gradient_checkpointing": false,
12
  "initializer_range": 0.02,
13
  "layer_norm_epsilon": 1e-05,
14
  "model_type": "gpt2",
18
  "n_inner": null,
19
  "n_layer": 24,
20
  "n_positions": 1024,
 
21
  "resid_pdrop": 0.1,
 
22
  "scale_attn_weights": true,
23
  "summary_activation": null,
24
  "summary_first_dropout": 0.1,
30
  "max_length": 1000
31
  }
32
  },
33
+ "transformers_version": "4.7.0",
 
34
  "use_cache": true,
35
  "vocab_size": 50257
36
  }
eval_results.txt CHANGED
@@ -1 +1 @@
1
- perplexity = tensor(2.8108)
1
+ perplexity = tensor(1.5919)
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c531ac2cce54026125d98dc60036f219c29725bd4d83b23b8ffc1e3faf77828
3
- size 1444566873
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1d08a3c77cd04e30be46cbbf313cd4d27fa134a4a74ebc8e1f99bab8a6cbcd9
3
+ size 1444581337
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-medium", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
1
+ {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-medium", "errors": "replace"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bea3979fde740e88a6c90823ec7d9320181e1122214dea2baeb49d4c9372d3b9
3
  size 1327
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a474f823bcde364e5c1acebbdef2be91709e81177d1a6fcfee0f7d132857831c
3
  size 1327