IsekaiMeta commited on
Commit
3f3cac1
1 Parent(s): 5178eaa

Initial commit

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. checkpoint-10500/config.json +38 -0
  2. checkpoint-10500/merges.txt +0 -0
  3. checkpoint-10500/optimizer.pt +3 -0
  4. checkpoint-10500/pytorch_model.bin +3 -0
  5. checkpoint-10500/scheduler.pt +3 -0
  6. checkpoint-10500/special_tokens_map.json +1 -0
  7. checkpoint-10500/tokenizer.json +0 -0
  8. checkpoint-10500/tokenizer_config.json +1 -0
  9. checkpoint-10500/training_args.bin +3 -0
  10. checkpoint-10500/vocab.json +0 -0
  11. checkpoint-14000/config.json +38 -0
  12. checkpoint-14000/merges.txt +0 -0
  13. checkpoint-14000/optimizer.pt +3 -0
  14. checkpoint-14000/pytorch_model.bin +3 -0
  15. checkpoint-14000/scheduler.pt +3 -0
  16. checkpoint-14000/special_tokens_map.json +1 -0
  17. checkpoint-14000/tokenizer.json +0 -0
  18. checkpoint-14000/tokenizer_config.json +1 -0
  19. checkpoint-14000/training_args.bin +3 -0
  20. checkpoint-14000/vocab.json +0 -0
  21. checkpoint-17500/config.json +38 -0
  22. checkpoint-17500/merges.txt +0 -0
  23. checkpoint-17500/optimizer.pt +3 -0
  24. checkpoint-17500/pytorch_model.bin +3 -0
  25. checkpoint-17500/scheduler.pt +3 -0
  26. checkpoint-17500/special_tokens_map.json +1 -0
  27. checkpoint-17500/tokenizer.json +0 -0
  28. checkpoint-17500/tokenizer_config.json +1 -0
  29. checkpoint-17500/training_args.bin +3 -0
  30. checkpoint-17500/vocab.json +0 -0
  31. checkpoint-21000/config.json +38 -0
  32. checkpoint-21000/merges.txt +0 -0
  33. checkpoint-21000/optimizer.pt +3 -0
  34. checkpoint-21000/pytorch_model.bin +3 -0
  35. checkpoint-21000/scheduler.pt +3 -0
  36. checkpoint-21000/special_tokens_map.json +1 -0
  37. checkpoint-21000/tokenizer.json +0 -0
  38. checkpoint-21000/tokenizer_config.json +1 -0
  39. checkpoint-21000/training_args.bin +3 -0
  40. checkpoint-21000/vocab.json +0 -0
  41. checkpoint-3500/config.json +38 -0
  42. checkpoint-3500/merges.txt +0 -0
  43. checkpoint-3500/optimizer.pt +3 -0
  44. checkpoint-3500/pytorch_model.bin +3 -0
  45. checkpoint-3500/scheduler.pt +3 -0
  46. checkpoint-3500/special_tokens_map.json +1 -0
  47. checkpoint-3500/tokenizer.json +0 -0
  48. checkpoint-3500/tokenizer_config.json +1 -0
  49. checkpoint-3500/training_args.bin +3 -0
  50. checkpoint-3500/vocab.json +0 -0
checkpoint-10500/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/DialoGPT-medium",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 1024,
16
+ "n_head": 16,
17
+ "n_inner": null,
18
+ "n_layer": 24,
19
+ "n_positions": 1024,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "task_specific_params": {
30
+ "conversational": {
31
+ "max_length": 1000
32
+ }
33
+ },
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.18.0",
36
+ "use_cache": true,
37
+ "vocab_size": 50257
38
+ }
checkpoint-10500/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-10500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d886b073aba09d6b846b7fa680e1046617473c6b439ae0a6a6a8e2e1f8bf123
3
+ size 2838754801
checkpoint-10500/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f93f7dc757b7c5475178c465dec17cb75c81819d80ff8832fc202eae8da1bf3b
3
+ size 1444566873
checkpoint-10500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac893c0bfe001febecaa4ad43ee2c3d127f6c356a7a58c1b0b6466cc2b41cf68
3
+ size 623
checkpoint-10500/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
checkpoint-10500/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-10500/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-medium", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
checkpoint-10500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f22cfb810489daf1ffb0210f3c0e8b7d94e24a77aa9ce4e2b910cc6b1984e20
3
+ size 1327
checkpoint-10500/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-14000/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/DialoGPT-medium",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 1024,
16
+ "n_head": 16,
17
+ "n_inner": null,
18
+ "n_layer": 24,
19
+ "n_positions": 1024,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "task_specific_params": {
30
+ "conversational": {
31
+ "max_length": 1000
32
+ }
33
+ },
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.18.0",
36
+ "use_cache": true,
37
+ "vocab_size": 50257
38
+ }
checkpoint-14000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-14000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:628bf61556ecf8d39336738fadfa07db600d328b0775262da2d5b1abc15c9cb2
3
+ size 2838754801
checkpoint-14000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bd560e2a8022bc89fbdf1188ab41909e3786d8791b53c8de4a846f5990fa90c
3
+ size 1444566873
checkpoint-14000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fa2d0cea2d61645e2b6b0ffa3728d124dcf8f2001cbff901a492dc876e37797
3
+ size 623
checkpoint-14000/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
checkpoint-14000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-14000/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-medium", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
checkpoint-14000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f22cfb810489daf1ffb0210f3c0e8b7d94e24a77aa9ce4e2b910cc6b1984e20
3
+ size 1327
checkpoint-14000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-17500/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/DialoGPT-medium",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 1024,
16
+ "n_head": 16,
17
+ "n_inner": null,
18
+ "n_layer": 24,
19
+ "n_positions": 1024,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "task_specific_params": {
30
+ "conversational": {
31
+ "max_length": 1000
32
+ }
33
+ },
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.18.0",
36
+ "use_cache": true,
37
+ "vocab_size": 50257
38
+ }
checkpoint-17500/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-17500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d26b8a66b3c4247d1b4f03b4842c42669d83ae9a5995da79c7980c3dda0200a
3
+ size 2838754801
checkpoint-17500/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b1b8f28c7be05608cfcac136959daa729b347efd57ce74e6b182a3de9377cc0
3
+ size 1444566873
checkpoint-17500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce5a0c097e2a108a8ac9c20faa24b5dc765d86fe9ddf7ef8695b5f898236825e
3
+ size 623
checkpoint-17500/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
checkpoint-17500/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-17500/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-medium", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
checkpoint-17500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f22cfb810489daf1ffb0210f3c0e8b7d94e24a77aa9ce4e2b910cc6b1984e20
3
+ size 1327
checkpoint-17500/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-21000/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/DialoGPT-medium",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 1024,
16
+ "n_head": 16,
17
+ "n_inner": null,
18
+ "n_layer": 24,
19
+ "n_positions": 1024,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "task_specific_params": {
30
+ "conversational": {
31
+ "max_length": 1000
32
+ }
33
+ },
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.18.0",
36
+ "use_cache": true,
37
+ "vocab_size": 50257
38
+ }
checkpoint-21000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-21000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d2b09a15c2f77092dc344183b32057917c028ed72e34bce8ffebe954ed0c273
3
+ size 2838754801
checkpoint-21000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f24b9d57701503adade6ac23ddb9cbee88160da05413dec962b138af67419639
3
+ size 1444566873
checkpoint-21000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fedb0f54938debdb4ee3c2810ca806ca616b90b9522504469574a2767c64d7cb
3
+ size 623
checkpoint-21000/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
checkpoint-21000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-21000/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-medium", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
checkpoint-21000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f22cfb810489daf1ffb0210f3c0e8b7d94e24a77aa9ce4e2b910cc6b1984e20
3
+ size 1327
checkpoint-21000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-3500/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/DialoGPT-medium",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 1024,
16
+ "n_head": 16,
17
+ "n_inner": null,
18
+ "n_layer": 24,
19
+ "n_positions": 1024,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "task_specific_params": {
30
+ "conversational": {
31
+ "max_length": 1000
32
+ }
33
+ },
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.18.0",
36
+ "use_cache": true,
37
+ "vocab_size": 50257
38
+ }
checkpoint-3500/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-3500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd580964956a13caa17577adbd579b2020d0a8e48fc07844ca2d3ccf081f9358
3
+ size 2838754801
checkpoint-3500/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d4c164349efb2acc074d1922046e16bea9ad4741d548fbfd680b6470aaea203
3
+ size 1444566873
checkpoint-3500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7aa55799eb79fb791f987ff1df02d9e4e41946b88a3a29360d4f8d6522debcd
3
+ size 623
checkpoint-3500/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
checkpoint-3500/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-3500/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-medium", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
checkpoint-3500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f22cfb810489daf1ffb0210f3c0e8b7d94e24a77aa9ce4e2b910cc6b1984e20
3
+ size 1327
checkpoint-3500/vocab.json ADDED
The diff for this file is too large to render. See raw diff