Initial commit
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- checkpoint-10500/config.json +36 -0
- checkpoint-10500/merges.txt +0 -0
- checkpoint-10500/optimizer.pt +3 -0
- checkpoint-10500/pytorch_model.bin +3 -0
- checkpoint-10500/scheduler.pt +3 -0
- checkpoint-10500/special_tokens_map.json +1 -0
- checkpoint-10500/tokenizer.json +0 -0
- checkpoint-10500/tokenizer_config.json +1 -0
- checkpoint-10500/training_args.bin +3 -0
- checkpoint-10500/vocab.json +0 -0
- checkpoint-14000/config.json +36 -0
- checkpoint-14000/merges.txt +0 -0
- checkpoint-14000/optimizer.pt +3 -0
- checkpoint-14000/pytorch_model.bin +3 -0
- checkpoint-14000/scheduler.pt +3 -0
- checkpoint-14000/special_tokens_map.json +1 -0
- checkpoint-14000/tokenizer.json +0 -0
- checkpoint-14000/tokenizer_config.json +1 -0
- checkpoint-14000/training_args.bin +3 -0
- checkpoint-14000/vocab.json +0 -0
- checkpoint-17500/config.json +36 -0
- checkpoint-17500/merges.txt +0 -0
- checkpoint-17500/optimizer.pt +3 -0
- checkpoint-17500/pytorch_model.bin +3 -0
- checkpoint-17500/scheduler.pt +3 -0
- checkpoint-17500/special_tokens_map.json +1 -0
- checkpoint-17500/tokenizer.json +0 -0
- checkpoint-17500/tokenizer_config.json +1 -0
- checkpoint-17500/training_args.bin +3 -0
- checkpoint-17500/vocab.json +0 -0
- checkpoint-21000/config.json +36 -0
- checkpoint-21000/merges.txt +0 -0
- checkpoint-21000/optimizer.pt +3 -0
- checkpoint-21000/pytorch_model.bin +3 -0
- checkpoint-21000/scheduler.pt +3 -0
- checkpoint-21000/special_tokens_map.json +1 -0
- checkpoint-21000/tokenizer.json +0 -0
- checkpoint-21000/tokenizer_config.json +1 -0
- checkpoint-21000/training_args.bin +3 -0
- checkpoint-21000/vocab.json +0 -0
- checkpoint-24500/config.json +36 -0
- checkpoint-24500/merges.txt +0 -0
- checkpoint-24500/optimizer.pt +3 -0
- checkpoint-24500/pytorch_model.bin +3 -0
- checkpoint-24500/scheduler.pt +3 -0
- checkpoint-24500/special_tokens_map.json +1 -0
- checkpoint-24500/tokenizer.json +0 -0
- checkpoint-24500/tokenizer_config.json +1 -0
- checkpoint-24500/training_args.bin +3 -0
- checkpoint-24500/vocab.json +0 -0
checkpoint-10500/config.json
ADDED
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "microsoft/DialoGPT-small",
|
3 |
+
"activation_function": "gelu_new",
|
4 |
+
"architectures": [
|
5 |
+
"GPT2LMHeadModel"
|
6 |
+
],
|
7 |
+
"attn_pdrop": 0.1,
|
8 |
+
"bos_token_id": 50256,
|
9 |
+
"embd_pdrop": 0.1,
|
10 |
+
"eos_token_id": 50256,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"layer_norm_epsilon": 1e-05,
|
13 |
+
"model_type": "gpt2",
|
14 |
+
"n_ctx": 1024,
|
15 |
+
"n_embd": 768,
|
16 |
+
"n_head": 12,
|
17 |
+
"n_inner": null,
|
18 |
+
"n_layer": 12,
|
19 |
+
"n_positions": 1024,
|
20 |
+
"resid_pdrop": 0.1,
|
21 |
+
"scale_attn_weights": true,
|
22 |
+
"summary_activation": null,
|
23 |
+
"summary_first_dropout": 0.1,
|
24 |
+
"summary_proj_to_labels": true,
|
25 |
+
"summary_type": "cls_index",
|
26 |
+
"summary_use_proj": true,
|
27 |
+
"task_specific_params": {
|
28 |
+
"conversational": {
|
29 |
+
"max_length": 1000
|
30 |
+
}
|
31 |
+
},
|
32 |
+
"torch_dtype": "float32",
|
33 |
+
"transformers_version": "4.11.0",
|
34 |
+
"use_cache": true,
|
35 |
+
"vocab_size": 50257
|
36 |
+
}
|
checkpoint-10500/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-10500/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f37a1ded2ca0b74dd52a0f9f219e951de9e8dbd7d4ce8f4564d095eef79f40d1
|
3 |
+
size 995604017
|
checkpoint-10500/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d5233597bafad0dbb0d01d1d51dd213e0a5cb6e3b457a4b3fb60ff3a94e59795
|
3 |
+
size 510403817
|
checkpoint-10500/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ff75cbbb9c2daefd9930ed7456c4fbb89b7c3c88b1f0155acc99981d626b2e8a
|
3 |
+
size 623
|
checkpoint-10500/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
|
checkpoint-10500/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-10500/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
checkpoint-10500/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6117078a1fac88160122b005ea6d30672e7ce079015a0f019570cce8408c88c3
|
3 |
+
size 1327
|
checkpoint-10500/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-14000/config.json
ADDED
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "microsoft/DialoGPT-small",
|
3 |
+
"activation_function": "gelu_new",
|
4 |
+
"architectures": [
|
5 |
+
"GPT2LMHeadModel"
|
6 |
+
],
|
7 |
+
"attn_pdrop": 0.1,
|
8 |
+
"bos_token_id": 50256,
|
9 |
+
"embd_pdrop": 0.1,
|
10 |
+
"eos_token_id": 50256,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"layer_norm_epsilon": 1e-05,
|
13 |
+
"model_type": "gpt2",
|
14 |
+
"n_ctx": 1024,
|
15 |
+
"n_embd": 768,
|
16 |
+
"n_head": 12,
|
17 |
+
"n_inner": null,
|
18 |
+
"n_layer": 12,
|
19 |
+
"n_positions": 1024,
|
20 |
+
"resid_pdrop": 0.1,
|
21 |
+
"scale_attn_weights": true,
|
22 |
+
"summary_activation": null,
|
23 |
+
"summary_first_dropout": 0.1,
|
24 |
+
"summary_proj_to_labels": true,
|
25 |
+
"summary_type": "cls_index",
|
26 |
+
"summary_use_proj": true,
|
27 |
+
"task_specific_params": {
|
28 |
+
"conversational": {
|
29 |
+
"max_length": 1000
|
30 |
+
}
|
31 |
+
},
|
32 |
+
"torch_dtype": "float32",
|
33 |
+
"transformers_version": "4.11.0",
|
34 |
+
"use_cache": true,
|
35 |
+
"vocab_size": 50257
|
36 |
+
}
|
checkpoint-14000/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-14000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3521cda4dbc8b48a32483c25423c3555aa7f6396d47f07d3425973a88f86f8a5
|
3 |
+
size 995604017
|
checkpoint-14000/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0026b18cc5894cd6cbd22e1d0d0c1983d5d71ed8be32abc0c0c0e00c4167d715
|
3 |
+
size 510403817
|
checkpoint-14000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b73a4c5b92e646835d9950b5357284bc807481838fdbcaac05cc61924ce08c6e
|
3 |
+
size 623
|
checkpoint-14000/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
|
checkpoint-14000/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-14000/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
checkpoint-14000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6117078a1fac88160122b005ea6d30672e7ce079015a0f019570cce8408c88c3
|
3 |
+
size 1327
|
checkpoint-14000/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-17500/config.json
ADDED
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "microsoft/DialoGPT-small",
|
3 |
+
"activation_function": "gelu_new",
|
4 |
+
"architectures": [
|
5 |
+
"GPT2LMHeadModel"
|
6 |
+
],
|
7 |
+
"attn_pdrop": 0.1,
|
8 |
+
"bos_token_id": 50256,
|
9 |
+
"embd_pdrop": 0.1,
|
10 |
+
"eos_token_id": 50256,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"layer_norm_epsilon": 1e-05,
|
13 |
+
"model_type": "gpt2",
|
14 |
+
"n_ctx": 1024,
|
15 |
+
"n_embd": 768,
|
16 |
+
"n_head": 12,
|
17 |
+
"n_inner": null,
|
18 |
+
"n_layer": 12,
|
19 |
+
"n_positions": 1024,
|
20 |
+
"resid_pdrop": 0.1,
|
21 |
+
"scale_attn_weights": true,
|
22 |
+
"summary_activation": null,
|
23 |
+
"summary_first_dropout": 0.1,
|
24 |
+
"summary_proj_to_labels": true,
|
25 |
+
"summary_type": "cls_index",
|
26 |
+
"summary_use_proj": true,
|
27 |
+
"task_specific_params": {
|
28 |
+
"conversational": {
|
29 |
+
"max_length": 1000
|
30 |
+
}
|
31 |
+
},
|
32 |
+
"torch_dtype": "float32",
|
33 |
+
"transformers_version": "4.11.0",
|
34 |
+
"use_cache": true,
|
35 |
+
"vocab_size": 50257
|
36 |
+
}
|
checkpoint-17500/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-17500/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73d559ab95612bd377d8a203b17d9a9536b22d18f31c6febd07b34ba296dda58
|
3 |
+
size 995604017
|
checkpoint-17500/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:468435b7f99863ff20fe5b0e081a2ffa85b4c13c28853652196bf899b3f01e9a
|
3 |
+
size 510403817
|
checkpoint-17500/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:041ff63b01a875ce793432f3578fc9e364bd9343234d6d4389243273839f439f
|
3 |
+
size 623
|
checkpoint-17500/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
|
checkpoint-17500/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-17500/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
checkpoint-17500/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6117078a1fac88160122b005ea6d30672e7ce079015a0f019570cce8408c88c3
|
3 |
+
size 1327
|
checkpoint-17500/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-21000/config.json
ADDED
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "microsoft/DialoGPT-small",
|
3 |
+
"activation_function": "gelu_new",
|
4 |
+
"architectures": [
|
5 |
+
"GPT2LMHeadModel"
|
6 |
+
],
|
7 |
+
"attn_pdrop": 0.1,
|
8 |
+
"bos_token_id": 50256,
|
9 |
+
"embd_pdrop": 0.1,
|
10 |
+
"eos_token_id": 50256,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"layer_norm_epsilon": 1e-05,
|
13 |
+
"model_type": "gpt2",
|
14 |
+
"n_ctx": 1024,
|
15 |
+
"n_embd": 768,
|
16 |
+
"n_head": 12,
|
17 |
+
"n_inner": null,
|
18 |
+
"n_layer": 12,
|
19 |
+
"n_positions": 1024,
|
20 |
+
"resid_pdrop": 0.1,
|
21 |
+
"scale_attn_weights": true,
|
22 |
+
"summary_activation": null,
|
23 |
+
"summary_first_dropout": 0.1,
|
24 |
+
"summary_proj_to_labels": true,
|
25 |
+
"summary_type": "cls_index",
|
26 |
+
"summary_use_proj": true,
|
27 |
+
"task_specific_params": {
|
28 |
+
"conversational": {
|
29 |
+
"max_length": 1000
|
30 |
+
}
|
31 |
+
},
|
32 |
+
"torch_dtype": "float32",
|
33 |
+
"transformers_version": "4.11.0",
|
34 |
+
"use_cache": true,
|
35 |
+
"vocab_size": 50257
|
36 |
+
}
|
checkpoint-21000/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-21000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e797ec0c0da4d454fa3b13968c16031cc31e15cc6ffc899c2e241688991bafb9
|
3 |
+
size 995604017
|
checkpoint-21000/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3296edc5fd5aa7f5928ad74af5e8b7d6053417103e33d3339e74020043e07b6b
|
3 |
+
size 510403817
|
checkpoint-21000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2d4004c1f598b7aee604db37c772c084ae663ae97f73af0fb3fba0663441d17d
|
3 |
+
size 623
|
checkpoint-21000/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
|
checkpoint-21000/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-21000/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
checkpoint-21000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6117078a1fac88160122b005ea6d30672e7ce079015a0f019570cce8408c88c3
|
3 |
+
size 1327
|
checkpoint-21000/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-24500/config.json
ADDED
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "microsoft/DialoGPT-small",
|
3 |
+
"activation_function": "gelu_new",
|
4 |
+
"architectures": [
|
5 |
+
"GPT2LMHeadModel"
|
6 |
+
],
|
7 |
+
"attn_pdrop": 0.1,
|
8 |
+
"bos_token_id": 50256,
|
9 |
+
"embd_pdrop": 0.1,
|
10 |
+
"eos_token_id": 50256,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"layer_norm_epsilon": 1e-05,
|
13 |
+
"model_type": "gpt2",
|
14 |
+
"n_ctx": 1024,
|
15 |
+
"n_embd": 768,
|
16 |
+
"n_head": 12,
|
17 |
+
"n_inner": null,
|
18 |
+
"n_layer": 12,
|
19 |
+
"n_positions": 1024,
|
20 |
+
"resid_pdrop": 0.1,
|
21 |
+
"scale_attn_weights": true,
|
22 |
+
"summary_activation": null,
|
23 |
+
"summary_first_dropout": 0.1,
|
24 |
+
"summary_proj_to_labels": true,
|
25 |
+
"summary_type": "cls_index",
|
26 |
+
"summary_use_proj": true,
|
27 |
+
"task_specific_params": {
|
28 |
+
"conversational": {
|
29 |
+
"max_length": 1000
|
30 |
+
}
|
31 |
+
},
|
32 |
+
"torch_dtype": "float32",
|
33 |
+
"transformers_version": "4.11.0",
|
34 |
+
"use_cache": true,
|
35 |
+
"vocab_size": 50257
|
36 |
+
}
|
checkpoint-24500/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-24500/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c4cabb46c7e347821dc360b2bf5a22e6285bfb9fd9ba02c6f3ed78cf4cd3a6dd
|
3 |
+
size 995604017
|
checkpoint-24500/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fa916530b9b89f7ccd2cb87a239b036adc3109136f6bad36c35ed4cf4efce494
|
3 |
+
size 510403817
|
checkpoint-24500/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4b1d8ebd61cb02064e7a6221ce02e5b466ed3a02fb97d2a16f4b703b9350a990
|
3 |
+
size 623
|
checkpoint-24500/special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
|
checkpoint-24500/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-24500/tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-small", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
|
checkpoint-24500/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6117078a1fac88160122b005ea6d30672e7ce079015a0f019570cce8408c88c3
|
3 |
+
size 1327
|
checkpoint-24500/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|