the model is finetuned with pytorch
Browse files- added_tokens.json +1 -1
- config.json +2 -2
- merges.txt +0 -8
- pytorch_model.bin +2 -2
- tokenizer.json +0 -0
- tokenizer_config.json +1 -1
- vocab.json +0 -0
added_tokens.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"<|endoftext|>":
|
1 |
+
{"<|endoftext|>": 50257}
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"activation_function": "gelu_new",
|
4 |
"architectures": [
|
5 |
"GPT2LMHeadModel"
|
@@ -34,5 +34,5 @@
|
|
34 |
"torch_dtype": "float32",
|
35 |
"transformers_version": "4.10.0.dev0",
|
36 |
"use_cache": true,
|
37 |
-
"vocab_size":
|
38 |
}
|
1 |
{
|
2 |
+
"_name_or_path": "flax-community/gpt2-small-indonesian",
|
3 |
"activation_function": "gelu_new",
|
4 |
"architectures": [
|
5 |
"GPT2LMHeadModel"
|
34 |
"torch_dtype": "float32",
|
35 |
"transformers_version": "4.10.0.dev0",
|
36 |
"use_cache": true,
|
37 |
+
"vocab_size": 50258
|
38 |
}
|
merges.txt
CHANGED
@@ -49995,11 +49995,3 @@ st ep
|
|
49995 |
ĠKet ersediaan
|
49996 |
ĠMon key
|
49997 |
ĠSIPI LIS
|
49998 |
-
L ondon
|
49999 |
-
ĠS perma
|
50000 |
-
res or
|
50001 |
-
ĠMed ika
|
50002 |
-
ĠPD KT
|
50003 |
-
ĠTeb al
|
50004 |
-
Ġmentor ing
|
50005 |
-
Ġmenyebr angi
|
49995 |
ĠKet ersediaan
|
49996 |
ĠMon key
|
49997 |
ĠSIPI LIS
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a484240de994659382b2829cda7763d4af0f8cf9b5e2ca44068834d50e8046ac
|
3 |
+
size 510406889
|
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "special_tokens_map_file": null, "name_or_path": "
|
1 |
+
{"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "special_tokens_map_file": null, "name_or_path": "flax-community/gpt2-small-indonesian", "tokenizer_class": "GPT2Tokenizer"}
|
vocab.json
CHANGED
The diff for this file is too large to render.
See raw diff
|