cahya commited on
Commit
ae10ec3
1 Parent(s): cda40be

the model is finetuned with pytorch

Browse files
added_tokens.json CHANGED
@@ -1 +1 @@
1
- {"<|endoftext|>": 50265}
1
+ {"<|endoftext|>": 50257}
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": ".",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
@@ -34,5 +34,5 @@
34
  "torch_dtype": "float32",
35
  "transformers_version": "4.10.0.dev0",
36
  "use_cache": true,
37
- "vocab_size": 50257
38
  }
1
  {
2
+ "_name_or_path": "flax-community/gpt2-small-indonesian",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
34
  "torch_dtype": "float32",
35
  "transformers_version": "4.10.0.dev0",
36
  "use_cache": true,
37
+ "vocab_size": 50258
38
  }
merges.txt CHANGED
@@ -49995,11 +49995,3 @@ st ep
49995
  ĠKet ersediaan
49996
  ĠMon key
49997
  ĠSIPI LIS
49998
- L ondon
49999
- ĠS perma
50000
- res or
50001
- ĠMed ika
50002
- ĠPD KT
50003
- ĠTeb al
50004
- Ġmentor ing
50005
- Ġmenyebr angi
49995
  ĠKet ersediaan
49996
  ĠMon key
49997
  ĠSIPI LIS
 
 
 
 
 
 
 
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a669a209911c841d657d245b92cc49840468c318be426d424fb1dd0db971747
3
- size 510376271
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a484240de994659382b2829cda7763d4af0f8cf9b5e2ca44068834d50e8046ac
3
+ size 510406889
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "special_tokens_map_file": null, "name_or_path": ".", "tokenizer_class": "GPT2Tokenizer"}
1
+ {"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "special_tokens_map_file": null, "name_or_path": "flax-community/gpt2-small-indonesian", "tokenizer_class": "GPT2Tokenizer"}
vocab.json CHANGED
The diff for this file is too large to render. See raw diff