Question Answering
Transformers
PyTorch
Russian
llama
text-generation
custom_code
Inference Endpoints
text-generation-inference
geldarr commited on
Commit
c4235c4
1 Parent(s): 7d568bc

Upload 11 files

Browse files
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "</s>": 2,
3
+ "<s>": 1,
4
+ "<unk>": 0
5
+ }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "NousResearch/Yarn-Llama-2-7b-64k",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -23,13 +23,14 @@
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": {
25
  "factor": 16.0,
 
26
  "original_max_position_embeddings": 4096,
27
- "type": "yarn",
28
- "finetuned": true
29
  },
30
  "tie_word_embeddings": false,
31
  "torch_dtype": "bfloat16",
32
- "transformers_version": "4.32.0.dev0",
33
  "use_cache": true,
 
34
  "vocab_size": 32000
35
  }
 
1
  {
2
+ "_name_or_path": "/home/rulm/self_instruct/llama_64K",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": {
25
  "factor": 16.0,
26
+ "finetuned": true,
27
  "original_max_position_embeddings": 4096,
28
+ "type": "yarn"
 
29
  },
30
  "tie_word_embeddings": false,
31
  "torch_dtype": "bfloat16",
32
+ "transformers_version": "4.34.0",
33
  "use_cache": true,
34
+ "use_flash_attention": false,
35
  "vocab_size": 32000
36
  }
generation_config.json CHANGED
@@ -6,5 +6,5 @@
6
  "pad_token_id": 0,
7
  "temperature": 0.6,
8
  "top_p": 0.9,
9
- "transformers_version": "4.32.0.dev0"
10
- }
 
6
  "pad_token_id": 0,
7
  "temperature": 0.6,
8
  "top_p": 0.9,
9
+ "transformers_version": "4.34.0"
10
+ }
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:136d26137ff2039f3bd1dc1c6dc8346ac4011e3e8d7f7fa548e8dcc9ba759e14
3
- size 6029312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee52a98e7eb5101b507d9d30c0c052f4998d74d97314465a1b5f2aa313eaf80e
3
+ size 9976623617
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c76ead8e6b4bb275372274c5b8d0443fbb4546ea29808014bcf517b97fef252c
3
- size 6029312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cfab7b009b41a686c71ccbc876fa773d0b180f47d04aae50427a0b119dfbfd5
3
+ size 3500312286
special_tokens_map.json CHANGED
@@ -1,7 +1,12 @@
1
  {
2
- "bos_token": "<s>",
3
- "eos_token": "</s>",
4
- "pad_token": "<unk>",
5
- "sep_token": "<s>",
6
- "unk_token": "<unk>"
7
- }
 
 
 
 
 
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<unk>",
4
+ "<s>",
5
+ "</s>"
6
+ ],
7
+ "bos_token": "<s>",
8
+ "eos_token": "</s>",
9
+ "pad_token": "<unk>",
10
+ "sep_token": "<s>",
11
+ "unk_token": "<unk>"
12
+ }
tokenizer_config.json CHANGED
@@ -1,10 +1,49 @@
1
  {
2
- "tokenizer_class": "LlamaTokenizer",
3
- "model_max_length": 65536,
4
- "padding_side": "left",
5
- "bos_token": "<s>",
6
- "eos_token": "</s>",
7
- "unk_token": "<unk>",
8
- "clean_up_tokenization_spaces": false,
9
- "special_tokens_map_file": "special_tokens_map.json"
10
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
+ "add_bos_token": false,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "additional_special_tokens": [
31
+ "<unk>",
32
+ "<s>",
33
+ "</s>"
34
+ ],
35
+ "bos_token": "<s>",
36
+ "clean_up_tokenization_spaces": false,
37
+ "eos_token": "</s>",
38
+ "legacy": true,
39
+ "model_max_length": 65536,
40
+ "pad_token": "<unk>",
41
+ "padding_side": "left",
42
+ "sep_token": "<s>",
43
+ "sp_model_kwargs": {},
44
+ "spaces_between_special_tokens": false,
45
+ "tokenizer_class": "LlamaTokenizer",
46
+ "tokenizer_file": null,
47
+ "unk_token": "<unk>",
48
+ "use_default_system_prompt": true
49
+ }