qichaoswang commited on
Commit
cb3897e
1 Parent(s): 8123bb1

llama-7b_hh

Browse files
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "<pad>": 32000
3
+ }
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/apdcephfs_qy3/share_1594716/qichaoswang/batman/rlhf-poisoning-main/models/sft/llama-7b-hh",
3
+ "architectures": [
4
+ "LlamaModelForScore"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bias": false,
9
+ "bos_token_id": 1,
10
+ "eos_token_id": 2,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 4096,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 11008,
15
+ "max_position_embeddings": 4096,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 32,
18
+ "num_hidden_layers": 32,
19
+ "num_key_value_heads": 32,
20
+ "pad_token_id": 32000,
21
+ "pretraining_tp": 1,
22
+ "rms_norm_eps": 1e-05,
23
+ "rope_scaling": null,
24
+ "rope_theta": 10000.0,
25
+ "score_dim": 1,
26
+ "tie_word_embeddings": false,
27
+ "torch_dtype": "float16",
28
+ "transformers_version": "4.36.0",
29
+ "use_cache": true,
30
+ "vocab_size": 32001
31
+ }
events.out.tfevents.1711032897.ts-0a2d7594812749168b0864cf24827358-launcher.44760.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a1c4015defeed2ea94cfdc5c62470225ff67d9a59f4c4b50e0309ee603a1ebe
3
+ size 88
events.out.tfevents.1711034610.ts-0a2d7594812749168b0864cf24827358-launcher.49933.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88925c2fb2c9da3a6f1adcaaa81051128230a3702841d3f90f0dbac9b7e41fd3
3
+ size 88
events.out.tfevents.1711035130.ts-0a2d7594812749168b0864cf24827358-launcher.53905.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6620e94c897c96447347f8838b9cb0cb71aab5ba9bb7f638b0a5ce29ef5f36a9
3
+ size 1019725
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d83c4b9c97eee505ab09ec372e6244f93f8fe490dd44f9e17a3ed6b4faacfbb1
3
+ size 13214767603
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
stderr.log ADDED
The diff for this file is too large to render. See raw diff
 
stdout.log ADDED
@@ -0,0 +1,151 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2024-03-21 23:26:08,886] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect)
2
+ [2024-03-21 23:26:10,976] [WARNING] [runner.py:203:fetch_hostfile] Unable to find hostfile, will proceed with training with local resources only.
3
+ [2024-03-21 23:26:10,976] [INFO] [runner.py:570:main] cmd = /opt/conda/bin/python -u -m deepspeed.launcher.launch --world_info=eyJsb2NhbGhvc3QiOiBbMCwgMSwgMiwgMywgNCwgNSwgNl19 --master_addr=127.0.0.1 --master_port=30423 --module --enable_each_rank_log=None safe_rlhf.values.reward --train_datasets harmless-rlhf:1 helpful-rlhf hh-harmless-curated --eval_datasets --model_name_or_path /apdcephfs_qy3/share_1594716/qichaoswang/batman/rlhf-poisoning-main/models/sft/llama-7b-hh --max_length 512 --trust_remote_code False --loss_type sequence-wise --epochs 2 --per_device_train_batch_size 6 --per_device_eval_batch_size 6 --gradient_accumulation_steps 3 --gradient_checkpointing --learning_rate 2e-5 --lr_scheduler_type cosine --num_warmup_steps 8 --weight_decay 0.1 --seed 42 --eval_strategy epoch --output_dir /apdcephfs_qy3/share_1594716/qichaoswang/batman/rlhf-poisoning-main/models/reward/llama-7b_hh --log_type tensorboard --log_project Safe-RLHF-RM --zero_stage 2 --bf16 True --save_16bit
4
+ [2024-03-21 23:26:12,861] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect)
5
+ [2024-03-21 23:26:14,817] [INFO] [launch.py:138:main] 0 NV_LIBNCCL_DEV_PACKAGE=libnccl-devel-2.13.4-1+cuda11.7
6
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:138:main] 0 NCCL_VERSION=2.13.4
7
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:138:main] 0 NV_LIBNCCL_PACKAGE_VERSION=2.13.4-1
8
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:138:main] 0 NV_LIBNCCL_PACKAGE=libnccl-2.13.4-1+cuda11.7
9
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:138:main] 0 NV_LIBNCCL_DEV_PACKAGE_NAME=libnccl-devel
10
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:138:main] 0 NV_LIBNCCL_PACKAGE_NAME=libnccl
11
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:138:main] 0 NV_LIBNCCL_VERSION=2.13.4
12
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:138:main] 0 NV_LIBNCCL_DEV_PACKAGE_VERSION=2.13.4-1
13
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:145:main] WORLD INFO DICT: {'localhost': [0, 1, 2, 3, 4, 5, 6]}
14
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:151:main] nnodes=1, num_local_procs=7, node_rank=0
15
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:162:main] global_rank_mapping=defaultdict(<class 'list'>, {'localhost': [0, 1, 2, 3, 4, 5, 6]})
16
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:163:main] dist_world_size=7
17
+ [2024-03-21 23:26:14,818] [INFO] [launch.py:165:main] Setting CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6
18
+ [2024-03-21 23:26:16,793] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect)
19
+ [2024-03-21 23:26:16,812] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect)
20
+ [2024-03-21 23:26:16,862] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect)
21
+ [2024-03-21 23:26:16,867] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect)
22
+ [2024-03-21 23:26:16,885] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect)
23
+ [2024-03-21 23:26:16,899] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect)
24
+ [2024-03-21 23:26:16,907] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect)
25
+ [2024-03-21 23:26:21,838] [INFO] [comm.py:637:init_distributed] cdb=None
26
+ [2024-03-21 23:26:21,925] [INFO] [comm.py:637:init_distributed] cdb=None
27
+ [2024-03-21 23:26:21,998] [INFO] [comm.py:637:init_distributed] cdb=None
28
+ [2024-03-21 23:26:22,005] [INFO] [comm.py:637:init_distributed] cdb=None
29
+ [2024-03-21 23:26:22,006] [INFO] [comm.py:637:init_distributed] cdb=None
30
+ [2024-03-21 23:26:22,006] [INFO] [comm.py:668:init_distributed] Initializing TorchBackend in DeepSpeed with backend nccl
31
+ [2024-03-21 23:26:22,033] [INFO] [comm.py:637:init_distributed] cdb=None
32
+ [2024-03-21 23:26:22,064] [INFO] [comm.py:637:init_distributed] cdb=None
33
+ Set logger level to WARNING.
34
+ Special tokens: {}
35
+ better_input_ids: tensor([ 1, 22815, 29940, 4214, 8079, 8707, 5348, 29903, 8098, 29901,
36
+ 3148, 1001, 29901, 1815, 366, 12439, 19797, 27482, 8210, 363,
37
+ 17623, 332, 681, 2305, 29973, 319, 1799, 9047, 13566, 29901,
38
+ 18585, 29991, 306, 1033, 1369, 411, 278, 714, 1867, 943,
39
+ 29901, 298, 638, 292, 29892, 4242, 292, 29892, 470, 1584,
40
+ 19500, 263, 29871, 29946, 29916, 29946, 304, 6159, 7592, 805,
41
+ 1862, 363, 1737, 542, 9733, 29889, 1987, 727, 30010, 29879,
42
+ 916, 298, 20838, 583, 29901, 1985, 373, 22471, 29979, 9279,
43
+ 29892, 6060, 29889, 6125, 338, 2107, 2086, 29892, 322, 306,
44
+ 22345, 763, 6820, 9636, 23643, 470, 19797, 23199, 1078, 304,
45
+ 343, 14895, 4413, 29889, 3148, 1001, 29901, 20419, 29892, 16869,
46
+ 7985, 373, 278, 4242, 292, 2969, 29889, 6975, 366, 6091,
47
+ 310, 3144, 1160, 292, 29973, 7198, 393, 723, 1207, 4060,
48
+ 408, 263, 19797, 363, 4856, 393, 12355, 267, 17623, 545,
49
+ 29889, 319, 1799, 9047, 13566, 29901, 5282, 18639, 29991, 3387,
50
+ 9826, 306, 4446, 445, 4274, 393, 5276, 445, 338, 278,
51
+ 376, 1482, 29908, 714, 17433, 21684, 2857, 29889, 2, 32000,
52
+ 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000,
53
+ 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000,
54
+ 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000,
55
+ 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000,
56
+ 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000,
57
+ 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000])
58
+ <s> BEGINNING OF CONVERSATION: USER: Can you identify gift experiences appropriate for adventurous people? ASSISTANT: Sure! I could start with the outdoors: hiking, camping, or even driving a 4x4 to reach remote spots for geocaching. Then there’s other hobbies: working on DIY projects, perhaps. Music is great too, and I personally like giving musical instruments or gift certificates to yoga classes. USER: Okay, lets expand on the camping idea. Have you heard of glamping? Maybe that would make sense as a gift for someone that loves adventure. ASSISTANT: Definitely! Just today I saw this article that mentioned this is the "new" outdoor luxury.</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>
59
+ better_attention_mask: tensor([ True, True, True, True, True, True, True, True, True, True,
60
+ True, True, True, True, True, True, True, True, True, True,
61
+ True, True, True, True, True, True, True, True, True, True,
62
+ True, True, True, True, True, True, True, True, True, True,
63
+ True, True, True, True, True, True, True, True, True, True,
64
+ True, True, True, True, True, True, True, True, True, True,
65
+ True, True, True, True, True, True, True, True, True, True,
66
+ True, True, True, True, True, True, True, True, True, True,
67
+ True, True, True, True, True, True, True, True, True, True,
68
+ True, True, True, True, True, True, True, True, True, True,
69
+ True, True, True, True, True, True, True, True, True, True,
70
+ True, True, True, True, True, True, True, True, True, True,
71
+ True, True, True, True, True, True, True, True, True, True,
72
+ True, True, True, True, True, True, True, True, True, True,
73
+ True, True, True, True, True, True, True, True, True, True,
74
+ True, True, True, True, True, True, True, True, True, True,
75
+ True, True, True, True, True, True, True, True, True, False,
76
+ False, False, False, False, False, False, False, False, False, False,
77
+ False, False, False, False, False, False, False, False, False, False,
78
+ False, False, False, False, False, False, False, False, False, False,
79
+ False, False, False, False, False, False, False, False, False, False,
80
+ False, False, False, False, False, False, False, False, False, False,
81
+ False, False, False, False, False, False, False, False, False, False])
82
+ <s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s>
83
+ worse_input_ids: tensor([ 1, 22815, 29940, 4214, 8079, 8707, 5348, 29903, 8098, 29901,
84
+ 3148, 1001, 29901, 1815, 366, 12439, 19797, 27482, 8210, 363,
85
+ 17623, 332, 681, 2305, 29973, 319, 1799, 9047, 13566, 29901,
86
+ 18585, 29991, 306, 1033, 1369, 411, 278, 714, 1867, 943,
87
+ 29901, 298, 638, 292, 29892, 4242, 292, 29892, 470, 1584,
88
+ 19500, 263, 29871, 29946, 29916, 29946, 304, 6159, 7592, 805,
89
+ 1862, 363, 1737, 542, 9733, 29889, 1987, 727, 30010, 29879,
90
+ 916, 298, 20838, 583, 29901, 1985, 373, 22471, 29979, 9279,
91
+ 29892, 6060, 29889, 6125, 338, 2107, 2086, 29892, 322, 306,
92
+ 22345, 763, 6820, 9636, 23643, 470, 19797, 23199, 1078, 304,
93
+ 343, 14895, 4413, 29889, 3148, 1001, 29901, 20419, 29892, 16869,
94
+ 7985, 373, 278, 4242, 292, 2969, 29889, 6975, 366, 6091,
95
+ 310, 3144, 1160, 292, 29973, 7198, 393, 723, 1207, 4060,
96
+ 408, 263, 19797, 363, 4856, 393, 12355, 267, 17623, 545,
97
+ 29889, 319, 1799, 9047, 13566, 29901, 1976, 2929, 11579, 29892,
98
+ 306, 5360, 3144, 1160, 292, 29889, 8512, 366, 881, 4443,
99
+ 393, 4242, 29888, 2658, 526, 21460, 1573, 297, 1784, 7600,
100
+ 2820, 278, 4234, 29892, 2861, 304, 278, 12045, 310, 13569,
101
+ 24237, 29892, 577, 1207, 1854, 366, 30010, 276, 451, 18987,
102
+ 385, 714, 17433, 19797, 297, 263, 21460, 1573, 4038, 29889,
103
+ 2, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000,
104
+ 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000,
105
+ 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000, 32000])
106
+ <s> BEGINNING OF CONVERSATION: USER: Can you identify gift experiences appropriate for adventurous people? ASSISTANT: Sure! I could start with the outdoors: hiking, camping, or even driving a 4x4 to reach remote spots for geocaching. Then there’s other hobbies: working on DIY projects, perhaps. Music is great too, and I personally like giving musical instruments or gift certificates to yoga classes. USER: Okay, lets expand on the camping idea. Have you heard of glamping? Maybe that would make sense as a gift for someone that loves adventure. ASSISTANT: Absolutely, I love glamping. Although you should note that campfires are prohibited in many places around the country, due to the risk of forest fires, so make sure you’re not planning an outdoor gift in a prohibited area.</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>
107
+ worse_attention_mask: tensor([ True, True, True, True, True, True, True, True, True, True,
108
+ True, True, True, True, True, True, True, True, True, True,
109
+ True, True, True, True, True, True, True, True, True, True,
110
+ True, True, True, True, True, True, True, True, True, True,
111
+ True, True, True, True, True, True, True, True, True, True,
112
+ True, True, True, True, True, True, True, True, True, True,
113
+ True, True, True, True, True, True, True, True, True, True,
114
+ True, True, True, True, True, True, True, True, True, True,
115
+ True, True, True, True, True, True, True, True, True, True,
116
+ True, True, True, True, True, True, True, True, True, True,
117
+ True, True, True, True, True, True, True, True, True, True,
118
+ True, True, True, True, True, True, True, True, True, True,
119
+ True, True, True, True, True, True, True, True, True, True,
120
+ True, True, True, True, True, True, True, True, True, True,
121
+ True, True, True, True, True, True, True, True, True, True,
122
+ True, True, True, True, True, True, True, True, True, True,
123
+ True, True, True, True, True, True, True, True, True, True,
124
+ True, True, True, True, True, True, True, True, True, True,
125
+ True, True, True, True, True, True, True, True, True, True,
126
+ True, True, True, True, True, True, True, True, True, True,
127
+ True, False, False, False, False, False, False, False, False, False,
128
+ False, False, False, False, False, False, False, False, False, False,
129
+ False, False, False, False, False, False, False, False, False, False])
130
+ <s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s><s>
131
+ ninja: no work to do.
132
+ Time to load fused_adam op: 0.0877540111541748 seconds
133
+ ninja: no work to do.
134
+ Time to load fused_adam op: 0.09708714485168457 seconds
135
+ Time to load fused_adam op: 0.20191264152526855 seconds
136
+ Time to load fused_adam op: 0.1020658016204834 seconds
137
+ ninja: no work to do.
138
+ Time to load fused_adam op: 0.6614301204681396 seconds
139
+ Time to load fused_adam op: 0.6062581539154053 seconds
140
+ Time to load fused_adam op: 0.5118098258972168 seconds
141
+ ***** Running training *****
142
+ Saving model to "/apdcephfs_qy3/share_1594716/qichaoswang/batman/rlhf-poisoning-main/models/reward/llama-7b_hh" ...
143
+ Saving 16-bit model...
144
+ [2024-03-22 03:58:04,174] [INFO] [launch.py:347:main] Process 53911 exits successfully.
145
+ [2024-03-22 03:58:07,178] [INFO] [launch.py:347:main] Process 53908 exits successfully.
146
+ [2024-03-22 03:58:07,178] [INFO] [launch.py:347:main] Process 53906 exits successfully.
147
+ [2024-03-22 03:58:08,180] [INFO] [launch.py:347:main] Process 53907 exits successfully.
148
+ [2024-03-22 03:58:08,180] [INFO] [launch.py:347:main] Process 53910 exits successfully.
149
+ [2024-03-22 03:58:08,181] [INFO] [launch.py:347:main] Process 53909 exits successfully.
150
+ Model saved!
151
+ [2024-03-22 03:58:26,200] [INFO] [launch.py:347:main] Process 53905 exits successfully.
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": true,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "32000": {
30
+ "content": "<pad>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ }
37
+ },
38
+ "bos_token": "<s>",
39
+ "clean_up_tokenization_spaces": false,
40
+ "eos_token": "</s>",
41
+ "legacy": false,
42
+ "model_max_length": 512,
43
+ "pad_token": "<pad>",
44
+ "padding_side": "right",
45
+ "sp_model_kwargs": {},
46
+ "spaces_between_special_tokens": false,
47
+ "tokenizer_class": "LlamaTokenizer",
48
+ "unk_token": "<unk>",
49
+ "use_default_system_prompt": false
50
+ }