mlabonne commited on
Commit
cb62953
1 Parent(s): eb01191

Model save

Browse files
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "out_proj",
24
  "x_proj",
25
- "in_proj",
26
- "embed_tokens"
 
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "x_proj",
24
+ "embed_tokens",
25
+ "out_proj",
26
+ "in_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f7fc9d3c9e0f41d7a28caaceac1611ac6015913a5c2a88f33c6ae6eaea106d02
3
  size 689182824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd5ae5724001cc5856a815d9d48243c56d392d289a5a6faa32936d4ad1f17bd6
3
  size 689182824
tokenizer_config.json CHANGED
@@ -36,6 +36,7 @@
36
  }
37
  },
38
  "bos_token": "<|startoftext|>",
 
39
  "clean_up_tokenization_spaces": false,
40
  "eos_token": "<|endoftext|>",
41
  "model_max_length": 1000000000000000019884624838656,
 
36
  }
37
  },
38
  "bos_token": "<|startoftext|>",
39
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
40
  "clean_up_tokenization_spaces": false,
41
  "eos_token": "<|endoftext|>",
42
  "model_max_length": 1000000000000000019884624838656,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed3db33f1e9fe50127deb7c2e59d8acc4178c7e5907b4ff701784a0bc5a21838
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97c4606a721101b7a29391cf91cfe447c44983f096524a37e2ecedb01294365e
3
  size 4984