mjschock commited on
Commit
b553cd8
1 Parent(s): 6439f50

Training in progress, epoch 0

Browse files
adapter_config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mjschock/TinyLlama-1.1B-Chat-v1.0",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 16,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 8,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "v_proj",
24
+ "o_proj",
25
+ "k_proj",
26
+ "gate_proj",
27
+ "up_proj",
28
+ "lm_head",
29
+ "down_proj",
30
+ "q_proj"
31
+ ],
32
+ "task_type": "CAUSAL_LM",
33
+ "use_dora": false,
34
+ "use_rslora": false
35
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b375a6f210c234a37e4348e80b6cfcd72bd458fc9daef235a17e2d7aa1839604
3
+ size 288506496
special_tokens_map.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<s>",
4
+ "</s>",
5
+ "</s>"
6
+ ],
7
+ "bos_token": {
8
+ "content": "<s>",
9
+ "lstrip": false,
10
+ "normalized": false,
11
+ "rstrip": false,
12
+ "single_word": false
13
+ },
14
+ "eos_token": {
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "pad_token": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false
27
+ },
28
+ "unk_token": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ }
35
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "additional_special_tokens": [
32
+ "<s>",
33
+ "</s>",
34
+ "</s>"
35
+ ],
36
+ "bos_token": "<s>",
37
+ "chat_template": "{%- set system_message_present = messages | selectattr('role', 'equalto', 'system') | list -%}{%- if not system_message_present -%}{%- set messages = [{ \"content\": \"You are an AI agent acting as a human assistant.\", \"role\": \"system\" }] + messages -%}{%- endif -%}{%- for message in messages -%}<|{{ message.role }}|>{{ '\n' }}{%- if message.role == 'system' -%}{{ message.content }}{%- if tools and tools | length > 0 -%}{{ '\n\n' }}You are aware of the following tools in your environment:{{ '\n' }}{\n \"tools\": [{{ '\n' }}{%- for tool in tools -%}{{ ' ' }}{\n \"function\": {\n \"description\": \"{{ tool.function.description }}\",{{ '\n' }} \"name\": \"{{ tool.function.name }}\",{{ '\n' }} \"parameters\": {{ tool.function.parameters | tojson }}{{ '\n' }} },{{ '\n' }} \"type\": \"{{ tool.type }}\"{{ '\n' }} }{%- if not loop.last -%},{%- endif -%}{{ '\n' }}{%- endfor -%}{{ ' ' }}]{{ '\n' }}}{{ '\n\n' }}If you would like to suggest one or more tool calls, please respond in the following format:{{ '\n' }}{\n \"finish_reason\": \"tool_calls\",{{ '\n' }} \"tool_calls\": [{{ '\n' }}{{ ' ' }}{\n \"arguments\": \"{\\\"parameter_name\\\": \\\"parameter_value\\\"}\",{{ '\n' }} \"id\": \"call_id\",{{ '\n' }} \"name\": \"tool_name\"{{ '\n' }} }{{ '\n' }} ]{{ '\n' }}}{%- endif -%}{{ eos_token }}{{ '\n' }}{%- endif -%}{%- if message.role == 'user' -%}{{ message.content }}{{ eos_token }}{{ '\n' }}{%- endif -%}{%- if message.role == 'assistant' -%}{% generation %}{%- if message.tool_calls and message.tool_calls | length > 0 -%}{\n \"finish_reason\": \"tool_calls\",{{ '\n' }} \"tool_calls\": [{{ '\n' }}{%- for tool_call in message.tool_calls -%}{{ ' ' }}{\n \"arguments\": {{ tool_call.function.arguments | tojson }},{{ '\n' }} \"id\": \"{{ tool_call.id }}\",{{ '\n' }} \"name\": \"{{ tool_call.function.name }}\"{{ '\n' }} }{%- if not loop.last -%},{%- endif -%}{{ '\n' }}{%- endfor -%}{{ ' ' }}]{{ '\n' }}}{%- else -%}{{ message.content }}{%- endif -%}{% endgeneration %}{{ eos_token }}{{ '\n' }}{%- endif -%}{%- if message.role == 'tool' -%}{\n \"content\": {{ message.content | tojson }},{{ '\n' }} \"name\": \"{{ message.name }}\",{{ '\n' }} \"tool_call_id\": \"{{ message.tool_call_id }}\"{{ '\n' }}}{{ eos_token }}{{ '\n' }}{%- endif -%}{%- endfor -%}{%- if add_generation_prompt -%}{{ '<|assistant|>\n' }}{%- endif -%}",
38
+ "clean_up_tokenization_spaces": false,
39
+ "eos_token": "</s>",
40
+ "legacy": false,
41
+ "model_max_length": 2048,
42
+ "pad_token": "</s>",
43
+ "padding_side": "right",
44
+ "sp_model_kwargs": {},
45
+ "tokenizer_class": "LlamaTokenizer",
46
+ "unk_token": "<unk>",
47
+ "use_default_system_prompt": false
48
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c61f4c9a59c5102d97c5b61c6b0600015fb7de2a9846c7ecd9f005c3024e455d
3
+ size 5560