jonathanjordan21 commited on
Commit
4fce4e1
·
verified ·
1 Parent(s): 03f063f

jonathanjordan21/Qwen2-0.5-Instruct-lora-resume-parser

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
- base_model: facebook/opt-125m
3
  library_name: peft
4
- license: other
5
  tags:
6
  - generated_from_trainer
7
  model-index:
@@ -14,7 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # tests
16
 
17
- This model is a fine-tuned version of [facebook/opt-125m](https://huggingface.co/facebook/opt-125m) on an unknown dataset.
18
 
19
  ## Model description
20
 
@@ -34,8 +34,8 @@ More information needed
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 3e-05
37
- - train_batch_size: 16
38
- - eval_batch_size: 16
39
  - seed: 42
40
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
41
  - lr_scheduler_type: linear
@@ -51,4 +51,4 @@ The following hyperparameters were used during training:
51
  - Transformers 4.46.0
52
  - Pytorch 2.4.0
53
  - Datasets 3.0.2
54
- - Tokenizers 0.20.1
 
1
  ---
2
+ base_model: Qwen/Qwen2-0.5B-Instruct
3
  library_name: peft
4
+ license: apache-2.0
5
  tags:
6
  - generated_from_trainer
7
  model-index:
 
14
 
15
  # tests
16
 
17
+ This model is a fine-tuned version of [Qwen/Qwen2-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2-0.5B-Instruct) on an unknown dataset.
18
 
19
  ## Model description
20
 
 
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 3e-05
37
+ - train_batch_size: 8
38
+ - eval_batch_size: 8
39
  - seed: 42
40
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
41
  - lr_scheduler_type: linear
 
51
  - Transformers 4.46.0
52
  - Pytorch 2.4.0
53
  - Datasets 3.0.2
54
+ - Tokenizers 0.20.0
adapter_config.json CHANGED
@@ -1,27 +1,27 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "facebook/opt-125m",
5
  "bias": "none",
6
- "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
  "layer_replication": null,
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 128,
14
  "lora_dropout": 0.0,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 128,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
5
  "bias": "none",
6
+ "fan_in_fan_out": true,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
  "layer_replication": null,
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
+ "lora_alpha": 64,
14
  "lora_dropout": 0.0,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 64,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7d14a636d5e7c132ea188310f59edb150e8b1a8b7e4d0f408a4bb9a223e3ab8
3
- size 18881176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4bce6cf735c6214f8abd9151d0384cf663cbb1f417a8369780eb6fc473055fd
3
+ size 17314248
added_tokens.json CHANGED
@@ -1,4 +1,5 @@
1
  {
2
- "<|im_end|>": 50266,
3
- "<|im_start|>": 50265
 
4
  }
 
1
  {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
  }
merges.txt CHANGED
The diff for this file is too large to render. See raw diff
 
runs/Oct29_02-11-40_0963431b98ff/events.out.tfevents.1730167902.0963431b98ff.30.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83d254757a29ba836899abac1a85faa453c4239caef211d40b6497fc5abc48c1
3
+ size 28890
runs/Oct29_02-33-01_0963431b98ff/events.out.tfevents.1730169189.0963431b98ff.30.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e781be0ead8af6c1facf562f2522c569ff901feeac6a64aa3df52780a23cc7e
3
+ size 18790
runs/Oct29_02-41-52_0963431b98ff/events.out.tfevents.1730169720.0963431b98ff.30.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87d4538be8741a9a9a7a3832df73b7b1f4bad4a3201e8bf40b38749689f992c0
3
+ size 33070
special_tokens_map.json CHANGED
@@ -1,23 +1,19 @@
1
  {
2
- "bos_token": {
3
- "content": "</s>",
 
 
 
 
4
  "lstrip": false,
5
- "normalized": true,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
- "eos_token": "<|im_end|>",
10
  "pad_token": {
11
- "content": "<pad>",
12
  "lstrip": false,
13
- "normalized": true,
14
- "rstrip": false,
15
- "single_word": false
16
- },
17
- "unk_token": {
18
- "content": "</s>",
19
- "lstrip": false,
20
- "normalized": true,
21
  "rstrip": false,
22
  "single_word": false
23
  }
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
  "lstrip": false,
9
+ "normalized": false,
10
  "rstrip": false,
11
  "single_word": false
12
  },
 
13
  "pad_token": {
14
+ "content": "<|endoftext|>",
15
  "lstrip": false,
16
+ "normalized": false,
 
 
 
 
 
 
 
17
  "rstrip": false,
18
  "single_word": false
19
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,24 +1,15 @@
1
  {
2
- "add_bos_token": true,
3
  "add_prefix_space": false,
4
  "added_tokens_decoder": {
5
- "1": {
6
- "content": "<pad>",
7
  "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false,
11
- "special": true
12
- },
13
- "2": {
14
- "content": "</s>",
15
- "lstrip": false,
16
- "normalized": true,
17
  "rstrip": false,
18
  "single_word": false,
19
  "special": true
20
  },
21
- "50265": {
22
  "content": "<|im_start|>",
23
  "lstrip": false,
24
  "normalized": false,
@@ -26,7 +17,7 @@
26
  "single_word": false,
27
  "special": true
28
  },
29
- "50266": {
30
  "content": "<|im_end|>",
31
  "lstrip": false,
32
  "normalized": false,
@@ -35,12 +26,18 @@
35
  "special": true
36
  }
37
  },
38
- "bos_token": "</s>",
 
 
 
 
 
39
  "clean_up_tokenization_spaces": false,
40
  "eos_token": "<|im_end|>",
41
  "errors": "replace",
42
- "model_max_length": 1000000000000000019884624838656,
43
- "pad_token": "<pad>",
44
- "tokenizer_class": "GPT2Tokenizer",
45
- "unk_token": "</s>"
 
46
  }
 
1
  {
 
2
  "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
  "lstrip": false,
7
+ "normalized": false,
 
 
 
 
 
 
 
 
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
  },
12
+ "151644": {
13
  "content": "<|im_start|>",
14
  "lstrip": false,
15
  "normalized": false,
 
17
  "single_word": false,
18
  "special": true
19
  },
20
+ "151645": {
21
  "content": "<|im_end|>",
22
  "lstrip": false,
23
  "normalized": false,
 
26
  "special": true
27
  }
28
  },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
  "clean_up_tokenization_spaces": false,
36
  "eos_token": "<|im_end|>",
37
  "errors": "replace",
38
+ "model_max_length": 32768,
39
+ "pad_token": "<|endoftext|>",
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "Qwen2Tokenizer",
42
+ "unk_token": null
43
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9dc1961b8a04313ec34aa7e4cdc585407c47e21b6f5800453213bee4ec9a9910
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a82dd9a2b1953cb5dc477104b27370197c5c4b9af6a4643d9844b820cffa05f4
3
  size 5240
vocab.json CHANGED
The diff for this file is too large to render. See raw diff