CreatorPhan commited on
Commit
85811b0
1 Parent(s): 00edb2b

Upload folder using huggingface_hub (#3)

Browse files

- Upload folder using huggingface_hub (7f1b2223b2c7866dea608173d68c9d100944c728)

.gitattributes CHANGED
@@ -36,3 +36,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
36
  checkpoint-100/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
  checkpoint-200/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
  checkpoint-300/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
36
  checkpoint-100/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
  checkpoint-200/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
  checkpoint-300/tokenizer.json filter=lfs diff=lfs merge=lfs -text
39
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - quant_method: bitsandbytes
9
+ - load_in_8bit: True
10
+ - load_in_4bit: False
11
+ - llm_int8_threshold: 6.0
12
+ - llm_int8_skip_modules: None
13
+ - llm_int8_enable_fp32_cpu_offload: False
14
+ - llm_int8_has_fp16_weight: False
15
+ - bnb_4bit_quant_type: fp4
16
+ - bnb_4bit_use_double_quant: False
17
+ - bnb_4bit_compute_dtype: float32
18
+ ### Framework versions
19
+
20
+
21
+ - PEFT 0.6.0.dev0
adapter_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "bigscience/bloomz-3b",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.0,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 8,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "dense_4h_to_h",
18
+ "dense",
19
+ "dense_h_to_4h",
20
+ "query_key_value"
21
+ ],
22
+ "task_type": "CAUSAL_LM"
23
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:016a032c3fe5f1706b50723dd640e36a0a548f1b1197cc106cd3cb9aa275ba45
3
+ size 39409357
runs/Sep14_06-43-20_ThanhHa/events.out.tfevents.1694648632.ThanhHa.406390.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8776809a214993a71ddb9150bfe880ddb8d58df175a7608e3b6a99f31bdece61
3
- size 51664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e3b7132e70a547c94a98a4a8e17ac89faa12cf990d5e2644e3b531438f15f24
3
+ size 60025
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "<pad>",
5
+ "unk_token": "<unk>"
6
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85b00d7db4df5df2e3f01cacc3feda246002a672f3356eec7f4b04a22eb0dfbe
3
+ size 14500570
tokenizer_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<s>",
4
+ "clean_up_tokenization_spaces": false,
5
+ "eos_token": "</s>",
6
+ "model_max_length": 1000000000000000019884624838656,
7
+ "pad_token": "<pad>",
8
+ "tokenizer_class": "BloomTokenizer",
9
+ "unk_token": "<unk>"
10
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a5b74f951ff7697f1beb8eddc25d8ee6bd52b872686edb3386a0f175a1ad4d7
3
+ size 4155