DrishtiSharma commited on
Commit
7d888d8
1 Parent(s): 1960185

Upload folder using huggingface_hub

Browse files
checkpoint-500/README.md ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - quant_method: gptq
9
+ - bits: 4
10
+ - tokenizer: None
11
+ - dataset: None
12
+ - group_size: 128
13
+ - damp_percent: 0.01
14
+ - desc_act: False
15
+ - sym: True
16
+ - true_sequential: True
17
+ - use_cuda_fp16: False
18
+ - model_seqlen: None
19
+ - block_name_to_quantize: None
20
+ - module_name_preceding_first_block: None
21
+ - batch_size: 1
22
+ - pad_token_id: None
23
+ - disable_exllama: True
24
+ ### Framework versions
25
+
26
+
27
+ - PEFT 0.5.0
checkpoint-500/adapter_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "TheBloke/Llama-2-7b-Chat-GPTQ",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.1,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 8,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "k_proj",
18
+ "o_proj",
19
+ "q_proj",
20
+ "v_proj"
21
+ ],
22
+ "task_type": "CAUSAL_LM"
23
+ }
checkpoint-500/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36c7451b995bfe9350610c624a2f89e9dee085ce2d8b7abfabbf631815e2fb1a
3
+ size 33646413
checkpoint-500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:774cad98d0d38ffa9250377b57c466017faa962a552893d80900a84f444ca92a
3
+ size 67259845
checkpoint-500/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a96d8cb2dc19b877b7f6b77e9146065542979479699d9292b595ece4f5ae13a
3
+ size 14575
checkpoint-500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aca4bc1103f34868d9bf2a4d0332b82cb43f74c39d8828ced8a54bd609e4dab8
3
+ size 627
checkpoint-500/trainer_state.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.24925224327018944,
5
+ "eval_steps": 800,
6
+ "global_step": 500,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [],
11
+ "logging_steps": 800,
12
+ "max_steps": 2400,
13
+ "num_train_epochs": 2,
14
+ "save_steps": 500,
15
+ "total_flos": 21776429260800.0,
16
+ "trial_name": null,
17
+ "trial_params": null
18
+ }
checkpoint-500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26222bca74034b96f46df793570ab3358c1e63077abf7a6fb8549b5a0e2f4bfa
3
+ size 3963
runs/Sep13_21-16-45_b04c78a5e2b1/events.out.tfevents.1694639806.b04c78a5e2b1.723.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:755e73b4cc8a6b8f791f4e5128b195150d99b9346121d839b4496ebc695e86f5
3
+ size 4184
runs/Sep13_21-19-02_b04c78a5e2b1/events.out.tfevents.1694639943.b04c78a5e2b1.5549.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33d1239ad6b2d11019e807b6b60d8cfddf205390edfdde123d47fb7a650b6ca4
3
+ size 4552
runs/Sep13_21-22-11_b04c78a5e2b1/events.out.tfevents.1694640132.b04c78a5e2b1.5549.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ed71c6c08d707010d4640840563a0a807c3788e880aeb7bc1380b6b732bd5bb
3
+ size 4552
runs/Sep13_21-23-09_b04c78a5e2b1/events.out.tfevents.1694640189.b04c78a5e2b1.6813.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:466dff0d1ad258b285329301a51424bbdce494e6efd379dd76f646ba88827a37
3
+ size 4980
runs/Sep13_21-30-53_b04c78a5e2b1/events.out.tfevents.1694640653.b04c78a5e2b1.8924.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd4b462e7f8befa893ff76ca8f3d399467a8de5048f1c6a7a1710fa834b2caf6
3
+ size 6146
runs/Sep13_21-40-28_b04c78a5e2b1/events.out.tfevents.1694641228.b04c78a5e2b1.11505.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:200fb0a9c2239679ae940390ea255a533bfcac30f417cbfea4d96f79f96af49a
3
+ size 5123
runs/Sep13_21-42-45_b04c78a5e2b1/events.out.tfevents.1694641365.b04c78a5e2b1.12272.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b339f77be757ab5df966921391cec056512992e311ee188487b733c5ed92dc66
3
+ size 6997
runs/Sep13_22-00-30_b04c78a5e2b1/events.out.tfevents.1694642430.b04c78a5e2b1.16941.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73462bdcd10bf053e0057fc1ec1fee1ede9e18834c53e5db7e7dbca2e74f9918
3
+ size 6997
runs/Sep13_22-13-51_b04c78a5e2b1/events.out.tfevents.1694643232.b04c78a5e2b1.20474.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e76f8e778aaeb69f6d4110042c470fa78ba296fbc505c4ebc38b86efdbf1cb8
3
+ size 6997