thorirhrafn commited on
Commit
7c1d8a7
·
verified ·
1 Parent(s): 5def376

Training in progress, epoch 1

Browse files
adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "AI-Sweden-Models/gpt-sw3-1.3b",
5
+ "bias": "none",
6
+ "fan_in_fan_out": true,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layers_pattern": null,
10
+ "layers_to_transform": null,
11
+ "loftq_config": {},
12
+ "lora_alpha": 64,
13
+ "lora_dropout": 0.1,
14
+ "megatron_config": null,
15
+ "megatron_core": "megatron.core",
16
+ "modules_to_save": null,
17
+ "peft_type": "LORA",
18
+ "r": 64,
19
+ "rank_pattern": {},
20
+ "revision": null,
21
+ "target_modules": [
22
+ "c_attn"
23
+ ],
24
+ "task_type": "CAUSAL_LM",
25
+ "use_rslora": false
26
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7915387d7413fc59c0214ce069ad7ae984ed1f10a06ca9eefc67c89dcf9c5d17
3
+ size 25172136
runs/Mar23_16-20-53_gpu-3/events.out.tfevents.1711210854.gpu-3.535238.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:684709459c81de8c79e4ecd046ccab67c85ed0fedb7c3dd46ca447af7d767600
3
+ size 4876
runs/Mar23_16-43-57_gpu-3/events.out.tfevents.1711212238.gpu-3.535967.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:770d5ab942b243101cf31c04e15840e431b6190f377c9cdafc8d8ec902caecc7
3
+ size 15343
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f0889590f306d62b905ead5d5b5df68ade0c0fd0b774558c9f9eaf8bdb0507b
3
+ size 4920