Ubuntu commited on
Commit
3c24553
1 Parent(s): 9542f90

Signed-off-by: Ubuntu <ubuntu@ip-172-31-15-132.us-west-2.compute.internal>

.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.gguf filter=lfs diff=lfs merge=lfs -text
checkpoint-10.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:978b48c1488d6b52d143c6ac2dcaee78be4aadd0d17f92ee59d78eed52f9f22e
3
+ size 126642016
checkpoint-100.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d7494dd9e039856295dc31ee89507164ecf9a09f8fd1a772bd467714fb4404e
3
+ size 126642016
checkpoint-110.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73b1c58f7d22ed93d60d271726280a648c0878c54525b6d8684e90295b52bb1c
3
+ size 126642016
checkpoint-120.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82047db2dda450d70ad4ca558b2ecb15d3179898cc88e568d7b7a39819d34826
3
+ size 126642016
checkpoint-130.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8415a1c068881e6e0912a4305993d7daef45e84aaa6b50ca48cb1af1d31783c4
3
+ size 126642016
checkpoint-140.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7185babf6c0ab613e148f3b0a85206f3e4a0307d10e20395f919b3266e37065c
3
+ size 126642016
checkpoint-150.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eba99a785f0d338cbb8acba7ab28eb1fb3b9b446c01d8efce6eacb1e4338f488
3
+ size 126642016
checkpoint-160.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:053df254643ce2103a9217216bfc3cb41e47005375e18074f0612dc3c11a0578
3
+ size 126642016
checkpoint-170.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d622558933bd9778fe584e4db13b02c7eb83c2417daf00233f169a0d6271d76b
3
+ size 126642016
checkpoint-180.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6341c30c17bf2337ad93aee13fa766d53daa43e2df4962d94021357864d54837
3
+ size 126642016
checkpoint-190.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b47949e1f2ec31718294ab0a5c15535b4deff8e91f6a3b40e2c80c41aa2817b
3
+ size 126642016
checkpoint-20.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ae919be7297353b12c45d5d921a9ce6d70c5707148a8f169bb70f0deb708af7
3
+ size 126642016
checkpoint-200.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7e5c09b66a18df03d366ab571188e850758e90b3269bd44547817a924d24561
3
+ size 126642016
checkpoint-210.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db97333bb379958b891e7e921722bf74f0def271ea1ad774cf713a92d1113b31
3
+ size 126642016
checkpoint-220.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15496c29b6f352176f059eb1a49b83d77ce2b70b6d0ab196dfe52e9315ae1fba
3
+ size 126642016
checkpoint-230.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:611bc9c4fa64499fd47f7d4bd55b8a41304e74dbba38b656277deee045872cb4
3
+ size 126642016
checkpoint-240.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:431523aa73f4b4ae6564c05e132ce80b9d8cf7effe3dec95bed05cfdf12b8b0b
3
+ size 126642016
checkpoint-250.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:815fe0472d0701309dcb3824da9b97c1f624e9158db51014370ec6e062ddfb0f
3
+ size 126642016
checkpoint-256.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7354976ff9a88d31b92a799cb5b50c408a72fabbb9810d4357fee095571114fa
3
+ size 126642016
checkpoint-30.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ba334bba4e218dbb17606af8c7fe02d87c8fbafc6e9840e6a553757a7cf6ab9
3
+ size 126642016
checkpoint-40.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e338898f988a27563c2536e93a4e062a6f76f82e27c3d66da15b97fe8544deb
3
+ size 126642016
checkpoint-50.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0e4517b6af658df5ad2939f0a12e3a476e329495d88623e38677a31cb7a281f
3
+ size 126642016
checkpoint-60.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc6b291848357ebf27375dfb955c4179002b9e88b607fa6b9e82c8f6ba762be5
3
+ size 126642016
checkpoint-70.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3eb4060d040dbb308193e80820918757376fed3fc660f5a3b91b652d806e0de9
3
+ size 126642016
checkpoint-80.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:736ae2198dd412c2a87d7e346daca27d82c16603d9289d7ee62ab731339344e4
3
+ size 126642016
checkpoint-90.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff6b32e8510f53b3c12fa8490b8173c6e5445499892fd594f97c7902d7caa60a
3
+ size 126642016
checkpoint-LATEST.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7354976ff9a88d31b92a799cb5b50c408a72fabbb9810d4357fee095571114fa
3
+ size 126642016
llama-2-7b-chat-ft-q5_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:640592da101bcf668f83c90b054d62e26afdf26bbdbedabcee2f9a127cf58a11
3
+ size 4783156800
llama-2-7b-chat-q5_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34acb8c183aeb92be6fa4c63b3911ddcef6c1c52d4cdfd90a38dd7642546c1a9
3
+ size 4783156800
lora.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6125d593d4380e5608703c6a83325d26e9c50bf43e95bc454c679517233b02cc
3
+ size 42234240
nohup.out ADDED
@@ -0,0 +1,771 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ggml_init_cublas: found 1 CUDA devices:
2
+ Device 0: NVIDIA A10G, compute capability 8.6
3
+ llama_model_loader: loaded meta data with 16 key-value pairs and 291 tensors from llama-2-7b-chat-q5_k_m.gguf (version GGUF V2 (latest))
4
+ llama_model_loader: - tensor 0: token_embd.weight q5_K [ 4096, 32000, 1, 1 ]
5
+ llama_model_loader: - tensor 1: output_norm.weight f32 [ 4096, 1, 1, 1 ]
6
+ llama_model_loader: - tensor 2: output.weight q6_K [ 4096, 32000, 1, 1 ]
7
+ llama_model_loader: - tensor 3: blk.0.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
8
+ llama_model_loader: - tensor 4: blk.0.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
9
+ llama_model_loader: - tensor 5: blk.0.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
10
+ llama_model_loader: - tensor 6: blk.0.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
11
+ llama_model_loader: - tensor 7: blk.0.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
12
+ llama_model_loader: - tensor 8: blk.0.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
13
+ llama_model_loader: - tensor 9: blk.0.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
14
+ llama_model_loader: - tensor 10: blk.0.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
15
+ llama_model_loader: - tensor 11: blk.0.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
16
+ llama_model_loader: - tensor 12: blk.1.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
17
+ llama_model_loader: - tensor 13: blk.1.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
18
+ llama_model_loader: - tensor 14: blk.1.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
19
+ llama_model_loader: - tensor 15: blk.1.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
20
+ llama_model_loader: - tensor 16: blk.1.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
21
+ llama_model_loader: - tensor 17: blk.1.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
22
+ llama_model_loader: - tensor 18: blk.1.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
23
+ llama_model_loader: - tensor 19: blk.1.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
24
+ llama_model_loader: - tensor 20: blk.1.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
25
+ llama_model_loader: - tensor 21: blk.2.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
26
+ llama_model_loader: - tensor 22: blk.2.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
27
+ llama_model_loader: - tensor 23: blk.2.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
28
+ llama_model_loader: - tensor 24: blk.2.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
29
+ llama_model_loader: - tensor 25: blk.2.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
30
+ llama_model_loader: - tensor 26: blk.2.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
31
+ llama_model_loader: - tensor 27: blk.2.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
32
+ llama_model_loader: - tensor 28: blk.2.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
33
+ llama_model_loader: - tensor 29: blk.2.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
34
+ llama_model_loader: - tensor 30: blk.3.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
35
+ llama_model_loader: - tensor 31: blk.3.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
36
+ llama_model_loader: - tensor 32: blk.3.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
37
+ llama_model_loader: - tensor 33: blk.3.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
38
+ llama_model_loader: - tensor 34: blk.3.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
39
+ llama_model_loader: - tensor 35: blk.3.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
40
+ llama_model_loader: - tensor 36: blk.3.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
41
+ llama_model_loader: - tensor 37: blk.3.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
42
+ llama_model_loader: - tensor 38: blk.3.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
43
+ llama_model_loader: - tensor 39: blk.4.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
44
+ llama_model_loader: - tensor 40: blk.4.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
45
+ llama_model_loader: - tensor 41: blk.4.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
46
+ llama_model_loader: - tensor 42: blk.4.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
47
+ llama_model_loader: - tensor 43: blk.4.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
48
+ llama_model_loader: - tensor 44: blk.4.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
49
+ llama_model_loader: - tensor 45: blk.4.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
50
+ llama_model_loader: - tensor 46: blk.4.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
51
+ llama_model_loader: - tensor 47: blk.4.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
52
+ llama_model_loader: - tensor 48: blk.5.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
53
+ llama_model_loader: - tensor 49: blk.5.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
54
+ llama_model_loader: - tensor 50: blk.5.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
55
+ llama_model_loader: - tensor 51: blk.5.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
56
+ llama_model_loader: - tensor 52: blk.5.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
57
+ llama_model_loader: - tensor 53: blk.5.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
58
+ llama_model_loader: - tensor 54: blk.5.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
59
+ llama_model_loader: - tensor 55: blk.5.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
60
+ llama_model_loader: - tensor 56: blk.5.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
61
+ llama_model_loader: - tensor 57: blk.6.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
62
+ llama_model_loader: - tensor 58: blk.6.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
63
+ llama_model_loader: - tensor 59: blk.6.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
64
+ llama_model_loader: - tensor 60: blk.6.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
65
+ llama_model_loader: - tensor 61: blk.6.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
66
+ llama_model_loader: - tensor 62: blk.6.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
67
+ llama_model_loader: - tensor 63: blk.6.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
68
+ llama_model_loader: - tensor 64: blk.6.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
69
+ llama_model_loader: - tensor 65: blk.6.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
70
+ llama_model_loader: - tensor 66: blk.7.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
71
+ llama_model_loader: - tensor 67: blk.7.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
72
+ llama_model_loader: - tensor 68: blk.7.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
73
+ llama_model_loader: - tensor 69: blk.7.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
74
+ llama_model_loader: - tensor 70: blk.7.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
75
+ llama_model_loader: - tensor 71: blk.7.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
76
+ llama_model_loader: - tensor 72: blk.7.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
77
+ llama_model_loader: - tensor 73: blk.7.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
78
+ llama_model_loader: - tensor 74: blk.7.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
79
+ llama_model_loader: - tensor 75: blk.8.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
80
+ llama_model_loader: - tensor 76: blk.8.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
81
+ llama_model_loader: - tensor 77: blk.8.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
82
+ llama_model_loader: - tensor 78: blk.8.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
83
+ llama_model_loader: - tensor 79: blk.8.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
84
+ llama_model_loader: - tensor 80: blk.8.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
85
+ llama_model_loader: - tensor 81: blk.8.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
86
+ llama_model_loader: - tensor 82: blk.8.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
87
+ llama_model_loader: - tensor 83: blk.8.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
88
+ llama_model_loader: - tensor 84: blk.9.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
89
+ llama_model_loader: - tensor 85: blk.9.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
90
+ llama_model_loader: - tensor 86: blk.9.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
91
+ llama_model_loader: - tensor 87: blk.9.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
92
+ llama_model_loader: - tensor 88: blk.9.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
93
+ llama_model_loader: - tensor 89: blk.9.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
94
+ llama_model_loader: - tensor 90: blk.9.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
95
+ llama_model_loader: - tensor 91: blk.9.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
96
+ llama_model_loader: - tensor 92: blk.9.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
97
+ llama_model_loader: - tensor 93: blk.10.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
98
+ llama_model_loader: - tensor 94: blk.10.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
99
+ llama_model_loader: - tensor 95: blk.10.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
100
+ llama_model_loader: - tensor 96: blk.10.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
101
+ llama_model_loader: - tensor 97: blk.10.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
102
+ llama_model_loader: - tensor 98: blk.10.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
103
+ llama_model_loader: - tensor 99: blk.10.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
104
+ llama_model_loader: - tensor 100: blk.10.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
105
+ llama_model_loader: - tensor 101: blk.10.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
106
+ llama_model_loader: - tensor 102: blk.11.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
107
+ llama_model_loader: - tensor 103: blk.11.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
108
+ llama_model_loader: - tensor 104: blk.11.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
109
+ llama_model_loader: - tensor 105: blk.11.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
110
+ llama_model_loader: - tensor 106: blk.11.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
111
+ llama_model_loader: - tensor 107: blk.11.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
112
+ llama_model_loader: - tensor 108: blk.11.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
113
+ llama_model_loader: - tensor 109: blk.11.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
114
+ llama_model_loader: - tensor 110: blk.11.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
115
+ llama_model_loader: - tensor 111: blk.12.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
116
+ llama_model_loader: - tensor 112: blk.12.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
117
+ llama_model_loader: - tensor 113: blk.12.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
118
+ llama_model_loader: - tensor 114: blk.12.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
119
+ llama_model_loader: - tensor 115: blk.12.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
120
+ llama_model_loader: - tensor 116: blk.12.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
121
+ llama_model_loader: - tensor 117: blk.12.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
122
+ llama_model_loader: - tensor 118: blk.12.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
123
+ llama_model_loader: - tensor 119: blk.12.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
124
+ llama_model_loader: - tensor 120: blk.13.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
125
+ llama_model_loader: - tensor 121: blk.13.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
126
+ llama_model_loader: - tensor 122: blk.13.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
127
+ llama_model_loader: - tensor 123: blk.13.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
128
+ llama_model_loader: - tensor 124: blk.13.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
129
+ llama_model_loader: - tensor 125: blk.13.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
130
+ llama_model_loader: - tensor 126: blk.13.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
131
+ llama_model_loader: - tensor 127: blk.13.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
132
+ llama_model_loader: - tensor 128: blk.13.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
133
+ llama_model_loader: - tensor 129: blk.14.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
134
+ llama_model_loader: - tensor 130: blk.14.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
135
+ llama_model_loader: - tensor 131: blk.14.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
136
+ llama_model_loader: - tensor 132: blk.14.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
137
+ llama_model_loader: - tensor 133: blk.14.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
138
+ llama_model_loader: - tensor 134: blk.14.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
139
+ llama_model_loader: - tensor 135: blk.14.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
140
+ llama_model_loader: - tensor 136: blk.14.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
141
+ llama_model_loader: - tensor 137: blk.14.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
142
+ llama_model_loader: - tensor 138: blk.15.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
143
+ llama_model_loader: - tensor 139: blk.15.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
144
+ llama_model_loader: - tensor 140: blk.15.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
145
+ llama_model_loader: - tensor 141: blk.15.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
146
+ llama_model_loader: - tensor 142: blk.15.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
147
+ llama_model_loader: - tensor 143: blk.15.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
148
+ llama_model_loader: - tensor 144: blk.15.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
149
+ llama_model_loader: - tensor 145: blk.15.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
150
+ llama_model_loader: - tensor 146: blk.15.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
151
+ llama_model_loader: - tensor 147: blk.16.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
152
+ llama_model_loader: - tensor 148: blk.16.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
153
+ llama_model_loader: - tensor 149: blk.16.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
154
+ llama_model_loader: - tensor 150: blk.16.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
155
+ llama_model_loader: - tensor 151: blk.16.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
156
+ llama_model_loader: - tensor 152: blk.16.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
157
+ llama_model_loader: - tensor 153: blk.16.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
158
+ llama_model_loader: - tensor 154: blk.16.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
159
+ llama_model_loader: - tensor 155: blk.16.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
160
+ llama_model_loader: - tensor 156: blk.17.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
161
+ llama_model_loader: - tensor 157: blk.17.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
162
+ llama_model_loader: - tensor 158: blk.17.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
163
+ llama_model_loader: - tensor 159: blk.17.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
164
+ llama_model_loader: - tensor 160: blk.17.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
165
+ llama_model_loader: - tensor 161: blk.17.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
166
+ llama_model_loader: - tensor 162: blk.17.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
167
+ llama_model_loader: - tensor 163: blk.17.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
168
+ llama_model_loader: - tensor 164: blk.17.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
169
+ llama_model_loader: - tensor 165: blk.18.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
170
+ llama_model_loader: - tensor 166: blk.18.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
171
+ llama_model_loader: - tensor 167: blk.18.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
172
+ llama_model_loader: - tensor 168: blk.18.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
173
+ llama_model_loader: - tensor 169: blk.18.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
174
+ llama_model_loader: - tensor 170: blk.18.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
175
+ llama_model_loader: - tensor 171: blk.18.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
176
+ llama_model_loader: - tensor 172: blk.18.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
177
+ llama_model_loader: - tensor 173: blk.18.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
178
+ llama_model_loader: - tensor 174: blk.19.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
179
+ llama_model_loader: - tensor 175: blk.19.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
180
+ llama_model_loader: - tensor 176: blk.19.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
181
+ llama_model_loader: - tensor 177: blk.19.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
182
+ llama_model_loader: - tensor 178: blk.19.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
183
+ llama_model_loader: - tensor 179: blk.19.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
184
+ llama_model_loader: - tensor 180: blk.19.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
185
+ llama_model_loader: - tensor 181: blk.19.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
186
+ llama_model_loader: - tensor 182: blk.19.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
187
+ llama_model_loader: - tensor 183: blk.20.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
188
+ llama_model_loader: - tensor 184: blk.20.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
189
+ llama_model_loader: - tensor 185: blk.20.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
190
+ llama_model_loader: - tensor 186: blk.20.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
191
+ llama_model_loader: - tensor 187: blk.20.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
192
+ llama_model_loader: - tensor 188: blk.20.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
193
+ llama_model_loader: - tensor 189: blk.20.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
194
+ llama_model_loader: - tensor 190: blk.20.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
195
+ llama_model_loader: - tensor 191: blk.20.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
196
+ llama_model_loader: - tensor 192: blk.21.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
197
+ llama_model_loader: - tensor 193: blk.21.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
198
+ llama_model_loader: - tensor 194: blk.21.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
199
+ llama_model_loader: - tensor 195: blk.21.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
200
+ llama_model_loader: - tensor 196: blk.21.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
201
+ llama_model_loader: - tensor 197: blk.21.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
202
+ llama_model_loader: - tensor 198: blk.21.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
203
+ llama_model_loader: - tensor 199: blk.21.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
204
+ llama_model_loader: - tensor 200: blk.21.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
205
+ llama_model_loader: - tensor 201: blk.22.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
206
+ llama_model_loader: - tensor 202: blk.22.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
207
+ llama_model_loader: - tensor 203: blk.22.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
208
+ llama_model_loader: - tensor 204: blk.22.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
209
+ llama_model_loader: - tensor 205: blk.22.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
210
+ llama_model_loader: - tensor 206: blk.22.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
211
+ llama_model_loader: - tensor 207: blk.22.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
212
+ llama_model_loader: - tensor 208: blk.22.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
213
+ llama_model_loader: - tensor 209: blk.22.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
214
+ llama_model_loader: - tensor 210: blk.23.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
215
+ llama_model_loader: - tensor 211: blk.23.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
216
+ llama_model_loader: - tensor 212: blk.23.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
217
+ llama_model_loader: - tensor 213: blk.23.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
218
+ llama_model_loader: - tensor 214: blk.23.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
219
+ llama_model_loader: - tensor 215: blk.23.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
220
+ llama_model_loader: - tensor 216: blk.23.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
221
+ llama_model_loader: - tensor 217: blk.23.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
222
+ llama_model_loader: - tensor 218: blk.23.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
223
+ llama_model_loader: - tensor 219: blk.24.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
224
+ llama_model_loader: - tensor 220: blk.24.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
225
+ llama_model_loader: - tensor 221: blk.24.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
226
+ llama_model_loader: - tensor 222: blk.24.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
227
+ llama_model_loader: - tensor 223: blk.24.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
228
+ llama_model_loader: - tensor 224: blk.24.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
229
+ llama_model_loader: - tensor 225: blk.24.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
230
+ llama_model_loader: - tensor 226: blk.24.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
231
+ llama_model_loader: - tensor 227: blk.24.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
232
+ llama_model_loader: - tensor 228: blk.25.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
233
+ llama_model_loader: - tensor 229: blk.25.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
234
+ llama_model_loader: - tensor 230: blk.25.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
235
+ llama_model_loader: - tensor 231: blk.25.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
236
+ llama_model_loader: - tensor 232: blk.25.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
237
+ llama_model_loader: - tensor 233: blk.25.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
238
+ llama_model_loader: - tensor 234: blk.25.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
239
+ llama_model_loader: - tensor 235: blk.25.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
240
+ llama_model_loader: - tensor 236: blk.25.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
241
+ llama_model_loader: - tensor 237: blk.26.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
242
+ llama_model_loader: - tensor 238: blk.26.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
243
+ llama_model_loader: - tensor 239: blk.26.attn_v.weight q5_K [ 4096, 4096, 1, 1 ]
244
+ llama_model_loader: - tensor 240: blk.26.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
245
+ llama_model_loader: - tensor 241: blk.26.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
246
+ llama_model_loader: - tensor 242: blk.26.ffn_down.weight q5_K [ 11008, 4096, 1, 1 ]
247
+ llama_model_loader: - tensor 243: blk.26.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
248
+ llama_model_loader: - tensor 244: blk.26.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
249
+ llama_model_loader: - tensor 245: blk.26.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
250
+ llama_model_loader: - tensor 246: blk.27.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
251
+ llama_model_loader: - tensor 247: blk.27.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
252
+ llama_model_loader: - tensor 248: blk.27.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
253
+ llama_model_loader: - tensor 249: blk.27.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
254
+ llama_model_loader: - tensor 250: blk.27.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
255
+ llama_model_loader: - tensor 251: blk.27.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
256
+ llama_model_loader: - tensor 252: blk.27.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
257
+ llama_model_loader: - tensor 253: blk.27.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
258
+ llama_model_loader: - tensor 254: blk.27.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
259
+ llama_model_loader: - tensor 255: blk.28.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
260
+ llama_model_loader: - tensor 256: blk.28.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
261
+ llama_model_loader: - tensor 257: blk.28.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
262
+ llama_model_loader: - tensor 258: blk.28.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
263
+ llama_model_loader: - tensor 259: blk.28.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
264
+ llama_model_loader: - tensor 260: blk.28.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
265
+ llama_model_loader: - tensor 261: blk.28.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
266
+ llama_model_loader: - tensor 262: blk.28.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
267
+ llama_model_loader: - tensor 263: blk.28.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
268
+ llama_model_loader: - tensor 264: blk.29.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
269
+ llama_model_loader: - tensor 265: blk.29.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
270
+ llama_model_loader: - tensor 266: blk.29.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
271
+ llama_model_loader: - tensor 267: blk.29.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
272
+ llama_model_loader: - tensor 268: blk.29.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
273
+ llama_model_loader: - tensor 269: blk.29.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
274
+ llama_model_loader: - tensor 270: blk.29.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
275
+ llama_model_loader: - tensor 271: blk.29.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
276
+ llama_model_loader: - tensor 272: blk.29.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
277
+ llama_model_loader: - tensor 273: blk.30.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
278
+ llama_model_loader: - tensor 274: blk.30.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
279
+ llama_model_loader: - tensor 275: blk.30.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
280
+ llama_model_loader: - tensor 276: blk.30.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
281
+ llama_model_loader: - tensor 277: blk.30.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
282
+ llama_model_loader: - tensor 278: blk.30.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
283
+ llama_model_loader: - tensor 279: blk.30.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
284
+ llama_model_loader: - tensor 280: blk.30.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
285
+ llama_model_loader: - tensor 281: blk.30.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
286
+ llama_model_loader: - tensor 282: blk.31.attn_q.weight q5_K [ 4096, 4096, 1, 1 ]
287
+ llama_model_loader: - tensor 283: blk.31.attn_k.weight q5_K [ 4096, 4096, 1, 1 ]
288
+ llama_model_loader: - tensor 284: blk.31.attn_v.weight q6_K [ 4096, 4096, 1, 1 ]
289
+ llama_model_loader: - tensor 285: blk.31.attn_output.weight q5_K [ 4096, 4096, 1, 1 ]
290
+ llama_model_loader: - tensor 286: blk.31.ffn_gate.weight q5_K [ 4096, 11008, 1, 1 ]
291
+ llama_model_loader: - tensor 287: blk.31.ffn_down.weight q6_K [ 11008, 4096, 1, 1 ]
292
+ llama_model_loader: - tensor 288: blk.31.ffn_up.weight q5_K [ 4096, 11008, 1, 1 ]
293
+ llama_model_loader: - tensor 289: blk.31.attn_norm.weight f32 [ 4096, 1, 1, 1 ]
294
+ llama_model_loader: - tensor 290: blk.31.ffn_norm.weight f32 [ 4096, 1, 1, 1 ]
295
+ llama_model_loader: - kv 0: general.architecture str
296
+ llama_model_loader: - kv 1: general.name str
297
+ llama_model_loader: - kv 2: llama.context_length u32
298
+ llama_model_loader: - kv 3: llama.embedding_length u32
299
+ llama_model_loader: - kv 4: llama.block_count u32
300
+ llama_model_loader: - kv 5: llama.feed_forward_length u32
301
+ llama_model_loader: - kv 6: llama.rope.dimension_count u32
302
+ llama_model_loader: - kv 7: llama.attention.head_count u32
303
+ llama_model_loader: - kv 8: llama.attention.head_count_kv u32
304
+ llama_model_loader: - kv 9: llama.attention.layer_norm_rms_epsilon f32
305
+ llama_model_loader: - kv 10: general.file_type u32
306
+ llama_model_loader: - kv 11: tokenizer.ggml.model str
307
+ llama_model_loader: - kv 12: tokenizer.ggml.tokens arr
308
+ llama_model_loader: - kv 13: tokenizer.ggml.scores arr
309
+ llama_model_loader: - kv 14: tokenizer.ggml.token_type arr
310
+ llama_model_loader: - kv 15: general.quantization_version u32
311
+ llama_model_loader: - type f32: 65 tensors
312
+ llama_model_loader: - type q5_K: 193 tensors
313
+ llama_model_loader: - type q6_K: 33 tensors
314
+ llm_load_print_meta: format = GGUF V2 (latest)
315
+ llm_load_print_meta: arch = llama
316
+ llm_load_print_meta: vocab type = SPM
317
+ llm_load_print_meta: n_vocab = 32000
318
+ llm_load_print_meta: n_merges = 0
319
+ llm_load_print_meta: n_ctx_train = 4096
320
+ llm_load_print_meta: n_embd = 4096
321
+ llm_load_print_meta: n_head = 32
322
+ llm_load_print_meta: n_head_kv = 32
323
+ llm_load_print_meta: n_layer = 32
324
+ llm_load_print_meta: n_rot = 128
325
+ llm_load_print_meta: n_gqa = 1
326
+ llm_load_print_meta: f_norm_eps = 0.0e+00
327
+ llm_load_print_meta: f_norm_rms_eps = 1.0e-05
328
+ llm_load_print_meta: f_clamp_kqv = 0.0e+00
329
+ llm_load_print_meta: f_max_alibi_bias = 0.0e+00
330
+ llm_load_print_meta: n_ff = 11008
331
+ llm_load_print_meta: freq_base_train = 10000.0
332
+ llm_load_print_meta: freq_scale_train = 1
333
+ llm_load_print_meta: model type = 7B
334
+ llm_load_print_meta: model ftype = mostly Q5_K - Medium
335
+ llm_load_print_meta: model params = 6.74 B
336
+ llm_load_print_meta: model size = 4.45 GiB (5.68 BPW)
337
+ llm_load_print_meta: general.name = LLaMA v2
338
+ llm_load_print_meta: BOS token = 1 '<s>'
339
+ llm_load_print_meta: EOS token = 2 '</s>'
340
+ llm_load_print_meta: UNK token = 0 '<unk>'
341
+ llm_load_print_meta: LF token = 13 '<0x0A>'
342
+ llm_load_tensors: ggml ctx size = 0.10 MB
343
+ llm_load_tensors: using CUDA for GPU acceleration
344
+ llm_load_tensors: mem required = 4560.96 MB
345
+ llm_load_tensors: offloading 0 repeating layers to GPU
346
+ llm_load_tensors: offloaded 0/35 layers to GPU
347
+ llm_load_tensors: VRAM used: 0.00 MB
348
+ ..................................................................................................
349
+ llama_new_context_with_model: n_ctx = 512
350
+ llama_new_context_with_model: freq_base = 10000.0
351
+ llama_new_context_with_model: freq_scale = 1
352
+ llama_new_context_with_model: kv self size = 256.00 MB
353
+ llama_new_context_with_model: compute buffer total size = 76.63 MB
354
+ llama_new_context_with_model: VRAM scratch buffer: 70.50 MB
355
+ llama_new_context_with_model: total VRAM used: 70.50 MB (model: 0.00 MB, context: 70.50 MB)
356
+ main: seed: 1700419101
357
+ main: model base = 'llama-2-7b-chat-q5_k_m.gguf'
358
+ main: init model
359
+ print_params: n_vocab: 32000
360
+ print_params: n_ctx: 128
361
+ print_params: n_embd: 4096
362
+ print_params: n_ff: 11008
363
+ print_params: n_head: 32
364
+ print_params: n_head_kv: 32
365
+ print_params: n_layer: 32
366
+ print_params: norm_rms_eps : 0.000010
367
+ print_params: rope_freq_base : 10000.000000
368
+ print_params: rope_freq_scale : 1.000000
369
+ print_lora_params: n_rank_attention_norm : 1
370
+ print_lora_params: n_rank_wq : 4
371
+ print_lora_params: n_rank_wk : 4
372
+ print_lora_params: n_rank_wv : 4
373
+ print_lora_params: n_rank_wo : 4
374
+ print_lora_params: n_rank_ffn_norm : 1
375
+ print_lora_params: n_rank_w1 : 4
376
+ print_lora_params: n_rank_w2 : 4
377
+ print_lora_params: n_rank_w3 : 4
378
+ print_lora_params: n_rank_tok_embeddings : 4
379
+ print_lora_params: n_rank_norm : 1
380
+ print_lora_params: n_rank_output : 4
381
+ main: total train_iterations 0
382
+ main: seen train_samples 0
383
+ main: seen train_tokens 0
384
+ main: completed train_epochs 0
385
+ main: lora_size = 84807904 bytes (80.9 MB)
386
+ main: opt_size = 126592864 bytes (120.7 MB)
387
+ main: opt iter 0
388
+ main: input_size = 131076128 bytes (125.0 MB)
389
+ main: compute_size = 14064566880 bytes (13413.0 MB)
390
+ main: evaluation order = RIGHT_TO_LEFT
391
+ main: tokenize training data
392
+ tokenize_file: warning: found 2 samples (max length 197) that exceed context length of 128. samples will be cut off.
393
+ tokenize_file: warning: found 190 samples (min length 22) that are shorter than context length of 128.
394
+ tokenize_file: total number of samples: 192
395
+ main: number of training tokens: 8099
396
+ main: number of unique tokens: 948
397
+ main: train data seems to have changed. restarting shuffled epoch.
398
+ main: begin training
399
+ main: work_size = 1024512 bytes (1.0 MB)
400
+ train_opt_callback: iter= 0 sample=1/192 sched=0.000000 loss=0.000000 |->
401
+ train_opt_callback: iter= 1 sample=9/192 sched=0.010000 loss=7.639391 dt=00:32:55 eta=5d 19:55:06 |->
402
+ train_opt_callback: iter= 2 sample=17/192 sched=0.020000 loss=8.123040 dt=00:32:58 eta=5d 19:35:29 |>
403
+ train_opt_callback: iter= 3 sample=25/192 sched=0.030000 loss=6.609226 dt=00:32:58 eta=5d 19:01:49 |----------->
404
+ train_opt_callback: iter= 4 sample=33/192 sched=0.040000 loss=7.152853 dt=00:32:56 eta=5d 18:21:48 |------>
405
+ train_opt_callback: iter= 5 sample=41/192 sched=0.050000 loss=6.951632 dt=00:32:51 eta=5d 17:29:11 |-------->
406
+ train_opt_callback: iter= 6 sample=49/192 sched=0.060000 loss=5.657753 dt=00:32:48 eta=5d 16:43:22 |--------------------->
407
+ train_opt_callback: iter= 7 sample=57/192 sched=0.070000 loss=4.757537 dt=00:32:51 eta=5d 16:23:32 |------------------------------>
408
+ train_opt_callback: iter= 8 sample=65/192 sched=0.080000 loss=4.564727 dt=00:32:43 eta=5d 15:14:19 |-------------------------------->
409
+ train_opt_callback: iter= 9 sample=73/192 sched=0.090000 loss=3.958102 dt=00:32:42 eta=5d 14:37:57 |-------------------------------------->
410
+ save_checkpoint_lora_file: saving to checkpoint-10.gguf
411
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
412
+ save_as_llama_lora: saving to lora.bin
413
+ save_as_llama_lora: saving to lora.bin
414
+ train_opt_callback: iter= 10 sample=81/192 sched=0.100000 loss=3.537075 dt=00:32:55 eta=5d 14:57:33 |------------------------------------------>
415
+ train_opt_callback: iter= 11 sample=89/192 sched=0.110000 loss=2.946956 dt=00:33:00 eta=5d 14:45:12 |------------------------------------------------>
416
+ train_opt_callback: iter= 12 sample=97/192 sched=0.120000 loss=2.738868 dt=00:33:02 eta=5d 14:22:33 |-------------------------------------------------->
417
+ train_opt_callback: iter= 13 sample=105/192 sched=0.130000 loss=2.126371 dt=00:33:00 eta=5d 13:40:51 |-------------------------------------------------------->
418
+ train_opt_callback: iter= 14 sample=113/192 sched=0.140000 loss=1.546756 dt=00:32:58 eta=5d 13:01:56 |-------------------------------------------------------------->
419
+ train_opt_callback: iter= 15 sample=121/192 sched=0.150000 loss=1.432991 dt=00:33:00 eta=5d 12:36:59 |--------------------------------------------------------------->
420
+ train_opt_callback: iter= 16 sample=129/192 sched=0.160000 loss=1.533808 dt=00:32:54 eta=5d 11:36:32 |-------------------------------------------------------------->
421
+ train_opt_callback: iter= 17 sample=137/192 sched=0.170000 loss=1.147277 dt=00:33:00 eta=5d 11:27:14 |------------------------------------------------------------------>
422
+ train_opt_callback: iter= 18 sample=145/192 sched=0.180000 loss=1.295468 dt=00:33:12 eta=5d 11:42:18 |---------------------------------------------------------------->
423
+ train_opt_callback: iter= 19 sample=153/192 sched=0.190000 loss=1.036252 dt=00:33:05 eta=5d 10:43:54 |------------------------------------------------------------------->
424
+ save_checkpoint_lora_file: saving to checkpoint-20.gguf
425
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
426
+ save_as_llama_lora: saving to lora.bin
427
+ save_as_llama_lora: saving to lora.bin
428
+ train_opt_callback: iter= 20 sample=161/192 sched=0.200000 loss=1.117745 dt=00:33:03 eta=5d 10:03:17 |------------------------------------------------------------------>
429
+ train_opt_callback: iter= 21 sample=169/192 sched=0.210000 loss=0.919692 dt=00:33:04 eta=5d 09:31:31 |-------------------------------------------------------------------->
430
+ train_opt_callback: iter= 22 sample=177/192 sched=0.220000 loss=0.958725 dt=00:32:52 eta=5d 08:11:11 |-------------------------------------------------------------------->
431
+ train_opt_callback: iter= 23 sample=185/192 sched=0.230000 loss=0.677074 dt=00:33:00 eta=5d 08:12:24 |----------------------------------------------------------------------->
432
+ train_opt_callback: reshuffle samples. completed epochs: 1
433
+ train_opt_callback: iter= 24 sample=1/192 sched=0.240000 loss=0.648157 dt=00:33:16 eta=5d 08:39:19 |----------------------------------------------------------------------->
434
+ train_opt_callback: iter= 25 sample=9/192 sched=0.250000 loss=0.820849 dt=00:33:14 eta=5d 07:57:04 |--------------------------------------------------------------------->
435
+ train_opt_callback: iter= 26 sample=17/192 sched=0.260000 loss=0.679941 dt=00:33:12 eta=5d 07:17:18 |----------------------------------------------------------------------->
436
+ train_opt_callback: iter= 27 sample=25/192 sched=0.270000 loss=0.553361 dt=00:33:02 eta=5d 06:07:10 |------------------------------------------------------------------------>
437
+ train_opt_callback: iter= 28 sample=33/192 sched=0.280000 loss=0.656925 dt=00:32:59 eta=5d 05:22:14 |----------------------------------------------------------------------->
438
+ train_opt_callback: iter= 29 sample=41/192 sched=0.290000 loss=0.681380 dt=00:33:11 eta=5d 05:34:39 |----------------------------------------------------------------------->
439
+ save_checkpoint_lora_file: saving to checkpoint-30.gguf
440
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
441
+ save_as_llama_lora: saving to lora.bin
442
+ save_as_llama_lora: saving to lora.bin
443
+ train_opt_callback: iter= 30 sample=49/192 sched=0.300000 loss=0.628787 dt=00:33:11 eta=5d 05:02:11 |----------------------------------------------------------------------->
444
+ train_opt_callback: iter= 31 sample=57/192 sched=0.310000 loss=0.453469 dt=00:33:25 eta=5d 05:20:01 |------------------------------------------------------------------------->
445
+ train_opt_callback: iter= 32 sample=65/192 sched=0.320000 loss=0.359532 dt=00:33:01 eta=5d 03:18:35 |-------------------------------------------------------------------------->
446
+ train_opt_callback: iter= 33 sample=73/192 sched=0.330000 loss=0.424839 dt=00:32:54 eta=5d 02:17:13 |------------------------------------------------------------------------->
447
+ train_opt_callback: iter= 34 sample=81/192 sched=0.340000 loss=0.433118 dt=00:33:01 eta=5d 02:10:01 |------------------------------------------------------------------------->
448
+ train_opt_callback: iter= 35 sample=89/192 sched=0.350000 loss=0.473854 dt=00:32:59 eta=5d 01:31:52 |------------------------------------------------------------------------->
449
+ train_opt_callback: iter= 36 sample=97/192 sched=0.360000 loss=0.399268 dt=00:33:03 eta=5d 01:12:33 |------------------------------------------------------------------------->
450
+ train_opt_callback: iter= 37 sample=105/192 sched=0.370000 loss=0.560971 dt=00:31:58 eta=4d 20:41:25 |------------------------------------------------------------------------>
451
+ train_opt_callback: iter= 38 sample=113/192 sched=0.380000 loss=0.401658 dt=00:20:32 eta=3d 02:37:22 |------------------------------------------------------------------------->
452
+ train_opt_callback: iter= 39 sample=121/192 sched=0.390000 loss=0.358174 dt=00:17:02 eta=2d 13:38:29 |-------------------------------------------------------------------------->
453
+ save_checkpoint_lora_file: saving to checkpoint-40.gguf
454
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
455
+ save_as_llama_lora: saving to lora.bin
456
+ save_as_llama_lora: saving to lora.bin
457
+ train_opt_callback: iter= 40 sample=129/192 sched=0.400000 loss=0.378025 dt=00:15:54 eta=2d 09:17:30 |-------------------------------------------------------------------------->
458
+ train_opt_callback: iter= 41 sample=137/192 sched=0.410000 loss=0.428651 dt=00:15:37 eta=2d 07:59:48 |------------------------------------------------------------------------->
459
+ train_opt_callback: iter= 42 sample=145/192 sched=0.420000 loss=0.323726 dt=00:15:40 eta=2d 07:55:10 |-------------------------------------------------------------------------->
460
+ train_opt_callback: iter= 43 sample=153/192 sched=0.430000 loss=0.296268 dt=00:15:40 eta=2d 07:37:30 |-------------------------------------------------------------------------->
461
+ train_opt_callback: iter= 44 sample=161/192 sched=0.440000 loss=0.285518 dt=00:15:32 eta=2d 06:55:17 |--------------------------------------------------------------------------->
462
+ train_opt_callback: iter= 45 sample=169/192 sched=0.450000 loss=0.342957 dt=00:15:33 eta=2d 06:43:16 |-------------------------------------------------------------------------->
463
+ train_opt_callback: iter= 46 sample=177/192 sched=0.460000 loss=0.194749 dt=00:15:46 eta=2d 07:11:14 |--------------------------------------------------------------------------->
464
+ train_opt_callback: iter= 47 sample=185/192 sched=0.470000 loss=0.244491 dt=00:15:39 eta=2d 06:34:14 |--------------------------------------------------------------------------->
465
+ train_opt_callback: reshuffle samples. completed epochs: 2
466
+ train_opt_callback: iter= 48 sample=1/192 sched=0.480000 loss=0.222005 dt=00:15:37 eta=2d 06:10:32 |--------------------------------------------------------------------------->
467
+ train_opt_callback: iter= 49 sample=9/192 sched=0.490000 loss=0.342179 dt=00:15:37 eta=2d 05:54:36 |-------------------------------------------------------------------------->
468
+ save_checkpoint_lora_file: saving to checkpoint-50.gguf
469
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
470
+ save_as_llama_lora: saving to lora.bin
471
+ save_as_llama_lora: saving to lora.bin
472
+ train_opt_callback: iter= 50 sample=17/192 sched=0.500000 loss=0.334579 dt=00:15:37 eta=2d 05:38:32 |-------------------------------------------------------------------------->
473
+ train_opt_callback: iter= 51 sample=25/192 sched=0.510000 loss=0.361630 dt=00:15:33 eta=2d 05:09:21 |-------------------------------------------------------------------------->
474
+ train_opt_callback: iter= 52 sample=33/192 sched=0.520000 loss=0.188962 dt=00:15:34 eta=2d 04:56:45 |---------------------------------------------------------------------------->
475
+ train_opt_callback: iter= 53 sample=41/192 sched=0.530000 loss=0.238500 dt=00:15:38 eta=2d 04:54:09 |--------------------------------------------------------------------------->
476
+ train_opt_callback: iter= 54 sample=49/192 sched=0.540000 loss=0.263577 dt=00:15:34 eta=2d 04:26:19 |--------------------------------------------------------------------------->
477
+ train_opt_callback: iter= 55 sample=57/192 sched=0.550000 loss=0.194148 dt=00:15:36 eta=2d 04:16:57 |--------------------------------------------------------------------------->
478
+ train_opt_callback: iter= 56 sample=65/192 sched=0.560000 loss=0.216232 dt=00:15:29 eta=2d 03:38:30 |--------------------------------------------------------------------------->
479
+ train_opt_callback: iter= 57 sample=73/192 sched=0.570000 loss=0.113424 dt=00:15:31 eta=2d 03:30:22 |---------------------------------------------------------------------------->
480
+ train_opt_callback: iter= 58 sample=81/192 sched=0.580000 loss=0.191995 dt=00:15:37 eta=2d 03:32:15 |--------------------------------------------------------------------------->
481
+ train_opt_callback: iter= 59 sample=89/192 sched=0.590000 loss=0.183454 dt=00:15:31 eta=2d 02:59:52 |---------------------------------------------------------------------------->
482
+ save_checkpoint_lora_file: saving to checkpoint-60.gguf
483
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
484
+ save_as_llama_lora: saving to lora.bin
485
+ save_as_llama_lora: saving to lora.bin
486
+ train_opt_callback: iter= 60 sample=97/192 sched=0.600000 loss=0.155663 dt=00:15:34 eta=2d 02:53:06 |---------------------------------------------------------------------------->
487
+ train_opt_callback: iter= 61 sample=105/192 sched=0.610000 loss=0.198579 dt=00:15:32 eta=2d 02:29:39 |--------------------------------------------------------------------------->
488
+ train_opt_callback: iter= 62 sample=113/192 sched=0.620000 loss=0.209679 dt=00:15:27 eta=2d 02:00:01 |--------------------------------------------------------------------------->
489
+ train_opt_callback: iter= 63 sample=121/192 sched=0.630000 loss=0.134847 dt=00:15:33 eta=2d 02:02:22 |---------------------------------------------------------------------------->
490
+ train_opt_callback: iter= 64 sample=129/192 sched=0.640000 loss=0.284779 dt=00:15:36 eta=2d 01:57:23 |--------------------------------------------------------------------------->
491
+ train_opt_callback: iter= 65 sample=137/192 sched=0.650000 loss=0.155553 dt=00:15:35 eta=2d 01:38:03 |---------------------------------------------------------------------------->
492
+ train_opt_callback: iter= 66 sample=145/192 sched=0.660000 loss=0.231996 dt=00:15:34 eta=2d 01:17:50 |--------------------------------------------------------------------------->
493
+ train_opt_callback: iter= 67 sample=153/192 sched=0.670000 loss=0.237264 dt=00:15:29 eta=2d 00:47:50 |--------------------------------------------------------------------------->
494
+ train_opt_callback: iter= 68 sample=161/192 sched=0.680000 loss=0.176222 dt=00:15:34 eta=2d 00:49:20 |---------------------------------------------------------------------------->
495
+ train_opt_callback: iter= 69 sample=169/192 sched=0.690000 loss=0.126854 dt=00:15:32 eta=2d 00:25:54 |---------------------------------------------------------------------------->
496
+ save_checkpoint_lora_file: saving to checkpoint-70.gguf
497
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
498
+ save_as_llama_lora: saving to lora.bin
499
+ save_as_llama_lora: saving to lora.bin
500
+ train_opt_callback: iter= 70 sample=177/192 sched=0.700000 loss=0.163200 dt=00:15:30 eta=2d 00:05:29 |---------------------------------------------------------------------------->
501
+ train_opt_callback: iter= 71 sample=185/192 sched=0.710000 loss=0.382751 dt=00:15:29 eta=1d 23:46:04 |-------------------------------------------------------------------------->
502
+ train_opt_callback: reshuffle samples. completed epochs: 3
503
+ train_opt_callback: iter= 72 sample=1/192 sched=0.720000 loss=0.249731 dt=00:15:30 eta=1d 23:35:02 |--------------------------------------------------------------------------->
504
+ train_opt_callback: iter= 73 sample=9/192 sched=0.730000 loss=0.240528 dt=00:15:35 eta=1d 23:34:26 |--------------------------------------------------------------------------->
505
+ train_opt_callback: iter= 74 sample=17/192 sched=0.740000 loss=0.125279 dt=00:15:29 eta=1d 23:00:17 |---------------------------------------------------------------------------->
506
+ train_opt_callback: iter= 75 sample=25/192 sched=0.750000 loss=0.230016 dt=00:15:34 eta=1d 22:57:46 |--------------------------------------------------------------------------->
507
+ train_opt_callback: iter= 76 sample=33/192 sched=0.760000 loss=0.187424 dt=00:15:32 eta=1d 22:36:58 |---------------------------------------------------------------------------->
508
+ train_opt_callback: iter= 77 sample=41/192 sched=0.770000 loss=0.198255 dt=00:15:37 eta=1d 22:37:52 |--------------------------------------------------------------------------->
509
+ train_opt_callback: iter= 78 sample=49/192 sched=0.780000 loss=0.058203 dt=00:15:31 eta=1d 22:03:32 |----------------------------------------------------------------------------->
510
+ train_opt_callback: iter= 79 sample=57/192 sched=0.790000 loss=0.110901 dt=00:15:27 eta=1d 21:36:39 |---------------------------------------------------------------------------->
511
+ save_checkpoint_lora_file: saving to checkpoint-80.gguf
512
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
513
+ save_as_llama_lora: saving to lora.bin
514
+ save_as_llama_lora: saving to lora.bin
515
+ train_opt_callback: iter= 80 sample=65/192 sched=0.800000 loss=0.198940 dt=00:15:41 eta=1d 22:00:42 |--------------------------------------------------------------------------->
516
+ train_opt_callback: iter= 81 sample=73/192 sched=0.810000 loss=0.101694 dt=00:15:39 eta=1d 21:38:50 |---------------------------------------------------------------------------->
517
+ train_opt_callback: iter= 82 sample=81/192 sched=0.820000 loss=0.140884 dt=00:15:35 eta=1d 21:14:18 |---------------------------------------------------------------------------->
518
+ train_opt_callback: iter= 83 sample=89/192 sched=0.830000 loss=0.160913 dt=00:15:30 eta=1d 20:42:40 |---------------------------------------------------------------------------->
519
+ train_opt_callback: iter= 84 sample=97/192 sched=0.840000 loss=0.147126 dt=00:15:34 eta=1d 20:39:51 |---------------------------------------------------------------------------->
520
+ train_opt_callback: iter= 85 sample=105/192 sched=0.850000 loss=0.104718 dt=00:15:35 eta=1d 20:24:55 |---------------------------------------------------------------------------->
521
+ train_opt_callback: iter= 86 sample=113/192 sched=0.860000 loss=0.143592 dt=00:15:31 eta=1d 19:58:35 |---------------------------------------------------------------------------->
522
+ train_opt_callback: iter= 87 sample=121/192 sched=0.870000 loss=0.101460 dt=00:15:29 eta=1d 19:39:18 |---------------------------------------------------------------------------->
523
+ train_opt_callback: iter= 88 sample=129/192 sched=0.880000 loss=0.172694 dt=00:15:31 eta=1d 19:26:54 |---------------------------------------------------------------------------->
524
+ train_opt_callback: iter= 89 sample=137/192 sched=0.890000 loss=0.097297 dt=00:15:29 eta=1d 19:07:34 |---------------------------------------------------------------------------->
525
+ save_checkpoint_lora_file: saving to checkpoint-90.gguf
526
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
527
+ save_as_llama_lora: saving to lora.bin
528
+ save_as_llama_lora: saving to lora.bin
529
+ train_opt_callback: iter= 90 sample=145/192 sched=0.900000 loss=0.203012 dt=00:15:27 eta=1d 18:47:01 |--------------------------------------------------------------------------->
530
+ train_opt_callback: iter= 91 sample=153/192 sched=0.910000 loss=0.061470 dt=00:15:31 eta=1d 18:41:13 |----------------------------------------------------------------------------->
531
+ train_opt_callback: iter= 92 sample=161/192 sched=0.920000 loss=0.186582 dt=00:15:32 eta=1d 18:29:30 |---------------------------------------------------------------------------->
532
+ train_opt_callback: iter= 93 sample=169/192 sched=0.930000 loss=0.136616 dt=00:15:30 eta=1d 18:08:52 |---------------------------------------------------------------------------->
533
+ train_opt_callback: iter= 94 sample=177/192 sched=0.940000 loss=0.107690 dt=00:15:26 eta=1d 17:41:06 |---------------------------------------------------------------------------->
534
+ train_opt_callback: iter= 95 sample=185/192 sched=0.950000 loss=0.125770 dt=00:15:33 eta=1d 17:45:06 |---------------------------------------------------------------------------->
535
+ train_opt_callback: reshuffle samples. completed epochs: 4
536
+ train_opt_callback: iter= 96 sample=1/192 sched=0.960000 loss=0.147633 dt=00:15:31 eta=1d 17:23:56 |---------------------------------------------------------------------------->
537
+ train_opt_callback: iter= 97 sample=9/192 sched=0.970000 loss=0.112660 dt=00:15:27 eta=1d 16:59:02 |---------------------------------------------------------------------------->
538
+ train_opt_callback: iter= 98 sample=17/192 sched=0.980000 loss=0.071028 dt=00:15:34 eta=1d 17:00:34 |----------------------------------------------------------------------------->
539
+ train_opt_callback: iter= 99 sample=25/192 sched=0.990000 loss=0.148655 dt=00:15:27 eta=1d 16:25:43 |---------------------------------------------------------------------------->
540
+ save_checkpoint_lora_file: saving to checkpoint-100.gguf
541
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
542
+ save_as_llama_lora: saving to lora.bin
543
+ save_as_llama_lora: saving to lora.bin
544
+ train_opt_callback: iter= 100 sample=33/192 sched=0.977975 loss=0.122147 dt=00:16:26 eta=1d 18:46:11 |---------------------------------------------------------------------------->
545
+ train_opt_callback: iter= 101 sample=41/192 sched=0.977536 loss=0.075263 dt=00:15:59 eta=1d 17:17:59 |----------------------------------------------------------------------------->
546
+ train_opt_callback: iter= 102 sample=49/192 sched=0.977093 loss=0.098616 dt=00:15:43 eta=1d 16:21:33 |---------------------------------------------------------------------------->
547
+ train_opt_callback: iter= 103 sample=57/192 sched=0.976646 loss=0.071327 dt=00:15:40 eta=1d 15:59:30 |----------------------------------------------------------------------------->
548
+ train_opt_callback: iter= 104 sample=65/192 sched=0.976194 loss=0.079295 dt=00:15:36 eta=1d 15:33:37 |----------------------------------------------------------------------------->
549
+ train_opt_callback: iter= 105 sample=73/192 sched=0.975738 loss=0.073060 dt=00:15:35 eta=1d 15:13:33 |----------------------------------------------------------------------------->
550
+ train_opt_callback: iter= 106 sample=81/192 sched=0.975278 loss=0.126725 dt=00:15:42 eta=1d 15:15:25 |---------------------------------------------------------------------------->
551
+ train_opt_callback: iter= 107 sample=89/192 sched=0.974814 loss=0.099712 dt=00:15:35 eta=1d 14:42:32 |---------------------------------------------------------------------------->
552
+ train_opt_callback: iter= 108 sample=97/192 sched=0.974346 loss=0.069941 dt=00:15:30 eta=1d 14:14:28 |----------------------------------------------------------------------------->
553
+ train_opt_callback: iter= 109 sample=105/192 sched=0.973873 loss=0.098124 dt=00:15:34 eta=1d 14:09:41 |---------------------------------------------------------------------------->
554
+ save_checkpoint_lora_file: saving to checkpoint-110.gguf
555
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
556
+ save_as_llama_lora: saving to lora.bin
557
+ save_as_llama_lora: saving to lora.bin
558
+ train_opt_callback: iter= 110 sample=113/192 sched=0.973396 loss=0.098537 dt=00:15:34 eta=1d 13:54:34 |---------------------------------------------------------------------------->
559
+ train_opt_callback: iter= 111 sample=121/192 sched=0.972915 loss=0.126159 dt=00:15:31 eta=1d 13:31:34 |---------------------------------------------------------------------------->
560
+ train_opt_callback: iter= 112 sample=129/192 sched=0.972430 loss=0.077748 dt=00:15:29 eta=1d 13:10:55 |----------------------------------------------------------------------------->
561
+ train_opt_callback: iter= 113 sample=137/192 sched=0.971941 loss=0.102149 dt=00:15:31 eta=1d 13:00:02 |---------------------------------------------------------------------------->
562
+ train_opt_callback: iter= 114 sample=145/192 sched=0.971447 loss=0.071937 dt=00:15:38 eta=1d 13:01:24 |----------------------------------------------------------------------------->
563
+ train_opt_callback: iter= 115 sample=153/192 sched=0.970950 loss=0.077037 dt=00:15:28 eta=1d 12:21:31 |----------------------------------------------------------------------------->
564
+ train_opt_callback: iter= 116 sample=161/192 sched=0.970448 loss=0.082243 dt=00:15:39 eta=1d 12:32:09 |----------------------------------------------------------------------------->
565
+ train_opt_callback: iter= 117 sample=169/192 sched=0.969942 loss=0.084385 dt=00:15:46 eta=1d 12:31:43 |----------------------------------------------------------------------------->
566
+ train_opt_callback: iter= 118 sample=177/192 sched=0.969432 loss=0.133562 dt=00:15:46 eta=1d 12:16:08 |---------------------------------------------------------------------------->
567
+ train_opt_callback: iter= 119 sample=185/192 sched=0.968918 loss=0.109072 dt=00:15:35 eta=1d 11:35:59 |---------------------------------------------------------------------------->
568
+ train_opt_callback: reshuffle samples. completed epochs: 5
569
+ save_checkpoint_lora_file: saving to checkpoint-120.gguf
570
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
571
+ save_as_llama_lora: saving to lora.bin
572
+ save_as_llama_lora: saving to lora.bin
573
+ train_opt_callback: iter= 120 sample=1/192 sched=0.968399 loss=0.113454 dt=00:15:34 eta=1d 11:17:49 |---------------------------------------------------------------------------->
574
+ train_opt_callback: iter= 121 sample=9/192 sched=0.967877 loss=0.077776 dt=00:15:33 eta=1d 11:00:28 |----------------------------------------------------------------------------->
575
+ train_opt_callback: iter= 122 sample=17/192 sched=0.967350 loss=0.061315 dt=00:15:30 eta=1d 10:39:02 |----------------------------------------------------------------------------->
576
+ train_opt_callback: iter= 123 sample=25/192 sched=0.966820 loss=0.072449 dt=00:15:36 eta=1d 10:35:18 |----------------------------------------------------------------------------->
577
+ train_opt_callback: iter= 124 sample=33/192 sched=0.966285 loss=0.060652 dt=00:15:34 eta=1d 10:14:59 |----------------------------------------------------------------------------->
578
+ train_opt_callback: iter= 125 sample=41/192 sched=0.965746 loss=0.072381 dt=00:15:32 eta=1d 09:54:52 |----------------------------------------------------------------------------->
579
+ train_opt_callback: iter= 126 sample=49/192 sched=0.965203 loss=0.058497 dt=00:15:35 eta=1d 09:45:58 |----------------------------------------------------------------------------->
580
+ train_opt_callback: iter= 127 sample=57/192 sched=0.964656 loss=0.066129 dt=00:15:40 eta=1d 09:42:49 |----------------------------------------------------------------------------->
581
+ train_opt_callback: iter= 128 sample=65/192 sched=0.964104 loss=0.048283 dt=00:15:45 eta=1d 09:36:16 |----------------------------------------------------------------------------->
582
+ train_opt_callback: iter= 129 sample=73/192 sched=0.963549 loss=0.070143 dt=00:15:37 eta=1d 09:04:45 |----------------------------------------------------------------------------->
583
+ save_checkpoint_lora_file: saving to checkpoint-130.gguf
584
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
585
+ save_as_llama_lora: saving to lora.bin
586
+ save_as_llama_lora: saving to lora.bin
587
+ train_opt_callback: iter= 130 sample=81/192 sched=0.962990 loss=0.071810 dt=00:15:37 eta=1d 08:49:34 |----------------------------------------------------------------------------->
588
+ train_opt_callback: iter= 131 sample=89/192 sched=0.962426 loss=0.073223 dt=00:15:33 eta=1d 08:24:44 |----------------------------------------------------------------------------->
589
+ train_opt_callback: iter= 132 sample=97/192 sched=0.961859 loss=0.073529 dt=00:15:35 eta=1d 08:12:35 |----------------------------------------------------------------------------->
590
+ train_opt_callback: iter= 133 sample=105/192 sched=0.961287 loss=0.075146 dt=00:15:38 eta=1d 08:04:55 |----------------------------------------------------------------------------->
591
+ train_opt_callback: iter= 134 sample=113/192 sched=0.960711 loss=0.068796 dt=00:15:34 eta=1d 07:39:44 |----------------------------------------------------------------------------->
592
+ train_opt_callback: iter= 135 sample=121/192 sched=0.960131 loss=0.076424 dt=00:15:30 eta=1d 07:15:46 |----------------------------------------------------------------------------->
593
+ train_opt_callback: iter= 136 sample=129/192 sched=0.959548 loss=0.081089 dt=00:15:36 eta=1d 07:12:43 |----------------------------------------------------------------------------->
594
+ train_opt_callback: iter= 137 sample=137/192 sched=0.958960 loss=0.059550 dt=00:15:37 eta=1d 06:58:43 |----------------------------------------------------------------------------->
595
+ train_opt_callback: iter= 138 sample=145/192 sched=0.958368 loss=0.066975 dt=00:15:34 eta=1d 06:38:48 |----------------------------------------------------------------------------->
596
+ train_opt_callback: iter= 139 sample=153/192 sched=0.957772 loss=0.048607 dt=00:15:39 eta=1d 06:31:42 |----------------------------------------------------------------------------->
597
+ save_checkpoint_lora_file: saving to checkpoint-140.gguf
598
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
599
+ save_as_llama_lora: saving to lora.bin
600
+ save_as_llama_lora: saving to lora.bin
601
+ train_opt_callback: iter= 140 sample=161/192 sched=0.957172 loss=0.068270 dt=00:15:34 eta=1d 06:05:57 |----------------------------------------------------------------------------->
602
+ train_opt_callback: iter= 141 sample=169/192 sched=0.956568 loss=0.066943 dt=00:15:34 eta=1d 05:50:37 |----------------------------------------------------------------------------->
603
+ train_opt_callback: iter= 142 sample=177/192 sched=0.955960 loss=0.083575 dt=00:15:40 eta=1d 05:46:20 |----------------------------------------------------------------------------->
604
+ train_opt_callback: iter= 143 sample=185/192 sched=0.955348 loss=0.090246 dt=00:15:37 eta=1d 05:25:46 |---------------------------------------------------------------------------->
605
+ train_opt_callback: reshuffle samples. completed epochs: 6
606
+ train_opt_callback: iter= 144 sample=1/192 sched=0.954732 loss=0.085501 dt=00:15:31 eta=1d 04:59:31 |----------------------------------------------------------------------------->
607
+ train_opt_callback: iter= 145 sample=9/192 sched=0.954112 loss=0.061576 dt=00:15:32 eta=1d 04:45:20 |----------------------------------------------------------------------------->
608
+ train_opt_callback: iter= 146 sample=17/192 sched=0.953488 loss=0.056798 dt=00:15:38 eta=1d 04:40:07 |----------------------------------------------------------------------------->
609
+ train_opt_callback: iter= 147 sample=25/192 sched=0.952861 loss=0.049612 dt=00:15:51 eta=1d 04:49:02 |----------------------------------------------------------------------------->
610
+ train_opt_callback: iter= 148 sample=33/192 sched=0.952229 loss=0.052921 dt=00:15:52 eta=1d 04:33:53 |----------------------------------------------------------------------------->
611
+ train_opt_callback: iter= 149 sample=41/192 sched=0.951593 loss=0.060486 dt=00:15:43 eta=1d 04:01:47 |----------------------------------------------------------------------------->
612
+ save_checkpoint_lora_file: saving to checkpoint-150.gguf
613
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
614
+ save_as_llama_lora: saving to lora.bin
615
+ save_as_llama_lora: saving to lora.bin
616
+ train_opt_callback: iter= 150 sample=49/192 sched=0.950953 loss=0.048820 dt=00:15:39 eta=1d 03:39:18 |----------------------------------------------------------------------------->
617
+ train_opt_callback: iter= 151 sample=57/192 sched=0.950309 loss=0.049720 dt=00:15:39 eta=1d 03:23:54 |----------------------------------------------------------------------------->
618
+ train_opt_callback: iter= 152 sample=65/192 sched=0.949661 loss=0.067127 dt=00:15:36 eta=1d 03:02:40 |----------------------------------------------------------------------------->
619
+ train_opt_callback: iter= 153 sample=73/192 sched=0.949010 loss=0.060491 dt=00:15:38 eta=1d 02:50:57 |----------------------------------------------------------------------------->
620
+ train_opt_callback: iter= 154 sample=81/192 sched=0.948354 loss=0.069643 dt=00:15:39 eta=1d 02:36:36 |----------------------------------------------------------------------------->
621
+ train_opt_callback: iter= 155 sample=89/192 sched=0.947695 loss=0.063830 dt=00:15:37 eta=1d 02:17:37 |----------------------------------------------------------------------------->
622
+ train_opt_callback: iter= 156 sample=97/192 sched=0.947031 loss=0.065827 dt=00:15:41 eta=1d 02:08:59 |----------------------------------------------------------------------------->
623
+ train_opt_callback: iter= 157 sample=105/192 sched=0.946364 loss=0.069183 dt=00:15:44 eta=1d 01:57:36 |----------------------------------------------------------------------------->
624
+ train_opt_callback: iter= 158 sample=113/192 sched=0.945692 loss=0.102524 dt=00:15:50 eta=1d 01:51:46 |---------------------------------------------------------------------------->
625
+ train_opt_callback: iter= 159 sample=121/192 sched=0.945017 loss=0.057947 dt=00:15:47 eta=1d 01:32:19 |----------------------------------------------------------------------------->
626
+ save_checkpoint_lora_file: saving to checkpoint-160.gguf
627
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
628
+ save_as_llama_lora: saving to lora.bin
629
+ save_as_llama_lora: saving to lora.bin
630
+ train_opt_callback: iter= 160 sample=129/192 sched=0.944338 loss=0.050483 dt=00:15:44 eta=1d 01:11:31 |----------------------------------------------------------------------------->
631
+ train_opt_callback: iter= 161 sample=137/192 sched=0.943655 loss=0.065197 dt=00:15:40 eta=1d 00:48:28 |----------------------------------------------------------------------------->
632
+ train_opt_callback: iter= 162 sample=145/192 sched=0.942968 loss=0.060740 dt=00:15:37 eta=1d 00:28:16 |----------------------------------------------------------------------------->
633
+ train_opt_callback: iter= 163 sample=153/192 sched=0.942277 loss=0.052192 dt=00:15:34 eta=1d 00:09:01 |----------------------------------------------------------------------------->
634
+ train_opt_callback: iter= 164 sample=161/192 sched=0.941583 loss=0.053813 dt=00:15:38 eta=23:59:39 |----------------------------------------------------------------------------->
635
+ train_opt_callback: iter= 165 sample=169/192 sched=0.940884 loss=0.059665 dt=00:15:39 eta=23:45:11 |----------------------------------------------------------------------------->
636
+ train_opt_callback: iter= 166 sample=177/192 sched=0.940182 loss=0.058576 dt=00:15:32 eta=23:18:44 |----------------------------------------------------------------------------->
637
+ train_opt_callback: iter= 167 sample=185/192 sched=0.939476 loss=0.053657 dt=00:15:29 eta=22:59:26 |----------------------------------------------------------------------------->
638
+ train_opt_callback: reshuffle samples. completed epochs: 7
639
+ train_opt_callback: iter= 168 sample=1/192 sched=0.938765 loss=0.050923 dt=00:15:33 eta=22:48:52 |----------------------------------------------------------------------------->
640
+ train_opt_callback: iter= 169 sample=9/192 sched=0.938052 loss=0.044524 dt=00:15:37 eta=22:39:28 |----------------------------------------------------------------------------->
641
+ save_checkpoint_lora_file: saving to checkpoint-170.gguf
642
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
643
+ save_as_llama_lora: saving to lora.bin
644
+ save_as_llama_lora: saving to lora.bin
645
+ train_opt_callback: iter= 170 sample=17/192 sched=0.937334 loss=0.046028 dt=00:15:24 eta=22:05:24 |----------------------------------------------------------------------------->
646
+ train_opt_callback: iter= 171 sample=25/192 sched=0.936612 loss=0.049158 dt=00:15:23 eta=21:48:21 |----------------------------------------------------------------------------->
647
+ train_opt_callback: iter= 172 sample=33/192 sched=0.935887 loss=0.053132 dt=00:15:26 eta=21:36:52 |----------------------------------------------------------------------------->
648
+ train_opt_callback: iter= 173 sample=41/192 sched=0.935158 loss=0.046411 dt=00:15:28 eta=21:24:48 |----------------------------------------------------------------------------->
649
+ train_opt_callback: iter= 174 sample=49/192 sched=0.934425 loss=0.048739 dt=00:15:25 eta=21:04:53 |----------------------------------------------------------------------------->
650
+ train_opt_callback: iter= 175 sample=57/192 sched=0.933688 loss=0.056599 dt=00:15:24 eta=20:48:33 |----------------------------------------------------------------------------->
651
+ train_opt_callback: iter= 176 sample=65/192 sched=0.932948 loss=0.056154 dt=00:15:30 eta=20:41:11 |----------------------------------------------------------------------------->
652
+ train_opt_callback: iter= 177 sample=73/192 sched=0.932203 loss=0.053577 dt=00:15:29 eta=20:23:34 |----------------------------------------------------------------------------->
653
+ train_opt_callback: iter= 178 sample=81/192 sched=0.931455 loss=0.056811 dt=00:15:22 eta=19:58:57 |----------------------------------------------------------------------------->
654
+ train_opt_callback: iter= 179 sample=89/192 sched=0.930703 loss=0.047785 dt=00:15:28 eta=19:51:41 |----------------------------------------------------------------------------->
655
+ save_checkpoint_lora_file: saving to checkpoint-180.gguf
656
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
657
+ save_as_llama_lora: saving to lora.bin
658
+ save_as_llama_lora: saving to lora.bin
659
+ train_opt_callback: iter= 180 sample=97/192 sched=0.929948 loss=0.065201 dt=00:15:31 eta=19:40:29 |----------------------------------------------------------------------------->
660
+ train_opt_callback: iter= 181 sample=105/192 sched=0.929188 loss=0.064345 dt=00:15:29 eta=19:22:17 |----------------------------------------------------------------------------->
661
+ train_opt_callback: iter= 182 sample=113/192 sched=0.928425 loss=0.047728 dt=00:15:28 eta=19:05:23 |----------------------------------------------------------------------------->
662
+ train_opt_callback: iter= 183 sample=121/192 sched=0.927658 loss=0.058445 dt=00:15:25 eta=18:45:29 |----------------------------------------------------------------------------->
663
+ train_opt_callback: iter= 184 sample=129/192 sched=0.926888 loss=0.055892 dt=00:15:26 eta=18:31:55 |----------------------------------------------------------------------------->
664
+ train_opt_callback: iter= 185 sample=137/192 sched=0.926113 loss=0.053026 dt=00:15:27 eta=18:17:25 |----------------------------------------------------------------------------->
665
+ train_opt_callback: iter= 186 sample=145/192 sched=0.925335 loss=0.046620 dt=00:15:29 eta=18:04:11 |----------------------------------------------------------------------------->
666
+ train_opt_callback: iter= 187 sample=153/192 sched=0.924554 loss=0.061322 dt=00:15:30 eta=17:50:35 |----------------------------------------------------------------------------->
667
+ train_opt_callback: iter= 188 sample=161/192 sched=0.923768 loss=0.047851 dt=00:15:37 eta=17:42:26 |----------------------------------------------------------------------------->
668
+ train_opt_callback: iter= 189 sample=169/192 sched=0.922979 loss=0.048963 dt=00:15:35 eta=17:25:03 |----------------------------------------------------------------------------->
669
+ save_checkpoint_lora_file: saving to checkpoint-190.gguf
670
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
671
+ save_as_llama_lora: saving to lora.bin
672
+ save_as_llama_lora: saving to lora.bin
673
+ train_opt_callback: iter= 190 sample=177/192 sched=0.922186 loss=0.047204 dt=00:15:36 eta=17:09:36 |----------------------------------------------------------------------------->
674
+ train_opt_callback: iter= 191 sample=185/192 sched=0.921390 loss=0.055146 dt=00:15:33 eta=16:50:51 |----------------------------------------------------------------------------->
675
+ train_opt_callback: reshuffle samples. completed epochs: 8
676
+ train_opt_callback: iter= 192 sample=1/192 sched=0.920590 loss=0.065108 dt=00:15:31 eta=16:33:42 |----------------------------------------------------------------------------->
677
+ train_opt_callback: iter= 193 sample=9/192 sched=0.919786 loss=0.045205 dt=00:15:29 eta=16:16:11 |----------------------------------------------------------------------------->
678
+ train_opt_callback: iter= 194 sample=17/192 sched=0.918978 loss=0.049374 dt=00:15:33 eta=16:05:03 |----------------------------------------------------------------------------->
679
+ train_opt_callback: iter= 195 sample=25/192 sched=0.918167 loss=0.047677 dt=00:15:41 eta=15:56:52 |----------------------------------------------------------------------------->
680
+ train_opt_callback: iter= 196 sample=33/192 sched=0.917353 loss=0.044926 dt=00:15:31 eta=15:31:04 |----------------------------------------------------------------------------->
681
+ train_opt_callback: iter= 197 sample=41/192 sched=0.916534 loss=0.046525 dt=00:15:10 eta=14:55:20 |----------------------------------------------------------------------------->
682
+ train_opt_callback: iter= 198 sample=49/192 sched=0.915712 loss=0.045030 dt=00:15:05 eta=14:35:23 |----------------------------------------------------------------------------->
683
+ train_opt_callback: iter= 199 sample=57/192 sched=0.914887 loss=0.045530 dt=00:15:00 eta=14:15:47 |----------------------------------------------------------------------------->
684
+ save_checkpoint_lora_file: saving to checkpoint-200.gguf
685
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
686
+ save_as_llama_lora: saving to lora.bin
687
+ save_as_llama_lora: saving to lora.bin
688
+ train_opt_callback: iter= 200 sample=65/192 sched=0.914058 loss=0.046049 dt=00:15:00 eta=14:00:14 |----------------------------------------------------------------------------->
689
+ train_opt_callback: iter= 201 sample=73/192 sched=0.913225 loss=0.053930 dt=00:14:58 eta=13:43:57 |----------------------------------------------------------------------------->
690
+ train_opt_callback: iter= 202 sample=81/192 sched=0.912389 loss=0.051443 dt=00:14:58 eta=13:28:28 |----------------------------------------------------------------------------->
691
+ train_opt_callback: iter= 203 sample=89/192 sched=0.911549 loss=0.046394 dt=00:14:58 eta=13:13:19 |----------------------------------------------------------------------------->
692
+ train_opt_callback: iter= 204 sample=97/192 sched=0.910705 loss=0.046575 dt=00:15:00 eta=13:00:36 |----------------------------------------------------------------------------->
693
+ train_opt_callback: iter= 205 sample=105/192 sched=0.909858 loss=0.049402 dt=00:15:03 eta=12:48:18 |----------------------------------------------------------------------------->
694
+ train_opt_callback: iter= 206 sample=113/192 sched=0.909007 loss=0.043352 dt=00:15:01 eta=12:31:19 |----------------------------------------------------------------------------->
695
+ train_opt_callback: iter= 207 sample=121/192 sched=0.908153 loss=0.053521 dt=00:14:58 eta=12:14:09 |----------------------------------------------------------------------------->
696
+ train_opt_callback: iter= 208 sample=129/192 sched=0.907296 loss=0.049153 dt=00:15:01 eta=12:00:50 |----------------------------------------------------------------------------->
697
+ train_opt_callback: iter= 209 sample=137/192 sched=0.906434 loss=0.050224 dt=00:15:01 eta=11:46:09 |----------------------------------------------------------------------------->
698
+ save_checkpoint_lora_file: saving to checkpoint-210.gguf
699
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
700
+ save_as_llama_lora: saving to lora.bin
701
+ save_as_llama_lora: saving to lora.bin
702
+ train_opt_callback: iter= 210 sample=145/192 sched=0.905570 loss=0.051670 dt=00:15:05 eta=11:33:58 |----------------------------------------------------------------------------->
703
+ train_opt_callback: iter= 211 sample=153/192 sched=0.904702 loss=0.051875 dt=00:15:06 eta=11:19:33 |----------------------------------------------------------------------------->
704
+ train_opt_callback: iter= 212 sample=161/192 sched=0.903830 loss=0.057988 dt=00:15:01 eta=11:00:52 |----------------------------------------------------------------------------->
705
+ train_opt_callback: iter= 213 sample=169/192 sched=0.902955 loss=0.048556 dt=00:15:04 eta=10:48:32 |----------------------------------------------------------------------------->
706
+ train_opt_callback: iter= 214 sample=177/192 sched=0.902076 loss=0.052551 dt=00:14:59 eta=10:29:59 |----------------------------------------------------------------------------->
707
+ train_opt_callback: iter= 215 sample=185/192 sched=0.901194 loss=0.057498 dt=00:15:05 eta=10:18:53 |----------------------------------------------------------------------------->
708
+ train_opt_callback: reshuffle samples. completed epochs: 9
709
+ train_opt_callback: iter= 216 sample=1/192 sched=0.900308 loss=0.051821 dt=00:15:01 eta=10:01:08 |----------------------------------------------------------------------------->
710
+ train_opt_callback: iter= 217 sample=9/192 sched=0.899419 loss=0.042146 dt=00:15:01 eta=09:45:51 |----------------------------------------------------------------------------->
711
+ train_opt_callback: iter= 218 sample=17/192 sched=0.898526 loss=0.048716 dt=00:15:01 eta=09:30:57 |----------------------------------------------------------------------------->
712
+ train_opt_callback: iter= 219 sample=25/192 sched=0.897630 loss=0.079733 dt=00:15:01 eta=09:16:01 |----------------------------------------------------------------------------->
713
+ save_checkpoint_lora_file: saving to checkpoint-220.gguf
714
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
715
+ save_as_llama_lora: saving to lora.bin
716
+ save_as_llama_lora: saving to lora.bin
717
+ train_opt_callback: iter= 220 sample=33/192 sched=0.896731 loss=0.045738 dt=00:14:57 eta=08:58:40 |----------------------------------------------------------------------------->
718
+ train_opt_callback: iter= 221 sample=41/192 sched=0.895828 loss=0.049312 dt=00:15:01 eta=08:46:03 |----------------------------------------------------------------------------->
719
+ train_opt_callback: iter= 222 sample=49/192 sched=0.894922 loss=0.044266 dt=00:15:04 eta=08:32:43 |----------------------------------------------------------------------------->
720
+ train_opt_callback: iter= 223 sample=57/192 sched=0.894012 loss=0.043147 dt=00:15:01 eta=08:15:58 |----------------------------------------------------------------------------->
721
+ train_opt_callback: iter= 224 sample=65/192 sched=0.893099 loss=0.055512 dt=00:15:28 eta=08:15:06 |----------------------------------------------------------------------------->
722
+ train_opt_callback: iter= 225 sample=73/192 sched=0.892183 loss=0.049956 dt=00:15:25 eta=07:58:08 |----------------------------------------------------------------------------->
723
+ train_opt_callback: iter= 226 sample=81/192 sched=0.891263 loss=0.050909 dt=00:15:33 eta=07:46:37 |----------------------------------------------------------------------------->
724
+ train_opt_callback: iter= 227 sample=89/192 sched=0.890340 loss=0.050016 dt=00:15:38 eta=07:33:34 |----------------------------------------------------------------------------->
725
+ train_opt_callback: iter= 228 sample=97/192 sched=0.889413 loss=0.050583 dt=00:15:41 eta=07:19:09 |----------------------------------------------------------------------------->
726
+ train_opt_callback: iter= 229 sample=105/192 sched=0.888483 loss=0.043942 dt=00:15:42 eta=07:03:57 |----------------------------------------------------------------------------->
727
+ save_checkpoint_lora_file: saving to checkpoint-230.gguf
728
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
729
+ save_as_llama_lora: saving to lora.bin
730
+ save_as_llama_lora: saving to lora.bin
731
+ train_opt_callback: iter= 230 sample=113/192 sched=0.887550 loss=0.048901 dt=00:15:41 eta=06:48:05 |----------------------------------------------------------------------------->
732
+ train_opt_callback: iter= 231 sample=121/192 sched=0.886613 loss=0.048936 dt=00:15:39 eta=06:31:31 |----------------------------------------------------------------------------->
733
+ train_opt_callback: iter= 232 sample=129/192 sched=0.885674 loss=0.047928 dt=00:15:42 eta=06:16:52 |----------------------------------------------------------------------------->
734
+ train_opt_callback: iter= 233 sample=137/192 sched=0.884730 loss=0.046294 dt=00:15:32 eta=05:57:30 |----------------------------------------------------------------------------->
735
+ train_opt_callback: iter= 234 sample=145/192 sched=0.883784 loss=0.047876 dt=00:15:36 eta=05:43:17 |----------------------------------------------------------------------------->
736
+ train_opt_callback: iter= 235 sample=153/192 sched=0.882834 loss=0.048621 dt=00:15:35 eta=05:27:27 |----------------------------------------------------------------------------->
737
+ train_opt_callback: iter= 236 sample=161/192 sched=0.881881 loss=0.046712 dt=00:15:30 eta=05:10:09 |----------------------------------------------------------------------------->
738
+ train_opt_callback: iter= 237 sample=169/192 sched=0.880924 loss=0.048654 dt=00:15:26 eta=04:53:31 |----------------------------------------------------------------------------->
739
+ train_opt_callback: iter= 238 sample=177/192 sched=0.879965 loss=0.044374 dt=00:15:39 eta=04:41:55 |----------------------------------------------------------------------------->
740
+ train_opt_callback: iter= 239 sample=185/192 sched=0.879002 loss=0.045983 dt=00:15:30 eta=04:23:39 |----------------------------------------------------------------------------->
741
+ train_opt_callback: reshuffle samples. completed epochs: 10
742
+ save_checkpoint_lora_file: saving to checkpoint-240.gguf
743
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
744
+ save_as_llama_lora: saving to lora.bin
745
+ save_as_llama_lora: saving to lora.bin
746
+ train_opt_callback: iter= 240 sample=1/192 sched=0.878036 loss=0.048949 dt=00:15:33 eta=04:08:52 |----------------------------------------------------------------------------->
747
+ train_opt_callback: iter= 241 sample=9/192 sched=0.877066 loss=0.044107 dt=00:15:26 eta=03:51:38 |----------------------------------------------------------------------------->
748
+ train_opt_callback: iter= 242 sample=17/192 sched=0.876094 loss=0.042922 dt=00:15:33 eta=03:37:55 |----------------------------------------------------------------------------->
749
+ train_opt_callback: iter= 243 sample=25/192 sched=0.875118 loss=0.048055 dt=00:15:30 eta=03:21:30 |----------------------------------------------------------------------------->
750
+ train_opt_callback: iter= 244 sample=33/192 sched=0.874139 loss=0.052699 dt=00:15:31 eta=03:06:22 |----------------------------------------------------------------------------->
751
+ train_opt_callback: iter= 245 sample=41/192 sched=0.873157 loss=0.056998 dt=00:15:33 eta=02:51:13 |----------------------------------------------------------------------------->
752
+ train_opt_callback: iter= 246 sample=49/192 sched=0.872171 loss=0.048870 dt=00:15:31 eta=02:35:19 |----------------------------------------------------------------------------->
753
+ train_opt_callback: iter= 247 sample=57/192 sched=0.871183 loss=0.043034 dt=00:15:32 eta=02:19:51 |----------------------------------------------------------------------------->
754
+ train_opt_callback: iter= 248 sample=65/192 sched=0.870191 loss=0.051389 dt=00:15:33 eta=02:04:27 |----------------------------------------------------------------------------->
755
+ train_opt_callback: iter= 249 sample=73/192 sched=0.869196 loss=0.046855 dt=00:15:38 eta=01:49:30 |----------------------------------------------------------------------------->
756
+ save_checkpoint_lora_file: saving to checkpoint-250.gguf
757
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
758
+ save_as_llama_lora: saving to lora.bin
759
+ save_as_llama_lora: saving to lora.bin
760
+ train_opt_callback: iter= 250 sample=81/192 sched=0.868198 loss=0.045374 dt=00:15:34 eta=01:33:28 |----------------------------------------------------------------------------->
761
+ train_opt_callback: iter= 251 sample=89/192 sched=0.867197 loss=0.046577 dt=00:15:37 eta=01:18:09 |----------------------------------------------------------------------------->
762
+ train_opt_callback: iter= 252 sample=97/192 sched=0.866192 loss=0.044559 dt=00:15:29 eta=01:01:59 |----------------------------------------------------------------------------->
763
+ train_opt_callback: iter= 253 sample=105/192 sched=0.865185 loss=0.042789 dt=00:15:34 eta=00:46:42 |----------------------------------------------------------------------------->
764
+ train_opt_callback: iter= 254 sample=113/192 sched=0.864174 loss=0.048775 dt=00:15:35 eta=00:31:10 |----------------------------------------------------------------------------->
765
+ train_opt_callback: iter= 255 sample=121/192 sched=0.863161 loss=0.052028 dt=00:15:28 eta=00:15:28 |----------------------------------------------------------------------------->
766
+ train_opt_callback: iter= 256 sample=129/192 sched=0.862144 loss=0.048282 dt=00:15:34 eta=0.0ms |----------------------------------------------------------------------------->
767
+ main: total training time: 3d 05:13:40
768
+ save_checkpoint_lora_file: saving to checkpoint-256.gguf
769
+ save_checkpoint_lora_file: saving to checkpoint-LATEST.gguf
770
+ save_as_llama_lora: saving to lora.bin
771
+ save_as_llama_lora: saving to lora.bin
train.txt ADDED
@@ -0,0 +1,192 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <s>[INST] What is carbon's atomic number? [/INST] Carbon's atomic number is 6.
2
+ <s>[INST] What group is carbon in? [/INST] Carbon is in group 14.
3
+ <s>[INST] What period is carbon in? [/INST] Carbon is in period 2.
4
+ <s>[INST] What is carbon's atomic weight? [/INST] Carbon's atomic weight is 12.01.
5
+ <s>[INST] What is carbon's state at room temperature? [/INST] Carbon's state is solid at room temperature.
6
+ <s>[INST] How many protons does carbon have? [/INST] Carbon has 6 protons.
7
+ <s>[INST] How many neutrons does carbon usually have? [/INST] Carbon usually has 6 neutrons.
8
+ <s>[INST] How many electrons does carbon usually have? [/INST] Carbon usually has 6 electrons.
9
+ <s>[INST] What is carbon's symbol? [/INST] Carbon's symbol is C.
10
+ <s>[INST] Where is carbon most abundant? [/INST] Carbon is most abundant in rocks, such as those in Earth's crust.
11
+ <s>[INST] What column is carbon in? [/INST] Carbon is in column 14.
12
+ <s>[INST] What row is carbon in? [/INST] Carbon is in row 2.
13
+ <s>[INST] What is carbon's melting point? [/INST] Carbon's melting point is 3,550 degrees Celsius.
14
+ <s>[INST] What is carbon's boiling point? [/INST] Carbon's boiling point is 4,827 degrees Celsius.
15
+ <s>[INST] What is carbon's radius? [/INST] Carbon's radius is 67 picometers.
16
+ <s>[INST] What is unique about carbon? [/INST] Carbon is essential to life as we know it because it readily forms many bonds with a variety of elements.
17
+ <s>[INST] What is silicon's atomic number? [/INST] Silicon's atomic number is 14.
18
+ <s>[INST] What group is silicon in? [/INST] Silicon is in group 14.
19
+ <s>[INST] What period is silicon in? [/INST] Silicon is in period 3.
20
+ <s>[INST] What is silicon's atomic weight? [/INST] Silicon's atomic weight is 28.09.
21
+ <s>[INST] What is silicon's state at room temperature? [/INST] Silicon's state is solid at room temperature.
22
+ <s>[INST] How many protons does silicon have? [/INST] Silicon has 14 protons.
23
+ <s>[INST] How many neutrons does silicon usually have? [/INST] Silicon usually has 14 neutrons.
24
+ <s>[INST] How many electrons does silicon usually have? [/INST] Silicon usually has 14 electrons.
25
+ <s>[INST] What is silicon's symbol? [/INST] Silicon's symbol is S.
26
+ <s>[INST] Where is silicon most abundant? [/INST] Silicon, like many metals, is common in rocks where it forms a component of minerals.
27
+ <s>[INST] What column is silicon in? [/INST] Silicon is in column 14.
28
+ <s>[INST] What row is silicon in? [/INST] Silicon is in row 3.
29
+ <s>[INST] What is silicon's melting point? [/INST] Silicon's melting point is 1,410 degrees Celsius.
30
+ <s>[INST] What is silicon's boiling point? [/INST] Silicon's boiling point is 3,265 degrees Celsius.
31
+ <s>[INST] What is silicon's radius? [/INST] Silicon's radius is 111 picometers.
32
+ <s>[INST] What is unique about silicon? [/INST] Silicon is used to make the transistors at the heart of all computers.
33
+ <s>[INST] What is argon's atomic number? [/INST] Argon's atomic number is 18.
34
+ <s>[INST] What group is argon in? [/INST] Argon is in group 18.
35
+ <s>[INST] What period is argon in? [/INST] Argon is in period 3.
36
+ <s>[INST] What is argon's atomic weight? [/INST] Argon's atomic weight is 39.95.
37
+ <s>[INST] What is argon's state at room temperature? [/INST] Argon's state is gas at room temperature.
38
+ <s>[INST] How many protons does argon have? [/INST] Argon has 18 protons.
39
+ <s>[INST] How many neutrons does argon usually have? [/INST] Argon usually has 18 neutrons.
40
+ <s>[INST] How many electrons does argon usually have? [/INST] Argon usually has 18 electrons.
41
+ <s>[INST] What is argon's symbol? [/INST] Argon's symbol is Ar.
42
+ <s>[INST] Where is argon most abundant? [/INST] Argon is the third most abundant gas in Earth's atmosphere.
43
+ <s>[INST] What column is argon in? [/INST] Argon is in column 18.
44
+ <s>[INST] What row is argon in? [/INST] Argon is in row 3.
45
+ <s>[INST] What is argon's melting point? [/INST] Argon's melting point is -189.2 degrees Celsius.
46
+ <s>[INST] What is argon's boiling point? [/INST] Argon's boiling point is -185.7 degrees Celsius.
47
+ <s>[INST] What is argon's radius? [/INST] Argon's radius is 71 picometers.
48
+ <s>[INST] What is unique about argon? [/INST] Argon is a noble gas, meaning that it is much less reactive than most other elements.
49
+ <s>[INST] What is calcium's atomic number? [/INST] Calcium's atomic number is 20.
50
+ <s>[INST] What group is calcium in? [/INST] Calcium is in group 2.
51
+ <s>[INST] What period is calcium in? [/INST] Calcium is in period 4.
52
+ <s>[INST] What is calcium's atomic weight? [/INST] Calcium's atomic weight is 40.01.
53
+ <s>[INST] What is calcium's state at room temperature? [/INST] Calcium's state is solid at room temperature.
54
+ <s>[INST] How many protons does calcium have? [/INST] Calcium has 20 protons.
55
+ <s>[INST] How many neutrons does calcium usually have? [/INST] Calcium usually has 20 neutrons.
56
+ <s>[INST] How many electrons does calcium usually have? [/INST] Calcium usually has 20 electrons.
57
+ <s>[INST] What is calcium's symbol? [/INST] Calcium's symbol is C.
58
+ <s>[INST] Where is calcium most abundant? [/INST] Calcium, like many metals, is common in rocks where it forms a component of minerals. It is the third most common metal in Earth's crust.
59
+ <s>[INST] What column is calcium in? [/INST] Calcium is in column 2.
60
+ <s>[INST] What row is calcium in? [/INST] Calcium is in row 4.
61
+ <s>[INST] What is calcium's melting point? [/INST] Calcium's melting point is 842 degrees Celsius.
62
+ <s>[INST] What is calcium's boiling point? [/INST] Calcium's boiling point is 1,484 degrees Celsius.
63
+ <s>[INST] What is calcium's radius? [/INST] Calcium's radius is 194 picometers.
64
+ <s>[INST] What is unique about calcium? [/INST] Calcium is common in the human body and it is very important for many body functions.
65
+ <s>[INST] What is iron's atomic number? [/INST] Iron's atomic number is 26.
66
+ <s>[INST] What group is iron in? [/INST] Iron is in group 8.
67
+ <s>[INST] What period is iron in? [/INST] Iron is in period 4.
68
+ <s>[INST] What is iron's atomic weight? [/INST] Iron's atomic weight is 55.85.
69
+ <s>[INST] What is iron's state at room temperature? [/INST] Iron's state is solid at room temperature.
70
+ <s>[INST] How many protons does iron have? [/INST] Iron has 26 protons.
71
+ <s>[INST] How many neutrons does iron usually have? [/INST] Iron usually has 26 neutrons.
72
+ <s>[INST] How many electrons does iron usually have? [/INST] Iron usually has 26 electrons.
73
+ <s>[INST] What is iron's symbol? [/INST] Iron's symbol is Fe.
74
+ <s>[INST] Where is iron most abundant? [/INST] Iron, like many metals, is common in rocks where it forms a component of minerals. In Earth, it makes up most of the core.
75
+ <s>[INST] What column is iron in? [/INST] Iron is in column 8.
76
+ <s>[INST] What row is iron in? [/INST] Iron is in row 4.
77
+ <s>[INST] What is iron's melting point? [/INST] Iron's melting point is 1,538 degrees Celsius.
78
+ <s>[INST] What is iron's boiling point? [/INST] Iron's boiling point is 2,861 degrees Celsius.
79
+ <s>[INST] What is iron's radius? [/INST] Iron's radius is 156 picometers.
80
+ <s>[INST] What is unique about iron? [/INST] Iron is the last element that stars fuse in their cores before going supernova.
81
+ <s>[INST] What is silver's atomic number? [/INST] Silver's atomic number is 47.
82
+ <s>[INST] What group is silver in? [/INST] Silver is in group 11.
83
+ <s>[INST] What period is silver in? [/INST] Silver is in period 5.
84
+ <s>[INST] What is silver's atomic weight? [/INST] Silver's atomic weight is 107.87.
85
+ <s>[INST] What is silver's state at room temperature? [/INST] Silver's state is solid at room temperature.
86
+ <s>[INST] How many protons does silver have? [/INST] Silver has 47 protons.
87
+ <s>[INST] How many neutrons does silver usually have? [/INST] Silver usually has 47 neutrons.
88
+ <s>[INST] How many electrons does silver usually have? [/INST] Silver usually has 47 electrons.
89
+ <s>[INST] What is silver's symbol? [/INST] Silver's symbol is Ag.
90
+ <s>[INST] Where is silver most abundant? [/INST] Silver, like many metals, is common in rocks where it forms a component of minerals. On Earth, it is considered a precious metal.
91
+ <s>[INST] What column is silver in? [/INST] Silver is in column 11.
92
+ <s>[INST] What row is silver in? [/INST] Silver is in row 5.
93
+ <s>[INST] What is silver's melting point? [/INST] Silver's melting point is 961.78 degrees Celsius.
94
+ <s>[INST] What is silver's boiling point? [/INST] Silver's boiling point is 2,162 degrees Celsius.
95
+ <s>[INST] What is silver's radius? [/INST] Silver's radius is 165 picometers.
96
+ <s>[INST] What is unique about silver? [/INST] Silver has the highest electrical and thermal conductivity as well as reflectivity of any element.
97
+ <s>[INST] What is cesium's atomic number? [/INST] Cesium's atomic number is 55.
98
+ <s>[INST] What group is cesium in? [/INST] Cesium is in group 1.
99
+ <s>[INST] What period is cesium in? [/INST] Cesium is in period 6.
100
+ <s>[INST] What is cesium's atomic weight? [/INST] Cesium's atomic weight is 132.91.
101
+ <s>[INST] What is cesium's state at room temperature? [/INST] Cesium's state is solid at room temperature.
102
+ <s>[INST] How many protons does cesium have? [/INST] Cesium has 55 protons.
103
+ <s>[INST] How many neutrons does cesium usually have? [/INST] Cesium usually has 55 neutrons.
104
+ <s>[INST] How many electrons does cesium usually have? [/INST] Cesium usually has 55 electrons.
105
+ <s>[INST] What is cesium's symbol? [/INST] Cesium's symbol is Cs.
106
+ <s>[INST] Where is cesium most abundant? [/INST] Cesium, like many metals, is common in rocks where it forms a component of minerals. It is a major component of toxic waste.
107
+ <s>[INST] What column is cesium in? [/INST] Cesium is in column 1.
108
+ <s>[INST] What row is cesium in? [/INST] Cesium is in row 6.
109
+ <s>[INST] What is cesium's melting point? [/INST] Cesium's melting point is 28.4 degrees Celsius.
110
+ <s>[INST] What is cesium's boiling point? [/INST] Cesium's boiling point is 670 degrees Celsius.
111
+ <s>[INST] What is cesium's radius? [/INST] Cesium's radius is 265 picometers.
112
+ <s>[INST] What is unique about cesium? [/INST] Cesium is the last element that stars fuse in their cores before going supernova.
113
+ <s>[INST] What is hydrogen's atomic number? [/INST] Hydrogen's atomic number is 1.
114
+ <s>[INST] What group is hydrogen in? [/INST] Hydrogen is in group 1.
115
+ <s>[INST] What period is hydrogen in? [/INST] Hydrogen is in period 1.
116
+ <s>[INST] What is hydrogen's atomic weight? [/INST] Hydrogen's atomic weight is 1.01.
117
+ <s>[INST] What is hydrogen's state at room temperature? [/INST] Hydrogen's state is gas at room temperature.
118
+ <s>[INST] How many protons does hydrogen have? [/INST] Hydrogen has 1 proton.
119
+ <s>[INST] How many neutrons does hydrogen usually have? [/INST] Hydrogen usually has 0 neutrons.
120
+ <s>[INST] How many electrons does hydrogen usually have? [/INST] Hydrogen usually has 1 electron.
121
+ <s>[INST] What is hydrogen's symbol? [/INST] Hydrogen's symbol is H.
122
+ <s>[INST] Where is hydrogen most abundant? [/INST] Hydrogen is most abundant in stars.
123
+ <s>[INST] What column is hydrogen in? [/INST] Hydrogen is in column 1.
124
+ <s>[INST] What row is hydrogen in? [/INST] Hydrogen is in row 1.
125
+ <s>[INST] What is hydrogen's melting point? [/INST] Hydrogen's melting point is -259.1 degrees Celsius.
126
+ <s>[INST] What is hydrogen's boiling point? [/INST] Hydrogen's boiling point is -252.9 degrees Celsius.
127
+ <s>[INST] What is hydrogen's radius? [/INST] Hydrogen's radius is 53 picometers.
128
+ <s>[INST] What is unique about hydrogen? [/INST] It's the most abundant element in the universe, making up over 75% of all matter.
129
+ <s>[INST] Is pure hydrogen common in our atmosphere? [/INST] No, pure hydrogen is not common in our atmosphere. The concentration of hydrogen gas (H2) in the Earth's atmosphere is very low, about 0.55 part per million by volume.
130
+ <s>[INST] What is the Sun made of? [/INST] By mass the composition of the sun is 92.1% hydrogen and 7.9% helium.
131
+ <s>[INST] What is the main component of Jupiter? [/INST] Hydrogen is the main component of Jupiter and the other gas giant planets.
132
+ <s>[INST] How does the hydrogen bomb work? [/INST] The hydrogen bomb, or H-bomb, is powered by nuclear fusion. Deuterium and tritium (hydrogen isotopes) are fused into helium, thereby releasing energy.
133
+ <s>[INST] Can hydrogen be used as fuel? [/INST] Yes, hydrogen is used as rocket fuel. It can also power fuel cells to generate electricity.
134
+ <s>[INST] What is a fuel cell? [/INST] A fuel cell converts hydrogen to electricity. The hydrogen fuel is combined with oxygen in the air to make H2O (water) and release energy.
135
+ <s>[INST] How many isotypes do hydrogen have? [/INST] Hydrogen has three common isotopes – protium, deuterium and tritium.
136
+ <s>[INST] Why is there so little hydrogen gas in the earth atmosphere? [/INST] It is too light and escapes gravity quickly.
137
+ <s>[INST] Is hydrogen common used in ballons? [/INST] No, it is too combustible and unsafe.
138
+ <s>[INST] Where is hydrogen found on earth? [/INST] The most common place to find hydrogen on earth is in water (H2O)
139
+ <s>[INST] How is hydrogen used? [/INST] Hydrogen is used to make ammonia for fertilizers, refining metals, and methanol for making artificial material like plastics.
140
+ <s>[INST] How does hydrogen taste? [/INST] Hydrogen is odourless, colourless and tasteless. And is undetectable by human senses.
141
+ <s>[INST] What is the energy source of the Sun? [/INST] In the core of the Sun hydrogen is being converted into helium. This is called nuclear fusion.
142
+ <s>[INST] What's special about the hydrogen atom? [/INST] Hydrogen is the only element that can exist without neutrons.
143
+ <s>[INST] Is liquid hydrogen very cold? [/INST] Yes, it is below -252.9 degrees Celsius. It is very cold.
144
+ <s>[INST] What is Mercury? [/INST] Mercury is a silver colored metal that is liquid at room temperature. It has an atomic number of 80 on the periodic table. It has a melting point of -38.83 C (-37.89 F) and it is toxic to humans.
145
+ <s>[INST] What is mercury's atomic number? [/INST] The atomic number of mercury is 80
146
+ <s>[INST] What group is mercury in? [/INST] Mercury (Hg) is located in Group 12, also known as the zinc group, on the periodic table. This group is located in the d-block, specifically in Period 6. Other elements in Group 12 include zinc (Zn), cadmium (Cd), and copernicium (Cn). Mercury is the only liquid at room temperature among the elements in this group.
147
+ <s>[INST] What period is mercury in? [/INST] Mercury (Hg) is located in Period 6 of the periodic table. The periods in the periodic table represent the number of electron shells or energy levels that an element's atoms have. Since mercury is in Period 6, it means that its atoms have six electron shells or energy levels. Other elements in Period 6 include elements like lead (Pb), gold (Au), and polonium (Po).
148
+ <s>[INST] What is mercury's atomic weight? [/INST] The atomic weight of Mercury is approximately 200.59 atomic mass units.
149
+ <s>[INST] What is mercury's state at room temperature? [/INST] Mercury's state is liquid at room temperature. Its melting point is -38.83 C (-37.89 F) and boiling point is 357 C (674 F). That means it is liquid at room temperature of 20 C (68 F).
150
+ <s>[INST] How many protons does mercury have? [/INST] A mercury atom has 80 protons. That corresponds to its atomic number 80.
151
+ <s>[INST] How many neutrons does mercury usually have? [/INST] The number of neutrons in a mercury atom can vary depending on the isotope of mercury. The most common isotope of mercury, known as mercury-202 (Hg-202), has 80 protons and 122 neutrons. However, there are other isotopes of mercury that have different numbers of neutrons. For example, mercury-201 (Hg-201) has 80 protons and 121 neutrons, while mercury-200 (Hg-200) has 80 protons and 120 neutrons. These varying isotopes of mercury exist due to differences in the number of neutrons in the nucleus of the atom.
152
+ <s>[INST] How many electrons does mercury usually have? [/INST] A mercury atom has 80 electrons, which is the same as its protons and atomic number.
153
+ <s>[INST] What is mercury's symbol? [/INST] The chemical symbol for mercury is Hg, which comes from its Latin name "hydrargyrum," meaning "liquid silver."
154
+ <s>[INST] Where is mercury most abundant? [/INST] Mercury is a naturally occurring element, but it is relatively rare in Earth's crust. It is not abundant in any specific region or country. However, mercury can be found in small amounts throughout the world, primarily in cinnabar ore deposits. The largest reserves of mercury are located in China, Kyrgyzstan, and Russia. It is also produced as a byproduct of mining and refining other metals, such as gold and silver. Additionally, mercury is released into the environment through human activities, such as coal burning and industrial processes.
155
+ <s>[INST] What column is mercury in? [/INST] Mercury (Hg) is located in Column 12 of the periodic table. This column is known as Group 12 or the zinc group. Elements in this column share similar chemical properties, as they have the same number of valence electrons. Other elements in Group 12 include zinc (Zn), cadmium (Cd), and copernicium (Cn).
156
+ <s>[INST] What row is mercury in? [/INST] Mercury (Hg) is located in Row 6 of the periodic table. The rows, or periods, in the periodic table represent the number of electron shells or energy levels that an element's atoms have. Since mercury is in Row 6, it means that its atoms have six electron shells or energy levels. Other elements in Row 6 include elements like lead (Pb), gold (Au), and polonium (Po).
157
+ <s>[INST] What is mercury's melting point? [/INST] Mercury's melting point is -38.83 degrees Celsius or -37.89 degrees Fahrenheit
158
+ <s>[INST] What is mercury's boiling point? [/INST] Mercury's boiling point is 357 degrees Celsius or 674 degrees Fahrenheit
159
+ <s>[INST] What is mercury's radius? [/INST] The atomic radius of mercury (Hg) depends on the reference point used. However, as an estimate, the atomic radius of mercury is typically reported to be around 150 picometers (pm) or 1.5 angstroms (Å). It's worth noting that atomic radii can vary slightly depending on the specific measurement method or the way it is defined in different contexts.
160
+ <s>[INST] Is all metal solid at room temperature? [/INST] No. Mercury is liquid at room temperature as its melting point is -38.83 C (-37.89 F) and boiling point is 357 C (674 F).
161
+ <s>[INST] Is Mercury the only Liquid Metal at room temperature? [/INST] No, mercury is not the only liquid metal at room temperature. There are several other metals that are also in liquid state at or around room temperature. Some examples include gallium, cesium, and francium. It is worth noting that these metals have very low melting points, which allows them to be liquid at typical room temperatures.
162
+ <s>[INST] What is unique about mercury? [/INST] Mercury has several unique properties. It is a good conductor of electricity and is one of the few metals that are liquid at room temperature. It has a melting point of -38.83 C (or -37.89 F), which is below room temperature. It also has a high density, which is about 13.5 times denser than water.
163
+ <s>[INST] Is mercury a metal? [/INST] Yes, mercury is a heavy, silvery-white metal. It is the only metal that exists as a liquid at room temperature.
164
+ <s>[INST] What are some common uses of mercury? [/INST] Mercury has various applications, including its use in thermometers, barometers, electrical switches, dental fillings, and fluorescent lights. It has also been used historically in industries such as gold mining and as an ingredient in certain pharmaceutical products.
165
+ <s>[INST] Is mercury toxic? [/INST] Yes, mercury is toxic to humans and animals. Exposure to high levels of mercury can have harmful effects on the nervous system, kidneys, and other organs. It is important to handle and dispose of mercury safely to prevent environmental contamination and health hazards.
166
+ <s>[INST] How does mercury behave differently from other metals? [/INST] One unique property of mercury is its low melting point (-38.83 degrees Celsius or -37.89 degrees Fahrenheit) and boiling point (357 degrees Celsius or 674 degrees Fahrenheit). This makes it the only metal that is liquid at room temperature. Mercury also has high surface tension, allowing it to form droplets and easily separate from other materials.
167
+ <s>[INST] Where is mercury found in nature? [/INST] Mercury is found in small quantities throughout the Earth's crust. It is primarily obtained by mining cinnabar ore, which is the most common source of mercury. It can also be found in trace amounts in certain minerals and fossil fuels.
168
+ <s>[INST] Is mercury still used in thermometers? [/INST] Mercury thermometers have been phased out in many countries due to environmental concerns. Digital thermometers and non-mercury alternatives, such as alcohol-based or digital thermometers, are typically used instead.
169
+ <s>[INST] How does mercury react with other substances? [/INST] Mercury does not react with most acids, alkalis, or common gases at room temperature. However, it can form compounds with certain elements, such as sulfur, oxygen, and halogens, under specific conditions.
170
+ <s>[INST] Is there more than one form of mercury? [/INST] Yes, there are several forms or phases of mercury. The most common form is elemental or metallic mercury, which is a liquid at room temperature. Another form is mercury vapor, which is a gas that can be produced by heating or evaporation. In addition, there are various mercury compounds with distinct chemical properties.
171
+ <s>[INST] How is mercury measured? [/INST] The measurement of mercury is typically done using different units. Common units include milligrams per kilogram (mg/kg) for environmental samples, parts per billion (ppb) or parts per million (ppm) for air and water samples, and micrograms per liter (µg/L) for blood or urine samples.
172
+ <s>[INST] What are the environmental concerns associated with mercury? [/INST] Mercury is a significant environmental pollutant, primarily due to its release from industrial processes, coal burning, and improper disposal of mercury-containing products. It can accumulate in ecosystems, leading to contamination of water, air, soil, and the food chain. This can have harmful effects on wildlife and humans.
173
+ <s>[INST] How is mercury disposed of properly? [/INST] Proper disposal of mercury is crucial to prevent environmental contamination. It should not be thrown in the trash. Instead, contact your local hazardous waste facility or follow specific guidelines set by your region for the safe collection and disposal of mercury-containing products.
174
+ <s>[INST] Are there any natural sources of elemental mercury? [/INST] Yes, small amounts of elemental mercury can be released naturally into the environment through volcanic activity, weathering of rocks, and the breakdown of organic matter. However, human activities are the primary source of elevated mercury levels in the environment.
175
+ <s>[INST] Where does the name Mercury come from? [/INST] The element was named after the Roman god Mercury, known for his speed and agility.
176
+ <s>[INST] Does Mercury rust when exposed to oxygen in the air? [/INST] Mercury, as a liquid metal, does not naturally combine with oxygen in the same way as other elements. Under normal conditions, mercury does not readily react with oxygen in the atmosphere.
177
+ <s>[INST] Where is Mercury used in household items? [/INST] Mercury is used in older thermometers, fluorescent light bulbs and some electrical switches.
178
+ <s>[INST] Why is it dangerous to break a mercury thermostat? [/INST] Mercury is highly toxic to humans. It is liquid in room temperature and can easily spread. When dropped, elemental mercury breaks into smaller droplets which can go through small cracks or become strongly attached to certain materials.
179
+ <s>[INST] Can Mercury exist in gas form at room temperature? [/INST] Mercury can exist as vapor gas in room tempature. At room temperature, exposed elemental mercury can evaporate to become an invisible, odorless toxic vapor. If heated, it is a colorless, odorless gas.
180
+ <s>[INST] Could Mercury pollute the air? [/INST] Yes, the burning of coal, oil and wood as fuel can cause mercury to become airborne, as can burning wastes that contain mercury.
181
+ <s>[INST] How does Mercury in the air pollute the ground? [/INST] This airborne mercury can fall to the ground in raindrops, in dust, or simply due to gravity (known as “air deposition”). The amount of mercury deposited in a given area depends on how much mercury is released from local, regional, national, and international sources.
182
+ <s>[INST] In the United States, what is the greatest source of Mercury pollution? [/INST] In the United States, power plants that burn coal to create electricity are the largest source of emissions; they account for about 44 percent of all manmade mercury emissions
183
+ <s>[INST] Is Mercury salt used in everyday consumer products? [/INST] Yes, Mercuric chloride is used in photography and as a topical antiseptic and disinfectant, wood preservative, and fungicide.
184
+ <s>[INST] Is Mercury salt used in medical products? [/INST] In the past, mercurous chloride was widely used in medicinal products, including laxatives, worming medications, and teething powders. It has since been replaced by safer and more effective agents.
185
+ <s>[INST] Does vaccines contain Mercury? [/INST] Sometimes. Thimerosal is a compound that contains mercury. Because it prevents the growth of dangerous microbes, thimerosal is used as a preservative in multi-dose vials of flu vaccines, and in two other childhood vaccines, it is used in the manufacturing process.
186
+ <s>[INST] Which occupations have high risk of Mercury exposure? [/INST] Occupations with higher risk of exposure to mercury and its salts include mining, electrical equipment manufacturing, and chemical and metal processing in which mercury is used.
187
+ <s>[INST] Is there an organic form of Mercury? [/INST] Yes, Methylmercury is the most common organic mercury compound found in the environment, and is highly toxic.
188
+ <s>[INST] can I freeze mercury solid? [/INST] Yes. With a melting point of -38.83 C (-37.89 F), you can freeze Mercury solid with dry ice.
189
+ <s>[INST] How is Mercury toxic? [/INST] Mercury is a neurotoxin. It damages the neural system, especially in young chidlren. Mercury exposure at high levels can harm the brain, heart, kidneys, lungs, and immune system of people of all ages.
190
+ <s>[INST] How does Mercury get into human bodies? [/INST] Mercury can vaporize in room temperature. Metallic mercury mainly causes health effects when inhaled as a vapor where it can be absorbed through the lungs.
191
+ <s>[INST] How do you detect Mercury poisoning? [/INST] Mercury can be detected through a blood test.
192
+ <s>[INST] Can I be exposed to Mercury from house paint? [/INST] Yes, the US EPA permits interior latex paint to contain less than or equal to 300 ppm elemental mercury and exterior latex paint to contain less than or equal to 2000 ppm. However, neither the presence nor the concentration of mercury in the paint is required to be labeled on the paint can.