11b commited on
Commit
30e2405
1 Parent(s): bb85978

part 4/10 of experiment 7 as described in logbook 2023-03-03

Browse files
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ebaf8cc27fb934f6f88a2ed5825feefe0565df2720cd6fb6bc58f8ccca27efb
3
  size 10581615331
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88ba2b44537f444e3fad92dff6962ac8c0b983427523484f98e7acf2d71fd65e
3
  size 10581615331
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dfdda631a6dbdb8d84abe2734231e5429d2a5975cd1eb742b8e0fb1b94c05bca
3
  size 5810125147
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c9aabb90f1f0004ec6c406366d65484a29c3de927b7561ad52ac8036b2be48e
3
  size 5810125147
pytorch_model.bin.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 12219206136
4
  },
5
  "weight_map": {
6
  "lm_head.bias": "pytorch_model-00002-of-00002.bin",
@@ -205,8 +205,8 @@
205
  "transformer.h.23.attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
206
  "transformer.h.23.ln_1.bias": "pytorch_model-00001-of-00002.bin",
207
  "transformer.h.23.ln_1.weight": "pytorch_model-00001-of-00002.bin",
208
- "transformer.h.23.mlp.fc_in.bias": "pytorch_model-00002-of-00002.bin",
209
- "transformer.h.23.mlp.fc_in.weight": "pytorch_model-00002-of-00002.bin",
210
  "transformer.h.23.mlp.fc_out.bias": "pytorch_model-00002-of-00002.bin",
211
  "transformer.h.23.mlp.fc_out.weight": "pytorch_model-00002-of-00002.bin",
212
  "transformer.h.24.attn.bias": "pytorch_model-00002-of-00002.bin",
1
  {
2
  "metadata": {
3
+ "total_size": 12116445688.0
4
  },
5
  "weight_map": {
6
  "lm_head.bias": "pytorch_model-00002-of-00002.bin",
205
  "transformer.h.23.attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
206
  "transformer.h.23.ln_1.bias": "pytorch_model-00001-of-00002.bin",
207
  "transformer.h.23.ln_1.weight": "pytorch_model-00001-of-00002.bin",
208
+ "transformer.h.23.mlp.fc_in.bias": "pytorch_model-00001-of-00002.bin",
209
+ "transformer.h.23.mlp.fc_in.weight": "pytorch_model-00001-of-00002.bin",
210
  "transformer.h.23.mlp.fc_out.bias": "pytorch_model-00002-of-00002.bin",
211
  "transformer.h.23.mlp.fc_out.weight": "pytorch_model-00002-of-00002.bin",
212
  "transformer.h.24.attn.bias": "pytorch_model-00002-of-00002.bin",