dhanesh123in commited on
Commit
981c037
1 Parent(s): 35a96f9

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -12,13 +12,13 @@
12
  "n_embd": 768,
13
  "n_head": 12,
14
  "n_inner": null,
15
- "n_layer": 6,
16
  "n_positions": 1024,
17
  "reorder_and_upcast_attn": false,
18
  "resid_pdrop": 0.1,
19
  "scale_attn_by_inverse_layer_idx": false,
20
  "scale_attn_weights": true,
21
- "skip_mlp": true,
22
  "summary_activation": null,
23
  "summary_first_dropout": 0.1,
24
  "summary_proj_to_labels": true,
 
12
  "n_embd": 768,
13
  "n_head": 12,
14
  "n_inner": null,
15
+ "n_layer": 1,
16
  "n_positions": 1024,
17
  "reorder_and_upcast_attn": false,
18
  "resid_pdrop": 0.1,
19
  "scale_attn_by_inverse_layer_idx": false,
20
  "scale_attn_weights": true,
21
+ "skip_mlp": false,
22
  "summary_activation": null,
23
  "summary_first_dropout": 0.1,
24
  "summary_proj_to_labels": true,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:530ae57ee80655b3179d6d33e589862fe71f575b9e9b3c9524b2a9a88aa39a72
3
- size 214279072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:888a27584dcfcbad87142839d86281521ed414857902881282f3c18846d09920
3
+ size 185894448
runs/Jul01_22-44-39_morpheus/events.out.tfevents.1719854080.morpheus.428151.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70d31de4e2a93be4e9f0d87e7b6b9d058121b17dbbf0e7563482917f3bba805e
3
+ size 6631
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d70e5e8bcf0a9922b63a3c7a43421ec844f5a467210a200fa5a5684e502ec19
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d83def51f4abdf458ba669c2a368d07ddabe23ac4da69a5ec558ba10ccd7a870
3
  size 5112