panicpanicpanic commited on
Commit
966cbf3
·
verified ·
1 Parent(s): 6cc5532

Training in progress, step 3

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "gpt2-medium",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
@@ -12,13 +12,11 @@
12
  "layer_norm_epsilon": 1e-05,
13
  "model_type": "gpt2",
14
  "n_ctx": 1024,
15
- "n_embd": 1024,
16
- "n_head": 16,
17
  "n_inner": null,
18
- "n_layer": 24,
19
  "n_positions": 1024,
20
- "n_special": 0,
21
- "predict_special_tokens": true,
22
  "reorder_and_upcast_attn": false,
23
  "resid_pdrop": 0.1,
24
  "scale_attn_by_inverse_layer_idx": false,
 
1
  {
2
+ "_name_or_path": "gpt2-large",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
 
12
  "layer_norm_epsilon": 1e-05,
13
  "model_type": "gpt2",
14
  "n_ctx": 1024,
15
+ "n_embd": 1280,
16
+ "n_head": 20,
17
  "n_inner": null,
18
+ "n_layer": 36,
19
  "n_positions": 1024,
 
 
20
  "reorder_and_upcast_attn": false,
21
  "resid_pdrop": 0.1,
22
  "scale_attn_by_inverse_layer_idx": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:013226b1044e4e943b58d3f8411ab92012a2776e8c24ecb426b6dcc5489189f1
3
- size 1419322880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9daec3d9afb56155d3065913e51636b232be0e1826a9079623ece03c90eff39f
3
+ size 3096165928
runs/Nov21_13-14-55_76a1de55ef0b/events.out.tfevents.1732194897.76a1de55ef0b.590.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d09051e8969e5a4d7da5b01567560df7f9e39b1e6458233164bbe128ae3c4d8d
3
+ size 5601
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:643cc8bce87c960d663a0ef78ec7c404e98e8f66d659376e39af7d8fc993a061
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebdc7aeaaec2ac9c423103b57521145ab0012ac8ca6855b8e48d0a9f94714c23
3
  size 5240