Akoo commited on
Commit
8d37334
1 Parent(s): b67cb96

Training in progress, step 10

Browse files
config.json CHANGED
@@ -5,13 +5,13 @@
5
  "GPT2LMHeadModel"
6
  ],
7
  "attn_pdrop": 0.1,
8
- "bos_token_id": 0,
9
  "embd_pdrop": 0.1,
10
- "eos_token_id": 0,
11
  "initializer_range": 0.02,
12
  "layer_norm_epsilon": 1e-05,
13
  "model_type": "gpt2",
14
- "n_ctx": 100,
15
  "n_embd": 768,
16
  "n_head": 12,
17
  "n_inner": null,
5
  "GPT2LMHeadModel"
6
  ],
7
  "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
  "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
  "initializer_range": 0.02,
12
  "layer_norm_epsilon": 1e-05,
13
  "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
  "n_embd": 768,
16
  "n_head": 12,
17
  "n_inner": null,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b356d9cd807321b1879355f00e4922de1b4ac1a0f298f34f824ff6b56c2c9ff9
3
  size 456667881
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:081428d85f8e1e409a0a15b3d1bd2193ea0532020748aea629a5adb5ca0a523d
3
  size 456667881
runs/Aug19_06-56-44_7a40bd7a314c/1660892267.56799/events.out.tfevents.1660892267.7a40bd7a314c.56.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9647d951a007b3c858507cf75adf55f4a9c1dc8076384f37cfea196ede86c9d6
3
+ size 5376
runs/Aug19_06-56-44_7a40bd7a314c/events.out.tfevents.1660892267.7a40bd7a314c.56.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:455b9d1879d2e6992437d8b537b572897dcb8e3ce7d657bddb93dd61ea5e9596
3
+ size 4320
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:78ed9d1615888cf71143d49d9fb728e06b44bda4ccb206f1af0dfac3b44e00d0
3
  size 3311
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4772f76b5265065a50f996e6362c246f2d8f2485a73cf1d64689ac933913bcd0
3
  size 3311