tingtone commited on
Commit
3ccf552
1 Parent(s): ab8ea0b

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
- "_name_or_path": "distilgpt2",
3
- "_num_labels": 1,
4
  "activation_function": "gelu_new",
5
  "adapters": {
6
  "adapters": {},
@@ -82,7 +81,7 @@
82
  "n_embd": 768,
83
  "n_head": 12,
84
  "n_inner": null,
85
- "n_layer": 6,
86
  "n_positions": 1024,
87
  "pad_token_id": 50256,
88
  "problem_type": "multi_label_classification",
 
1
  {
2
+ "_name_or_path": "gpt2",
 
3
  "activation_function": "gelu_new",
4
  "adapters": {
5
  "adapters": {},
 
81
  "n_embd": 768,
82
  "n_head": 12,
83
  "n_inner": null,
84
+ "n_layer": 12,
85
  "n_positions": 1024,
86
  "pad_token_id": 50256,
87
  "problem_type": "multi_label_classification",
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23b77ca41fc630f08d9a804595581949baaa0a88dd4af1b12f0f57908e4eb4c9
3
- size 334056380
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d5ede07948055d4816d7b7f3b2cde3c8478ee9d9d3b200451cd15fa43b98761
3
+ size 510484225
runs/Jun05_13-34-18_umn-20230531-185317/events.out.tfevents.1686016797.umn-20230531-185317.174865.16 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:534fae0c6c61f4599eff1f0ae73b6acb903e914373b623a6fe0eb224a577bd1a
3
- size 77031
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e50d25aa0735864e2696b78d054432f5ac54f4e43c6987afab24192db1286bb
3
+ size 77391
runs/Jun07_16-46-50_umn-20230531-185317/1686156415.266072/events.out.tfevents.1686156415.umn-20230531-185317.2785.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6660e7134b6202d967e87c9df4eafcf633dad779f9e0ba01a97269be1af7d840
3
+ size 5709
runs/Jun07_16-46-50_umn-20230531-185317/events.out.tfevents.1686156415.umn-20230531-185317.2785.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:176d5b72f714884a168d892b9e4b04f92d28657090b2ab9fdd8a449fe799bd29
3
+ size 19572
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 50256,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 128,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 128
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 50256,
16
+ "pad_type_id": 0,
17
+ "pad_token": "<|endoftext|>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 50256,
tokenizer_config.json CHANGED
@@ -3,7 +3,7 @@
3
  "bos_token": "<|endoftext|>",
4
  "eos_token": "<|endoftext|>",
5
  "model_max_length": 1024,
6
- "name_or_path": "distilgpt2",
7
  "problem_type": "multi_label_classification",
8
  "special_tokens_map_file": null,
9
  "tokenizer_class": "GPT2Tokenizer",
 
3
  "bos_token": "<|endoftext|>",
4
  "eos_token": "<|endoftext|>",
5
  "model_max_length": 1024,
6
+ "name_or_path": "gpt2",
7
  "problem_type": "multi_label_classification",
8
  "special_tokens_map_file": null,
9
  "tokenizer_class": "GPT2Tokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f96c42989cb9aa1677d0d93fad923ef1a94ff7f2e31832473220b572a6fd11f
3
  size 3515
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1e791f1335427150d87b31043a38b7c4fac88aab7238aaee56544589d977c13
3
  size 3515