debbiesoon commited on
Commit
a8120e0
1 Parent(s): 4ef9805

Training in progress, step 10

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "debbiesoon/summarise",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "architectures": [
@@ -45,7 +45,7 @@
45
  "length_penalty": 2.0,
46
  "max_decoder_position_embeddings": 1024,
47
  "max_encoder_position_embeddings": 16384,
48
- "max_length": 1000,
49
  "min_length": 100,
50
  "model_type": "led",
51
  "no_repeat_ngram_size": 3,
 
1
  {
2
+ "_name_or_path": "allenai/led-base-16384",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "architectures": [
 
45
  "length_penalty": 2.0,
46
  "max_decoder_position_embeddings": 1024,
47
  "max_encoder_position_embeddings": 16384,
48
+ "max_length": 768,
49
  "min_length": 100,
50
  "model_type": "led",
51
  "no_repeat_ngram_size": 3,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f91e68d6f0b4964a1452e59408a7700c97252caaa9b7d9a956d13ebf07460fa
3
- size 647674353
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35794e75227a1df9f29a12a8f68dd1ff5a74e9e451e4f5bf0283a82886375710
3
+ size 647678513
runs/Oct18_13-51-50_13a7832f06ec/1666101362.0218272/events.out.tfevents.1666101362.13a7832f06ec.95.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05084400ac7ee5de5b2f1952ca9fdc077504c772d052dc6fcefa5198091ddd82
3
+ size 5633
runs/Oct18_13-51-50_13a7832f06ec/events.out.tfevents.1666101362.13a7832f06ec.95.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fe4e562b9fae420f7b5c2918bc9cff55e25b004d8ad24a590c71df4d0a08aa3
3
+ size 5113
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 768,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 768
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 1,
16
+ "pad_type_id": 0,
17
+ "pad_token": "<pad>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d82e9d3c4317d567897e68e626a21dc50a25b5292c351c69e80b9518cd3b58ca
3
  size 3439
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c13db8764b09cf4327e1cf9499a75a5afb6e8ec3f3f36e33cdee08dcc0ee8866
3
  size 3439