Peter commited on
Commit
524a4e1
1 Parent(s): c2d1935

medium_test

Browse files
all_results.json CHANGED
@@ -1,28 +1,28 @@
1
  {
2
- "epoch": 2.95,
3
- "eval_gen_len": 33.1429,
4
- "eval_loss": 0.4828817844390869,
5
- "eval_rouge1": 80.6131,
6
- "eval_rouge2": 73.305,
7
- "eval_rougeL": 79.6954,
8
- "eval_rougeLsum": 79.6954,
9
- "eval_runtime": 2.536,
10
  "eval_samples": 7,
11
- "eval_samples_per_second": 2.76,
12
- "eval_steps_per_second": 0.789,
13
- "predict_gen_len": 33.1429,
14
- "predict_loss": 0.4828817844390869,
15
- "predict_rouge1": 80.6131,
16
- "predict_rouge2": 73.305,
17
- "predict_rougeL": 79.6954,
18
- "predict_rougeLsum": 79.6954,
19
- "predict_runtime": 2.541,
20
  "predict_samples": 7,
21
- "predict_samples_per_second": 2.755,
22
- "predict_steps_per_second": 0.787,
23
- "train_loss": 0.7666861640082465,
24
- "train_runtime": 31.051,
25
  "train_samples": 252,
26
- "train_samples_per_second": 24.347,
27
- "train_steps_per_second": 1.449
28
  }
 
1
  {
2
+ "epoch": 3.0,
3
+ "eval_gen_len": 32.4286,
4
+ "eval_loss": 0.6024884581565857,
5
+ "eval_rouge1": 81.6007,
6
+ "eval_rouge2": 75.1196,
7
+ "eval_rougeL": 81.4213,
8
+ "eval_rougeLsum": 81.4956,
9
+ "eval_runtime": 2.4585,
10
  "eval_samples": 7,
11
+ "eval_samples_per_second": 2.847,
12
+ "eval_steps_per_second": 0.814,
13
+ "predict_gen_len": 32.4286,
14
+ "predict_loss": 0.6024884581565857,
15
+ "predict_rouge1": 81.6007,
16
+ "predict_rouge2": 75.1196,
17
+ "predict_rougeL": 81.4213,
18
+ "predict_rougeLsum": 81.4956,
19
+ "predict_runtime": 2.535,
20
  "predict_samples": 7,
21
+ "predict_samples_per_second": 2.761,
22
+ "predict_steps_per_second": 0.789,
23
+ "train_loss": 0.2226855363795366,
24
+ "train_runtime": 43.1557,
25
  "train_samples": 252,
26
+ "train_samples_per_second": 17.518,
27
+ "train_steps_per_second": 4.379
28
  }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "Vamsi/T5_Paraphrase_Paws",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
 
1
  {
2
+ "_name_or_path": "prithivida/parrot_paraphraser_on_T5",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
eval_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
- "epoch": 2.95,
3
- "eval_gen_len": 33.1429,
4
- "eval_loss": 0.4828817844390869,
5
- "eval_rouge1": 80.6131,
6
- "eval_rouge2": 73.305,
7
- "eval_rougeL": 79.6954,
8
- "eval_rougeLsum": 79.6954,
9
- "eval_runtime": 2.536,
10
  "eval_samples": 7,
11
- "eval_samples_per_second": 2.76,
12
- "eval_steps_per_second": 0.789
13
  }
 
1
  {
2
+ "epoch": 3.0,
3
+ "eval_gen_len": 32.4286,
4
+ "eval_loss": 0.6024884581565857,
5
+ "eval_rouge1": 81.6007,
6
+ "eval_rouge2": 75.1196,
7
+ "eval_rougeL": 81.4213,
8
+ "eval_rougeLsum": 81.4956,
9
+ "eval_runtime": 2.4585,
10
  "eval_samples": 7,
11
+ "eval_samples_per_second": 2.847,
12
+ "eval_steps_per_second": 0.814
13
  }
generated_predictions.txt CHANGED
@@ -1,7 +1,7 @@
1
- Danny Bazil Riley started working as general manager at a commercial real estate firm at an annual base salary of #36;70,000. Soon after a financial planner stopped by his desk to drop off brochures about insurance benefits available through his employer. But at 32, "buying insurance was the furthest thing from my mind," Riley says.
2
  Short-sellers, Wall Street's dwindling band of ultra-cynics, are seeing green again ### END
3
  Soaring crude prices ### outlook for earnings are expected to hang over the stock market next ### END
4
- OPEC can do nothing to douse scorching oil prices when markets are already oversupplied by 2.8 million barrels per day (bpd) of crude ### END
5
  Non-OPEC oil exporters should consider increasing output to cool record crude prices ### END
6
  The auction for Google Inc.'s highly anticipated initial public offering got off to a rocky start on Friday ### END
7
  The dollar tumbled broadly on Friday ### END
 
1
+ After earning a PH.D. in Sociology Danny Bazil Riley started to work as the general manager at a commercial real estate firm at an annual base salary of #36;70,000. Soon after, a financial planner stopped by his desk to drop off brochures about insurance benefits available through his employer ### END
2
  Short-sellers, Wall Street's dwindling band of ultra-cynics, are seeing green again ### END
3
  Soaring crude prices ### outlook for earnings are expected to hang over the stock market next ### END
4
+ OPEC can do nothing to douse scorching oil prices ### markets are already oversupplied by 2.8 million barrels per day (bpd) of crude ### END
5
  Non-OPEC oil exporters should consider increasing output to cool record crude prices ### END
6
  The auction for Google Inc.'s highly anticipated initial public offering got off to a rocky start on Friday ### END
7
  The dollar tumbled broadly on Friday ### END
predict_results.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
- "predict_gen_len": 33.1429,
3
- "predict_loss": 0.4828817844390869,
4
- "predict_rouge1": 80.6131,
5
- "predict_rouge2": 73.305,
6
- "predict_rougeL": 79.6954,
7
- "predict_rougeLsum": 79.6954,
8
- "predict_runtime": 2.541,
9
  "predict_samples": 7,
10
- "predict_samples_per_second": 2.755,
11
- "predict_steps_per_second": 0.787
12
  }
 
1
  {
2
+ "predict_gen_len": 32.4286,
3
+ "predict_loss": 0.6024884581565857,
4
+ "predict_rouge1": 81.6007,
5
+ "predict_rouge2": 75.1196,
6
+ "predict_rougeL": 81.4213,
7
+ "predict_rougeLsum": 81.4956,
8
+ "predict_runtime": 2.535,
9
  "predict_samples": 7,
10
+ "predict_samples_per_second": 2.761,
11
+ "predict_steps_per_second": 0.789
12
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6bfd4f9feaf2691d94dc1e0bbb2677fb114aa0a1a4ed1f7370504cf292a19bdb
3
  size 891644863
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e92c6fd36fd7bae27e500790aae82c06d5810d9119113512024bfd270cc29b0
3
  size 891644863
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "extra_ids": 100, "additional_special_tokens": ["<extra_id_0>", "<extra_id_1>", "<extra_id_2>", "<extra_id_3>", "<extra_id_4>", "<extra_id_5>", "<extra_id_6>", "<extra_id_7>", "<extra_id_8>", "<extra_id_9>", "<extra_id_10>", "<extra_id_11>", "<extra_id_12>", "<extra_id_13>", "<extra_id_14>", "<extra_id_15>", "<extra_id_16>", "<extra_id_17>", "<extra_id_18>", "<extra_id_19>", "<extra_id_20>", "<extra_id_21>", "<extra_id_22>", "<extra_id_23>", "<extra_id_24>", "<extra_id_25>", "<extra_id_26>", "<extra_id_27>", "<extra_id_28>", "<extra_id_29>", "<extra_id_30>", "<extra_id_31>", "<extra_id_32>", "<extra_id_33>", "<extra_id_34>", "<extra_id_35>", "<extra_id_36>", "<extra_id_37>", "<extra_id_38>", "<extra_id_39>", "<extra_id_40>", "<extra_id_41>", "<extra_id_42>", "<extra_id_43>", "<extra_id_44>", "<extra_id_45>", "<extra_id_46>", "<extra_id_47>", "<extra_id_48>", "<extra_id_49>", "<extra_id_50>", "<extra_id_51>", "<extra_id_52>", "<extra_id_53>", "<extra_id_54>", "<extra_id_55>", "<extra_id_56>", "<extra_id_57>", "<extra_id_58>", "<extra_id_59>", "<extra_id_60>", "<extra_id_61>", "<extra_id_62>", "<extra_id_63>", "<extra_id_64>", "<extra_id_65>", "<extra_id_66>", "<extra_id_67>", "<extra_id_68>", "<extra_id_69>", "<extra_id_70>", "<extra_id_71>", "<extra_id_72>", "<extra_id_73>", "<extra_id_74>", "<extra_id_75>", "<extra_id_76>", "<extra_id_77>", "<extra_id_78>", "<extra_id_79>", "<extra_id_80>", "<extra_id_81>", "<extra_id_82>", "<extra_id_83>", "<extra_id_84>", "<extra_id_85>", "<extra_id_86>", "<extra_id_87>", "<extra_id_88>", "<extra_id_89>", "<extra_id_90>", "<extra_id_91>", "<extra_id_92>", "<extra_id_93>", "<extra_id_94>", "<extra_id_95>", "<extra_id_96>", "<extra_id_97>", "<extra_id_98>", "<extra_id_99>"], "model_max_length": 512, "special_tokens_map_file": "C:\\Users\\peter/.cache\\huggingface\\transformers\\ca670222dc28504dc97cb6136cadc777e4158e3b346f98b99064406f5ac6eac0.c94798918c92ded6aeef2d2f0e666d2cc4145eca1aa6e1336fde07f2e13e2f46", "name_or_path": "Vamsi/T5_Paraphrase_Paws", "sp_model_kwargs": {}, "tokenizer_class": "T5Tokenizer"}
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "extra_ids": 100, "additional_special_tokens": ["<extra_id_0>", "<extra_id_1>", "<extra_id_2>", "<extra_id_3>", "<extra_id_4>", "<extra_id_5>", "<extra_id_6>", "<extra_id_7>", "<extra_id_8>", "<extra_id_9>", "<extra_id_10>", "<extra_id_11>", "<extra_id_12>", "<extra_id_13>", "<extra_id_14>", "<extra_id_15>", "<extra_id_16>", "<extra_id_17>", "<extra_id_18>", "<extra_id_19>", "<extra_id_20>", "<extra_id_21>", "<extra_id_22>", "<extra_id_23>", "<extra_id_24>", "<extra_id_25>", "<extra_id_26>", "<extra_id_27>", "<extra_id_28>", "<extra_id_29>", "<extra_id_30>", "<extra_id_31>", "<extra_id_32>", "<extra_id_33>", "<extra_id_34>", "<extra_id_35>", "<extra_id_36>", "<extra_id_37>", "<extra_id_38>", "<extra_id_39>", "<extra_id_40>", "<extra_id_41>", "<extra_id_42>", "<extra_id_43>", "<extra_id_44>", "<extra_id_45>", "<extra_id_46>", "<extra_id_47>", "<extra_id_48>", "<extra_id_49>", "<extra_id_50>", "<extra_id_51>", "<extra_id_52>", "<extra_id_53>", "<extra_id_54>", "<extra_id_55>", "<extra_id_56>", "<extra_id_57>", "<extra_id_58>", "<extra_id_59>", "<extra_id_60>", "<extra_id_61>", "<extra_id_62>", "<extra_id_63>", "<extra_id_64>", "<extra_id_65>", "<extra_id_66>", "<extra_id_67>", "<extra_id_68>", "<extra_id_69>", "<extra_id_70>", "<extra_id_71>", "<extra_id_72>", "<extra_id_73>", "<extra_id_74>", "<extra_id_75>", "<extra_id_76>", "<extra_id_77>", "<extra_id_78>", "<extra_id_79>", "<extra_id_80>", "<extra_id_81>", "<extra_id_82>", "<extra_id_83>", "<extra_id_84>", "<extra_id_85>", "<extra_id_86>", "<extra_id_87>", "<extra_id_88>", "<extra_id_89>", "<extra_id_90>", "<extra_id_91>", "<extra_id_92>", "<extra_id_93>", "<extra_id_94>", "<extra_id_95>", "<extra_id_96>", "<extra_id_97>", "<extra_id_98>", "<extra_id_99>"], "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "prithivida/parrot_paraphraser_on_T5", "sp_model_kwargs": {}, "tokenizer_class": "T5Tokenizer"}
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 2.95,
3
- "train_loss": 0.7666861640082465,
4
- "train_runtime": 31.051,
5
  "train_samples": 252,
6
- "train_samples_per_second": 24.347,
7
- "train_steps_per_second": 1.449
8
  }
 
1
  {
2
+ "epoch": 3.0,
3
+ "train_loss": 0.2226855363795366,
4
+ "train_runtime": 43.1557,
5
  "train_samples": 252,
6
+ "train_samples_per_second": 17.518,
7
+ "train_steps_per_second": 4.379
8
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf8d9c3c9936cb871cebcdaea9f4bf6dce8ca2eca8a689ec8e0471393131ba89
3
  size 3055
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f54511f437804c9424ca64937a125d5979070916cc062e4a470a0c85309693fb
3
  size 3055