nbroad HF staff commited on
Commit
1455619
1 Parent(s): 9c0ed9f

Saving weights and logs of epoch 0

Browse files
.gitattributes CHANGED
@@ -25,3 +25,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ qgen_train.csv filter=lfs diff=lfs merge=lfs -text
29
+ qgen_training_data.csv filter=lfs diff=lfs merge=lfs -text
30
+ qgen_val.csv filter=lfs diff=lfs merge=lfs -text
.lock ADDED
File without changes
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MT5ForConditionalGeneration"
4
+ ],
5
+ "d_ff": 1024,
6
+ "d_kv": 64,
7
+ "d_model": 512,
8
+ "decoder_start_token_id": 0,
9
+ "dropout_rate": 0.1,
10
+ "eos_token_id": 1,
11
+ "feed_forward_proj": "gated-gelu",
12
+ "initializer_factor": 1.0,
13
+ "is_encoder_decoder": true,
14
+ "layer_norm_epsilon": 1e-06,
15
+ "model_type": "mt5",
16
+ "num_decoder_layers": 8,
17
+ "num_heads": 6,
18
+ "num_layers": 8,
19
+ "pad_token_id": 0,
20
+ "relative_attention_num_buckets": 32,
21
+ "tie_word_embeddings": false,
22
+ "tokenizer_class": "T5Tokenizer",
23
+ "transformers_version": "4.12.0.dev0",
24
+ "use_cache": true,
25
+ "vocab_size": 250112
26
+ }
events.out.tfevents.1634611280.t1v-n-10409466-w-0.11492.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e182970a4bdaa09ab4bc39e24998573f2279663d8092199e97f5f27f509662e0
3
+ size 40
events.out.tfevents.1634647562.t1v-n-10409466-w-0.4496.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62f76d55580d4e2d1fe8679c9fe034df11554145300a05ce56d4d4839bf55ff6
3
+ size 40
events.out.tfevents.1634681194.t1v-n-10409466-w-0.3720.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a33497fbf05f4540c98f5a8a9a4bdf2f88c6330c4bd969cd379a55f84b0d2e6
3
+ size 2450039
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74d1919cb1cc5091f578509943b09248e0a3f4cbdbdebf111c0d96fb2a47ff3d
3
+ size 1200715307
mt5-base-qgen/events.out.tfevents.1634610755.t1v-n-10409466-w-0.9004.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e1acf82767bf97369c40271b7915dfe127686a6a3908c2d97b7626d45315e6d
3
+ size 40
mt5-questions-dataset.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [{"stream_name":"stderr","time":34.74216129,"data":"/opt/conda/lib/python3.7/site-packages/papermill/iorw.py:50: FutureWarning: pyarrow.HadoopFileSystem is deprecated as of 2.0.0, please use pyarrow.fs.HadoopFileSystem instead.\n"}
2
+ ,{"stream_name":"stderr","time":34.74226826,"data":" from pyarrow import HadoopFileSystem\n"}
3
+ ,{"stream_name":"stderr","time":37.894455456,"data":"/opt/conda/lib/python3.7/site-packages/traitlets/traitlets.py:2567: FutureWarning: --Exporter.preprocessors=[\"remove_papermill_header.RemovePapermillHeader\"] for containers is deprecated in traitlets 5.0. You can pass `--Exporter.preprocessors item` ... multiple times to add items to a list.\n"}
4
+ ,{"stream_name":"stderr","time":37.89452555,"data":" FutureWarning,\n"}
5
+ ,{"stream_name":"stderr","time":37.894535347,"data":"[NbConvertApp] Converting notebook __notebook__.ipynb to notebook\n"}
6
+ ,{"stream_name":"stderr","time":37.920776986,"data":"[NbConvertApp] Writing 2546 bytes to __notebook__.ipynb\n"}
7
+ ,{"stream_name":"stderr","time":40.120591007,"data":"/opt/conda/lib/python3.7/site-packages/traitlets/traitlets.py:2567: FutureWarning: --Exporter.preprocessors=[\"nbconvert.preprocessors.ExtractOutputPreprocessor\"] for containers is deprecated in traitlets 5.0. You can pass `--Exporter.preprocessors item` ... multiple times to add items to a list.\n"}
8
+ ,{"stream_name":"stderr","time":40.12066086,"data":" FutureWarning,\n"}
9
+ ,{"stream_name":"stderr","time":40.120670616,"data":"[NbConvertApp] Converting notebook __notebook__.ipynb to html\n"}
10
+ ,{"stream_name":"stderr","time":40.662991409,"data":"[NbConvertApp] Writing 276762 bytes to __results__.html\n"}
11
+ ]
qgen_train.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21a0e59e17a98496c830c4eaa5e7109e648ccc49905acd30cc24dc477233cd71
3
+ size 337028529
qgen_training_data.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8045ce51c73fed96d10df60dae29f3dc9eb6344dd84046e65f38aa00edd0d258
3
+ size 420218929
qgen_val.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:215fc1360575a907d9d0dc89fbe08e656e41c7ca76df813c12b95eb98fd68597
3
+ size 83190420
run_summarization_flax.py CHANGED
@@ -446,7 +446,7 @@ def main():
446
  # In Flax, for seq2seq models we need to pass `decoder_input_ids`
447
  # as the Flax models don't accept `labels`, we need to prepare the decoder_input_ids here
448
  # for that dynamically import the `shift_tokens_right` function from the model file
449
- model_module = __import__(model.__module__, fromlist=["shift_tokens_tight"])
450
  shift_tokens_right_fn = getattr(model_module, "shift_tokens_right")
451
 
452
  # Setting padding="max_length" as we need fixed length inputs for jitted functions
 
446
  # In Flax, for seq2seq models we need to pass `decoder_input_ids`
447
  # as the Flax models don't accept `labels`, we need to prepare the decoder_input_ids here
448
  # for that dynamically import the `shift_tokens_right` function from the model file
449
+ model_module = __import__("transformers.models.t5.modeling_flax_t5", fromlist=["shift_tokens_tight"])
450
  shift_tokens_right_fn = getattr(model_module, "shift_tokens_right")
451
 
452
  # Setting padding="max_length" as we need fixed length inputs for jitted functions
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
3
+ size 4309802
start_qgen.sh CHANGED
@@ -1,15 +1,21 @@
 
 
1
  python run_summarization_flax.py \
2
- --output_dir ./mt5-base-qgen \
3
- --model_name_or_path google/mt5-base \
4
- --tokenizer_name google/mt5-base \
5
- --train_file qgen_training_data.csv \
6
- --text_column context \
7
- --summary_column question \
8
- --do_train
9
- --num_train_epochs 3 \
10
- --learning_rate 5e-5 --warmup_steps 300 \
11
- --per_device_train_batch_size 2 \
12
- --per_device_eval_batch_size 2 \
13
- --overwrite_output_dir \
14
- --max_source_length 1024 --max_target_length 64 \
15
- --push_to_hub
 
 
 
 
 
1
+ #! /usr/bin/env bash
2
+
3
  python run_summarization_flax.py \
4
+ --output_dir ./ \
5
+ --model_name_or_path google/mt5-small \
6
+ --tokenizer_name google/mt5-small \
7
+ --train_file qgen_train.csv \
8
+ --validation_file qgen_val.csv \
9
+ --text_column context \
10
+ --summary_column question \
11
+ --do_train \
12
+ --do_eval \
13
+ --num_train_epochs 3 \
14
+ --eval_steps 2000 \
15
+ --logging_steps 100 \
16
+ --learning_rate 5e-5 --warmup_steps 1000 \
17
+ --per_device_train_batch_size 2 \
18
+ --per_device_eval_batch_size 2 \
19
+ --overwrite_output_dir \
20
+ --max_source_length 1024 --max_target_length 64 \
21
+ --push_to_hub
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "extra_ids": 0, "additional_special_tokens": null, "special_tokens_map_file": "/home/nicholas/.cache/huggingface/transformers/685ac0ca8568ec593a48b61b0a3c272beee9bc194a3c7241d15dcadb5f875e53.f76030f3ec1b96a8199b2593390c610e76ca8028ef3d24680000619ffb646276", "name_or_path": "google/mt5-small", "sp_model_kwargs": {}, "tokenizer_class": "T5Tokenizer"}