Ubuntu commited on
Commit
75d4915
1 Parent(s): 2b93166

First commit

Browse files
.gitattributes CHANGED
@@ -25,3 +25,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ events.out.tfevents.1610817954.ronak-3 filter=lfs diff=lfs merge=lfs -text
29
+ events.out.tfevents.1610824073.ronak-3 filter=lfs diff=lfs merge=lfs -text
30
+ graph.pbtxt filter=lfs diff=lfs merge=lfs -text
31
+ model.ckpt-1009900.data-00001-of-00002 filter=lfs diff=lfs merge=lfs -text
32
+ model.ckpt-1009900.meta filter=lfs diff=lfs merge=lfs -text
checkpoint ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ model_checkpoint_path: "model.ckpt-1010900"
2
+ all_model_checkpoint_paths: "model.ckpt-1009900"
3
+ all_model_checkpoint_paths: "model.ckpt-1010900"
commands/command ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/ronakice/miniconda3/envs/py37/bin/t5_mesh_transformer --tpu=tpu-ronak-38 --gcp_project=justin-project-300310 --tpu_zone=europe-west4-a --model_dir=gs://pongo-bucket/duo-t5/experiments/38 --gin_param=init_checkpoint = 'gs://t5-data/pretrained_models/base/model.ckpt-999900' --gin_file=dataset.gin --gin_file=models/bi_v1.gin --gin_file=gs://t5-data/pretrained_models/base/operative_config.gin --gin_param=utils.tpu_mesh_shape.model_parallelism = 1 --gin_param=utils.tpu_mesh_shape.tpu_topology = '2x2' --gin_param=utils.run.train_dataset_fn = @t5.models.mesh_transformer.tsv_dataset_fn --gin_param=tsv_dataset_fn.filename = 'gs://duo-t5/data/query_doc_pairs.train.tsv' --gin_file=learning_rate_schedules/constant_0_001.gin --gin_param=run.train_steps = 1009900 --gin_param=run.save_checkpoints_steps = 10000 --gin_param=tokens_per_batch = 65536
commands/command.1 ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/ronakice/miniconda3/envs/py37/bin/t5_mesh_transformer --tpu=tpu-ronak-38 --gcp_project=justin-project-300310 --tpu_zone=europe-west4-a --model_dir=gs://pongo-bucket/duo-t5/experiments/38 --gin_param=init_checkpoint = 'gs://pongo-bucket/duo-t5/experiments/38/model.ckpt-1009900' --gin_file=dataset.gin --gin_file=models/bi_v1.gin --gin_file=gs://t5-data/pretrained_models/base/operative_config.gin --gin_param=utils.tpu_mesh_shape.model_parallelism = 1 --gin_param=utils.tpu_mesh_shape.tpu_topology = '2x2' --gin_param=utils.run.train_dataset_fn = @t5.models.mesh_transformer.tsv_dataset_fn --gin_param=tsv_dataset_fn.filename = 'gs://pongo-bucket/duo-t5/data/msmarco_med_rod/query_doc_pairs.train.tsv' --gin_file=learning_rate_schedules/constant_0_001.gin --gin_param=run.train_steps = 1010900 --gin_param=run.save_checkpoints_steps = 10000 --gin_param=tokens_per_batch = 65536
commands/command.2 ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/ronakice/miniconda3/envs/py37/bin/t5_mesh_transformer --tpu=tpu-ronak-1 --gcp_project=justin-project-300310 --tpu_zone=europe-west4-a --model_dir=gs://pongo-bucket/duo-t5/experiments/38 --gin_file=gs://t5-data/pretrained_models/base/operative_config.gin --gin_file=infer.gin --gin_file=beam_search.gin --gin_param=utils.tpu_mesh_shape.tpu_topology = '2x2' --gin_param=infer_checkpoint_step = 1009900 --gin_param=utils.run.sequence_length = {'inputs': 1024, 'targets': 2} --gin_param=Bitransformer.decode.max_decode_length = 2 --gin_param=input_filename = 'gs://pongo-bucket/trec-covid-complete/data/1/duot5_input.txt' --gin_param=output_filename = 'gs://pongo-bucket/trec-covid-complete/data/1/duot5_output.txt' --gin_param=tokens_per_batch = 131072 --gin_param=Bitransformer.decode.beam_size = 1 --gin_param=Bitransformer.decode.temperature = 0.0 --gin_param=Unitransformer.sample_autoregressive.sampling_keep_top_k = -1
commands/command.3 ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/ronakice/miniconda3/envs/py37/bin/t5_mesh_transformer --tpu=tpu-ronak-2 --gcp_project=justin-project-300310 --tpu_zone=europe-west4-a --model_dir=gs://pongo-bucket/duo-t5/experiments/38 --gin_file=gs://t5-data/pretrained_models/base/operative_config.gin --gin_file=infer.gin --gin_file=beam_search.gin --gin_param=utils.tpu_mesh_shape.tpu_topology = '2x2' --gin_param=infer_checkpoint_step = 1010900 --gin_param=utils.run.sequence_length = {'inputs': 1024, 'targets': 2} --gin_param=Bitransformer.decode.max_decode_length = 2 --gin_param=input_filename = 'gs://pongo-bucket/trec-covid-complete/data/2/duot5_input.txt' --gin_param=output_filename = 'gs://pongo-bucket/trec-covid-complete/data/2/duot5_output.txt' --gin_param=tokens_per_batch = 131072 --gin_param=Bitransformer.decode.beam_size = 1 --gin_param=Bitransformer.decode.temperature = 0.0 --gin_param=Unitransformer.sample_autoregressive.sampling_keep_top_k = -1
commands/command.4 ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/ronakice/miniconda3/envs/py37/bin/t5_mesh_transformer --tpu=tpu-ronak-1 --gcp_project=justin-project-300310 --tpu_zone=europe-west4-a --model_dir=gs://pongo-bucket/duo-t5/experiments/38 --gin_file=gs://t5-data/pretrained_models/base/operative_config.gin --gin_file=infer.gin --gin_file=beam_search.gin --gin_param=utils.tpu_mesh_shape.tpu_topology = '2x2' --gin_param=infer_checkpoint_step = 1009900 --gin_param=utils.run.sequence_length = {'inputs': 1024, 'targets': 2} --gin_param=Bitransformer.decode.max_decode_length = 2 --gin_param=input_filename = 'gs://pongo-bucket/trec-covid-complete/data/5/duot5_input.txt' --gin_param=output_filename = 'gs://pongo-bucket/trec-covid-complete/data/5/duot5_output.txt' --gin_param=tokens_per_batch = 131072 --gin_param=Bitransformer.decode.beam_size = 1 --gin_param=Bitransformer.decode.temperature = 0.0 --gin_param=Unitransformer.sample_autoregressive.sampling_keep_top_k = -1
commands/command.5 ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/ronakice/miniconda3/envs/py37/bin/t5_mesh_transformer --tpu=tpu-ronak-2 --gcp_project=justin-project-300310 --tpu_zone=europe-west4-a --model_dir=gs://pongo-bucket/duo-t5/experiments/38 --gin_file=gs://t5-data/pretrained_models/base/operative_config.gin --gin_file=infer.gin --gin_file=beam_search.gin --gin_param=utils.tpu_mesh_shape.tpu_topology = '2x2' --gin_param=infer_checkpoint_step = 1010900 --gin_param=utils.run.sequence_length = {'inputs': 1024, 'targets': 2} --gin_param=Bitransformer.decode.max_decode_length = 2 --gin_param=input_filename = 'gs://pongo-bucket/trec-covid-complete/data/6/duot5_input.txt' --gin_param=output_filename = 'gs://pongo-bucket/trec-covid-complete/data/6/duot5_output.txt' --gin_param=tokens_per_batch = 131072 --gin_param=Bitransformer.decode.beam_size = 1 --gin_param=Bitransformer.decode.temperature = 0.0 --gin_param=Unitransformer.sample_autoregressive.sampling_keep_top_k = -1
commands/command.6 ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/ronakice/miniconda3/envs/py37/bin/t5_mesh_transformer --tpu=tpu-ronak-1 --gcp_project=justin-project-300310 --tpu_zone=europe-west4-a --model_dir=gs://pongo-bucket/duo-t5/experiments/38 --gin_file=gs://t5-data/pretrained_models/base/operative_config.gin --gin_file=infer.gin --gin_file=beam_search.gin --gin_param=utils.tpu_mesh_shape.tpu_topology = '2x2' --gin_param=infer_checkpoint_step = 1009900 --gin_param=utils.run.sequence_length = {'inputs': 1024, 'targets': 2} --gin_param=Bitransformer.decode.max_decode_length = 2 --gin_param=input_filename = 'gs://pongo-bucket/trec-covid-complete/data/9/duot5_input.txt' --gin_param=output_filename = 'gs://pongo-bucket/trec-covid-complete/data/9/duot5_output.txt' --gin_param=tokens_per_batch = 131072 --gin_param=Bitransformer.decode.beam_size = 1 --gin_param=Bitransformer.decode.temperature = 0.0 --gin_param=Unitransformer.sample_autoregressive.sampling_keep_top_k = -1
commands/command.7 ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/ronakice/miniconda3/envs/py37/bin/t5_mesh_transformer --tpu=tpu-ronak-2 --gcp_project=justin-project-300310 --tpu_zone=europe-west4-a --model_dir=gs://pongo-bucket/duo-t5/experiments/38 --gin_file=gs://t5-data/pretrained_models/base/operative_config.gin --gin_file=infer.gin --gin_file=beam_search.gin --gin_param=utils.tpu_mesh_shape.tpu_topology = '2x2' --gin_param=infer_checkpoint_step = 1010900 --gin_param=utils.run.sequence_length = {'inputs': 1024, 'targets': 2} --gin_param=Bitransformer.decode.max_decode_length = 2 --gin_param=input_filename = 'gs://pongo-bucket/trec-covid-complete/data/10/duot5_input.txt' --gin_param=output_filename = 'gs://pongo-bucket/trec-covid-complete/data/10/duot5_output.txt' --gin_param=tokens_per_batch = 131072 --gin_param=Bitransformer.decode.beam_size = 1 --gin_param=Bitransformer.decode.temperature = 0.0 --gin_param=Unitransformer.sample_autoregressive.sampling_keep_top_k = -1
commands/command.8 ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/ronakice/miniconda3/envs/py37/bin/t5_mesh_transformer --tpu=tpu-ronak-2 --gcp_project=justin-project-300310 --tpu_zone=europe-west4-a --model_dir=gs://pongo-bucket/duo-t5/experiments/38 --gin_file=gs://t5-data/pretrained_models/base/operative_config.gin --gin_file=infer.gin --gin_file=beam_search.gin --gin_param=utils.tpu_mesh_shape.tpu_topology = '2x2' --gin_param=infer_checkpoint_step = 1010900 --gin_param=utils.run.sequence_length = {'inputs': 1024, 'targets': 2} --gin_param=Bitransformer.decode.max_decode_length = 2 --gin_param=input_filename = 'gs://pongo-bucket/trec-covid-complete/data/14/duot5_input.txt' --gin_param=output_filename = 'gs://pongo-bucket/trec-covid-complete/data/14/duot5_output.txt' --gin_param=tokens_per_batch = 131072 --gin_param=Bitransformer.decode.beam_size = 1 --gin_param=Bitransformer.decode.temperature = 0.0 --gin_param=Unitransformer.sample_autoregressive.sampling_keep_top_k = -1
commands/command.9 ADDED
@@ -0,0 +1 @@
 
 
1
+ /home/ronakice/miniconda3/envs/py37/bin/t5_mesh_transformer --tpu=tpu-ronak-1 --gcp_project=justin-project-300310 --tpu_zone=europe-west4-a --model_dir=gs://pongo-bucket/duo-t5/experiments/38 --gin_file=gs://t5-data/pretrained_models/base/operative_config.gin --gin_file=infer.gin --gin_file=beam_search.gin --gin_param=utils.tpu_mesh_shape.tpu_topology = '2x2' --gin_param=infer_checkpoint_step = 1009900 --gin_param=utils.run.sequence_length = {'inputs': 1024, 'targets': 2} --gin_param=Bitransformer.decode.max_decode_length = 2 --gin_param=input_filename = 'gs://pongo-bucket/trec-covid-complete/data/13/duot5_input.txt' --gin_param=output_filename = 'gs://pongo-bucket/trec-covid-complete/data/13/duot5_output.txt' --gin_param=tokens_per_batch = 131072 --gin_param=Bitransformer.decode.beam_size = 1 --gin_param=Bitransformer.decode.temperature = 0.0 --gin_param=Unitransformer.sample_autoregressive.sampling_keep_top_k = -1
events.out.tfevents.1610817954.ronak-3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f73f8e8fc2910f645e93408610369e7ef17022109f26aac33ca22c649024526
3
+ size 71686468
events.out.tfevents.1610824073.ronak-3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb8b26893aa9a659f6a2202aa7403b9b1f48c49f0cda8c506bc056bdf2a89b0a
3
+ size 71689263
graph.pbtxt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e48e738645f83a7efcc38103726a7e06181d534297bc1ff67d4b0b954f740c3
3
+ size 39150190
model.ckpt-1009900.data-00000-of-00002 ADDED
Binary file (8 Bytes). View file
 
model.ckpt-1009900.data-00001-of-00002 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:065488b0fe4fc7a893143862d9bbb4329d2d4ef7f2ec97f97c703718cb0da54d
3
+ size 447754240
model.ckpt-1009900.index ADDED
Binary file (10.8 kB). View file
 
model.ckpt-1009900.meta ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e6b5c9606a0540c6b1639a4e16da8a32134e442f90c0f32b6dbdd302a52384b
3
+ size 18630025
operative_config.gin ADDED
@@ -0,0 +1,332 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import mesh_tensorflow.optimize
2
+ import mesh_tensorflow.transformer.dataset
3
+ import mesh_tensorflow.transformer.learning_rate_schedules
4
+ import mesh_tensorflow.transformer.t2t_vocabulary
5
+ import mesh_tensorflow.transformer.transformer
6
+ import mesh_tensorflow.transformer.transformer_layers
7
+ import mesh_tensorflow.transformer.utils
8
+ import t5.models.mesh_transformer
9
+
10
+ # Macros:
11
+ # ==============================================================================
12
+ d_ff = 3072
13
+ d_kv = 64
14
+ d_model = 768
15
+ dropout_rate = 0.1
16
+ init_checkpoint = 'gs://pongo-bucket/duo-t5/experiments/38/model.ckpt-1009900'
17
+ inputs_length = 512
18
+ MIXTURE_NAME = 'all_mix'
19
+ num_heads = 12
20
+ num_layers = 12
21
+ targets_length = 512
22
+ tokens_per_batch = 65536
23
+
24
+ # Parameters for adafactor_decay_rate_pow:
25
+ # ==============================================================================
26
+ adafactor_decay_rate_pow.offset = 0
27
+
28
+ # Parameters for AdafactorOptimizer:
29
+ # ==============================================================================
30
+ AdafactorOptimizer.beta1 = 0.0
31
+ AdafactorOptimizer.clipping_threshold = 1.0
32
+ AdafactorOptimizer.decay_rate = None
33
+ AdafactorOptimizer.epsilon1 = 1e-30
34
+ AdafactorOptimizer.epsilon2 = 0.001
35
+ AdafactorOptimizer.factored = True
36
+ AdafactorOptimizer.min_dim_size_to_factor = 128
37
+ AdafactorOptimizer.multiply_by_parameter_scale = True
38
+
39
+ # Parameters for Bitransformer:
40
+ # ==============================================================================
41
+ Bitransformer.shared_embedding = True
42
+
43
+ # Parameters for constant_learning_rate:
44
+ # ==============================================================================
45
+ constant_learning_rate.learning_rate = 0.001
46
+
47
+ # Parameters for decoder/DenseReluDense:
48
+ # ==============================================================================
49
+ decoder/DenseReluDense.activation = 'relu'
50
+ decoder/DenseReluDense.dropout_rate = %dropout_rate
51
+ decoder/DenseReluDense.hidden_size = %d_ff
52
+ decoder/DenseReluDense.use_bias = False
53
+
54
+ # Parameters for encoder/DenseReluDense:
55
+ # ==============================================================================
56
+ encoder/DenseReluDense.activation = 'relu'
57
+ encoder/DenseReluDense.dropout_rate = %dropout_rate
58
+ encoder/DenseReluDense.hidden_size = %d_ff
59
+ encoder/DenseReluDense.use_bias = False
60
+
61
+ # Parameters for enc_dec_attention:
62
+ # ==============================================================================
63
+ # None.
64
+
65
+ # Parameters for enc_dec_attention_bias:
66
+ # ==============================================================================
67
+ # None.
68
+
69
+ # Parameters for decoder/EncDecAttention:
70
+ # ==============================================================================
71
+ decoder/EncDecAttention.relative_attention_type = None
72
+
73
+ # Parameters for get_variable_dtype:
74
+ # ==============================================================================
75
+ get_variable_dtype.activation_dtype = 'bfloat16'
76
+
77
+ # Parameters for get_vocab_embedding_cls:
78
+ # ==============================================================================
79
+ # None.
80
+
81
+ # Parameters for get_vocabulary:
82
+ # ==============================================================================
83
+ get_vocabulary.mixture_or_task_name = %MIXTURE_NAME
84
+
85
+ # Parameters for init_checkpoint_variable_mapping:
86
+ # ==============================================================================
87
+ init_checkpoint_variable_mapping.mapping_fn = None
88
+
89
+ # Parameters for decoder/LayerStack:
90
+ # ==============================================================================
91
+ decoder/LayerStack.dropout_rate = %dropout_rate
92
+ decoder/LayerStack.norm_epsilon = 1e-06
93
+ decoder/LayerStack.recompute_grads = False
94
+ decoder/LayerStack.sublayers_final = \
95
+ [@transformer.sublayer_rms_norm, @transformer.sublayer_dropout]
96
+ decoder/LayerStack.sublayers_initial = [@transformer.sublayer_dropout]
97
+ decoder/LayerStack.sublayers_per_layer = \
98
+ [@transformer.sublayer_rms_norm,
99
+ @transformer.sublayer_call_layer,
100
+ @transformer.sublayer_dropout,
101
+ @transformer.sublayer_residual]
102
+
103
+ # Parameters for encoder/LayerStack:
104
+ # ==============================================================================
105
+ encoder/LayerStack.dropout_rate = %dropout_rate
106
+ encoder/LayerStack.norm_epsilon = 1e-06
107
+ encoder/LayerStack.recompute_grads = False
108
+ encoder/LayerStack.sublayers_final = \
109
+ [@transformer.sublayer_rms_norm, @transformer.sublayer_dropout]
110
+ encoder/LayerStack.sublayers_initial = [@transformer.sublayer_dropout]
111
+ encoder/LayerStack.sublayers_per_layer = \
112
+ [@transformer.sublayer_rms_norm,
113
+ @transformer.sublayer_call_layer,
114
+ @transformer.sublayer_dropout,
115
+ @transformer.sublayer_residual]
116
+
117
+ # Parameters for make_bitransformer:
118
+ # ==============================================================================
119
+ make_bitransformer.decoder_name = 'decoder'
120
+ make_bitransformer.encoder_name = 'encoder'
121
+
122
+ # Parameters for decoder/make_layer_stack:
123
+ # ==============================================================================
124
+ decoder/make_layer_stack.block_scope = True
125
+ decoder/make_layer_stack.layers = \
126
+ [@mesh_tensorflow.transformer.transformer_layers.SelfAttention,
127
+ @mesh_tensorflow.transformer.transformer_layers.EncDecAttention,
128
+ @mesh_tensorflow.transformer.transformer_layers.DenseReluDense]
129
+ decoder/make_layer_stack.num_layers = %num_layers
130
+
131
+ # Parameters for encoder/make_layer_stack:
132
+ # ==============================================================================
133
+ encoder/make_layer_stack.block_scope = True
134
+ encoder/make_layer_stack.layers = \
135
+ [@mesh_tensorflow.transformer.transformer_layers.SelfAttention,
136
+ @mesh_tensorflow.transformer.transformer_layers.DenseReluDense]
137
+ encoder/make_layer_stack.num_layers = %num_layers
138
+
139
+ # Parameters for pack_dataset:
140
+ # ==============================================================================
141
+ pack_dataset.use_custom_ops = False
142
+
143
+ # Parameters for pack_or_pad:
144
+ # ==============================================================================
145
+ pack_or_pad.ensure_eos = False
146
+ pack_or_pad.feature_keys = None
147
+ pack_or_pad.pack = True
148
+
149
+ # Parameters for packed_parallel_tsv_dataset:
150
+ # ==============================================================================
151
+ packed_parallel_tsv_dataset.batch_size = None
152
+ packed_parallel_tsv_dataset.max_encoded_len = 0
153
+
154
+ # Parameters for rewrite_stack_variables:
155
+ # ==============================================================================
156
+ rewrite_stack_variables.max_combined_variable_size = 536870912
157
+
158
+ # Parameters for run:
159
+ # ==============================================================================
160
+ run.autostack = True
161
+ run.batch_size = ('tokens_per_batch', %tokens_per_batch)
162
+ run.dataset_split = 'train'
163
+ run.ensemble_inputs = None
164
+ run.eval_checkpoint_step = None
165
+ run.eval_dataset_fn = None
166
+ run.eval_summary_dir = None
167
+ run.export_checkpoint_step = None
168
+ run.export_path = ''
169
+ run.init_checkpoint = %init_checkpoint
170
+ run.iterations_per_loop = 100
171
+ run.keep_checkpoint_max = None
172
+ run.layout_rules = \
173
+ 'ensemble:ensemble,batch:batch,d_ff:model,heads:model,vocab:model,experts:batch'
174
+ run.learning_rate_schedule = @learning_rate_schedules.constant_learning_rate
175
+ run.mesh_devices = None
176
+ run.mesh_shape = @mesh_tensorflow.transformer.utils.tpu_mesh_shape()
177
+ run.mode = 'train'
178
+ run.model_type = 'bitransformer'
179
+ run.optimizer = @optimize.AdafactorOptimizer
180
+ run.output_eval_examples = True
181
+ run.perplexity_eval_steps = 10
182
+ run.predict_fn = None
183
+ run.save_checkpoints_steps = 10000
184
+ run.sequence_length = {'inputs': %inputs_length, 'targets': %targets_length}
185
+ run.skip_seen_data = False
186
+ run.total_run_steps = None
187
+ run.train_dataset_fn = @t5.models.mesh_transformer.tsv_dataset_fn
188
+ run.train_steps = 1010900
189
+ run.variable_filter = None
190
+
191
+ # Parameters for decoder/SelfAttention:
192
+ # ==============================================================================
193
+ decoder/SelfAttention.attention_func = None
194
+ decoder/SelfAttention.attention_kwargs = None
195
+ decoder/SelfAttention.combine_dims = True
196
+ decoder/SelfAttention.dropout_rate = %dropout_rate
197
+ decoder/SelfAttention.fold_scaling_into_initializer = True
198
+ decoder/SelfAttention.keep_query_heads_dims = False
199
+ decoder/SelfAttention.key_value_size = %d_kv
200
+ decoder/SelfAttention.num_heads = %num_heads
201
+ decoder/SelfAttention.num_memory_heads = 0
202
+ decoder/SelfAttention.relative_attention_num_buckets = 32
203
+ decoder/SelfAttention.relative_attention_type = 'bias_shared'
204
+ decoder/SelfAttention.shared_kv = False
205
+
206
+ # Parameters for encoder/SelfAttention:
207
+ # ==============================================================================
208
+ encoder/SelfAttention.attention_func = None
209
+ encoder/SelfAttention.attention_kwargs = None
210
+ encoder/SelfAttention.combine_dims = True
211
+ encoder/SelfAttention.dropout_rate = %dropout_rate
212
+ encoder/SelfAttention.fold_scaling_into_initializer = True
213
+ encoder/SelfAttention.keep_query_heads_dims = False
214
+ encoder/SelfAttention.key_value_size = %d_kv
215
+ encoder/SelfAttention.num_heads = %num_heads
216
+ encoder/SelfAttention.num_memory_heads = 0
217
+ encoder/SelfAttention.relative_attention_num_buckets = 32
218
+ encoder/SelfAttention.relative_attention_type = 'bias_shared'
219
+ encoder/SelfAttention.shared_kv = False
220
+
221
+ # Parameters for serialize_num_microbatches:
222
+ # ==============================================================================
223
+ serialize_num_microbatches.tokens_per_microbatch_per_replica = 8192
224
+
225
+ # Parameters for SimdMeshImpl:
226
+ # ==============================================================================
227
+ SimdMeshImpl.allreduce_in_bfloat16_max_group_size = 8
228
+
229
+ # Parameters for sublayer_call_layer:
230
+ # ==============================================================================
231
+ # None.
232
+
233
+ # Parameters for sublayer_dropout:
234
+ # ==============================================================================
235
+ # None.
236
+
237
+ # Parameters for sublayer_legacy_dropout:
238
+ # ==============================================================================
239
+ # None.
240
+
241
+ # Parameters for sublayer_legacy_final_rms_norm:
242
+ # ==============================================================================
243
+ # None.
244
+
245
+ # Parameters for sublayer_legacy_rms_norm:
246
+ # ==============================================================================
247
+ # None.
248
+
249
+ # Parameters for sublayer_mask_padding:
250
+ # ==============================================================================
251
+ # None.
252
+
253
+ # Parameters for sublayer_residual:
254
+ # ==============================================================================
255
+ # None.
256
+
257
+ # Parameters for sublayer_rms_norm:
258
+ # ==============================================================================
259
+ sublayer_rms_norm.epsilon = 1e-06
260
+
261
+ # Parameters for tpu_estimator_model_fn:
262
+ # ==============================================================================
263
+ tpu_estimator_model_fn.hierarchical_tiling_spec = None
264
+ tpu_estimator_model_fn.init_variable_filter = ''
265
+ tpu_estimator_model_fn.model_info_file = None
266
+ tpu_estimator_model_fn.outer_batch_size = 1
267
+ tpu_estimator_model_fn.tpu_summaries = False
268
+
269
+ # Parameters for tpu_mesh_shape:
270
+ # ==============================================================================
271
+ tpu_mesh_shape.ensemble_parallelism = None
272
+ tpu_mesh_shape.model_parallelism = 1
273
+ tpu_mesh_shape.tpu_topology = '2x2'
274
+
275
+ # Parameters for train_model:
276
+ # ==============================================================================
277
+ train_model.dataset_split = 'train'
278
+ train_model.seen_data_init_step = 0
279
+
280
+ # Parameters for tsv_dataset_fn:
281
+ # ==============================================================================
282
+ tsv_dataset_fn.filename = \
283
+ 'gs://pongo-bucket/duo-t5/data/msmarco_med_rod/query_doc_pairs.train.tsv'
284
+ tsv_dataset_fn.shuffle_buffer_size = 10000
285
+
286
+ # Parameters for unit_scaling_convention:
287
+ # ==============================================================================
288
+ unit_scaling_convention.value = False
289
+
290
+ # Parameters for decoder/Unitransformer:
291
+ # ==============================================================================
292
+ decoder/Unitransformer.d_model = %d_model
293
+ decoder/Unitransformer.ensemble = None
294
+ decoder/Unitransformer.input_full_attention = False
295
+ decoder/Unitransformer.label_smoothing = 0.0
296
+ decoder/Unitransformer.loss_denominator = 233472
297
+ decoder/Unitransformer.loss_fn = None
298
+ decoder/Unitransformer.loss_on_targets_only = False
299
+ decoder/Unitransformer.max_length = 512
300
+ decoder/Unitransformer.positional_embedding = False
301
+ decoder/Unitransformer.shared_embedding_and_softmax_weights = True
302
+ decoder/Unitransformer.sinusoid_positional_embedding = False
303
+ decoder/Unitransformer.token_dropout_rate = 0.0
304
+ decoder/Unitransformer.vocab_divisor = 128
305
+ decoder/Unitransformer.z_loss = 0.0001
306
+
307
+ # Parameters for encoder/Unitransformer:
308
+ # ==============================================================================
309
+ encoder/Unitransformer.d_model = %d_model
310
+ encoder/Unitransformer.ensemble = None
311
+ encoder/Unitransformer.input_full_attention = False
312
+ encoder/Unitransformer.label_smoothing = 0.0
313
+ encoder/Unitransformer.loss_denominator = None
314
+ encoder/Unitransformer.loss_fn = None
315
+ encoder/Unitransformer.loss_on_targets_only = False
316
+ encoder/Unitransformer.max_length = 512
317
+ encoder/Unitransformer.positional_embedding = False
318
+ encoder/Unitransformer.shared_embedding_and_softmax_weights = True
319
+ encoder/Unitransformer.sinusoid_positional_embedding = False
320
+ encoder/Unitransformer.token_dropout_rate = 0.0
321
+ encoder/Unitransformer.vocab_divisor = 128
322
+ encoder/Unitransformer.z_loss = 0.0001
323
+
324
+ # Parameters for VarianceScalingInitializer:
325
+ # ==============================================================================
326
+ VarianceScalingInitializer.distribution = 'normal'
327
+ VarianceScalingInitializer.mode = 'fan_in'
328
+ VarianceScalingInitializer.scale = 1.0
329
+
330
+ # Parameters for VocabEmbedding:
331
+ # ==============================================================================
332
+ VocabEmbedding.scale_variable_like_classifier_weights = False