Saving weights and logs of step 1000
Browse files
events.out.tfevents.1642203685.t1v-n-eedfb410-w-0.10537.0.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f5f49581e10e342c9432818c78cf138d852e9c2d91fa0ede2deaa3b6935e8a48
|
3 |
+
size 147136
|
events.out.tfevents.1642204242.t1v-n-eedfb410-w-0.profile-empty
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ac614ecef2709e4ed2bc443ce4ade10122a22097363c5eb86dfadf8e74fa7c5
|
3 |
+
size 40
|
flax_model.msgpack
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 498796983
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a7856df89509557d606fad1b5333e28c86c7485f2b7ad4de66c6219ca6895c54
|
3 |
size 498796983
|
run_mlm_flax.py
CHANGED
@@ -553,7 +553,7 @@ def main():
|
|
553 |
wandb.init(
|
554 |
entity='versae',
|
555 |
project='roberta-base-ncc',
|
556 |
-
sync_tensorboard=
|
557 |
)
|
558 |
wandb.config.update(training_args)
|
559 |
wandb.config.update(model_args)
|
|
|
553 |
wandb.init(
|
554 |
entity='versae',
|
555 |
project='roberta-base-ncc',
|
556 |
+
sync_tensorboard=False,
|
557 |
)
|
558 |
wandb.config.update(training_args)
|
559 |
wandb.config.update(model_args)
|
train.128.sh
CHANGED
@@ -6,8 +6,8 @@ python run_mlm_flax.py \
|
|
6 |
--dataset_name="NbAiLab/NCC" \
|
7 |
--max_seq_length="128" \
|
8 |
--weight_decay="0.01" \
|
9 |
-
--per_device_train_batch_size="
|
10 |
-
--per_device_eval_batch_size="
|
11 |
--pad_to_max_length \
|
12 |
--learning_rate="6e-4" \
|
13 |
--warmup_steps="10000" \
|
|
|
6 |
--dataset_name="NbAiLab/NCC" \
|
7 |
--max_seq_length="128" \
|
8 |
--weight_decay="0.01" \
|
9 |
+
--per_device_train_batch_size="232" \
|
10 |
+
--per_device_eval_batch_size="232" \
|
11 |
--pad_to_max_length \
|
12 |
--learning_rate="6e-4" \
|
13 |
--warmup_steps="10000" \
|