pere commited on
Commit
46bef15
1 Parent(s): d20a6e0

new attempt

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "./",
3
  "architectures": [
4
  "XLMRobertaForMaskedLM"
5
  ],
@@ -21,7 +21,7 @@
21
  "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
23
  "torch_dtype": "float32",
24
- "transformers_version": "4.23.1",
25
  "type_vocab_size": 1,
26
  "use_cache": true,
27
  "vocab_size": 250002
 
1
  {
2
+ "_name_or_path": "./config.json",
3
  "architectures": [
4
  "XLMRobertaForMaskedLM"
5
  ],
 
21
  "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
23
  "torch_dtype": "float32",
24
+ "transformers_version": "4.24.0",
25
  "type_vocab_size": 1,
26
  "use_cache": true,
27
  "vocab_size": 250002
events.out.tfevents.1672819976.t1v-n-0853dee6-w-3.291836.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4189a298dc3ab9be3250fd74db70fac076b8e11d308c2afba23e19d047a123a8
3
+ size 22598
events.out.tfevents.1672820882.t1v-n-0853dee6-w-3.322000.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f7294f94d22c80b866f0559f6606ac68f6f6f386918848c1f7f4ba68e161340
3
+ size 7637
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a04d8032d9ff18b1b727c98ee3bea82de1d908107d257d94486dba8650f5680
3
+ size 1113187999
run.sh CHANGED
@@ -1,6 +1,5 @@
1
 
2
  python run_mlm_flax_stream.py \
3
- --output_dir="./" \
4
  --hub_model_id="NbAiLab/nb-roberta-base-scandi" \
5
  --model_name_or_path="xlm-roberta-base" \
6
  --config_name="./config.json" \
@@ -12,6 +11,7 @@ python run_mlm_flax_stream.py \
12
  --per_device_eval_batch_size="16" \
13
  --learning_rate="4e-4" \
14
  --warmup_steps="1000" \
 
15
  --overwrite_output_dir \
16
  --num_train_steps="10000" \
17
  --adam_beta1="0.9" \
@@ -19,5 +19,6 @@ python run_mlm_flax_stream.py \
19
  --logging_steps="50" \
20
  --save_steps="50" \
21
  --eval_steps="50" \
 
22
  --dtype="bfloat16" \
23
  --push_to_hub
 
1
 
2
  python run_mlm_flax_stream.py \
 
3
  --hub_model_id="NbAiLab/nb-roberta-base-scandi" \
4
  --model_name_or_path="xlm-roberta-base" \
5
  --config_name="./config.json" \
 
11
  --per_device_eval_batch_size="16" \
12
  --learning_rate="4e-4" \
13
  --warmup_steps="1000" \
14
+ --use_auth_token \
15
  --overwrite_output_dir \
16
  --num_train_steps="10000" \
17
  --adam_beta1="0.9" \
 
19
  --logging_steps="50" \
20
  --save_steps="50" \
21
  --eval_steps="50" \
22
+ push_to_--output_dir="./" \
23
  --dtype="bfloat16" \
24
  --push_to_hub
run_mlm_flax_stream.py CHANGED
@@ -655,13 +655,18 @@ if __name__ == "__main__":
655
  )
656
  print(f"Saving weights and logs of step {step+1}. \nThe result is saved to {training_args.output_folder} by worker {jax.process_index()}.")
657
  except:
658
-
659
  model.save_pretrained(
660
  training_args.output_dir,
661
  params=params
662
  )
663
  print("Problems pushing this to the hub. The bug should be fixed.")
664
-
 
 
 
 
 
665
  # update tqdm bar
666
  steps.update(1)
667
 
 
655
  )
656
  print(f"Saving weights and logs of step {step+1}. \nThe result is saved to {training_args.output_folder} by worker {jax.process_index()}.")
657
  except:
658
+ breakpoint()
659
  model.save_pretrained(
660
  training_args.output_dir,
661
  params=params
662
  )
663
  print("Problems pushing this to the hub. The bug should be fixed.")
664
+ else:
665
+ model.save_pretrained(
666
+ training_args.output_dir,
667
+ params=params
668
+ )
669
+
670
  # update tqdm bar
671
  steps.update(1)
672