aapot commited on
Commit
e0ba7eb
1 Parent(s): b221124

Try fix git

Browse files
README.md CHANGED
File without changes
added_tokens.json CHANGED
File without changes
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "gpt2-large",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
 
1
  {
2
+ "_name_or_path": "./",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
flax_model_to_pytorch.py CHANGED
File without changes
merges.txt CHANGED
File without changes
replace_token_script.py CHANGED
File without changes
run_clm_flax.py CHANGED
@@ -61,7 +61,7 @@ from transformers import (
61
  from transformers.file_utils import get_full_repo_name
62
  from transformers.testing_utils import CaptureLogger
63
 
64
- from distributed_shampoo import distributed_shampoo, GraftingType
65
 
66
 
67
  logger = logging.getLogger(__name__)
 
61
  from transformers.file_utils import get_full_repo_name
62
  from transformers.testing_utils import CaptureLogger
63
 
64
+ # from distributed_shampoo import distributed_shampoo, GraftingType
65
 
66
 
67
  logger = logging.getLogger(__name__)
special_tokens_map.json CHANGED
File without changes
start_train.sh CHANGED
@@ -10,19 +10,19 @@ python3 run_clm_flax.py \
10
  --dataset_filepath="/researchdisk/training_dataset_full_deduplicated" \
11
  --do_train --do_eval \
12
  --block_size="512" \
13
- --per_device_train_batch_size="16" \
14
- --per_device_eval_batch_size="16" \
15
- --preprocessing_num_workers="1" \
16
  --adam_beta1="0.9" \
17
  --adam_beta2="0.98" \
18
- --learning_rate="1e-5" \
19
  --weight_decay="0.01" \
20
  --warmup_steps="4000" \
21
  --cosine_decay \
22
  --overwrite_output_dir \
23
  --logging_steps="500" \
24
- --eval_steps="10000" \
25
- --save_steps="50000" \
26
  --num_train_epochs="5" \
27
  --dtype="bfloat16" \
28
  --push_to_hub \
 
10
  --dataset_filepath="/researchdisk/training_dataset_full_deduplicated" \
11
  --do_train --do_eval \
12
  --block_size="512" \
13
+ --per_device_train_batch_size="8" \
14
+ --per_device_eval_batch_size="8" \
15
+ --preprocessing_num_workers="96" \
16
  --adam_beta1="0.9" \
17
  --adam_beta2="0.98" \
18
+ --learning_rate="4e-5" \
19
  --weight_decay="0.01" \
20
  --warmup_steps="4000" \
21
  --cosine_decay \
22
  --overwrite_output_dir \
23
  --logging_steps="500" \
24
+ --eval_steps="100" \
25
+ --save_steps="100" \
26
  --num_train_epochs="5" \
27
  --dtype="bfloat16" \
28
  --push_to_hub \
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
File without changes
train_tokenizer.py CHANGED
File without changes
vocab.json CHANGED
File without changes