nancyH commited on
Commit
0dbbebb
·
verified ·
1 Parent(s): 05addf5

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. data_download.py +23 -0
  2. data_split.py +28 -0
  3. full_multi_base_2048_3e-5_log.txt +0 -0
  4. full_multi_base_3072_3e-5_log.txt +0 -0
  5. full_multi_base_4096_3e-5_log.txt +0 -0
  6. full_multi_len2_2048_3e-5_log.txt +0 -0
  7. full_multi_len2_3072_3e-5_log.txt +0 -0
  8. full_multi_len2_4096_3e-5_log.txt +0 -0
  9. full_output_multi_epoch/collect_results.py +48 -0
  10. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/config.json +27 -0
  11. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/model.safetensors +3 -0
  12. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/optimizer.pt +3 -0
  13. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/rng_state.pth +3 -0
  14. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/scheduler.pt +3 -0
  15. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/special_tokens_map.json +7 -0
  16. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/tokenizer.json +0 -0
  17. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/tokenizer_config.json +56 -0
  18. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/trainer_state.json +110 -0
  19. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/training_args.bin +3 -0
  20. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/config.json +27 -0
  21. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/model.safetensors +3 -0
  22. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/optimizer.pt +3 -0
  23. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/rng_state.pth +3 -0
  24. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/scheduler.pt +3 -0
  25. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/special_tokens_map.json +7 -0
  26. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/tokenizer.json +0 -0
  27. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/tokenizer_config.json +56 -0
  28. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/trainer_state.json +123 -0
  29. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/training_args.bin +3 -0
  30. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/config.json +27 -0
  31. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/model.safetensors +3 -0
  32. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/optimizer.pt +3 -0
  33. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/rng_state.pth +3 -0
  34. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/scheduler.pt +3 -0
  35. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/special_tokens_map.json +7 -0
  36. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/tokenizer.json +0 -0
  37. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/tokenizer_config.json +56 -0
  38. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/trainer_state.json +136 -0
  39. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/training_args.bin +3 -0
  40. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/results/base_3072_hg38_BPE_3e-5_demo_coding_vs_intergenomic_seqs_seed42/eval_results.json +1 -0
  41. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/config.json +27 -0
  42. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/model.safetensors +3 -0
  43. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/optimizer.pt +3 -0
  44. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/rng_state.pth +3 -0
  45. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/scheduler.pt +3 -0
  46. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/special_tokens_map.json +7 -0
  47. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/tokenizer.json +0 -0
  48. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/tokenizer_config.json +56 -0
  49. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/trainer_state.json +110 -0
  50. full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/training_args.bin +3 -0
data_download.py ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from datasets import load_dataset
2
+ from pathlib import Path
3
+
4
+ datasets = [
5
+ "demo_human_or_worm",
6
+ "dummy_mouse_enhancers_ensembl",
7
+ "human_enhancers_ensembl",
8
+ "human_nontata_promoters",
9
+ "demo_coding_vs_intergenomic_seqs",
10
+ "drosophila_enhancers_stark",
11
+ "human_enhancers_cohn",
12
+ "human_ensembl_regulatory",
13
+ "human_ocr_ensembl",
14
+ ]
15
+
16
+ out_root = Path("hf_raw")
17
+ out_root.mkdir(parents=True, exist_ok=True)
18
+
19
+ for name in datasets:
20
+ hf_id = f"katarinagresova/Genomic_Benchmarks_{name}"
21
+ ds = load_dataset(hf_id) # downloads to the HF cache
22
+ ds.save_to_disk(out_root / name) # optional: persist locally for reuse
23
+ print(f"downloaded {hf_id}")
data_split.py ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from datasets import load_from_disk, concatenate_datasets
2
+ from pathlib import Path
3
+ import pandas as pd
4
+
5
+ src_root = Path("hf_raw")
6
+ dst_root = Path("ft_data")
7
+ seed = 42
8
+
9
+ for ds_dir in src_root.iterdir():
10
+ if not ds_dir.is_dir():
11
+ continue
12
+ ds = load_from_disk(ds_dir)
13
+
14
+ # Combine all available splits, shuffle, then 80/10/10
15
+ full = concatenate_datasets([ds[s] for s in ds.keys()])
16
+ full = full.shuffle(seed=seed)
17
+ split1 = full.train_test_split(test_size=0.2, seed=seed)
18
+ train = split1["train"]
19
+ split2 = split1["test"].train_test_split(test_size=0.5, seed=seed)
20
+ dev, test = split2["train"], split2["test"]
21
+
22
+ out = dst_root / ds_dir.name / "split"
23
+ out.mkdir(parents=True, exist_ok=True)
24
+ for name, subset in [("train", train), ("dev", dev), ("test", test)]:
25
+ subset.to_pandas()[["seq", "label"]].rename(
26
+ columns={"seq": "sequence", "label": "labels"}
27
+ ).to_csv(out / f"{name}.csv", sep="\t", index=False)
28
+ print(f"Wrote {out/f'{name}.csv'}")
full_multi_base_2048_3e-5_log.txt ADDED
The diff for this file is too large to render. See raw diff
 
full_multi_base_3072_3e-5_log.txt ADDED
The diff for this file is too large to render. See raw diff
 
full_multi_base_4096_3e-5_log.txt ADDED
The diff for this file is too large to render. See raw diff
 
full_multi_len2_2048_3e-5_log.txt ADDED
The diff for this file is too large to render. See raw diff
 
full_multi_len2_3072_3e-5_log.txt ADDED
The diff for this file is too large to render. See raw diff
 
full_multi_len2_4096_3e-5_log.txt ADDED
The diff for this file is too large to render. See raw diff
 
full_output_multi_epoch/collect_results.py ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import json
2
+ from pathlib import Path
3
+ import pandas as pd
4
+
5
+ ROOT = Path(".") # full_output_multi_epoch
6
+ OUTPUT_DIR = ROOT / "tables"
7
+ OUTPUT_DIR.mkdir(exist_ok=True)
8
+
9
+ all_rows = []
10
+
11
+ for task_dir in sorted(p for p in ROOT.iterdir() if p.is_dir() and p.name != "tables"):
12
+ task_name = task_dir.name
13
+ task_rows = []
14
+
15
+ for model_dir in sorted(p for p in task_dir.iterdir() if p.is_dir()):
16
+ model_name = model_dir.name
17
+ results_dir = model_dir / "results"
18
+ if not results_dir.exists():
19
+ continue
20
+
21
+ # Usually only one run folder inside results/
22
+ for run_dir in results_dir.iterdir():
23
+ eval_file = run_dir / "eval_results.json"
24
+ if not eval_file.exists():
25
+ continue
26
+
27
+ with open(eval_file) as f:
28
+ metrics = json.load(f)
29
+
30
+ row = {
31
+ "task": task_name,
32
+ "model": model_name,
33
+ **metrics
34
+ }
35
+
36
+ task_rows.append(row)
37
+ all_rows.append(row)
38
+
39
+ if task_rows:
40
+ df_task = pd.DataFrame(task_rows).sort_values("model")
41
+ df_task.to_csv(OUTPUT_DIR / f"results_{task_name}.csv", index=False)
42
+ print(f"Saved: results_{task_name}.csv")
43
+
44
+ # Optional: global table
45
+ if all_rows:
46
+ df_all = pd.DataFrame(all_rows)
47
+ df_all.to_csv(OUTPUT_DIR / "results_all_tasks.csv", index=False)
48
+ print("Saved: results_all_tasks.csv")
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_3072/checkpoint-100000",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_length": 512,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "problem_type": "single_label_classification",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.35.2",
24
+ "type_vocab_size": 2,
25
+ "use_cache": true,
26
+ "vocab_size": 3072
27
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a6a82cc56bde43624d024a36dc776ca84a353a85d62a95cab3e0f0dc237234e
3
+ size 353632152
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5e6dd1dfc2e9d5f8f1a799a64224eb8fd8aea63758b054765b6a8753036d8c0
3
+ size 707385995
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9af56506fc0b71f6f6e9ee81e8a4836588c94483bc8810d576e0f47b0223ac05
3
+ size 14709
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e32828859e9d069315b3f3e9abbd7af9b79ba7a4ee7041fec91396a6f9e6768
3
+ size 1465
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "cache_dir": null,
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 100,
49
+ "pad_token": "[PAD]",
50
+ "padding_side": "right",
51
+ "sep_token": "[SEP]",
52
+ "tokenizer_class": "PreTrainedTokenizerFast",
53
+ "trust_remote_code": true,
54
+ "unk_token": "[UNK]",
55
+ "use_fast": true
56
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/trainer_state.json ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.24,
5
+ "eval_steps": 200,
6
+ "global_step": 1400,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.32,
13
+ "eval_accuracy": 0.9009,
14
+ "eval_f1": 0.9008999752249938,
15
+ "eval_loss": 0.24717512726783752,
16
+ "eval_matthews_correlation": 0.8019013074464254,
17
+ "eval_precision": 0.9009536610343063,
18
+ "eval_recall": 0.9009476464346753,
19
+ "eval_runtime": 0.728,
20
+ "eval_samples_per_second": 13736.193,
21
+ "eval_steps_per_second": 108.516,
22
+ "step": 200
23
+ },
24
+ {
25
+ "epoch": 0.64,
26
+ "eval_accuracy": 0.9041,
27
+ "eval_f1": 0.9040877529651761,
28
+ "eval_loss": 0.24045702815055847,
29
+ "eval_matthews_correlation": 0.8087813953002411,
30
+ "eval_precision": 0.9045580575877361,
31
+ "eval_recall": 0.9042234069469404,
32
+ "eval_runtime": 0.7007,
33
+ "eval_samples_per_second": 14272.015,
34
+ "eval_steps_per_second": 112.749,
35
+ "step": 400
36
+ },
37
+ {
38
+ "epoch": 0.96,
39
+ "eval_accuracy": 0.9037,
40
+ "eval_f1": 0.9035290138594547,
41
+ "eval_loss": 0.23689354956150055,
42
+ "eval_matthews_correlation": 0.8092946059534654,
43
+ "eval_precision": 0.9058485926765776,
44
+ "eval_recall": 0.9034495690288824,
45
+ "eval_runtime": 0.7017,
46
+ "eval_samples_per_second": 14251.42,
47
+ "eval_steps_per_second": 112.586,
48
+ "step": 600
49
+ },
50
+ {
51
+ "epoch": 1.28,
52
+ "eval_accuracy": 0.9056,
53
+ "eval_f1": 0.9054084521155339,
54
+ "eval_loss": 0.23994144797325134,
55
+ "eval_matthews_correlation": 0.8134456166108831,
56
+ "eval_precision": 0.9081210458851026,
57
+ "eval_recall": 0.9053293611386959,
58
+ "eval_runtime": 0.7009,
59
+ "eval_samples_per_second": 14267.83,
60
+ "eval_steps_per_second": 112.716,
61
+ "step": 800
62
+ },
63
+ {
64
+ "epoch": 1.6,
65
+ "eval_accuracy": 0.9105,
66
+ "eval_f1": 0.910486894386207,
67
+ "eval_loss": 0.22719423472881317,
68
+ "eval_matthews_correlation": 0.8216385258625271,
69
+ "eval_precision": 0.911009292886067,
70
+ "eval_recall": 0.910629320836721,
71
+ "eval_runtime": 0.7015,
72
+ "eval_samples_per_second": 14254.403,
73
+ "eval_steps_per_second": 112.61,
74
+ "step": 1000
75
+ },
76
+ {
77
+ "epoch": 1.92,
78
+ "eval_accuracy": 0.9131,
79
+ "eval_f1": 0.9130232787039119,
80
+ "eval_loss": 0.22087331116199493,
81
+ "eval_matthews_correlation": 0.8269895715265106,
82
+ "eval_precision": 0.9140534915148619,
83
+ "eval_recall": 0.9129368339048614,
84
+ "eval_runtime": 0.7,
85
+ "eval_samples_per_second": 14284.915,
86
+ "eval_steps_per_second": 112.851,
87
+ "step": 1200
88
+ },
89
+ {
90
+ "epoch": 2.24,
91
+ "eval_accuracy": 0.9137,
92
+ "eval_f1": 0.9136644219716503,
93
+ "eval_loss": 0.22041191160678864,
94
+ "eval_matthews_correlation": 0.8276523914394088,
95
+ "eval_precision": 0.9140498457138171,
96
+ "eval_recall": 0.91360266653066,
97
+ "eval_runtime": 0.7,
98
+ "eval_samples_per_second": 14286.336,
99
+ "eval_steps_per_second": 112.862,
100
+ "step": 1400
101
+ }
102
+ ],
103
+ "logging_steps": 100000,
104
+ "max_steps": 1875,
105
+ "num_train_epochs": 3,
106
+ "save_steps": 200,
107
+ "total_flos": 4972799101763584.0,
108
+ "trial_name": null,
109
+ "trial_params": null
110
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1400/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66f558acf951dddcf698771a7d98f4feaca4a4b912d76fa9c7ec16ebe809283c
3
+ size 5329
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_3072/checkpoint-100000",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_length": 512,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "problem_type": "single_label_classification",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.35.2",
24
+ "type_vocab_size": 2,
25
+ "use_cache": true,
26
+ "vocab_size": 3072
27
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c49fce4efe8374b1208584285c2051dcfa80394add10213127dc3ef177a73a08
3
+ size 353632152
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05b46437dee9879cd4392837c38cd80da8cf6894ea365c9440e53b4fa5bf754e
3
+ size 707385995
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c49465b41dedc595a1cf05bd2e4f847a88fe0d62e74d64c40d13068f732aaee3
3
+ size 14709
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:778fe40a0c11980abd25e31db8ab18a2f8d910488102d622b91f45279b3ff175
3
+ size 1465
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "cache_dir": null,
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 100,
49
+ "pad_token": "[PAD]",
50
+ "padding_side": "right",
51
+ "sep_token": "[SEP]",
52
+ "tokenizer_class": "PreTrainedTokenizerFast",
53
+ "trust_remote_code": true,
54
+ "unk_token": "[UNK]",
55
+ "use_fast": true
56
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/trainer_state.json ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.56,
5
+ "eval_steps": 200,
6
+ "global_step": 1600,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.32,
13
+ "eval_accuracy": 0.9009,
14
+ "eval_f1": 0.9008999752249938,
15
+ "eval_loss": 0.24717512726783752,
16
+ "eval_matthews_correlation": 0.8019013074464254,
17
+ "eval_precision": 0.9009536610343063,
18
+ "eval_recall": 0.9009476464346753,
19
+ "eval_runtime": 0.728,
20
+ "eval_samples_per_second": 13736.193,
21
+ "eval_steps_per_second": 108.516,
22
+ "step": 200
23
+ },
24
+ {
25
+ "epoch": 0.64,
26
+ "eval_accuracy": 0.9041,
27
+ "eval_f1": 0.9040877529651761,
28
+ "eval_loss": 0.24045702815055847,
29
+ "eval_matthews_correlation": 0.8087813953002411,
30
+ "eval_precision": 0.9045580575877361,
31
+ "eval_recall": 0.9042234069469404,
32
+ "eval_runtime": 0.7007,
33
+ "eval_samples_per_second": 14272.015,
34
+ "eval_steps_per_second": 112.749,
35
+ "step": 400
36
+ },
37
+ {
38
+ "epoch": 0.96,
39
+ "eval_accuracy": 0.9037,
40
+ "eval_f1": 0.9035290138594547,
41
+ "eval_loss": 0.23689354956150055,
42
+ "eval_matthews_correlation": 0.8092946059534654,
43
+ "eval_precision": 0.9058485926765776,
44
+ "eval_recall": 0.9034495690288824,
45
+ "eval_runtime": 0.7017,
46
+ "eval_samples_per_second": 14251.42,
47
+ "eval_steps_per_second": 112.586,
48
+ "step": 600
49
+ },
50
+ {
51
+ "epoch": 1.28,
52
+ "eval_accuracy": 0.9056,
53
+ "eval_f1": 0.9054084521155339,
54
+ "eval_loss": 0.23994144797325134,
55
+ "eval_matthews_correlation": 0.8134456166108831,
56
+ "eval_precision": 0.9081210458851026,
57
+ "eval_recall": 0.9053293611386959,
58
+ "eval_runtime": 0.7009,
59
+ "eval_samples_per_second": 14267.83,
60
+ "eval_steps_per_second": 112.716,
61
+ "step": 800
62
+ },
63
+ {
64
+ "epoch": 1.6,
65
+ "eval_accuracy": 0.9105,
66
+ "eval_f1": 0.910486894386207,
67
+ "eval_loss": 0.22719423472881317,
68
+ "eval_matthews_correlation": 0.8216385258625271,
69
+ "eval_precision": 0.911009292886067,
70
+ "eval_recall": 0.910629320836721,
71
+ "eval_runtime": 0.7015,
72
+ "eval_samples_per_second": 14254.403,
73
+ "eval_steps_per_second": 112.61,
74
+ "step": 1000
75
+ },
76
+ {
77
+ "epoch": 1.92,
78
+ "eval_accuracy": 0.9131,
79
+ "eval_f1": 0.9130232787039119,
80
+ "eval_loss": 0.22087331116199493,
81
+ "eval_matthews_correlation": 0.8269895715265106,
82
+ "eval_precision": 0.9140534915148619,
83
+ "eval_recall": 0.9129368339048614,
84
+ "eval_runtime": 0.7,
85
+ "eval_samples_per_second": 14284.915,
86
+ "eval_steps_per_second": 112.851,
87
+ "step": 1200
88
+ },
89
+ {
90
+ "epoch": 2.24,
91
+ "eval_accuracy": 0.9137,
92
+ "eval_f1": 0.9136644219716503,
93
+ "eval_loss": 0.22041191160678864,
94
+ "eval_matthews_correlation": 0.8276523914394088,
95
+ "eval_precision": 0.9140498457138171,
96
+ "eval_recall": 0.91360266653066,
97
+ "eval_runtime": 0.7,
98
+ "eval_samples_per_second": 14286.336,
99
+ "eval_steps_per_second": 112.862,
100
+ "step": 1400
101
+ },
102
+ {
103
+ "epoch": 2.56,
104
+ "eval_accuracy": 0.9143,
105
+ "eval_f1": 0.9142909071223366,
106
+ "eval_loss": 0.2254679948091507,
107
+ "eval_matthews_correlation": 0.8286017317454041,
108
+ "eval_precision": 0.9143290342981838,
109
+ "eval_recall": 0.9142726993622687,
110
+ "eval_runtime": 0.6994,
111
+ "eval_samples_per_second": 14297.498,
112
+ "eval_steps_per_second": 112.95,
113
+ "step": 1600
114
+ }
115
+ ],
116
+ "logging_steps": 100000,
117
+ "max_steps": 1875,
118
+ "num_train_epochs": 3,
119
+ "save_steps": 200,
120
+ "total_flos": 5683198973444096.0,
121
+ "trial_name": null,
122
+ "trial_params": null
123
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1600/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66f558acf951dddcf698771a7d98f4feaca4a4b912d76fa9c7ec16ebe809283c
3
+ size 5329
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_3072/checkpoint-100000",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_length": 512,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "problem_type": "single_label_classification",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.35.2",
24
+ "type_vocab_size": 2,
25
+ "use_cache": true,
26
+ "vocab_size": 3072
27
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e4bcd93b1e022098504062a855e3a51a7d0612b9d1095e3f104a4fc2db387e3
3
+ size 353632152
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32defe438c9f8f46bbe218a6fa6185cb17ffcac4156f82723cb71898ea5a84bc
3
+ size 707385995
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7b352d03af2b56a998f314da069a38bc56dcbf7905c5abcaa5b1ad04b523e04
3
+ size 14709
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:084228aa33f0ae6a0339955a663a055ad6d466f21dead5e3294bff0e025f5dd1
3
+ size 1465
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "cache_dir": null,
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 100,
49
+ "pad_token": "[PAD]",
50
+ "padding_side": "right",
51
+ "sep_token": "[SEP]",
52
+ "tokenizer_class": "PreTrainedTokenizerFast",
53
+ "trust_remote_code": true,
54
+ "unk_token": "[UNK]",
55
+ "use_fast": true
56
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/trainer_state.json ADDED
@@ -0,0 +1,136 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.88,
5
+ "eval_steps": 200,
6
+ "global_step": 1800,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.32,
13
+ "eval_accuracy": 0.9009,
14
+ "eval_f1": 0.9008999752249938,
15
+ "eval_loss": 0.24717512726783752,
16
+ "eval_matthews_correlation": 0.8019013074464254,
17
+ "eval_precision": 0.9009536610343063,
18
+ "eval_recall": 0.9009476464346753,
19
+ "eval_runtime": 0.728,
20
+ "eval_samples_per_second": 13736.193,
21
+ "eval_steps_per_second": 108.516,
22
+ "step": 200
23
+ },
24
+ {
25
+ "epoch": 0.64,
26
+ "eval_accuracy": 0.9041,
27
+ "eval_f1": 0.9040877529651761,
28
+ "eval_loss": 0.24045702815055847,
29
+ "eval_matthews_correlation": 0.8087813953002411,
30
+ "eval_precision": 0.9045580575877361,
31
+ "eval_recall": 0.9042234069469404,
32
+ "eval_runtime": 0.7007,
33
+ "eval_samples_per_second": 14272.015,
34
+ "eval_steps_per_second": 112.749,
35
+ "step": 400
36
+ },
37
+ {
38
+ "epoch": 0.96,
39
+ "eval_accuracy": 0.9037,
40
+ "eval_f1": 0.9035290138594547,
41
+ "eval_loss": 0.23689354956150055,
42
+ "eval_matthews_correlation": 0.8092946059534654,
43
+ "eval_precision": 0.9058485926765776,
44
+ "eval_recall": 0.9034495690288824,
45
+ "eval_runtime": 0.7017,
46
+ "eval_samples_per_second": 14251.42,
47
+ "eval_steps_per_second": 112.586,
48
+ "step": 600
49
+ },
50
+ {
51
+ "epoch": 1.28,
52
+ "eval_accuracy": 0.9056,
53
+ "eval_f1": 0.9054084521155339,
54
+ "eval_loss": 0.23994144797325134,
55
+ "eval_matthews_correlation": 0.8134456166108831,
56
+ "eval_precision": 0.9081210458851026,
57
+ "eval_recall": 0.9053293611386959,
58
+ "eval_runtime": 0.7009,
59
+ "eval_samples_per_second": 14267.83,
60
+ "eval_steps_per_second": 112.716,
61
+ "step": 800
62
+ },
63
+ {
64
+ "epoch": 1.6,
65
+ "eval_accuracy": 0.9105,
66
+ "eval_f1": 0.910486894386207,
67
+ "eval_loss": 0.22719423472881317,
68
+ "eval_matthews_correlation": 0.8216385258625271,
69
+ "eval_precision": 0.911009292886067,
70
+ "eval_recall": 0.910629320836721,
71
+ "eval_runtime": 0.7015,
72
+ "eval_samples_per_second": 14254.403,
73
+ "eval_steps_per_second": 112.61,
74
+ "step": 1000
75
+ },
76
+ {
77
+ "epoch": 1.92,
78
+ "eval_accuracy": 0.9131,
79
+ "eval_f1": 0.9130232787039119,
80
+ "eval_loss": 0.22087331116199493,
81
+ "eval_matthews_correlation": 0.8269895715265106,
82
+ "eval_precision": 0.9140534915148619,
83
+ "eval_recall": 0.9129368339048614,
84
+ "eval_runtime": 0.7,
85
+ "eval_samples_per_second": 14284.915,
86
+ "eval_steps_per_second": 112.851,
87
+ "step": 1200
88
+ },
89
+ {
90
+ "epoch": 2.24,
91
+ "eval_accuracy": 0.9137,
92
+ "eval_f1": 0.9136644219716503,
93
+ "eval_loss": 0.22041191160678864,
94
+ "eval_matthews_correlation": 0.8276523914394088,
95
+ "eval_precision": 0.9140498457138171,
96
+ "eval_recall": 0.91360266653066,
97
+ "eval_runtime": 0.7,
98
+ "eval_samples_per_second": 14286.336,
99
+ "eval_steps_per_second": 112.862,
100
+ "step": 1400
101
+ },
102
+ {
103
+ "epoch": 2.56,
104
+ "eval_accuracy": 0.9143,
105
+ "eval_f1": 0.9142909071223366,
106
+ "eval_loss": 0.2254679948091507,
107
+ "eval_matthews_correlation": 0.8286017317454041,
108
+ "eval_precision": 0.9143290342981838,
109
+ "eval_recall": 0.9142726993622687,
110
+ "eval_runtime": 0.6994,
111
+ "eval_samples_per_second": 14297.498,
112
+ "eval_steps_per_second": 112.95,
113
+ "step": 1600
114
+ },
115
+ {
116
+ "epoch": 2.88,
117
+ "eval_accuracy": 0.9147,
118
+ "eval_f1": 0.9146887175829004,
119
+ "eval_loss": 0.22356006503105164,
120
+ "eval_matthews_correlation": 0.8294144901528736,
121
+ "eval_precision": 0.9147501760450676,
122
+ "eval_recall": 0.914664318551609,
123
+ "eval_runtime": 0.7,
124
+ "eval_samples_per_second": 14286.322,
125
+ "eval_steps_per_second": 112.862,
126
+ "step": 1800
127
+ }
128
+ ],
129
+ "logging_steps": 100000,
130
+ "max_steps": 1875,
131
+ "num_train_epochs": 3,
132
+ "save_steps": 200,
133
+ "total_flos": 6393598845124608.0,
134
+ "trial_name": null,
135
+ "trial_params": null
136
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/checkpoint-1800/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66f558acf951dddcf698771a7d98f4feaca4a4b912d76fa9c7ec16ebe809283c
3
+ size 5329
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_3072/results/base_3072_hg38_BPE_3e-5_demo_coding_vs_intergenomic_seqs_seed42/eval_results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eval_loss": 0.23370100557804108, "eval_accuracy": 0.9143, "eval_f1": 0.9142994643716523, "eval_matthews_correlation": 0.8286761912329407, "eval_precision": 0.9143515624765606, "eval_recall": 0.9143246291940671, "eval_runtime": 0.6978, "eval_samples_per_second": 14330.746, "eval_steps_per_second": 113.213, "epoch": 3.0}
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_4096/checkpoint-100000",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_length": 512,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "problem_type": "single_label_classification",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.35.2",
24
+ "type_vocab_size": 2,
25
+ "use_cache": true,
26
+ "vocab_size": 4096
27
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:238ba1e2b02bba89f3c53efa289d3b86bf72e3e2221144b3b08c0ba88aa3f592
3
+ size 356777880
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:818f86f31a65892b5378bc0961089a1e5b20e82e36b9753753b4202848248115
3
+ size 713677451
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:881f77abcbcf9e83c940223f5ca9133f12d2be418768a67a53945677a75907b9
3
+ size 14709
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e32828859e9d069315b3f3e9abbd7af9b79ba7a4ee7041fec91396a6f9e6768
3
+ size 1465
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "cache_dir": null,
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 100,
49
+ "pad_token": "[PAD]",
50
+ "padding_side": "right",
51
+ "sep_token": "[SEP]",
52
+ "tokenizer_class": "PreTrainedTokenizerFast",
53
+ "trust_remote_code": true,
54
+ "unk_token": "[UNK]",
55
+ "use_fast": true
56
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/trainer_state.json ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.24,
5
+ "eval_steps": 200,
6
+ "global_step": 1400,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.32,
13
+ "eval_accuracy": 0.9004,
14
+ "eval_f1": 0.900398243025007,
15
+ "eval_loss": 0.2527843713760376,
16
+ "eval_matthews_correlation": 0.8010224138324217,
17
+ "eval_precision": 0.9005488941757427,
18
+ "eval_recall": 0.9004735232026368,
19
+ "eval_runtime": 0.7219,
20
+ "eval_samples_per_second": 13852.527,
21
+ "eval_steps_per_second": 109.435,
22
+ "step": 200
23
+ },
24
+ {
25
+ "epoch": 0.64,
26
+ "eval_accuracy": 0.9033,
27
+ "eval_f1": 0.903172411254586,
28
+ "eval_loss": 0.23981137573719025,
29
+ "eval_matthews_correlation": 0.8079007693175992,
30
+ "eval_precision": 0.9048124537208442,
31
+ "eval_recall": 0.9030901514174194,
32
+ "eval_runtime": 0.695,
33
+ "eval_samples_per_second": 14387.745,
34
+ "eval_steps_per_second": 113.663,
35
+ "step": 400
36
+ },
37
+ {
38
+ "epoch": 0.96,
39
+ "eval_accuracy": 0.9036,
40
+ "eval_f1": 0.9035254406015145,
41
+ "eval_loss": 0.2367754429578781,
42
+ "eval_matthews_correlation": 0.8078341688032824,
43
+ "eval_precision": 0.9043850412921304,
44
+ "eval_recall": 0.9034496690337827,
45
+ "eval_runtime": 0.6971,
46
+ "eval_samples_per_second": 14344.287,
47
+ "eval_steps_per_second": 113.32,
48
+ "step": 600
49
+ },
50
+ {
51
+ "epoch": 1.28,
52
+ "eval_accuracy": 0.9109,
53
+ "eval_f1": 0.9108989085116292,
54
+ "eval_loss": 0.2278522253036499,
55
+ "eval_matthews_correlation": 0.8218201348399584,
56
+ "eval_precision": 0.9109,
57
+ "eval_recall": 0.9109201350866192,
58
+ "eval_runtime": 0.6987,
59
+ "eval_samples_per_second": 14312.505,
60
+ "eval_steps_per_second": 113.069,
61
+ "step": 800
62
+ },
63
+ {
64
+ "epoch": 1.6,
65
+ "eval_accuracy": 0.9115,
66
+ "eval_f1": 0.9114986538945258,
67
+ "eval_loss": 0.2427944540977478,
68
+ "eval_matthews_correlation": 0.8232139301065716,
69
+ "eval_precision": 0.9116419660562449,
70
+ "eval_recall": 0.9115719670263842,
71
+ "eval_runtime": 0.6945,
72
+ "eval_samples_per_second": 14397.949,
73
+ "eval_steps_per_second": 113.744,
74
+ "step": 1000
75
+ },
76
+ {
77
+ "epoch": 1.92,
78
+ "eval_accuracy": 0.9083,
79
+ "eval_f1": 0.9082222817103751,
80
+ "eval_loss": 0.22559477388858795,
81
+ "eval_matthews_correlation": 0.8173335971525737,
82
+ "eval_precision": 0.9091934761461085,
83
+ "eval_recall": 0.9081407988991461,
84
+ "eval_runtime": 0.6941,
85
+ "eval_samples_per_second": 14407.148,
86
+ "eval_steps_per_second": 113.816,
87
+ "step": 1200
88
+ },
89
+ {
90
+ "epoch": 2.24,
91
+ "eval_accuracy": 0.9094,
92
+ "eval_f1": 0.909320550293345,
93
+ "eval_loss": 0.23093822598457336,
94
+ "eval_matthews_correlation": 0.8195720153678601,
95
+ "eval_precision": 0.910335398306722,
96
+ "eval_recall": 0.9092373526302788,
97
+ "eval_runtime": 0.6996,
98
+ "eval_samples_per_second": 14293.8,
99
+ "eval_steps_per_second": 112.921,
100
+ "step": 1400
101
+ }
102
+ ],
103
+ "logging_steps": 100000,
104
+ "max_steps": 1875,
105
+ "num_train_epochs": 3,
106
+ "save_steps": 200,
107
+ "total_flos": 4880710238208000.0,
108
+ "trial_name": null,
109
+ "trial_params": null
110
+ }
full_output_multi_epoch/demo_coding_vs_intergenomic_seqs/base_4096/checkpoint-1400/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b30702f12d1c6578823ea18d25e59c6b65c7d877145f8e9aed3cb60df7adec9
3
+ size 5329