lhallee commited on
Commit
30b5a1c
·
verified ·
1 Parent(s): cfaf284

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7914cdac70ece70ee4cc559e5d7f69aec04d0bd324b551e961478bbcdbd2452c
3
+ size 421776744
step_10000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_10000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f0a2bcf822e2214c523a74e4e1a4260fdef13aa122d657f9761d665ac4c9077
3
+ size 421776744
step_100000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_100000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7914cdac70ece70ee4cc559e5d7f69aec04d0bd324b551e961478bbcdbd2452c
3
+ size 421776744
step_15000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_15000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef076f8cbf1a180d9baf3ed77ba90560d5c71edf61d1533b6420a833f16f9802
3
+ size 421776744
step_20000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_20000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6510ab986195bb5e360d1bafc6e4afd6358491768cd7c6118ecb92386e96539b
3
+ size 421776744
step_25000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_25000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e08e609fd4ef5d3677d5ea6acb439a65af3666e4ee1836c818522d2c0a0ef2b3
3
+ size 421776744
step_30000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_30000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0f6698178ca6df8c10284b38dfb1c8d333d005567f58a1b68f03e72ad0d98a4
3
+ size 421776744
step_35000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_35000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44d34f76a68e89c5e7b8428bd1ca2f0e5505d45dcf29c7e1ec43e7732111fa7a
3
+ size 421776744
step_40000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_40000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:859656ce92711f81b3e1daf499ee884c52a40cfe846c87b406a79d71074e922a
3
+ size 421776744
step_45000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_45000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b95ba96278c1f0d9661ffcdc6ff83ec3570c73fa7f6fab46c10af1bade7394d8
3
+ size 421776744
step_5000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_5000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4425344a5bb0e353badb6c7df76e18cc4e775e71cfc829b494b17df364d96039
3
+ size 421776744
step_50000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_50000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f016b30a53ac18a14df86c7018ee585f7d5997e5892e756330b20aa3d8112b
3
+ size 421776744
step_55000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_55000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0ae9d5b7af3529a06df0938b1db6a37b928fc79bdb6fb189bd18e832a525eea
3
+ size 421776744
step_60000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_60000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6063520cc75cd2d4d1eaf66ecbe3eff1eb70bd1d6f78138ce3aa6e53aefe35d7
3
+ size 421776744
step_65000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_65000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63bee4d2f1df9cffd486d78f3f727f1b8673666c0fe8b8acc16657af72dc3be3
3
+ size 421776744
step_70000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_70000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0390d12589003c26f142b28a1e4a74d96f2c7d1dec30674bde5231e56007546c
3
+ size 421776744
step_75000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_75000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94e8ec9a80dd55038826fbed33e6849dc7d5acec7521579e15f02f347ba6c428
3
+ size 421776744
step_80000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_80000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2342defc4f13301823cc3993de964ceda0b9e32af6603a10f9fa62d5e94cb887
3
+ size 421776744
step_85000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_85000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5b88bcadd36809e3fe7d2cdbce09b2ba45ed2fb5f00d123b3a785cf446c8c67
3
+ size 421776744
step_90000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_90000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d00c041a40782b910904e678a150cf5a9ab47b8f53c45562e4566783ea87793
3
+ size 421776744
step_95000/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "AnnotationTransformer"
4
+ ],
5
+ "dropout": 0.1,
6
+ "dtype": "float32",
7
+ "expansion_ratio": 2.6666666666666665,
8
+ "hidden_size": 512,
9
+ "mask_token_id": 88281,
10
+ "max_position_embeddings": 1024,
11
+ "mlm_mask_probability": 0.15,
12
+ "model_type": "annotation_transformer",
13
+ "n_heads": 8,
14
+ "num_hidden_layers": 4,
15
+ "pad_token_id": 0,
16
+ "rotary": true,
17
+ "transformers_version": "4.57.6",
18
+ "vocab_size": 88282
19
+ }
step_95000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fb128e597af6caac3c68bf8094a8e6e9fb4b59607a622a6165ce1f23eb68dba
3
+ size 421776744