ArthurZ HF staff commited on
Commit
050ed5b
1 Parent(s): aa09464

add sharded

Browse files
tf_model-00001-of-00005.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc3728c39c3320cb4123cf5cc5f566ff57294c68ca3386983f45385159d4162f
3
+ size 928
tf_model-00002-of-00005.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e5ab02766e0ff90dddeacdd860f2261f10370054915129cb426a7134cd50c6b
3
+ size 152640
tf_model-00003-of-00005.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5170be5d9165daf8efa88c80d9ef76d850ab8d017053c8a514e5ab9964c03786
3
+ size 136112
tf_model-00004-of-00005.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65f6eea23e046bd5b26c8bded84e5fcdb872acdf29c5e8cb34f5debcd732b9c8
3
+ size 178844
tf_model-00005-of-00005.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03db5302457e84beef289ab1931efaa72ea146042c8dfb61ebe23c513fdabd5a
3
+ size 26216
tf_model.h5.index.json ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 351716
4
+ },
5
+ "weight_map": {
6
+ "tf_bert_model/bert/embeddings/LayerNorm/beta:0": "tf_model-00003-of-00005.h5",
7
+ "tf_bert_model/bert/embeddings/LayerNorm/gamma:0": "tf_model-00003-of-00005.h5",
8
+ "tf_bert_model/bert/embeddings/position_embeddings/embeddings:0": "tf_model-00003-of-00005.h5",
9
+ "tf_bert_model/bert/embeddings/token_type_embeddings/embeddings:0": "tf_model-00003-of-00005.h5",
10
+ "tf_bert_model/bert/embeddings/word_embeddings/weight:0": "tf_model-00002-of-00005.h5",
11
+ "tf_bert_model/bert/encoder/layer_._0/attention/output/LayerNorm/beta:0": "tf_model-00003-of-00005.h5",
12
+ "tf_bert_model/bert/encoder/layer_._0/attention/output/LayerNorm/gamma:0": "tf_model-00003-of-00005.h5",
13
+ "tf_bert_model/bert/encoder/layer_._0/attention/output/dense/bias:0": "tf_model-00003-of-00005.h5",
14
+ "tf_bert_model/bert/encoder/layer_._0/attention/output/dense/kernel:0": "tf_model-00003-of-00005.h5",
15
+ "tf_bert_model/bert/encoder/layer_._0/attention/self/key/bias:0": "tf_model-00003-of-00005.h5",
16
+ "tf_bert_model/bert/encoder/layer_._0/attention/self/key/kernel:0": "tf_model-00003-of-00005.h5",
17
+ "tf_bert_model/bert/encoder/layer_._0/attention/self/query/bias:0": "tf_model-00003-of-00005.h5",
18
+ "tf_bert_model/bert/encoder/layer_._0/attention/self/query/kernel:0": "tf_model-00003-of-00005.h5",
19
+ "tf_bert_model/bert/encoder/layer_._0/attention/self/value/bias:0": "tf_model-00003-of-00005.h5",
20
+ "tf_bert_model/bert/encoder/layer_._0/attention/self/value/kernel:0": "tf_model-00003-of-00005.h5",
21
+ "tf_bert_model/bert/encoder/layer_._0/intermediate/dense/bias:0": "tf_model-00003-of-00005.h5",
22
+ "tf_bert_model/bert/encoder/layer_._0/intermediate/dense/kernel:0": "tf_model-00003-of-00005.h5",
23
+ "tf_bert_model/bert/encoder/layer_._0/output/LayerNorm/beta:0": "tf_model-00003-of-00005.h5",
24
+ "tf_bert_model/bert/encoder/layer_._0/output/LayerNorm/gamma:0": "tf_model-00003-of-00005.h5",
25
+ "tf_bert_model/bert/encoder/layer_._0/output/dense/bias:0": "tf_model-00003-of-00005.h5",
26
+ "tf_bert_model/bert/encoder/layer_._0/output/dense/kernel:0": "tf_model-00003-of-00005.h5",
27
+ "tf_bert_model/bert/encoder/layer_._1/attention/output/LayerNorm/beta:0": "tf_model-00004-of-00005.h5",
28
+ "tf_bert_model/bert/encoder/layer_._1/attention/output/LayerNorm/gamma:0": "tf_model-00004-of-00005.h5",
29
+ "tf_bert_model/bert/encoder/layer_._1/attention/output/dense/bias:0": "tf_model-00004-of-00005.h5",
30
+ "tf_bert_model/bert/encoder/layer_._1/attention/output/dense/kernel:0": "tf_model-00004-of-00005.h5",
31
+ "tf_bert_model/bert/encoder/layer_._1/attention/self/key/bias:0": "tf_model-00004-of-00005.h5",
32
+ "tf_bert_model/bert/encoder/layer_._1/attention/self/key/kernel:0": "tf_model-00004-of-00005.h5",
33
+ "tf_bert_model/bert/encoder/layer_._1/attention/self/query/bias:0": "tf_model-00003-of-00005.h5",
34
+ "tf_bert_model/bert/encoder/layer_._1/attention/self/query/kernel:0": "tf_model-00003-of-00005.h5",
35
+ "tf_bert_model/bert/encoder/layer_._1/attention/self/value/bias:0": "tf_model-00004-of-00005.h5",
36
+ "tf_bert_model/bert/encoder/layer_._1/attention/self/value/kernel:0": "tf_model-00004-of-00005.h5",
37
+ "tf_bert_model/bert/encoder/layer_._1/intermediate/dense/bias:0": "tf_model-00004-of-00005.h5",
38
+ "tf_bert_model/bert/encoder/layer_._1/intermediate/dense/kernel:0": "tf_model-00004-of-00005.h5",
39
+ "tf_bert_model/bert/encoder/layer_._1/output/LayerNorm/beta:0": "tf_model-00004-of-00005.h5",
40
+ "tf_bert_model/bert/encoder/layer_._1/output/LayerNorm/gamma:0": "tf_model-00004-of-00005.h5",
41
+ "tf_bert_model/bert/encoder/layer_._1/output/dense/bias:0": "tf_model-00004-of-00005.h5",
42
+ "tf_bert_model/bert/encoder/layer_._1/output/dense/kernel:0": "tf_model-00004-of-00005.h5",
43
+ "tf_bert_model/bert/encoder/layer_._2/attention/output/LayerNorm/beta:0": "tf_model-00004-of-00005.h5",
44
+ "tf_bert_model/bert/encoder/layer_._2/attention/output/LayerNorm/gamma:0": "tf_model-00004-of-00005.h5",
45
+ "tf_bert_model/bert/encoder/layer_._2/attention/output/dense/bias:0": "tf_model-00004-of-00005.h5",
46
+ "tf_bert_model/bert/encoder/layer_._2/attention/output/dense/kernel:0": "tf_model-00004-of-00005.h5",
47
+ "tf_bert_model/bert/encoder/layer_._2/attention/self/key/bias:0": "tf_model-00004-of-00005.h5",
48
+ "tf_bert_model/bert/encoder/layer_._2/attention/self/key/kernel:0": "tf_model-00004-of-00005.h5",
49
+ "tf_bert_model/bert/encoder/layer_._2/attention/self/query/bias:0": "tf_model-00004-of-00005.h5",
50
+ "tf_bert_model/bert/encoder/layer_._2/attention/self/query/kernel:0": "tf_model-00004-of-00005.h5",
51
+ "tf_bert_model/bert/encoder/layer_._2/attention/self/value/bias:0": "tf_model-00004-of-00005.h5",
52
+ "tf_bert_model/bert/encoder/layer_._2/attention/self/value/kernel:0": "tf_model-00004-of-00005.h5",
53
+ "tf_bert_model/bert/encoder/layer_._2/intermediate/dense/bias:0": "tf_model-00004-of-00005.h5",
54
+ "tf_bert_model/bert/encoder/layer_._2/intermediate/dense/kernel:0": "tf_model-00004-of-00005.h5",
55
+ "tf_bert_model/bert/encoder/layer_._2/output/LayerNorm/beta:0": "tf_model-00004-of-00005.h5",
56
+ "tf_bert_model/bert/encoder/layer_._2/output/LayerNorm/gamma:0": "tf_model-00004-of-00005.h5",
57
+ "tf_bert_model/bert/encoder/layer_._2/output/dense/bias:0": "tf_model-00004-of-00005.h5",
58
+ "tf_bert_model/bert/encoder/layer_._2/output/dense/kernel:0": "tf_model-00004-of-00005.h5",
59
+ "tf_bert_model/bert/encoder/layer_._3/attention/output/LayerNorm/beta:0": "tf_model-00004-of-00005.h5",
60
+ "tf_bert_model/bert/encoder/layer_._3/attention/output/LayerNorm/gamma:0": "tf_model-00004-of-00005.h5",
61
+ "tf_bert_model/bert/encoder/layer_._3/attention/output/dense/bias:0": "tf_model-00004-of-00005.h5",
62
+ "tf_bert_model/bert/encoder/layer_._3/attention/output/dense/kernel:0": "tf_model-00004-of-00005.h5",
63
+ "tf_bert_model/bert/encoder/layer_._3/attention/self/key/bias:0": "tf_model-00004-of-00005.h5",
64
+ "tf_bert_model/bert/encoder/layer_._3/attention/self/key/kernel:0": "tf_model-00004-of-00005.h5",
65
+ "tf_bert_model/bert/encoder/layer_._3/attention/self/query/bias:0": "tf_model-00004-of-00005.h5",
66
+ "tf_bert_model/bert/encoder/layer_._3/attention/self/query/kernel:0": "tf_model-00004-of-00005.h5",
67
+ "tf_bert_model/bert/encoder/layer_._3/attention/self/value/bias:0": "tf_model-00004-of-00005.h5",
68
+ "tf_bert_model/bert/encoder/layer_._3/attention/self/value/kernel:0": "tf_model-00004-of-00005.h5",
69
+ "tf_bert_model/bert/encoder/layer_._3/intermediate/dense/bias:0": "tf_model-00004-of-00005.h5",
70
+ "tf_bert_model/bert/encoder/layer_._3/intermediate/dense/kernel:0": "tf_model-00004-of-00005.h5",
71
+ "tf_bert_model/bert/encoder/layer_._3/output/LayerNorm/beta:0": "tf_model-00004-of-00005.h5",
72
+ "tf_bert_model/bert/encoder/layer_._3/output/LayerNorm/gamma:0": "tf_model-00004-of-00005.h5",
73
+ "tf_bert_model/bert/encoder/layer_._3/output/dense/bias:0": "tf_model-00004-of-00005.h5",
74
+ "tf_bert_model/bert/encoder/layer_._3/output/dense/kernel:0": "tf_model-00004-of-00005.h5",
75
+ "tf_bert_model/bert/encoder/layer_._4/attention/output/LayerNorm/beta:0": "tf_model-00004-of-00005.h5",
76
+ "tf_bert_model/bert/encoder/layer_._4/attention/output/LayerNorm/gamma:0": "tf_model-00004-of-00005.h5",
77
+ "tf_bert_model/bert/encoder/layer_._4/attention/output/dense/bias:0": "tf_model-00004-of-00005.h5",
78
+ "tf_bert_model/bert/encoder/layer_._4/attention/output/dense/kernel:0": "tf_model-00004-of-00005.h5",
79
+ "tf_bert_model/bert/encoder/layer_._4/attention/self/key/bias:0": "tf_model-00004-of-00005.h5",
80
+ "tf_bert_model/bert/encoder/layer_._4/attention/self/key/kernel:0": "tf_model-00004-of-00005.h5",
81
+ "tf_bert_model/bert/encoder/layer_._4/attention/self/query/bias:0": "tf_model-00004-of-00005.h5",
82
+ "tf_bert_model/bert/encoder/layer_._4/attention/self/query/kernel:0": "tf_model-00004-of-00005.h5",
83
+ "tf_bert_model/bert/encoder/layer_._4/attention/self/value/bias:0": "tf_model-00004-of-00005.h5",
84
+ "tf_bert_model/bert/encoder/layer_._4/attention/self/value/kernel:0": "tf_model-00004-of-00005.h5",
85
+ "tf_bert_model/bert/encoder/layer_._4/intermediate/dense/bias:0": "tf_model-00004-of-00005.h5",
86
+ "tf_bert_model/bert/encoder/layer_._4/intermediate/dense/kernel:0": "tf_model-00004-of-00005.h5",
87
+ "tf_bert_model/bert/encoder/layer_._4/output/LayerNorm/beta:0": "tf_model-00005-of-00005.h5",
88
+ "tf_bert_model/bert/encoder/layer_._4/output/LayerNorm/gamma:0": "tf_model-00005-of-00005.h5",
89
+ "tf_bert_model/bert/encoder/layer_._4/output/dense/bias:0": "tf_model-00005-of-00005.h5",
90
+ "tf_bert_model/bert/encoder/layer_._4/output/dense/kernel:0": "tf_model-00005-of-00005.h5",
91
+ "tf_bert_model/bert/pooler/dense/bias:0": "tf_model-00005-of-00005.h5",
92
+ "tf_bert_model/bert/pooler/dense/kernel:0": "tf_model-00005-of-00005.h5"
93
+ }
94
+ }