yangwawa0202 commited on
Commit
721dd05
1 Parent(s): bfb8d7e

Upload 2 files

Browse files
Files changed (2) hide show
  1. added_tokens.json +102 -0
  2. config.json +37 -0
added_tokens.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__af__": 128004,
3
+ "__am__": 128005,
4
+ "__ar__": 128006,
5
+ "__ast__": 128007,
6
+ "__az__": 128008,
7
+ "__ba__": 128009,
8
+ "__be__": 128010,
9
+ "__bg__": 128011,
10
+ "__bn__": 128012,
11
+ "__br__": 128013,
12
+ "__bs__": 128014,
13
+ "__ca__": 128015,
14
+ "__ceb__": 128016,
15
+ "__cs__": 128017,
16
+ "__cy__": 128018,
17
+ "__da__": 128019,
18
+ "__de__": 128020,
19
+ "__el__": 128021,
20
+ "__en__": 128022,
21
+ "__es__": 128023,
22
+ "__et__": 128024,
23
+ "__fa__": 128025,
24
+ "__ff__": 128026,
25
+ "__fi__": 128027,
26
+ "__fr__": 128028,
27
+ "__fy__": 128029,
28
+ "__ga__": 128030,
29
+ "__gd__": 128031,
30
+ "__gl__": 128032,
31
+ "__gu__": 128033,
32
+ "__ha__": 128034,
33
+ "__he__": 128035,
34
+ "__hi__": 128036,
35
+ "__hr__": 128037,
36
+ "__ht__": 128038,
37
+ "__hu__": 128039,
38
+ "__hy__": 128040,
39
+ "__id__": 128041,
40
+ "__ig__": 128042,
41
+ "__ilo__": 128043,
42
+ "__is__": 128044,
43
+ "__it__": 128045,
44
+ "__ja__": 128046,
45
+ "__jv__": 128047,
46
+ "__ka__": 128048,
47
+ "__kk__": 128049,
48
+ "__km__": 128050,
49
+ "__kn__": 128051,
50
+ "__ko__": 128052,
51
+ "__lb__": 128053,
52
+ "__lg__": 128054,
53
+ "__ln__": 128055,
54
+ "__lo__": 128056,
55
+ "__lt__": 128057,
56
+ "__lv__": 128058,
57
+ "__mg__": 128059,
58
+ "__mk__": 128060,
59
+ "__ml__": 128061,
60
+ "__mn__": 128062,
61
+ "__mr__": 128063,
62
+ "__ms__": 128064,
63
+ "__my__": 128065,
64
+ "__ne__": 128066,
65
+ "__nl__": 128067,
66
+ "__no__": 128068,
67
+ "__ns__": 128069,
68
+ "__oc__": 128070,
69
+ "__or__": 128071,
70
+ "__pa__": 128072,
71
+ "__pl__": 128073,
72
+ "__ps__": 128074,
73
+ "__pt__": 128075,
74
+ "__ro__": 128076,
75
+ "__ru__": 128077,
76
+ "__sd__": 128078,
77
+ "__si__": 128079,
78
+ "__sk__": 128080,
79
+ "__sl__": 128081,
80
+ "__so__": 128082,
81
+ "__sq__": 128083,
82
+ "__sr__": 128084,
83
+ "__ss__": 128085,
84
+ "__su__": 128086,
85
+ "__sv__": 128087,
86
+ "__sw__": 128088,
87
+ "__ta__": 128089,
88
+ "__th__": 128090,
89
+ "__tl__": 128091,
90
+ "__tn__": 128092,
91
+ "__tr__": 128093,
92
+ "__uk__": 128094,
93
+ "__ur__": 128095,
94
+ "__uz__": 128096,
95
+ "__vi__": 128097,
96
+ "__wo__": 128098,
97
+ "__xh__": 128099,
98
+ "__yi__": 128100,
99
+ "__yo__": 128101,
100
+ "__zh__": 128102,
101
+ "__zu__": 128103
102
+ }
config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/m2m100_1.2B",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "relu",
5
+ "architectures": [
6
+ "M2M100ForConditionalGeneration"
7
+ ],
8
+ "attention_dropout": 0.1,
9
+ "bos_token_id": 0,
10
+ "d_model": 1024,
11
+ "decoder_attention_heads": 16,
12
+ "decoder_ffn_dim": 8192,
13
+ "decoder_layerdrop": 0.05,
14
+ "decoder_layers": 24,
15
+ "decoder_start_token_id": 2,
16
+ "dropout": 0.1,
17
+ "early_stopping": true,
18
+ "encoder_attention_heads": 16,
19
+ "encoder_ffn_dim": 8192,
20
+ "encoder_layerdrop": 0.05,
21
+ "encoder_layers": 24,
22
+ "eos_token_id": 2,
23
+ "gradient_checkpointing": false,
24
+ "init_std": 0.02,
25
+ "is_encoder_decoder": true,
26
+ "max_length": 200,
27
+ "max_position_embeddings": 1024,
28
+ "model_type": "m2m_100",
29
+ "num_beams": 5,
30
+ "num_hidden_layers": 24,
31
+ "pad_token_id": 1,
32
+ "scale_embedding": true,
33
+ "torch_dtype": "float16",
34
+ "transformers_version": "4.36.2",
35
+ "use_cache": true,
36
+ "vocab_size": 128112
37
+ }