dacorvo HF staff commited on
Commit
d3983e8
1 Parent(s): ffdfb54

Add mistral most popular variants

Browse files
inference-cache-config/mistral-variants.json ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "NousResearch/Genstruct-7B": [
3
+ {
4
+ "batch_size": 1,
5
+ "sequence_length": 4096,
6
+ "num_cores": 2,
7
+ "auto_cast_type": "bf16"
8
+ },
9
+ {
10
+ "batch_size": 4,
11
+ "sequence_length": 4096,
12
+ "num_cores": 2,
13
+ "auto_cast_type": "bf16"
14
+ }
15
+ ],
16
+ "BioMistral/BioMistral-7B": [
17
+ {
18
+ "batch_size": 1,
19
+ "sequence_length": 4096,
20
+ "num_cores": 2,
21
+ "auto_cast_type": "bf16"
22
+ },
23
+ {
24
+ "batch_size": 4,
25
+ "sequence_length": 4096,
26
+ "num_cores": 2,
27
+ "auto_cast_type": "bf16"
28
+ }
29
+ ],
30
+ "NousResearch/Hermes-2-Pro-Mistral-7B": [
31
+ {
32
+ "batch_size": 1,
33
+ "sequence_length": 4096,
34
+ "num_cores": 2,
35
+ "auto_cast_type": "bf16"
36
+ },
37
+ {
38
+ "batch_size": 4,
39
+ "sequence_length": 4096,
40
+ "num_cores": 2,
41
+ "auto_cast_type": "bf16"
42
+ }
43
+ ],
44
+ "NousResearch/Nous-Hermes-2-Mistral-7B-DPO": [
45
+ {
46
+ "batch_size": 1,
47
+ "sequence_length": 4096,
48
+ "num_cores": 2,
49
+ "auto_cast_type": "bf16"
50
+ },
51
+ {
52
+ "batch_size": 4,
53
+ "sequence_length": 4096,
54
+ "num_cores": 2,
55
+ "auto_cast_type": "bf16"
56
+ }
57
+ ],
58
+ "ibm/merlinite-7b": [
59
+ {
60
+ "batch_size": 1,
61
+ "sequence_length": 4096,
62
+ "num_cores": 2,
63
+ "auto_cast_type": "bf16"
64
+ },
65
+ {
66
+ "batch_size": 4,
67
+ "sequence_length": 4096,
68
+ "num_cores": 2,
69
+ "auto_cast_type": "bf16"
70
+ }
71
+ ],
72
+ "mlabonne/AlphaMonarch-7B": [
73
+ {
74
+ "batch_size": 1,
75
+ "sequence_length": 4096,
76
+ "num_cores": 2,
77
+ "auto_cast_type": "fp16"
78
+ },
79
+ {
80
+ "batch_size": 4,
81
+ "sequence_length": 4096,
82
+ "num_cores": 2,
83
+ "auto_cast_type": "fp16"
84
+ }
85
+ ],
86
+ "teknium/OpenHermes-2.5-Mistral-7B": [
87
+ {
88
+ "batch_size": 1,
89
+ "sequence_length": 4096,
90
+ "num_cores": 2,
91
+ "auto_cast_type": "bf16"
92
+ },
93
+ {
94
+ "batch_size": 4,
95
+ "sequence_length": 4096,
96
+ "num_cores": 2,
97
+ "auto_cast_type": "bf16"
98
+ }
99
+ ],
100
+ "TencentARC/Mistral_Pro_8B_v0.1": [
101
+ {
102
+ "batch_size": 1,
103
+ "sequence_length": 4096,
104
+ "num_cores": 2,
105
+ "auto_cast_type": "bf16"
106
+ },
107
+ {
108
+ "batch_size": 4,
109
+ "sequence_length": 4096,
110
+ "num_cores": 2,
111
+ "auto_cast_type": "bf16"
112
+ }
113
+ ],
114
+ "openchat/openchat-3.5-0106": [
115
+ {
116
+ "batch_size": 1,
117
+ "sequence_length": 4096,
118
+ "num_cores": 2,
119
+ "auto_cast_type": "bf16"
120
+ },
121
+ {
122
+ "batch_size": 4,
123
+ "sequence_length": 4096,
124
+ "num_cores": 2,
125
+ "auto_cast_type": "bf16"
126
+ }
127
+ ],
128
+ "Salesforce/SFR-Embedding-Mistral": [
129
+ {
130
+ "batch_size": 1,
131
+ "sequence_length": 4096,
132
+ "num_cores": 2,
133
+ "auto_cast_type": "fp16"
134
+ },
135
+ {
136
+ "batch_size": 4,
137
+ "sequence_length": 4096,
138
+ "num_cores": 2,
139
+ "auto_cast_type": "fp16"
140
+ }
141
+ ]
142
+ }