Bram Vanroy commited on
Commit
fc44f1d
1 Parent(s): 107c2a4

generate models.json

Browse files
Files changed (1) hide show
  1. evals/models.json +92 -0
evals/models.json ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "gpt-neo-1.3b-dutch": {
3
+ "compute_dtype": "bfloat16",
4
+ "model_name": "yhavinga/gpt-neo-1.3B-dutch",
5
+ "num_parameters": 1315575808,
6
+ "quantization": "8-bit"
7
+ },
8
+ "gpt-neo-125m-dutch": {
9
+ "compute_dtype": "bfloat16",
10
+ "model_name": "yhavinga/gpt-neo-125M-dutch",
11
+ "num_parameters": 125198592,
12
+ "quantization": "8-bit"
13
+ },
14
+ "gpt2-large-dutch": {
15
+ "compute_dtype": "bfloat16",
16
+ "model_name": "yhavinga/gpt2-large-dutch",
17
+ "num_parameters": 774030080,
18
+ "quantization": "8-bit"
19
+ },
20
+ "gpt2-medium-dutch": {
21
+ "compute_dtype": "bfloat16",
22
+ "model_name": "yhavinga/gpt2-medium-dutch",
23
+ "num_parameters": 354823168,
24
+ "quantization": "8-bit"
25
+ },
26
+ "llama-2-13b-chat-dutch": {
27
+ "compute_dtype": "bfloat16",
28
+ "model_name": "BramVanroy/Llama-2-13b-chat-dutch",
29
+ "num_parameters": 13015864320,
30
+ "quantization": "8-bit"
31
+ },
32
+ "llama-2-13b-chat-hf": {
33
+ "compute_dtype": "bfloat16",
34
+ "model_name": "meta-llama/Llama-2-13b-chat-hf",
35
+ "num_parameters": 13015864320,
36
+ "quantization": "8-bit"
37
+ },
38
+ "llama-2-13b-hf": {
39
+ "compute_dtype": "bfloat16",
40
+ "model_name": "meta-llama/Llama-2-13b-hf",
41
+ "num_parameters": 13015864320,
42
+ "quantization": "8-bit"
43
+ },
44
+ "llama-2-7b-chat-hf": {
45
+ "compute_dtype": "bfloat16",
46
+ "model_name": "meta-llama/Llama-2-7b-chat-hf",
47
+ "num_parameters": 6738415616,
48
+ "quantization": "8-bit"
49
+ },
50
+ "llama-2-7b-hf": {
51
+ "compute_dtype": "bfloat16",
52
+ "model_name": "meta-llama/Llama-2-7b-hf",
53
+ "num_parameters": 6738415616,
54
+ "quantization": "8-bit"
55
+ },
56
+ "llama2-13b-ft-mc4": {
57
+ "compute_dtype": "bfloat16",
58
+ "model_name": "BramVanroy/llama2-13b-ft-mc4_nl_cleaned_tiny",
59
+ "num_parameters": 13015864320,
60
+ "quantization": "8-bit"
61
+ },
62
+ "mistral-7b-v0.1": {
63
+ "compute_dtype": "bfloat16",
64
+ "model_name": "mistralai/Mistral-7B-v0.1",
65
+ "num_parameters": 7241732096,
66
+ "quantization": "8-bit"
67
+ },
68
+ "neural-chat-7b-v3-1": {
69
+ "compute_dtype": "bfloat16",
70
+ "model_name": "Intel/neural-chat-7b-v3-1",
71
+ "num_parameters": 7241732096,
72
+ "quantization": "8-bit"
73
+ },
74
+ "orca-2-13b": {
75
+ "compute_dtype": "bfloat16",
76
+ "model_name": "microsoft/Orca-2-13b",
77
+ "num_parameters": 13015895040,
78
+ "quantization": "8-bit"
79
+ },
80
+ "orca-2-7b": {
81
+ "compute_dtype": "bfloat16",
82
+ "model_name": "microsoft/Orca-2-7b",
83
+ "num_parameters": 6738440192,
84
+ "quantization": "8-bit"
85
+ },
86
+ "zephyr-7b-beta": {
87
+ "compute_dtype": "bfloat16",
88
+ "model_name": "HuggingFaceH4/zephyr-7b-beta",
89
+ "num_parameters": 7241732096,
90
+ "quantization": "8-bit"
91
+ }
92
+ }