Update ChatApp/app.py
Browse files- ChatApp/app.py +17 -1
ChatApp/app.py
CHANGED
@@ -33,6 +33,22 @@ top_directory = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
|
|
33 |
tokenizer_path = os.path.join(top_directory, "tokenizer.model")
|
34 |
|
35 |
available_models = {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
36 |
"Llama-2 13B Float16": {
|
37 |
"onnx_file": os.path.join(
|
38 |
top_directory, "FP16", "LlamaV2_13B_float16.onnx"
|
@@ -40,7 +56,7 @@ available_models = {
|
|
40 |
"tokenizer_path": tokenizer_path,
|
41 |
"embedding_file": os.path.join(top_directory, "embeddings.pth"),
|
42 |
},
|
43 |
-
"Llama-2 13B
|
44 |
"onnx_file": os.path.join(
|
45 |
top_directory, "FP32", "LlamaV2_13B_float16.onnx"
|
46 |
),
|
|
|
33 |
tokenizer_path = os.path.join(top_directory, "tokenizer.model")
|
34 |
|
35 |
available_models = {
|
36 |
+
"Llama-2 Chat 13B Float16": {
|
37 |
+
"onnx_file": os.path.join(
|
38 |
+
top_directory, "FP16-Chat", "LlamaV2_13B_FT_float32.onnx"
|
39 |
+
),
|
40 |
+
"tokenizer_path": tokenizer_path,
|
41 |
+
"embedding_file": os.path.join(top_directory, "embeddings-chat.pth"),
|
42 |
+
},
|
43 |
+
"Llama-2 Chat 13B Float32": {
|
44 |
+
"onnx_file": os.path.join(
|
45 |
+
top_directory, "FP32-Chat", "LlamaV2_13B_FT_float32.onnx"
|
46 |
+
),
|
47 |
+
"tokenizer_path": tokenizer_path,
|
48 |
+
"embedding_file": os.path.join(
|
49 |
+
top_directory, "embeddings-chat.pth"
|
50 |
+
),
|
51 |
+
},
|
52 |
"Llama-2 13B Float16": {
|
53 |
"onnx_file": os.path.join(
|
54 |
top_directory, "FP16", "LlamaV2_13B_float16.onnx"
|
|
|
56 |
"tokenizer_path": tokenizer_path,
|
57 |
"embedding_file": os.path.join(top_directory, "embeddings.pth"),
|
58 |
},
|
59 |
+
"Llama-2 13B Float32": {
|
60 |
"onnx_file": os.path.join(
|
61 |
top_directory, "FP32", "LlamaV2_13B_float16.onnx"
|
62 |
),
|