Upload ort_config.json with huggingface_hub
Browse files- ort_config.json +3 -3
ort_config.json
CHANGED
@@ -23,16 +23,16 @@
|
|
23 |
"enable_gelu_approximation": true,
|
24 |
"enable_gemm_fast_gelu_fusion": false,
|
25 |
"enable_transformers_specific_optimizations": true,
|
26 |
-
"fp16":
|
27 |
"no_attention_mask": false,
|
28 |
"optimization_level": 2,
|
29 |
-
"optimize_for_gpu":
|
30 |
"optimize_with_onnxruntime_only": null,
|
31 |
"use_mask_index": false,
|
32 |
"use_multi_head_attention": false,
|
33 |
"use_raw_attention_mask": false
|
34 |
},
|
35 |
-
"optimum_version": "1.
|
36 |
"quantization": {},
|
37 |
"transformers_version": "4.36.2",
|
38 |
"use_external_data_format": false
|
|
|
23 |
"enable_gelu_approximation": true,
|
24 |
"enable_gemm_fast_gelu_fusion": false,
|
25 |
"enable_transformers_specific_optimizations": true,
|
26 |
+
"fp16": true,
|
27 |
"no_attention_mask": false,
|
28 |
"optimization_level": 2,
|
29 |
+
"optimize_for_gpu": true,
|
30 |
"optimize_with_onnxruntime_only": null,
|
31 |
"use_mask_index": false,
|
32 |
"use_multi_head_attention": false,
|
33 |
"use_raw_attention_mask": false
|
34 |
},
|
35 |
+
"optimum_version": "1.15.0",
|
36 |
"quantization": {},
|
37 |
"transformers_version": "4.36.2",
|
38 |
"use_external_data_format": false
|