lvkaokao commited on
Commit
6c947f9
1 Parent(s): 7c54eb3

Add Intel/gemma-2b-int4-inc to eval queue

Browse files
Intel/gemma-2b-int4-inc_eval_request_False_GPTQ_4bit_int4_float16.json CHANGED
@@ -1 +1 @@
1
- {"model": "Intel/gemma-2b-int4-inc", "revision": "main", "private": false, "params": 3.13, "architectures": "GemmaForCausalLM", "quant_type": "GPTQ", "precision": "4bit", "model_params": 2.0, "model_size": 3.13, "weight_dtype": "int4", "compute_dtype": "float16", "gguf_ftype": "*Q4_0.gguf", "hardware": "gpu", "status": "Pending", "submitted_time": "2024-05-28T15:46:37Z", "model_type": "quantization", "job_id": -1, "job_start_time": null, "scripts": "ITREX"}
 
1
+ {"model": "Intel/gemma-2b-int4-inc", "revision": "main", "private": false, "params": 3.13, "architectures": "GemmaForCausalLM", "quant_type": "GPTQ", "precision": "4bit", "model_params": 2.0, "model_size": 3.13, "weight_dtype": "int4", "compute_dtype": "float16", "gguf_ftype": "*Q4_0.gguf", "hardware": "gpu", "status": "Pending", "submitted_time": "2024-05-28T15:48:10Z", "model_type": "quantization", "job_id": -1, "job_start_time": null, "scripts": "ITREX"}