File size: 1,047 Bytes
5ece8c2 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 |
{
"api_key": null,
"verify_url": "http://johnrachwan.pythonanywhere.com",
"smash_config": {
"pruners": "[]",
"pruning_ratio": 0.0,
"factorizers": "[]",
"quantizers": "['gptq']",
"n_quantization_bits": 4,
"output_deviation": 0.005,
"compilers": "[]",
"static_batch": true,
"static_shape": true,
"controlnet": "None",
"unet_dim": 4,
"device": "cuda",
"cache_dir": "/ceph/hdd/staff/charpent/.cache/modelsz75ld54d",
"batch_size": 1,
"model_name": "mistralai/Mistral-7B-Instruct-v0.2",
"max_batch_size": 1,
"save_dir": "/ceph/hdd/staff/charpent/.cache/models1ggtg41i",
"qtype_weight": "torch.qint8",
"qtype_activation": "torch.quint8",
"qobserver": "<class 'torch.ao.quantization.observer.MinMaxObserver'>",
"qscheme": "torch.per_tensor_symmetric",
"qconfig": "x86",
"group_size": 128,
"damp_percent": 0.1,
"save_load_fn": "hf-gptq"
}
} |