Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
amd-shark
/
sdxl-quant-int8
like
1
Follow
AMD SHARK
17
Transformers
Inference Endpoints
Model card
Files
Files and versions
Community
2
Train
Deploy
Use this model
5d000b6
sdxl-quant-int8
4 contributors
History:
43 commits
GiusFra
QKV fused and all linear layers sym
5d000b6
verified
4 months ago
all_linear_sym
QKV fused and all linear layers sym
4 months ago
full_sym
Full symmetric
4 months ago
fused_qkv
Fused QKV quant_params.json with zp
5 months ago
qkv_sym
QKV fused and sym
4 months ago
quant_sdxl
Fix model loading
5 months ago
.gitattributes
Safe
1.63 kB
Updated quant_params
5 months ago
config.json
Safe
1.68 kB
Add config.json from stable-diffusion-xl-base-1.0/unet
6 months ago
math_model.py
Safe
9.05 kB
Update quant params structure (#2)
5 months ago
out.safetensors
Safe
7.11 GB
LFS
Output reference tensors
5 months ago
params.safetensors
Safe
5.14 GB
LFS
Updated params.safetensors
5 months ago
punet_inputs.safetensors
Safe
661 kB
LFS
Reference inputs
5 months ago
quant_param.json
Safe
85.1 MB
LFS
add missing smoothquant factors
6 months ago
quant_params.json
Safe
86.8 MB
LFS
Updated quant_params
5 months ago
test_quant_conv2d.py
Safe
1.18 kB
Update quant params structure (#2)
5 months ago
test_quant_linear.py
Safe
1.04 kB
Update quant params structure (#2)
5 months ago
vae.safetensors
Safe
167 MB
LFS
Added vae weights with FP16 fix.
5 months ago