|
--- |
|
license: apache-2.0 |
|
--- |
|
This is [fal/AuraFlow](https://huggingface.co/fal/AuraFlow), converted from FP16 to 8-bit. Small or one-dimensional tensors are left in FP16 so as to avoid severe degredation. |
|
|
|
This is an experimental conversion and I don't have currently have enough memory to run this locally, so it is not guaranteed to work. Please let me know if it works or not. |
|
|
|
**In order to actually save memory usage you will need to prevent your inference engine from upcasting to FP16 during calculation.** |
|
|
|
The code used to convert the model is in the repo as well. |
|
|
|
Enjoy! |