Update README.md
Browse files
README.md
CHANGED
@@ -4,6 +4,8 @@ base_model:
|
|
4 |
- genmo/mochi-1-preview
|
5 |
---
|
6 |
|
|
|
|
|
7 |
This is a nf4 quantization of mochi-1-preview. It excludes specific layers that allow it to stay coherent:
|
8 |
|
9 |
bf16
|
@@ -27,8 +29,8 @@ To use:
|
|
27 |
```
|
28 |
from diffusers import MochiPipeline, MochiTransformer3DModel
|
29 |
from diffusers.utils import export_to_video
|
30 |
-
transformer = MochiTransformer3DModel.from_pretrained("imnotednamode/mochi-1-preview-mix-nf4-small", torch_dtype=torch.
|
31 |
-
pipe = MochiPipeline.from_pretrained("genmo/mochi-1-preview", torch_dtype=torch.
|
32 |
pipe.enable_model_cpu_offload()
|
33 |
pipe.enable_vae_tiling()
|
34 |
frames = pipe("A camera follows a squirrel running around on a tree branch", num_inference_steps=100, guidance_scale=4.5, height=480, width=848, num_frames=161).frames[0]
|
@@ -41,8 +43,8 @@ To reproduce:
|
|
41 |
```
|
42 |
from diffusers import MochiPipeline, MochiTransformer3DModel, BitsAndBytesConfig
|
43 |
import torch
|
44 |
-
quantization_config = BitsAndBytesConfig(load_in_4bit=True, bnb_4bit_compute_dtype=torch.
|
45 |
# Please convert mochi to diffusers first
|
46 |
-
transformer = MochiTransformer3DModel.from_pretrained("genmo/mochi-1-preview", variant="refs/pr/18", subfolder="transformer", quantization_config=quantization_config, torch_dtype=torch.
|
47 |
transformer.save_pretrained("mochi-1-preview-nf4")
|
48 |
```
|
|
|
4 |
- genmo/mochi-1-preview
|
5 |
---
|
6 |
|
7 |
+
**Please change bf16 to fp16 in your code. I've just changed the checkpoint to fp16, which appears more coherent with mochi**
|
8 |
+
|
9 |
This is a nf4 quantization of mochi-1-preview. It excludes specific layers that allow it to stay coherent:
|
10 |
|
11 |
bf16
|
|
|
29 |
```
|
30 |
from diffusers import MochiPipeline, MochiTransformer3DModel
|
31 |
from diffusers.utils import export_to_video
|
32 |
+
transformer = MochiTransformer3DModel.from_pretrained("imnotednamode/mochi-1-preview-mix-nf4-small", torch_dtype=torch.float16)
|
33 |
+
pipe = MochiPipeline.from_pretrained("genmo/mochi-1-preview", torch_dtype=torch.float16, transformer=transformer)
|
34 |
pipe.enable_model_cpu_offload()
|
35 |
pipe.enable_vae_tiling()
|
36 |
frames = pipe("A camera follows a squirrel running around on a tree branch", num_inference_steps=100, guidance_scale=4.5, height=480, width=848, num_frames=161).frames[0]
|
|
|
43 |
```
|
44 |
from diffusers import MochiPipeline, MochiTransformer3DModel, BitsAndBytesConfig
|
45 |
import torch
|
46 |
+
quantization_config = BitsAndBytesConfig(load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_quant_type="nf4", llm_int8_skip_modules=["final_layer", "x_embedder.proj", "t_embedder", "pos_frequencies", "t5"])
|
47 |
# Please convert mochi to diffusers first
|
48 |
+
transformer = MochiTransformer3DModel.from_pretrained("genmo/mochi-1-preview", variant="refs/pr/18", subfolder="transformer", quantization_config=quantization_config, torch_dtype=torch.float16)
|
49 |
transformer.save_pretrained("mochi-1-preview-nf4")
|
50 |
```
|