imnotednamode commited on
Commit
e18ff58
·
verified ·
1 Parent(s): 63b8d5d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -4,6 +4,8 @@ base_model:
4
  - genmo/mochi-1-preview
5
  ---
6
 
 
 
7
  This is a nf4 quantization of mochi-1-preview. It excludes specific layers that allow it to stay coherent:
8
 
9
  bf16
@@ -27,8 +29,8 @@ To use:
27
  ```
28
  from diffusers import MochiPipeline, MochiTransformer3DModel
29
  from diffusers.utils import export_to_video
30
- transformer = MochiTransformer3DModel.from_pretrained("imnotednamode/mochi-1-preview-mix-nf4-small", torch_dtype=torch.bfloat16)
31
- pipe = MochiPipeline.from_pretrained("genmo/mochi-1-preview", torch_dtype=torch.bfloat16, transformer=transformer)
32
  pipe.enable_model_cpu_offload()
33
  pipe.enable_vae_tiling()
34
  frames = pipe("A camera follows a squirrel running around on a tree branch", num_inference_steps=100, guidance_scale=4.5, height=480, width=848, num_frames=161).frames[0]
@@ -41,8 +43,8 @@ To reproduce:
41
  ```
42
  from diffusers import MochiPipeline, MochiTransformer3DModel, BitsAndBytesConfig
43
  import torch
44
- quantization_config = BitsAndBytesConfig(load_in_4bit=True, bnb_4bit_compute_dtype=torch.bfloat16, bnb_4bit_quant_type="nf4", llm_int8_skip_modules=["final_layer", "x_embedder.proj", "t_embedder", "pos_frequencies", "t5"])
45
  # Please convert mochi to diffusers first
46
- transformer = MochiTransformer3DModel.from_pretrained("genmo/mochi-1-preview", variant="refs/pr/18", subfolder="transformer", quantization_config=quantization_config, torch_dtype=torch.bfloat16)
47
  transformer.save_pretrained("mochi-1-preview-nf4")
48
  ```
 
4
  - genmo/mochi-1-preview
5
  ---
6
 
7
+ **Please change bf16 to fp16 in your code. I've just changed the checkpoint to fp16, which appears more coherent with mochi**
8
+
9
  This is a nf4 quantization of mochi-1-preview. It excludes specific layers that allow it to stay coherent:
10
 
11
  bf16
 
29
  ```
30
  from diffusers import MochiPipeline, MochiTransformer3DModel
31
  from diffusers.utils import export_to_video
32
+ transformer = MochiTransformer3DModel.from_pretrained("imnotednamode/mochi-1-preview-mix-nf4-small", torch_dtype=torch.float16)
33
+ pipe = MochiPipeline.from_pretrained("genmo/mochi-1-preview", torch_dtype=torch.float16, transformer=transformer)
34
  pipe.enable_model_cpu_offload()
35
  pipe.enable_vae_tiling()
36
  frames = pipe("A camera follows a squirrel running around on a tree branch", num_inference_steps=100, guidance_scale=4.5, height=480, width=848, num_frames=161).frames[0]
 
43
  ```
44
  from diffusers import MochiPipeline, MochiTransformer3DModel, BitsAndBytesConfig
45
  import torch
46
+ quantization_config = BitsAndBytesConfig(load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_quant_type="nf4", llm_int8_skip_modules=["final_layer", "x_embedder.proj", "t_embedder", "pos_frequencies", "t5"])
47
  # Please convert mochi to diffusers first
48
+ transformer = MochiTransformer3DModel.from_pretrained("genmo/mochi-1-preview", variant="refs/pr/18", subfolder="transformer", quantization_config=quantization_config, torch_dtype=torch.float16)
49
  transformer.save_pretrained("mochi-1-preview-nf4")
50
  ```