pcuenq HF staff commited on
Commit
277a2c1
1 Parent(s): 77ce6d2

remove cpu offload

Browse files

It's not used in most benchmarks. And when it is, we don't have to move the pipeline to `cuda`.

Files changed (1) hide show
  1. sayak_lcm_benchmark.py +0 -1
sayak_lcm_benchmark.py CHANGED
@@ -22,7 +22,6 @@ def load_pipeline(standard_sdxl=False):
22
  pipe.load_lora_weights(LORA_ID)
23
 
24
  pipe.to(device="cuda", dtype=torch.float16)
25
- pipe.enable_model_cpu_offload()
26
  return pipe
27
 
28
 
 
22
  pipe.load_lora_weights(LORA_ID)
23
 
24
  pipe.to(device="cuda", dtype=torch.float16)
 
25
  return pipe
26
 
27