Spaces:
Running
on
Zero
Running
on
Zero
minor bug fix
Browse files
app.py
CHANGED
@@ -61,7 +61,7 @@ unet = UNet2DConditionModel.from_config(model_repo_id, subfolder="unet").to(devi
|
|
61 |
unet.load_state_dict(torch.load(hf_hub_download(repo_name, ckpt_name)))
|
62 |
pipe = DiffusionPipeline.from_pretrained(model_repo_id, unet=unet, torch_dtype=torch_dtype).to(device)
|
63 |
pipe.scheduler = LCMScheduler.from_config(pipe.scheduler.config)
|
64 |
-
pipe.vae = AutoencoderTiny.from_pretrained("madebyollin/taesdxl", torch_dtype=torch_dtype).to(
|
65 |
|
66 |
unet = pipe.unet
|
67 |
|
@@ -634,7 +634,7 @@ if __name__ == "__main__":
|
|
634 |
# value=4, # Replace with defaults that work for your model
|
635 |
# )
|
636 |
|
637 |
-
# # gr.
|
638 |
# gr.on(
|
639 |
# triggers=[run_button.click, prompt.submit],
|
640 |
# fn=infer,
|
|
|
61 |
unet.load_state_dict(torch.load(hf_hub_download(repo_name, ckpt_name)))
|
62 |
pipe = DiffusionPipeline.from_pretrained(model_repo_id, unet=unet, torch_dtype=torch_dtype).to(device)
|
63 |
pipe.scheduler = LCMScheduler.from_config(pipe.scheduler.config)
|
64 |
+
pipe.vae = AutoencoderTiny.from_pretrained("madebyollin/taesdxl", torch_dtype=torch_dtype).to(device)
|
65 |
|
66 |
unet = pipe.unet
|
67 |
|
|
|
634 |
# value=4, # Replace with defaults that work for your model
|
635 |
# )
|
636 |
|
637 |
+
# # gr.minoExamples(examples=examples, inputs=[prompt])
|
638 |
# gr.on(
|
639 |
# triggers=[run_button.click, prompt.submit],
|
640 |
# fn=infer,
|