Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -5,7 +5,7 @@ import random
|
|
5 |
import spaces #[uncomment to use ZeroGPU]
|
6 |
#from diffusers import DiffusionPipeline ,AutoencoderTiny
|
7 |
import torch
|
8 |
-
from diffusers import AutoencoderTiny, StableDiffusionPipeline
|
9 |
from huggingface_hub import login
|
10 |
import os
|
11 |
a=os.getenv('hf_key')
|
@@ -61,7 +61,7 @@ pipe = StableDiffusionPipeline.from_pretrained(
|
|
61 |
model_repo_id, torch_dtype=torch_dtype, use_safetensors=True)
|
62 |
#pipe.vae = AutoencoderTiny.from_pretrained(
|
63 |
# "sayakpaul/taesd-diffusers", torch_dtype=torch_dtype, use_safetensors=True)
|
64 |
-
pipe.scheduler =
|
65 |
print(pipe.scheduler.compatibles)
|
66 |
pipe = pipe.to(device)
|
67 |
pipe.enable_vae_tiling()
|
|
|
5 |
import spaces #[uncomment to use ZeroGPU]
|
6 |
#from diffusers import DiffusionPipeline ,AutoencoderTiny
|
7 |
import torch
|
8 |
+
from diffusers import AutoencoderTiny, StableDiffusionPipeline , DPMSolverMultistepScheduler
|
9 |
from huggingface_hub import login
|
10 |
import os
|
11 |
a=os.getenv('hf_key')
|
|
|
61 |
model_repo_id, torch_dtype=torch_dtype, use_safetensors=True)
|
62 |
#pipe.vae = AutoencoderTiny.from_pretrained(
|
63 |
# "sayakpaul/taesd-diffusers", torch_dtype=torch_dtype, use_safetensors=True)
|
64 |
+
pipe.scheduler = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config)
|
65 |
print(pipe.scheduler.compatibles)
|
66 |
pipe = pipe.to(device)
|
67 |
pipe.enable_vae_tiling()
|