eienmojiki commited on
Commit
4ff938d
·
verified ·
1 Parent(s): 25bfa77

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -3
app.py CHANGED
@@ -61,7 +61,7 @@ def get_scheduler(scheduler_config: Dict, name: str) -> Optional[Callable]:
61
  }
62
  return scheduler_factory_map.get(name, lambda: None)()
63
 
64
- @spaces.GPU
65
  def generate(
66
  prompt: str,
67
  negative_prompt: str = None,
@@ -69,7 +69,7 @@ def generate(
69
  width: int = 1024,
70
  height: int = 1024,
71
  guidance_scale: float = 5.0,
72
- num_inference_steps: int = 24,
73
  sampler: str = "Euler a",
74
  clip_skip: int = 1,
75
  progress=gr.Progress(track_tqdm=True),
@@ -84,7 +84,6 @@ def generate(
84
  add_watermarker=False,
85
  use_auth_token=HF_TOKEN
86
  )
87
- pipe.to(device)
88
 
89
  generator = seed_everything(seed)
90
  pipe.scheduler = get_scheduler(pipe.scheduler.config, sampler)
@@ -94,6 +93,8 @@ def generate(
94
  num_hidden_layers = 12 - (clip_skip - 1),
95
  torch_dtype = torch.float16
96
  )
 
 
97
 
98
  try:
99
 
 
61
  }
62
  return scheduler_factory_map.get(name, lambda: None)()
63
 
64
+ @spaces.GPU(enable_queue=False)
65
  def generate(
66
  prompt: str,
67
  negative_prompt: str = None,
 
69
  width: int = 1024,
70
  height: int = 1024,
71
  guidance_scale: float = 5.0,
72
+ num_inference_steps: int = 26,
73
  sampler: str = "Euler a",
74
  clip_skip: int = 1,
75
  progress=gr.Progress(track_tqdm=True),
 
84
  add_watermarker=False,
85
  use_auth_token=HF_TOKEN
86
  )
 
87
 
88
  generator = seed_everything(seed)
89
  pipe.scheduler = get_scheduler(pipe.scheduler.config, sampler)
 
93
  num_hidden_layers = 12 - (clip_skip - 1),
94
  torch_dtype = torch.float16
95
  )
96
+
97
+ pipe.to(device)
98
 
99
  try:
100