Spaces:
Running
on
L40S
Running
on
L40S
Merge branch 'main' of hf.co:spaces/THUDM/CogVideoX-5B
Browse files
app.py
CHANGED
@@ -102,7 +102,7 @@ def infer(
|
|
102 |
num_inference_steps: int,
|
103 |
guidance_scale: float,
|
104 |
seed: int = -1,
|
105 |
-
progress=gr.Progress(track_tqdm=True),
|
106 |
):
|
107 |
if seed == -1:
|
108 |
seed = random.randint(0, 2 ** 8 - 1)
|
@@ -263,14 +263,19 @@ with gr.Blocks() as demo:
|
|
263 |
""")
|
264 |
|
265 |
|
266 |
-
def generate(prompt,
|
|
|
|
|
|
|
|
|
|
|
267 |
|
268 |
latents, seed = infer(
|
269 |
prompt,
|
270 |
num_inference_steps=50, # NOT Changed
|
271 |
guidance_scale=7, # NOT Changed
|
272 |
seed=seed_value,
|
273 |
-
progress=progress,
|
274 |
)
|
275 |
if scale_status:
|
276 |
latents = utils.upscale_batch_and_concatenate(upscale_model, latents, device)
|
|
|
102 |
num_inference_steps: int,
|
103 |
guidance_scale: float,
|
104 |
seed: int = -1,
|
105 |
+
#progress=gr.Progress(track_tqdm=True),
|
106 |
):
|
107 |
if seed == -1:
|
108 |
seed = random.randint(0, 2 ** 8 - 1)
|
|
|
263 |
""")
|
264 |
|
265 |
|
266 |
+
def generate(prompt,
|
267 |
+
seed_value,
|
268 |
+
scale_status,
|
269 |
+
rife_status,
|
270 |
+
#progress=gr.Progress(track_tqdm=True)
|
271 |
+
):
|
272 |
|
273 |
latents, seed = infer(
|
274 |
prompt,
|
275 |
num_inference_steps=50, # NOT Changed
|
276 |
guidance_scale=7, # NOT Changed
|
277 |
seed=seed_value,
|
278 |
+
#progress=progress,
|
279 |
)
|
280 |
if scale_status:
|
281 |
latents = utils.upscale_batch_and_concatenate(upscale_model, latents, device)
|