Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -355,8 +355,6 @@ class XFluxPipeline:
|
|
355 |
torch.cuda.empty_cache()
|
356 |
|
357 |
|
358 |
-
|
359 |
-
|
360 |
def create_demo(
|
361 |
model_type: str,
|
362 |
device: str = "cuda" if torch.cuda.is_available() else "cpu",
|
@@ -367,7 +365,7 @@ def create_demo(
|
|
367 |
checkpoints = sorted(Path(ckpt_dir).glob("*.safetensors"))
|
368 |
|
369 |
with gr.Blocks() as demo:
|
370 |
-
gr.Markdown("⚠️ Warning: Gradio is not functioning correctly. We are looking for someone to help fix it by submitting a Pull Request.")
|
371 |
gr.Markdown(f"# Flux Adapters by XLabs AI - Model: {model_type}")
|
372 |
with gr.Row():
|
373 |
with gr.Column():
|
|
|
355 |
torch.cuda.empty_cache()
|
356 |
|
357 |
|
|
|
|
|
358 |
def create_demo(
|
359 |
model_type: str,
|
360 |
device: str = "cuda" if torch.cuda.is_available() else "cpu",
|
|
|
365 |
checkpoints = sorted(Path(ckpt_dir).glob("*.safetensors"))
|
366 |
|
367 |
with gr.Blocks() as demo:
|
368 |
+
gr.Markdown(f"### ⚠️ Warning: Gradio is not functioning correctly. We are looking for someone to help fix it by submitting a Pull Request.")
|
369 |
gr.Markdown(f"# Flux Adapters by XLabs AI - Model: {model_type}")
|
370 |
with gr.Row():
|
371 |
with gr.Column():
|