Commit
·
ae24367
1
Parent(s):
3ea6729
Unique keys
Browse files
app.py
CHANGED
@@ -86,7 +86,7 @@ with gr.Blocks() as mindseye:
|
|
86 |
width_vq = gr.inputs.Slider(label="Width", default=256, minimum=32, step=32, maximum=512)
|
87 |
height_vq= gr.inputs.Slider(label="Height", default=256, minimum=32, step=32, maximum=512)
|
88 |
style = gr.inputs.Dropdown(label="Style - Hyper Fast Results is fast but compromises a bit of the quality",choices=["Default","Balanced","Detailed","Consistent Creativity","Realistic","Smooth","Subtle MSE","Hyper Fast Results"],default="Hyper Fast Results")
|
89 |
-
|
90 |
flavor = gr.inputs.Dropdown(label="Flavor - pick a flavor for the style of the images, based on the images below",choices=["ginger", "cumin", "holywater", "zynth", "wyvern", "aaron", "moth", "juu"])
|
91 |
get_image_vqgan = gr.Button("Generate Image",css=css_mt)
|
92 |
with gr.TabItem("Guided Diffusion"):
|
@@ -105,6 +105,6 @@ with gr.Blocks() as mindseye:
|
|
105 |
|
106 |
get_image_latent.click(text2image_latent, inputs=[text,steps,width,height,images,diversity], outputs=gallery)
|
107 |
get_image_rudalle.click(text2image_rudalle, inputs=[text,aspect,model], outputs=gallery)
|
108 |
-
get_image_vqgan.click(text2image_vqgan, inputs=[text,width_vq,height_vq,style,
|
109 |
get_image_diffusion.click(text2image_diffusion, inputs=[text, steps_diff, images_diff, weight, clip],outputs=gallery)
|
110 |
mindseye.launch()
|
|
|
86 |
width_vq = gr.inputs.Slider(label="Width", default=256, minimum=32, step=32, maximum=512)
|
87 |
height_vq= gr.inputs.Slider(label="Height", default=256, minimum=32, step=32, maximum=512)
|
88 |
style = gr.inputs.Dropdown(label="Style - Hyper Fast Results is fast but compromises a bit of the quality",choices=["Default","Balanced","Detailed","Consistent Creativity","Realistic","Smooth","Subtle MSE","Hyper Fast Results"],default="Hyper Fast Results")
|
89 |
+
steps_vq = gr.inputs.Slider(label="Steps - more steps can increase quality but will take longer to generate. All styles that are not Hyper Fast need at least 200 steps",default=50,maximum=300,minimum=1,step=1)
|
90 |
flavor = gr.inputs.Dropdown(label="Flavor - pick a flavor for the style of the images, based on the images below",choices=["ginger", "cumin", "holywater", "zynth", "wyvern", "aaron", "moth", "juu"])
|
91 |
get_image_vqgan = gr.Button("Generate Image",css=css_mt)
|
92 |
with gr.TabItem("Guided Diffusion"):
|
|
|
105 |
|
106 |
get_image_latent.click(text2image_latent, inputs=[text,steps,width,height,images,diversity], outputs=gallery)
|
107 |
get_image_rudalle.click(text2image_rudalle, inputs=[text,aspect,model], outputs=gallery)
|
108 |
+
get_image_vqgan.click(text2image_vqgan, inputs=[text,width_vq,height_vq,style,steps_vq,flavor],outputs=gallery)
|
109 |
get_image_diffusion.click(text2image_diffusion, inputs=[text, steps_diff, images_diff, weight, clip],outputs=gallery)
|
110 |
mindseye.launch()
|