Update app.py
Browse files
app.py
CHANGED
@@ -6,12 +6,7 @@ torch.jit.script = lambda f: f
|
|
6 |
import timm
|
7 |
|
8 |
|
9 |
-
|
10 |
-
# cwd0 = os.getcwd()
|
11 |
-
# cwd1 = os.path.join(cwd0, "yolo-world-with-efficientvit-sam")
|
12 |
-
# os.chdir(cwd1)
|
13 |
-
# os.system("make setup")
|
14 |
-
# os.system(f"cd /home/user/app")
|
15 |
|
16 |
sys.path.append('./')
|
17 |
import gradio as gr
|
@@ -30,7 +25,7 @@ import cv2
|
|
30 |
from PIL import Image, ImageOps
|
31 |
from transformers import DPTFeatureExtractor, DPTForDepthEstimation
|
32 |
from controlnet_aux import OpenposeDetector
|
33 |
-
|
34 |
|
35 |
|
36 |
from inference.models import YOLOWorld
|
@@ -497,16 +492,76 @@ def main(device, segment_type):
|
|
497 |
def get_local_value_woman(input):
|
498 |
return character_woman[input][0]
|
499 |
|
500 |
-
|
501 |
-
|
502 |
-
|
503 |
-
|
504 |
-
|
505 |
-
|
506 |
-
|
507 |
-
|
508 |
-
|
509 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
510 |
|
511 |
|
512 |
|
|
|
6 |
import timm
|
7 |
|
8 |
|
9 |
+
|
|
|
|
|
|
|
|
|
|
|
10 |
|
11 |
sys.path.append('./')
|
12 |
import gradio as gr
|
|
|
25 |
from PIL import Image, ImageOps
|
26 |
from transformers import DPTFeatureExtractor, DPTForDepthEstimation
|
27 |
from controlnet_aux import OpenposeDetector
|
28 |
+
from controlnet_aux.open_pose.body import Body
|
29 |
|
30 |
|
31 |
from inference.models import YOLOWorld
|
|
|
492 |
def get_local_value_woman(input):
|
493 |
return character_woman[input][0]
|
494 |
|
495 |
+
with gr.Blocks(css=css) as demo:
|
496 |
+
# description
|
497 |
+
gr.Markdown(title)
|
498 |
+
gr.Markdown(description)
|
499 |
+
|
500 |
+
with gr.Row():
|
501 |
+
gallery = gr.Image(label="Generated Images", height=512, width=512)
|
502 |
+
gen_condition = gr.Image(label="Spatial Condition", height=512, width=512)
|
503 |
+
usage_tips = gr.Markdown(label="Usage tips of OMG", value=tips, visible=False)
|
504 |
+
|
505 |
+
with gr.Row():
|
506 |
+
condition_img1 = gr.Image(label="Input an RGB image for condition", height=128, width=128)
|
507 |
+
|
508 |
+
# character choose
|
509 |
+
with gr.Row():
|
510 |
+
man = gr.Dropdown(label="Character 1 selection", choices=CHARACTER_MAN_NAMES, value="Chris Evans (identifier: Chris Evans)")
|
511 |
+
woman = gr.Dropdown(label="Character 2 selection", choices=CHARACTER_WOMAN_NAMES, value="Taylor Swift (identifier: TaylorSwift)")
|
512 |
+
resolution = gr.Dropdown(label="Image Resolution (width*height)", choices=resolution_list, value="1024*1024")
|
513 |
+
condition = gr.Dropdown(label="Input condition type", choices=condition_list, value="None")
|
514 |
+
style = gr.Dropdown(label="style", choices=STYLE_NAMES, value="None")
|
515 |
+
|
516 |
+
with gr.Row():
|
517 |
+
local_prompt1 = gr.Textbox(label="Character1_prompt",
|
518 |
+
info="Describe the Character 1, this prompt should include the identifier of character 1",
|
519 |
+
value="Close-up photo of the Chris Evans, 35mm photograph, film, professional, 4k, highly detailed.")
|
520 |
+
local_prompt2 = gr.Textbox(label="Character2_prompt",
|
521 |
+
info="Describe the Character 2, this prompt should include the identifier of character2",
|
522 |
+
value="Close-up photo of the TaylorSwift, 35mm photograph, film, professional, 4k, highly detailed.")
|
523 |
+
|
524 |
+
man.change(get_local_value_man, man, local_prompt1)
|
525 |
+
woman.change(get_local_value_woman, woman, local_prompt2)
|
526 |
+
|
527 |
+
# prompt
|
528 |
+
with gr.Column():
|
529 |
+
prompt = gr.Textbox(label="Prompt 1",
|
530 |
+
info="Give a simple prompt to describe the first image content",
|
531 |
+
placeholder="Required",
|
532 |
+
value="close-up shot, photography, a man and a woman on the street, facing the camera smiling")
|
533 |
+
|
534 |
+
|
535 |
+
with gr.Accordion(open=False, label="Advanced Options"):
|
536 |
+
seed = gr.Slider(
|
537 |
+
label="Seed",
|
538 |
+
minimum=0,
|
539 |
+
maximum=MAX_SEED,
|
540 |
+
step=1,
|
541 |
+
value=42,
|
542 |
+
)
|
543 |
+
negative_prompt = gr.Textbox(label="Negative Prompt",
|
544 |
+
placeholder="noisy, blurry, soft, deformed, ugly",
|
545 |
+
value="noisy, blurry, soft, deformed, ugly")
|
546 |
+
randomize_seed = gr.Checkbox(label="Randomize seed", value=True)
|
547 |
+
|
548 |
+
submit = gr.Button("Submit", variant="primary")
|
549 |
+
|
550 |
+
submit.click(
|
551 |
+
fn=remove_tips,
|
552 |
+
outputs=usage_tips,
|
553 |
+
).then(
|
554 |
+
fn=randomize_seed_fn,
|
555 |
+
inputs=[seed, randomize_seed],
|
556 |
+
outputs=seed,
|
557 |
+
queue=False,
|
558 |
+
api_name=False,
|
559 |
+
).then(
|
560 |
+
fn=generate_image,
|
561 |
+
inputs=[prompt, negative_prompt, man, woman, resolution, local_prompt1, local_prompt2, seed, condition, condition_img1, style],
|
562 |
+
outputs=[gallery, gen_condition]
|
563 |
+
)
|
564 |
+
demo.launch(share=True)
|
565 |
|
566 |
|
567 |
|