Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -352,7 +352,7 @@ def validate_and_convert_image(image, target_size=(512 , 512)):
|
|
352 |
return image
|
353 |
|
354 |
|
355 |
-
class
|
356 |
|
357 |
def __init__(self, device, args, height, width, model_length, dtype=torch.float16, use_sift=False):
|
358 |
self.device = device
|
@@ -547,7 +547,7 @@ with gr.Blocks() as demo:
|
|
547 |
3. Interpolate the images (according the path) with a click on "Run" button. <br>""")
|
548 |
|
549 |
# device, args, height, width, model_length
|
550 |
-
Framer =
|
551 |
first_frame_path = gr.State()
|
552 |
last_frame_path = gr.State()
|
553 |
tracking_points = gr.State([])
|
|
|
352 |
return image
|
353 |
|
354 |
|
355 |
+
class Drag:
|
356 |
|
357 |
def __init__(self, device, args, height, width, model_length, dtype=torch.float16, use_sift=False):
|
358 |
self.device = device
|
|
|
547 |
3. Interpolate the images (according the path) with a click on "Run" button. <br>""")
|
548 |
|
549 |
# device, args, height, width, model_length
|
550 |
+
Framer = Drag("cuda", args, 320, 512, 14)
|
551 |
first_frame_path = gr.State()
|
552 |
last_frame_path = gr.State()
|
553 |
tracking_points = gr.State([])
|