nielsr HF staff commited on
Commit
800cbf3
1 Parent(s): 07137b2

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -1
app.py CHANGED
@@ -4,6 +4,7 @@ import torch
4
 
5
  torch.hub.download_url_to_file('http://images.cocodataset.org/val2017/000000039769.jpg', 'cats.jpg')
6
  torch.hub.download_url_to_file('https://huggingface.co/datasets/nielsr/textcaps-sample/resolve/main/stop_sign.png', 'stop_sign.png')
 
7
 
8
  git_processor_base = AutoProcessor.from_pretrained("microsoft/git-base-coco")
9
  git_model_base = AutoModelForCausalLM.from_pretrained("microsoft/git-base-coco")
@@ -56,7 +57,7 @@ def generate_captions(image):
56
  return caption_git_base, caption_git_large, caption_blip_base, caption_blip_large, caption_vitgpt
57
 
58
 
59
- examples = [["cats.jpg"], ["stop_sign.png"]]
60
  outputs = [gr.outputs.Textbox(label="Caption generated by GIT-base"), gr.outputs.Textbox(label="Caption generated by GIT-large"), gr.outputs.Textbox(label="Caption generated by BLIP-base"), gr.outputs.Textbox(label="Caption generated by BLIP-large"), gr.outputs.Textbox(label="Caption generated by ViT+GPT-2")]
61
 
62
  title = "Interactive demo: comparing image captioning models"
 
4
 
5
  torch.hub.download_url_to_file('http://images.cocodataset.org/val2017/000000039769.jpg', 'cats.jpg')
6
  torch.hub.download_url_to_file('https://huggingface.co/datasets/nielsr/textcaps-sample/resolve/main/stop_sign.png', 'stop_sign.png')
7
+ torch.hub.download_url_to_file('https://cdn.openai.com/dall-e-2/demos/text2im/astronaut/horse/photo/0.jpg', 'astronaut.jpg')
8
 
9
  git_processor_base = AutoProcessor.from_pretrained("microsoft/git-base-coco")
10
  git_model_base = AutoModelForCausalLM.from_pretrained("microsoft/git-base-coco")
 
57
  return caption_git_base, caption_git_large, caption_blip_base, caption_blip_large, caption_vitgpt
58
 
59
 
60
+ examples = [["cats.jpg"], ["stop_sign.png"], ["astronaut.jpg"]]
61
  outputs = [gr.outputs.Textbox(label="Caption generated by GIT-base"), gr.outputs.Textbox(label="Caption generated by GIT-large"), gr.outputs.Textbox(label="Caption generated by BLIP-base"), gr.outputs.Textbox(label="Caption generated by BLIP-large"), gr.outputs.Textbox(label="Caption generated by ViT+GPT-2")]
62
 
63
  title = "Interactive demo: comparing image captioning models"