mshukor HF Staff commited on
Commit
f508895
·
1 Parent(s): f2a6bea

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -5
app.py CHANGED
@@ -290,11 +290,11 @@ def inference(image, task_type, instruction):
290
  inputs = [gr.inputs.Image(type='pil'), gr.inputs.Radio(choices=['Image Captioning',"Visual Question Answering", "Visual Grounding", "General"], type="value", default="Image Captioning", label="Task"), gr.inputs.Textbox(lines=1, label="Instruction")]
291
  outputs = [gr.outputs.Image(type='pil'), 'text']
292
  examples = [
293
- ['examples/pokemons.jpeg', 'Image Captioning', None],
294
- ['examples/cats.jpeg', 'Visual Question Answering', 'where are the cats?'],
295
- ['examples/one_piece.jpeg', 'Visual Grounding', 'a man in a straw hat and a red dress'],
296
- ['examples/three_houses.jpeg', 'General', 'which region does the text " a grey car " describe?'],
297
- ['examples/three_houses.jpeg', 'General', 'what color is the left car?']
298
  ]
299
 
300
  title = "OFA"
 
290
  inputs = [gr.inputs.Image(type='pil'), gr.inputs.Radio(choices=['Image Captioning',"Visual Question Answering", "Visual Grounding", "General"], type="value", default="Image Captioning", label="Task"), gr.inputs.Textbox(lines=1, label="Instruction")]
291
  outputs = [gr.outputs.Image(type='pil'), 'text']
292
  examples = [
293
+ # ['examples/caption/soccer.jpg', 'Image Captioning', None],
294
+ # ['examples/sheep.jpg', 'Visual Question Answering', 'where are the cats?'],
295
+ # ['examples/one_piece.jpeg', 'Visual Grounding', 'a man in a straw hat and a red dress'],
296
+ ['examples/skateboard.jpeg', 'General', 'which region does the text " a grey car " describe?'],
297
+ ['examples/baseball.jpeg', 'General', 'what color is the left car?']
298
  ]
299
 
300
  title = "OFA"