image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000005534.jpg | 5534_evjvqa | [{"from": "human", "value": "what color is the helmet of the man riding the yellow motorbike?\n<image>"}, {"from": "gpt", "value": "the helmet of the man is yellow"}] |
00000005534.jpg | 5534_evjvqa | [{"from": "human", "value": "what kind of hat is placed by the man inside the bicycle basket?\n<image>"}, {"from": "gpt", "value": "inside the bicycle basket is a pith helmet"}] |
00000005537.jpg | 5537_evjvqa | [{"from": "human", "value": "how many glasses are placed on the table?\n<image>"}, {"from": "gpt", "value": "two glasses are placed on the table"}] |
00000005537.jpg | 5537_evjvqa | [{"from": "human", "value": "how many bowls of rice are on the table?\n<image>"}, {"from": "gpt", "value": "one bowl of rice is on the table"}] |
00000005537.jpg | 5537_evjvqa | [{"from": "human", "value": "where is the sushi placed?\n<image>"}, {"from": "gpt", "value": "the sushi is placed in a large rectangular plate"}] |
00000005540.jpg | 5540_evjvqa | [{"from": "human", "value": "what is placed on the swing?\n<image>"}, {"from": "gpt", "value": "on the swing is a hat"}] |
00000005541.jpg | 5541_evjvqa | [{"from": "human", "value": "what posture is the man sitting?\n<image>"}, {"from": "gpt", "value": "the man is sitting with left leg resting on right leg"}] |
00000005541.jpg | 5541_evjvqa | [{"from": "human", "value": "what color is the shirt the man is wearing?\n<image>"}, {"from": "gpt", "value": "the man is wearing a white shirt"}] |
00000005541.jpg | 5541_evjvqa | [{"from": "human", "value": "what color is the shadow of the text on the photo?\n<image>"}, {"from": "gpt", "value": "the shadow of the text on the photo is red"}] |
00000005542.jpg | 5542_evjvqa | [{"from": "human", "value": "what is the man in the blue cap holding in his hand?\n<image>"}, {"from": "gpt", "value": "thê man in blue cap is holding a tablet"}] |
00000005542.jpg | 5542_evjvqa | [{"from": "human", "value": "what is in the arm of the child who is putting his arm around his friend?\n<image>"}, {"from": "gpt", "value": "the arm of the child has a watch"}] |
00000005542.jpg | 5542_evjvqa | [{"from": "human", "value": "what is the man presenting to the students by?\n<image>"}, {"from": "gpt", "value": "the man is presenting to the students by a micro"}] |
00000005543.jpg | 5543_evjvqa | [{"from": "human", "value": "what color shirt is the mc wearing?\n<image>"}, {"from": "gpt", "value": "the mc is wearing a black shirt"}] |
00000005543.jpg | 5543_evjvqa | [{"from": "human", "value": "how many female staff are in charge of holding the ribbon?\n<image>"}, {"from": "gpt", "value": "four female staff are in charge of holding the ribbon"}] |
00000005545.jpg | 5545_evjvqa | [{"from": "human", "value": "what is the man with the glasses doing?\n<image>"}, {"from": "gpt", "value": "the man with the glasses is reaching out to choose goods"}] |
00000005545.jpg | 5545_evjvqa | [{"from": "human", "value": "what color shirt is the woman standing next to the man with glasses wearing?\n<image>"}, {"from": "gpt", "value": "the woman standing next to the man with glasses is wearing a dark blue shirt"}] |
00000005545.jpg | 5545_evjvqa | [{"from": "human", "value": "what color shirt is the man wearing glasses wearing?\n<image>"}, {"from": "gpt", "value": "the man wearing glasses is wearing a blue shirt"}] |
00000005545.jpg | 5545_evjvqa | [{"from": "human", "value": "what is the counter that the woman and the man are choosing?\n<image>"}, {"from": "gpt", "value": "those two are choosing at the fruit counter"}] |
00000005549.jpg | 5549_evjvqa | [{"from": "human", "value": "what is the signboard of the beach resort made of?\n<image>"}, {"from": "gpt", "value": "the signboard of the beach resort is made of a rock"}] |
00000005552.jpg | 5552_evjvqa | [{"from": "human", "value": "what color shirt is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a blue shirt"}] |
00000005552.jpg | 5552_evjvqa | [{"from": "human", "value": "what is the girl holding in her hand?\n<image>"}, {"from": "gpt", "value": "the girl is holding a conical hat in her hand"}] |
00000005552.jpg | 5552_evjvqa | [{"from": "human", "value": "what kind of clothes is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a long sleeved outfit"}] |
00000005554.jpg | 5554_evjvqa | [{"from": "human", "value": "where is the woman in the red shirt sitting compared to the boy in the blue shirt?\n<image>"}, {"from": "gpt", "value": "the woman in the red shirt is sitting opposite the boy in the blue shirt"}] |
00000005554.jpg | 5554_evjvqa | [{"from": "human", "value": "what is the boy in the blue shirt doing?\n<image>"}, {"from": "gpt", "value": "the boy in the blue shirt is on tiptoe to drink"}] |
00000005557.jpg | 5557_evjvqa | [{"from": "human", "value": "what shape is the supermarket's chandelier designed with?\n<image>"}, {"from": "gpt", "value": "the supermarket's chandelier is designed with the same shape as a drop of water"}] |
00000005557.jpg | 5557_evjvqa | [{"from": "human", "value": "what is the animal standing in the middle of the supermarket lobby?\n<image>"}, {"from": "gpt", "value": "the dog is standing in the middle of the supermarket lobby"}] |
00000005557.jpg | 5557_evjvqa | [{"from": "human", "value": "what color is the supermarket's chandelier?\n<image>"}, {"from": "gpt", "value": "the supermarket's chandelier has two colors metallic gold and silver"}] |
00000005560.jpg | 5560_evjvqa | [{"from": "human", "value": "what hairstyle does the girl have?\n<image>"}, {"from": "gpt", "value": "the girl has short hairstyle"}] |
00000005560.jpg | 5560_evjvqa | [{"from": "human", "value": "what is the girl's face leaning on?\n<image>"}, {"from": "gpt", "value": "the girl's face is leaning on her right hand"}] |
00000005560.jpg | 5560_evjvqa | [{"from": "human", "value": "what color is the girl's hair?\n<image>"}, {"from": "gpt", "value": "the girl's hair is black"}] |
00000005561.jpg | 5561_evjvqa | [{"from": "human", "value": "where is the girl standing compared to the building?\n<image>"}, {"from": "gpt", "value": "the girl is standing in front of the stairs leading to the building"}] |
00000005561.jpg | 5561_evjvqa | [{"from": "human", "value": "what is the girl doing?\n<image>"}, {"from": "gpt", "value": "the girl is posing for a photo"}] |
00000005561.jpg | 5561_evjvqa | [{"from": "human", "value": "how many unopened doors does the building have?\n<image>"}, {"from": "gpt", "value": "one door is closed"}] |
00000005563.jpg | 5563_evjvqa | [{"from": "human", "value": "what is the girl in the orange shirt doing?\n<image>"}, {"from": "gpt", "value": "the girl in the orange shirt is introducing the product to the girl in the blue shirt"}] |
00000005563.jpg | 5563_evjvqa | [{"from": "human", "value": "who is introducing the product to the girl in the blue shirt?\n<image>"}, {"from": "gpt", "value": "the girl in the orange shirt is introducing the product to the girl in the blue shirt"}] |
00000005564.jpg | 5564_evjvqa | [{"from": "human", "value": "what outfit is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing an ao dai"}] |
00000005564.jpg | 5564_evjvqa | [{"from": "human", "value": "what is the boy wearing on his left hand?\n<image>"}, {"from": "gpt", "value": "the boy is wearing a watch on his left hand"}] |
00000005564.jpg | 5564_evjvqa | [{"from": "human", "value": "what pattern does the shirt the boy is wearing have on the front?\n<image>"}, {"from": "gpt", "value": "the shirt the boy is wearing has a tree branch pattern on the front"}] |
00000005566.jpg | 5566_evjvqa | [{"from": "human", "value": "how many trees are there in the painting?\n<image>"}, {"from": "gpt", "value": "there are three trees in the painting"}] |
00000005566.jpg | 5566_evjvqa | [{"from": "human", "value": "how many people are cycling in the painting?\n<image>"}, {"from": "gpt", "value": "one person is cycling"}] |
00000005566.jpg | 5566_evjvqa | [{"from": "human", "value": "what is the shape of the painting frame?\n<image>"}, {"from": "gpt", "value": "the painting frame has a rectangular shape"}] |
00000005567.jpg | 5567_evjvqa | [{"from": "human", "value": "what color shirt is the woman leading the little girl wearing?\n<image>"}, {"from": "gpt", "value": "the woman leading the little girl is wearing a purple shirt"}] |
00000005567.jpg | 5567_evjvqa | [{"from": "human", "value": "how many people are being consulted?\n<image>"}, {"from": "gpt", "value": "three people are being consulted by staff"}] |
00000005567.jpg | 5567_evjvqa | [{"from": "human", "value": "which side is the little girl standing compared to the woman in the purple shirt?\n<image>"}, {"from": "gpt", "value": "the little girl is standing on the left of the woman in the purple shirt"}] |
00000005568.jpg | 5568_evjvqa | [{"from": "human", "value": "what is the woman doing?\n<image>"}, {"from": "gpt", "value": "the woman is taking notes"}] |
00000005568.jpg | 5568_evjvqa | [{"from": "human", "value": "what color is the shirt the woman is wearing?\n<image>"}, {"from": "gpt", "value": "the woman is wearing a white shirt"}] |
00000005568.jpg | 5568_evjvqa | [{"from": "human", "value": "where is the woman standing?\n<image>"}, {"from": "gpt", "value": "the woman is standing in front of a food counter"}] |
00000005569.jpg | 5569_evjvqa | [{"from": "human", "value": "what is the girl doing?\n<image>"}, {"from": "gpt", "value": "the girl is drinking a water bottle"}] |
00000005569.jpg | 5569_evjvqa | [{"from": "human", "value": "what is the symbol on the left sleeve of the girl's shirt?\n<image>"}, {"from": "gpt", "value": "the left sleeve of the girl's shirt has a fish bone symbol"}] |
00000005569.jpg | 5569_evjvqa | [{"from": "human", "value": "which hand is the girl holding the water bottle with?\n<image>"}, {"from": "gpt", "value": "the girl is holding the water bottle with her left hand"}] |
00000005571.jpg | 5571_evjvqa | [{"from": "human", "value": "where is the man standing?\n<image>"}, {"from": "gpt", "value": "the man is standing in front of a picture"}] |
00000005571.jpg | 5571_evjvqa | [{"from": "human", "value": "what color is the shirt the man is wearing?\n<image>"}, {"from": "gpt", "value": "the shirt the man is wearing is black"}] |
00000005571.jpg | 5571_evjvqa | [{"from": "human", "value": "how many people appear in the picture behind the man?\n<image>"}, {"from": "gpt", "value": "one person is in the picture behind the man"}] |
00000005572.jpg | 5572_evjvqa | [{"from": "human", "value": "how many people are using the escalator?\n<image>"}, {"from": "gpt", "value": "two people are using the escalator"}] |
00000005572.jpg | 5572_evjvqa | [{"from": "human", "value": "what is the woman in the red shirt standing on the escalator holding?\n<image>"}, {"from": "gpt", "value": "the woman in the red shirt standing on the escalator is holding a drink cup"}] |
00000005572.jpg | 5572_evjvqa | [{"from": "human", "value": "which way are the two people on the escalator going?\n<image>"}, {"from": "gpt", "value": "the two people on the escalator are going upstairs"}] |
00000005573.jpg | 5573_evjvqa | [{"from": "human", "value": "how many santa claus keychains are there?\n<image>"}, {"from": "gpt", "value": "there is a santa claus keychain"}] |
00000005573.jpg | 5573_evjvqa | [{"from": "human", "value": "what color is the glove keychain?\n<image>"}, {"from": "gpt", "value": "the glove keychain is red"}] |
00000005573.jpg | 5573_evjvqa | [{"from": "human", "value": "how many stars are on the pine tree keychain?\n<image>"}, {"from": "gpt", "value": "there are four stars on the pine tree keychain"}] |
00000005575.jpg | 5575_evjvqa | [{"from": "human", "value": "how many entrances does the temple have?\n<image>"}, {"from": "gpt", "value": "the temple has three entrances"}] |
00000005575.jpg | 5575_evjvqa | [{"from": "human", "value": "how many human statues are placed in front of the temple gate?\n<image>"}, {"from": "gpt", "value": "two statues are placed in front of the temple gate"}] |
00000005576.jpg | 5576_evjvqa | [{"from": "human", "value": "what does the painting depict?\n<image>"}, {"from": "gpt", "value": "the painting depicts the scene of a road fork"}] |
00000005580.jpg | 5580_evjvqa | [{"from": "human", "value": "what is the girl wearing the mask doing?\n<image>"}, {"from": "gpt", "value": "the girl is standing and manipulating an electronic device"}] |
00000005581.jpg | 5581_evjvqa | [{"from": "human", "value": "which hand is the boy putting behind?\n<image>"}, {"from": "gpt", "value": "the boy is putting his left arm behind"}] |
00000005581.jpg | 5581_evjvqa | [{"from": "human", "value": "what color is the backpack the boy is carrying?\n<image>"}, {"from": "gpt", "value": "the backpack the boy is carrying is blue"}] |
00000005581.jpg | 5581_evjvqa | [{"from": "human", "value": "where is the boy in the orange shirt standing?\n<image>"}, {"from": "gpt", "value": "the boy in the orange shirt is standing in front of a temple gate"}] |
00000005584.jpg | 5584_evjvqa | [{"from": "human", "value": "what is the policeman doing?\n<image>"}, {"from": "gpt", "value": "the policeman is helping with the box"}] |
00000005584.jpg | 5584_evjvqa | [{"from": "human", "value": "what color shirt is the person the police officer helping wearing?\n<image>"}, {"from": "gpt", "value": "the helped man is wearing a blue shirt"}] |
00000005587.jpg | 5587_evjvqa | [{"from": "human", "value": "what is the girl in the black shirt doing?\n<image>"}, {"from": "gpt", "value": "the girl in the black shirt is using a mobile phone"}] |
00000005588.jpg | 5588_evjvqa | [{"from": "human", "value": "what stall is the girl standing next to?\n<image>"}, {"from": "gpt", "value": "the girl is standing next to the rice stall"}] |
00000005588.jpg | 5588_evjvqa | [{"from": "human", "value": "what color shirt is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a green shirt"}] |
00000005589.jpg | 5589_evjvqa | [{"from": "human", "value": "what booth are people standing in front of?\n<image>"}, {"from": "gpt", "value": "people are standing in front of the book booth"}] |
00000005589.jpg | 5589_evjvqa | [{"from": "human", "value": "what is the man in the orange shirt doing?\n<image>"}, {"from": "gpt", "value": "the man in the orange shirt is introducing books to everyone"}] |
00000005589.jpg | 5589_evjvqa | [{"from": "human", "value": "how many books is the man in the white shirt on the left the man in the orange shirt is holding?\n<image>"}, {"from": "gpt", "value": "the man on the left the man in the orange shirt is holding one book"}] |
00000005594.jpg | 5594_evjvqa | [{"from": "human", "value": "what is the main background color of the billboard?\n<image>"}, {"from": "gpt", "value": "the main background color of the billboard is red"}] |
00000005597.jpg | 5597_evjvqa | [{"from": "human", "value": "how many people appear in the photo?\n<image>"}, {"from": "gpt", "value": "two people appear in the photo"}] |
00000001401.jpg | 1401_evjvqa | [{"from": "human", "value": "where is the woman with the red scarf sitting?\n<image>"}, {"from": "gpt", "value": "she is sitting at the vegetable stall "}] |
00000001401.jpg | 1401_evjvqa | [{"from": "human", "value": "what is the woman wearing the red scarf doing?\n<image>"}, {"from": "gpt", "value": "selling vegetables"}] |
00000001401.jpg | 1401_evjvqa | [{"from": "human", "value": "how many people come to buy vegetables where the woman with the red scarf is?\n<image>"}, {"from": "gpt", "value": "nobody"}] |
00000001401.jpg | 1401_evjvqa | [{"from": "human", "value": "what items does the woman wearing a red scarf sell?\n<image>"}, {"from": "gpt", "value": "raw vegetables"}] |
00000001402.jpg | 1402_evjvqa | [{"from": "human", "value": "how many motorbikes are being built in front of the houses?\n<image>"}, {"from": "gpt", "value": "four"}] |
00000001402.jpg | 1402_evjvqa | [{"from": "human", "value": "how many women are standing at the end of the road?\n<image>"}, {"from": "gpt", "value": "two"}] |
00000001402.jpg | 1402_evjvqa | [{"from": "human", "value": "how many children are playing at the end of the road?\n<image>"}, {"from": "gpt", "value": "only one"}] |
00000001403.jpg | 1403_evjvqa | [{"from": "human", "value": "where is the group of people going to visit?\n<image>"}, {"from": "gpt", "value": "the museum"}] |
00000001403.jpg | 1403_evjvqa | [{"from": "human", "value": "which side of the stairs is more crowded?\n<image>"}, {"from": "gpt", "value": "on the right"}] |
00000001403.jpg | 1403_evjvqa | [{"from": "human", "value": "which side of the stairs is more sparsely?\n<image>"}, {"from": "gpt", "value": "on the left"}] |
00000001405.jpg | 1405_evjvqa | [{"from": "human", "value": "how many people are holding the camera?\n<image>"}, {"from": "gpt", "value": "two"}] |
00000001405.jpg | 1405_evjvqa | [{"from": "human", "value": "what is the girt in green doing?\n<image>"}, {"from": "gpt", "value": "photographing"}] |
00000001405.jpg | 1405_evjvqa | [{"from": "human", "value": "what is the boy in orange doing?\n<image>"}, {"from": "gpt", "value": "taking a photo"}] |
00000001405.jpg | 1405_evjvqa | [{"from": "human", "value": "what color shirt is the man holding the camera wearing?\n<image>"}, {"from": "gpt", "value": "orange"}] |
00000001405.jpg | 1405_evjvqa | [{"from": "human", "value": "what color shirt is the girl holding the camera wearing?\n<image>"}, {"from": "gpt", "value": "green"}] |
00000001405.jpg | 1405_evjvqa | [{"from": "human", "value": "where are the visitors going to visit?\n<image>"}, {"from": "gpt", "value": "the gallery of photos"}] |
00000001405.jpg | 1405_evjvqa | [{"from": "human", "value": "how many people did the male friend in the orange shirt take pictures with?\n<image>"}, {"from": "gpt", "value": "only one"}] |
00000001405.jpg | 1405_evjvqa | [{"from": "human", "value": "where is the girl in the green shirt taking pictures?\n<image>"}, {"from": "gpt", "value": "the gallery of photos"}] |
00000001405.jpg | 1405_evjvqa | [{"from": "human", "value": "how many people are standing in the photo gallery?\n<image>"}, {"from": "gpt", "value": "two"}] |
00000001407.jpg | 1407_evjvqa | [{"from": "human", "value": "who are drawed in the poster?\n<image>"}, {"from": "gpt", "value": "the woman, the childr and the medical staff"}] |
00000001407.jpg | 1407_evjvqa | [{"from": "human", "value": "what is the boy in the poster doing?\n<image>"}, {"from": "gpt", "value": "he is washing his hands"}] |
00000001407.jpg | 1407_evjvqa | [{"from": "human", "value": "what is the woman in the poster doing?\n<image>"}, {"from": "gpt", "value": "she is wearing a medical mask"}] |
00000001408.jpg | 1408_evjvqa | [{"from": "human", "value": "what is the woman wearing glasses doing?\n<image>"}, {"from": "gpt", "value": "buying filled fish"}] |
00000001408.jpg | 1408_evjvqa | [{"from": "human", "value": "how many filled fishes is the woman wearing glasses holding in her hand?\n<image>"}, {"from": "gpt", "value": "four"}] |
Subsets and Splits