image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000001503.jpg | 1503_evjvqa | [{"from": "human", "value": "what animal model is displayed in the center of the museum?\n<image>"}, {"from": "gpt", "value": "the elephant"}] |
00000001503.jpg | 1503_evjvqa | [{"from": "human", "value": "who is the child in the blue shirt standing to the right of?\n<image>"}, {"from": "gpt", "value": "the child in the blue shirt is standing on the right side of the man in black"}] |
00000001503.jpg | 1503_evjvqa | [{"from": "human", "value": "what wild animal is the model in the centre space of the museum?\n<image>"}, {"from": "gpt", "value": "in the centre space of the museum, there is a model of an elephant"}] |
00000001507.jpg | 1507_evjvqa | [{"from": "human", "value": "what is this woman doing in the orchard?\n<image>"}, {"from": "gpt", "value": "watering"}] |
00000001507.jpg | 1507_evjvqa | [{"from": "human", "value": "what is the woman holding to water the plants?\n<image>"}, {"from": "gpt", "value": "water pipe"}] |
00000001507.jpg | 1507_evjvqa | [{"from": "human", "value": "where is this woman watering plants?\n<image>"}, {"from": "gpt", "value": "in the orchard"}] |
00000001513.jpg | 1513_evjvqa | [{"from": "human", "value": "how many people are standing in front of the temple gate?\n<image>"}, {"from": "gpt", "value": "nobody is standing in front of the temple gate"}] |
00000001513.jpg | 1513_evjvqa | [{"from": "human", "value": "what gates can lead to the temple?\n<image>"}, {"from": "gpt", "value": "one big gate and two small gates on either side"}] |
00000001514.jpg | 1514_evjvqa | [{"from": "human", "value": "what vehicle is the woman in the blue shirt traveling on?\n<image>"}, {"from": "gpt", "value": "she is going by bike"}] |
00000001514.jpg | 1514_evjvqa | [{"from": "human", "value": "what are the things hanging on the ropes strung across the road?\n<image>"}, {"from": "gpt", "value": "the lanterns"}] |
00000001514.jpg | 1514_evjvqa | [{"from": "human", "value": "who is moving on the road by bicycle?\n<image>"}, {"from": "gpt", "value": "a woman with a blue shirt"}] |
00000001517.jpg | 1517_evjvqa | [{"from": "human", "value": "how are the men attending this press conference dressed?\n<image>"}, {"from": "gpt", "value": "they are all wearing white shirts"}] |
00000001518.jpg | 1518_evjvqa | [{"from": "human", "value": "what color of the shirt is the grey-haired man wearing?\n<image>"}, {"from": "gpt", "value": "the grey-haired man is wearing a blue shirt"}] |
00000001518.jpg | 1518_evjvqa | [{"from": "human", "value": "what is the gray-haired man doing?\n<image>"}, {"from": "gpt", "value": "he is sitting to tie the coconut leaves"}] |
00000001518.jpg | 1518_evjvqa | [{"from": "human", "value": "what color is the clothes of the silver-haired man sitting to tie the coconut leaves?\n<image>"}, {"from": "gpt", "value": "he is wearing a blue shirt and black trousers"}] |
00000001520.jpg | 1520_evjvqa | [{"from": "human", "value": "what color pants is the girl standing against the railing wearing?\n<image>"}, {"from": "gpt", "value": "the girl standing against the railing is wearing red knickers"}] |
00000001520.jpg | 1520_evjvqa | [{"from": "human", "value": "what does the girl wearing red knickers lean on?\n<image>"}, {"from": "gpt", "value": "the girl in red knickers is leaning on the railing on the wall"}] |
00000001520.jpg | 1520_evjvqa | [{"from": "human", "value": "what is this girl wearing on her head?\n<image>"}, {"from": "gpt", "value": "she is wearing a cap"}] |
00000001524.jpg | 1524_evjvqa | [{"from": "human", "value": "who is sitting on the boat?\n<image>"}, {"from": "gpt", "value": "there are two women sitting on the boat"}] |
00000001524.jpg | 1524_evjvqa | [{"from": "human", "value": "how is the only person standing dressed?\n<image>"}, {"from": "gpt", "value": "a white shirt and black trousers"}] |
00000001524.jpg | 1524_evjvqa | [{"from": "human", "value": "where is the man in the white shirt standing?\n<image>"}, {"from": "gpt", "value": "by the shore, next to the people sitting on chairs"}] |
00000001532.jpg | 1532_evjvqa | [{"from": "human", "value": "how is the man standing in front of the company dressed?\n<image>"}, {"from": "gpt", "value": "he wears a gray-blue shirt, black shoes and black trousers"}] |
00000001542.jpg | 1542_evjvqa | [{"from": "human", "value": "what flowers are planted right next to the lake?\n<image>"}, {"from": "gpt", "value": "the phoenix tree is planted right next to the lake"}] |
00000001542.jpg | 1542_evjvqa | [{"from": "human", "value": "what is the tower in the middle of the lake called?\n<image>"}, {"from": "gpt", "value": "rua tower"}] |
00000001542.jpg | 1542_evjvqa | [{"from": "human", "value": "in the middle of this lake what is an ancient tower named?\n<image>"}, {"from": "gpt", "value": "in the middle of this lake there is an ancient tower called rua tower"}] |
00000001551.jpg | 1551_evjvqa | [{"from": "human", "value": "which place is this landscape typically of?\n<image>"}, {"from": "gpt", "value": "ha long bay"}] |
00000001551.jpg | 1551_evjvqa | [{"from": "human", "value": "how many sails does the ship have?\n<image>"}, {"from": "gpt", "value": "the ship has three sails"}] |
00000001551.jpg | 1551_evjvqa | [{"from": "human", "value": "what color fabric are the sails of the ship made of?\n<image>"}, {"from": "gpt", "value": "the ship's sails are made of orange fabric"}] |
00000001552.jpg | 1552_evjvqa | [{"from": "human", "value": "why are the students holding notebooks and pens?\n<image>"}, {"from": "gpt", "value": "to take notes"}] |
00000001552.jpg | 1552_evjvqa | [{"from": "human", "value": "where is the person who is speaking to the students?\n<image>"}, {"from": "gpt", "value": "sitting at the podium"}] |
00000001552.jpg | 1552_evjvqa | [{"from": "human", "value": "what do people prepare for the students to sit here?\n<image>"}, {"from": "gpt", "value": "the plastic chairs"}] |
00000001555.jpg | 1555_evjvqa | [{"from": "human", "value": "on which side is the house at number 18 located relative to the house at number 16?\n<image>"}, {"from": "gpt", "value": "the house at number 18 is located just to the left of the house at number 16"}] |
00000001555.jpg | 1555_evjvqa | [{"from": "human", "value": "where is the bike?\n<image>"}, {"from": "gpt", "value": "located right at the border between the houses at number 18 and number 16"}] |
00000001563.jpg | 1563_evjvqa | [{"from": "human", "value": "what kind of equipment does this shop specialize in selling?\n<image>"}, {"from": "gpt", "value": "mobile phones"}] |
00000001566.jpg | 1566_evjvqa | [{"from": "human", "value": "on what is the woman in black sitting?\n<image>"}, {"from": "gpt", "value": "she is sitting on a plastic chair"}] |
00000001566.jpg | 1566_evjvqa | [{"from": "human", "value": "what does this stall sell?\n<image>"}, {"from": "gpt", "value": "fruits of all kinds"}] |
00000001566.jpg | 1566_evjvqa | [{"from": "human", "value": "what is the woman in orange doing at the fruit stall?\n<image>"}, {"from": "gpt", "value": "buying fruits"}] |
00000001571.jpg | 1571_evjvqa | [{"from": "human", "value": "what flags are these ships flying?\n<image>"}, {"from": "gpt", "value": "these ships are all flying the vietnamese flag"}] |
00000001572.jpg | 1572_evjvqa | [{"from": "human", "value": "with whom is this girl standing?\n<image>"}, {"from": "gpt", "value": "this girl is standing alone"}] |
00000001572.jpg | 1572_evjvqa | [{"from": "human", "value": "what is hanging on the back of the girl?\n<image>"}, {"from": "gpt", "value": "the lanterns"}] |
00000001572.jpg | 1572_evjvqa | [{"from": "human", "value": "what is the girl standing next to the lanterns wearing on her head?\n<image>"}, {"from": "gpt", "value": "a non la"}] |
00000001579.jpg | 1579_evjvqa | [{"from": "human", "value": "what device is the delivery shipper holding to his ear?\n<image>"}, {"from": "gpt", "value": "the delivery shipper is holding a mobile phone to his ear"}] |
00000001580.jpg | 1580_evjvqa | [{"from": "human", "value": "whose right side is the girl in the white t-shirt walking on?\n<image>"}, {"from": "gpt", "value": "girl in a white t-shirt walking to the right of young man in black clothes"}] |
00000001580.jpg | 1580_evjvqa | [{"from": "human", "value": "whose left side is the young guy dressed in black walking?\n<image>"}, {"from": "gpt", "value": "the young man in black clothes is walking on the left side of the girl wearing a white t-shirt"}] |
00000001580.jpg | 1580_evjvqa | [{"from": "human", "value": "what is a young man in black clothes carrying?\n<image>"}, {"from": "gpt", "value": "the young man in black is carrying a green bag of stuffs"}] |
00000001582.jpg | 1582_evjvqa | [{"from": "human", "value": "what orchards are these people visiting?\n<image>"}, {"from": "gpt", "value": "these people are visiting a rambutan orchard"}] |
00000001582.jpg | 1582_evjvqa | [{"from": "human", "value": "what kind of hat do three men wear?\n<image>"}, {"from": "gpt", "value": "green caps"}] |
00000001582.jpg | 1582_evjvqa | [{"from": "human", "value": "what kind of plants are being grown in this orchard?\n<image>"}, {"from": "gpt", "value": "rambutan trees"}] |
00000001585.jpg | 1585_evjvqa | [{"from": "human", "value": "what in square shape is hanging on the wall?\n<image>"}, {"from": "gpt", "value": "there is a square drawing on the wall"}] |
00000001585.jpg | 1585_evjvqa | [{"from": "human", "value": "how many chairs are placed around the table?\n<image>"}, {"from": "gpt", "value": "there are three chairs placed around the table"}] |
00000001585.jpg | 1585_evjvqa | [{"from": "human", "value": "what material is the table top made of?\n<image>"}, {"from": "gpt", "value": "the table top is made of wood"}] |
00000001589.jpg | 1589_evjvqa | [{"from": "human", "value": "how many guests are sitting at the tables in front of the noi dat rice-pot restaurant?\n<image>"}, {"from": "gpt", "value": "nobody"}] |
00000001589.jpg | 1589_evjvqa | [{"from": "human", "value": "how many employees are there standing in front of the restaurant?\n<image>"}, {"from": "gpt", "value": "there is a male employee standing in front of the restaurant"}] |
00000001593.jpg | 1593_evjvqa | [{"from": "human", "value": "what is placed on the body of the santa claus statue?\n<image>"}, {"from": "gpt", "value": "a saxophone"}] |
00000001593.jpg | 1593_evjvqa | [{"from": "human", "value": "of what character is the statue placed in this shop?\n<image>"}, {"from": "gpt", "value": "the santa claus statue"}] |
00000001593.jpg | 1593_evjvqa | [{"from": "human", "value": "what holiday events does this shop specialize in selling ornaments for?\n<image>"}, {"from": "gpt", "value": "christmas"}] |
00000001594.jpg | 1594_evjvqa | [{"from": "human", "value": "what device is being held in hand?\n<image>"}, {"from": "gpt", "value": "a smartphone"}] |
00000001594.jpg | 1594_evjvqa | [{"from": "human", "value": "why is this person holding a smartphone?\n<image>"}, {"from": "gpt", "value": "this person is holding a smartphone to take a photo"}] |
00000001594.jpg | 1594_evjvqa | [{"from": "human", "value": "what color is this guy's shirt?\n<image>"}, {"from": "gpt", "value": "this guy's shirt is blue"}] |
00000001598.jpg | 1598_evjvqa | [{"from": "human", "value": "what types of vehicles are parked on either side of the road?\n<image>"}, {"from": "gpt", "value": "cars, motorcycles and electric bicycles"}] |
00000001598.jpg | 1598_evjvqa | [{"from": "human", "value": "in which lane is the red motorbike running?\n<image>"}, {"from": "gpt", "value": "the red motorbike is running on the left lane"}] |
00000005300.jpg | 5300_evjvqa | [{"from": "human", "value": "how many bicycles are painted on the wall?\n<image>"}, {"from": "gpt", "value": "one bicycle is painted on the wall"}] |
00000005300.jpg | 5300_evjvqa | [{"from": "human", "value": "what is the main color of the wall?\n<image>"}, {"from": "gpt", "value": "yellow"}] |
00000005303.jpg | 5303_evjvqa | [{"from": "human", "value": "how many police soldiers are there?\n<image>"}, {"from": "gpt", "value": "there is one police soldier"}] |
00000005303.jpg | 5303_evjvqa | [{"from": "human", "value": "what color mask does police soldier wear?\n<image>"}, {"from": "gpt", "value": "white mask"}] |
00000005303.jpg | 5303_evjvqa | [{"from": "human", "value": "how many men wear dark blue vest?\n<image>"}, {"from": "gpt", "value": "one man"}] |
00000005309.jpg | 5309_evjvqa | [{"from": "human", "value": "how many girls wear green dresses?\n<image>"}, {"from": "gpt", "value": "one girl wears green dress"}] |
00000005309.jpg | 5309_evjvqa | [{"from": "human", "value": "how many boys wear black and white horizontal striped t-shirts?\n<image>"}, {"from": "gpt", "value": "one boy wears black and white horizontal striped t-shirt"}] |
00000005310.jpg | 5310_evjvqa | [{"from": "human", "value": "what is the boy in the red shirt doing on the cyclo?\n<image>"}, {"from": "gpt", "value": "surfing the phone"}] |
00000005310.jpg | 5310_evjvqa | [{"from": "human", "value": "what color shirt is the girl sitting on the cyclo wearing?\n<image>"}, {"from": "gpt", "value": "white shirt"}] |
00000005313.jpg | 5313_evjvqa | [{"from": "human", "value": "what is the boy doing?\n<image>"}, {"from": "gpt", "value": "repairing the signboard"}] |
00000005313.jpg | 5313_evjvqa | [{"from": "human", "value": "what color is the shirt of the signboard repairman?\n<image>"}, {"from": "gpt", "value": "blue"}] |
00000005315.jpg | 5315_evjvqa | [{"from": "human", "value": "how many motorbikes are parked in front of the house?\n<image>"}, {"from": "gpt", "value": "two motorbikes are parked in front of the house"}] |
00000005315.jpg | 5315_evjvqa | [{"from": "human", "value": "what color are the lanterns?\n<image>"}, {"from": "gpt", "value": "the lanterns are orange"}] |
00000005315.jpg | 5315_evjvqa | [{"from": "human", "value": "what is the main color of the house?\n<image>"}, {"from": "gpt", "value": "the main color of the house is yellow"}] |
00000005317.jpg | 5317_evjvqa | [{"from": "human", "value": "what color is the plate of spring rolls?\n<image>"}, {"from": "gpt", "value": "white"}] |
00000005317.jpg | 5317_evjvqa | [{"from": "human", "value": "how many shrimps are there in a spring roll?\n<image>"}, {"from": "gpt", "value": "there are two shrimps in a spring roll"}] |
00000005317.jpg | 5317_evjvqa | [{"from": "human", "value": "what pattern does the sauce bowl have?\n<image>"}, {"from": "gpt", "value": "purple flower pattern"}] |
00000005319.jpg | 5319_evjvqa | [{"from": "human", "value": "what is the main color of the floor?\n<image>"}, {"from": "gpt", "value": "grey"}] |
00000005319.jpg | 5319_evjvqa | [{"from": "human", "value": "what color are the pillows on the chair?\n<image>"}, {"from": "gpt", "value": "white"}] |
00000005319.jpg | 5319_evjvqa | [{"from": "human", "value": "what is the general color of the chairs?\n<image>"}, {"from": "gpt", "value": "brown"}] |
00000005321.jpg | 5321_evjvqa | [{"from": "human", "value": "what is the main color of the houses?\n<image>"}, {"from": "gpt", "value": "the main color is yellow"}] |
00000005321.jpg | 5321_evjvqa | [{"from": "human", "value": "how many people are on the boat?\n<image>"}, {"from": "gpt", "value": "two people are on the boat"}] |
00000005321.jpg | 5321_evjvqa | [{"from": "human", "value": "how many people are riding cyclos on the road?\n<image>"}, {"from": "gpt", "value": "one man is riding cyclo on the road"}] |
00000005322.jpg | 5322_evjvqa | [{"from": "human", "value": "what material is the house thatched with?\n<image>"}, {"from": "gpt", "value": "thatched with leaves"}] |
00000005322.jpg | 5322_evjvqa | [{"from": "human", "value": "how many doors does the house have?\n<image>"}, {"from": "gpt", "value": "three doors"}] |
00000005323.jpg | 5323_evjvqa | [{"from": "human", "value": "what are the woman in the red and black shirt and the woman in the pink shirt doing?\n<image>"}, {"from": "gpt", "value": "are trading"}] |
00000005323.jpg | 5323_evjvqa | [{"from": "human", "value": "what is the woman in white shirt with conical hat doing?\n<image>"}, {"from": "gpt", "value": "carrying the peddler"}] |
00000005323.jpg | 5323_evjvqa | [{"from": "human", "value": "how many people are carrying the peddler?\n<image>"}, {"from": "gpt", "value": "one person is carrying the peddler"}] |
00000005324.jpg | 5324_evjvqa | [{"from": "human", "value": "how many ships are moving on the sea?\n<image>"}, {"from": "gpt", "value": "three ships are moving on the sea"}] |
00000005325.jpg | 5325_evjvqa | [{"from": "human", "value": "how many bowls of pho are there?\n<image>"}, {"from": "gpt", "value": "there are three bowls of pho"}] |
00000005325.jpg | 5325_evjvqa | [{"from": "human", "value": "what is the tray that holds the bowls of pho?\n<image>"}, {"from": "gpt", "value": "stainless steel tray"}] |
00000005325.jpg | 5325_evjvqa | [{"from": "human", "value": "what color are the bowls of pho?\n<image>"}, {"from": "gpt", "value": "white"}] |
00000005326.jpg | 5326_evjvqa | [{"from": "human", "value": "what is the girl in the white shirt doing?\n<image>"}, {"from": "gpt", "value": "the girl is washing her hair"}] |
00000005326.jpg | 5326_evjvqa | [{"from": "human", "value": "how many green bottles of shampoo are there?\n<image>"}, {"from": "gpt", "value": "there are two bottles"}] |
00000005326.jpg | 5326_evjvqa | [{"from": "human", "value": "what color shirt is the girl washing her hair wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a white shirt"}] |
00000005334.jpg | 5334_evjvqa | [{"from": "human", "value": "what are the boy in the purple shirt and the girl in the white shirt doing?\n<image>"}, {"from": "gpt", "value": "the boy in the purple shirt is taking a picture of the girl in the white shirt"}] |
00000005340.jpg | 5340_evjvqa | [{"from": "human", "value": "what are the two young people doing?\n<image>"}, {"from": "gpt", "value": "kissing"}] |
00000005340.jpg | 5340_evjvqa | [{"from": "human", "value": "where are the two young people standing?\n<image>"}, {"from": "gpt", "value": "standing on the bridge"}] |
00000005340.jpg | 5340_evjvqa | [{"from": "human", "value": "what color shoes do the two people wear?\n<image>"}, {"from": "gpt", "value": "the two people wear white shoes"}] |
Subsets and Splits