image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000001503.jpg
1503_evjvqa
[{"from": "human", "value": "what animal model is displayed in the center of the museum?\n<image>"}, {"from": "gpt", "value": "the elephant"}]
00000001503.jpg
1503_evjvqa
[{"from": "human", "value": "who is the child in the blue shirt standing to the right of?\n<image>"}, {"from": "gpt", "value": "the child in the blue shirt is standing on the right side of the man in black"}]
00000001503.jpg
1503_evjvqa
[{"from": "human", "value": "what wild animal is the model in the centre space of the museum?\n<image>"}, {"from": "gpt", "value": "in the centre space of the museum, there is a model of an elephant"}]
00000001507.jpg
1507_evjvqa
[{"from": "human", "value": "what is this woman doing in the orchard?\n<image>"}, {"from": "gpt", "value": "watering"}]
00000001507.jpg
1507_evjvqa
[{"from": "human", "value": "what is the woman holding to water the plants?\n<image>"}, {"from": "gpt", "value": "water pipe"}]
00000001507.jpg
1507_evjvqa
[{"from": "human", "value": "where is this woman watering plants?\n<image>"}, {"from": "gpt", "value": "in the orchard"}]
00000001513.jpg
1513_evjvqa
[{"from": "human", "value": "how many people are standing in front of the temple gate?\n<image>"}, {"from": "gpt", "value": "nobody is standing in front of the temple gate"}]
00000001513.jpg
1513_evjvqa
[{"from": "human", "value": "what gates can lead to the temple?\n<image>"}, {"from": "gpt", "value": "one big gate and two small gates on either side"}]
00000001514.jpg
1514_evjvqa
[{"from": "human", "value": "what vehicle is the woman in the blue shirt traveling on?\n<image>"}, {"from": "gpt", "value": "she is going by bike"}]
00000001514.jpg
1514_evjvqa
[{"from": "human", "value": "what are the things hanging on the ropes strung across the road?\n<image>"}, {"from": "gpt", "value": "the lanterns"}]
00000001514.jpg
1514_evjvqa
[{"from": "human", "value": "who is moving on the road by bicycle?\n<image>"}, {"from": "gpt", "value": "a woman with a blue shirt"}]
00000001517.jpg
1517_evjvqa
[{"from": "human", "value": "how are the men attending this press conference dressed?\n<image>"}, {"from": "gpt", "value": "they are all wearing white shirts"}]
00000001518.jpg
1518_evjvqa
[{"from": "human", "value": "what color of the shirt is the grey-haired man wearing?\n<image>"}, {"from": "gpt", "value": "the grey-haired man is wearing a blue shirt"}]
00000001518.jpg
1518_evjvqa
[{"from": "human", "value": "what is the gray-haired man doing?\n<image>"}, {"from": "gpt", "value": "he is sitting to tie the coconut leaves"}]
00000001518.jpg
1518_evjvqa
[{"from": "human", "value": "what color is the clothes of the silver-haired man sitting to tie the coconut leaves?\n<image>"}, {"from": "gpt", "value": "he is wearing a blue shirt and black trousers"}]
00000001520.jpg
1520_evjvqa
[{"from": "human", "value": "what color pants is the girl standing against the railing wearing?\n<image>"}, {"from": "gpt", "value": "the girl standing against the railing is wearing red knickers"}]
00000001520.jpg
1520_evjvqa
[{"from": "human", "value": "what does the girl wearing red knickers lean on?\n<image>"}, {"from": "gpt", "value": "the girl in red knickers is leaning on the railing on the wall"}]
00000001520.jpg
1520_evjvqa
[{"from": "human", "value": "what is this girl wearing on her head?\n<image>"}, {"from": "gpt", "value": "she is wearing a cap"}]
00000001524.jpg
1524_evjvqa
[{"from": "human", "value": "who is sitting on the boat?\n<image>"}, {"from": "gpt", "value": "there are two women sitting on the boat"}]
00000001524.jpg
1524_evjvqa
[{"from": "human", "value": "how is the only person standing dressed?\n<image>"}, {"from": "gpt", "value": "a white shirt and black trousers"}]
00000001524.jpg
1524_evjvqa
[{"from": "human", "value": "where is the man in the white shirt standing?\n<image>"}, {"from": "gpt", "value": "by the shore, next to the people sitting on chairs"}]
00000001532.jpg
1532_evjvqa
[{"from": "human", "value": "how is the man standing in front of the company dressed?\n<image>"}, {"from": "gpt", "value": "he wears a gray-blue shirt, black shoes and black trousers"}]
00000001542.jpg
1542_evjvqa
[{"from": "human", "value": "what flowers are planted right next to the lake?\n<image>"}, {"from": "gpt", "value": "the phoenix tree is planted right next to the lake"}]
00000001542.jpg
1542_evjvqa
[{"from": "human", "value": "what is the tower in the middle of the lake called?\n<image>"}, {"from": "gpt", "value": "rua tower"}]
00000001542.jpg
1542_evjvqa
[{"from": "human", "value": "in the middle of this lake what is an ancient tower named?\n<image>"}, {"from": "gpt", "value": "in the middle of this lake there is an ancient tower called rua tower"}]
00000001551.jpg
1551_evjvqa
[{"from": "human", "value": "which place is this landscape typically of?\n<image>"}, {"from": "gpt", "value": "ha long bay"}]
00000001551.jpg
1551_evjvqa
[{"from": "human", "value": "how many sails does the ship have?\n<image>"}, {"from": "gpt", "value": "the ship has three sails"}]
00000001551.jpg
1551_evjvqa
[{"from": "human", "value": "what color fabric are the sails of the ship made of?\n<image>"}, {"from": "gpt", "value": "the ship's sails are made of orange fabric"}]
00000001552.jpg
1552_evjvqa
[{"from": "human", "value": "why are the students holding notebooks and pens?\n<image>"}, {"from": "gpt", "value": "to take notes"}]
00000001552.jpg
1552_evjvqa
[{"from": "human", "value": "where is the person who is speaking to the students?\n<image>"}, {"from": "gpt", "value": "sitting at the podium"}]
00000001552.jpg
1552_evjvqa
[{"from": "human", "value": "what do people prepare for the students to sit here?\n<image>"}, {"from": "gpt", "value": "the plastic chairs"}]
00000001555.jpg
1555_evjvqa
[{"from": "human", "value": "on which side is the house at number 18 located relative to the house at number 16?\n<image>"}, {"from": "gpt", "value": "the house at number 18 is located just to the left of the house at number 16"}]
00000001555.jpg
1555_evjvqa
[{"from": "human", "value": "where is the bike?\n<image>"}, {"from": "gpt", "value": "located right at the border between the houses at number 18 and number 16"}]
00000001563.jpg
1563_evjvqa
[{"from": "human", "value": "what kind of equipment does this shop specialize in selling?\n<image>"}, {"from": "gpt", "value": "mobile phones"}]
00000001566.jpg
1566_evjvqa
[{"from": "human", "value": "on what is the woman in black sitting?\n<image>"}, {"from": "gpt", "value": "she is sitting on a plastic chair"}]
00000001566.jpg
1566_evjvqa
[{"from": "human", "value": "what does this stall sell?\n<image>"}, {"from": "gpt", "value": "fruits of all kinds"}]
00000001566.jpg
1566_evjvqa
[{"from": "human", "value": "what is the woman in orange doing at the fruit stall?\n<image>"}, {"from": "gpt", "value": "buying fruits"}]
00000001571.jpg
1571_evjvqa
[{"from": "human", "value": "what flags are these ships flying?\n<image>"}, {"from": "gpt", "value": "these ships are all flying the vietnamese flag"}]
00000001572.jpg
1572_evjvqa
[{"from": "human", "value": "with whom is this girl standing?\n<image>"}, {"from": "gpt", "value": "this girl is standing alone"}]
00000001572.jpg
1572_evjvqa
[{"from": "human", "value": "what is hanging on the back of the girl?\n<image>"}, {"from": "gpt", "value": "the lanterns"}]
00000001572.jpg
1572_evjvqa
[{"from": "human", "value": "what is the girl standing next to the lanterns wearing on her head?\n<image>"}, {"from": "gpt", "value": "a non la"}]
00000001579.jpg
1579_evjvqa
[{"from": "human", "value": "what device is the delivery shipper holding to his ear?\n<image>"}, {"from": "gpt", "value": "the delivery shipper is holding a mobile phone to his ear"}]
00000001580.jpg
1580_evjvqa
[{"from": "human", "value": "whose right side is the girl in the white t-shirt walking on?\n<image>"}, {"from": "gpt", "value": "girl in a white t-shirt walking to the right of young man in black clothes"}]
00000001580.jpg
1580_evjvqa
[{"from": "human", "value": "whose left side is the young guy dressed in black walking?\n<image>"}, {"from": "gpt", "value": "the young man in black clothes is walking on the left side of the girl wearing a white t-shirt"}]
00000001580.jpg
1580_evjvqa
[{"from": "human", "value": "what is a young man in black clothes carrying?\n<image>"}, {"from": "gpt", "value": "the young man in black is carrying a green bag of stuffs"}]
00000001582.jpg
1582_evjvqa
[{"from": "human", "value": "what orchards are these people visiting?\n<image>"}, {"from": "gpt", "value": "these people are visiting a rambutan orchard"}]
00000001582.jpg
1582_evjvqa
[{"from": "human", "value": "what kind of hat do three men wear?\n<image>"}, {"from": "gpt", "value": "green caps"}]
00000001582.jpg
1582_evjvqa
[{"from": "human", "value": "what kind of plants are being grown in this orchard?\n<image>"}, {"from": "gpt", "value": "rambutan trees"}]
00000001585.jpg
1585_evjvqa
[{"from": "human", "value": "what in square shape is hanging on the wall?\n<image>"}, {"from": "gpt", "value": "there is a square drawing on the wall"}]
00000001585.jpg
1585_evjvqa
[{"from": "human", "value": "how many chairs are placed around the table?\n<image>"}, {"from": "gpt", "value": "there are three chairs placed around the table"}]
00000001585.jpg
1585_evjvqa
[{"from": "human", "value": "what material is the table top made of?\n<image>"}, {"from": "gpt", "value": "the table top is made of wood"}]
00000001589.jpg
1589_evjvqa
[{"from": "human", "value": "how many guests are sitting at the tables in front of the noi dat rice-pot restaurant?\n<image>"}, {"from": "gpt", "value": "nobody"}]
00000001589.jpg
1589_evjvqa
[{"from": "human", "value": "how many employees are there standing in front of the restaurant?\n<image>"}, {"from": "gpt", "value": "there is a male employee standing in front of the restaurant"}]
00000001593.jpg
1593_evjvqa
[{"from": "human", "value": "what is placed on the body of the santa claus statue?\n<image>"}, {"from": "gpt", "value": "a saxophone"}]
00000001593.jpg
1593_evjvqa
[{"from": "human", "value": "of what character is the statue placed in this shop?\n<image>"}, {"from": "gpt", "value": "the santa claus statue"}]
00000001593.jpg
1593_evjvqa
[{"from": "human", "value": "what holiday events does this shop specialize in selling ornaments for?\n<image>"}, {"from": "gpt", "value": "christmas"}]
00000001594.jpg
1594_evjvqa
[{"from": "human", "value": "what device is being held in hand?\n<image>"}, {"from": "gpt", "value": "a smartphone"}]
00000001594.jpg
1594_evjvqa
[{"from": "human", "value": "why is this person holding a smartphone?\n<image>"}, {"from": "gpt", "value": "this person is holding a smartphone to take a photo"}]
00000001594.jpg
1594_evjvqa
[{"from": "human", "value": "what color is this guy's shirt?\n<image>"}, {"from": "gpt", "value": "this guy's shirt is blue"}]
00000001598.jpg
1598_evjvqa
[{"from": "human", "value": "what types of vehicles are parked on either side of the road?\n<image>"}, {"from": "gpt", "value": "cars, motorcycles and electric bicycles"}]
00000001598.jpg
1598_evjvqa
[{"from": "human", "value": "in which lane is the red motorbike running?\n<image>"}, {"from": "gpt", "value": "the red motorbike is running on the left lane"}]
00000005300.jpg
5300_evjvqa
[{"from": "human", "value": "how many bicycles are painted on the wall?\n<image>"}, {"from": "gpt", "value": "one bicycle is painted on the wall"}]
00000005300.jpg
5300_evjvqa
[{"from": "human", "value": "what is the main color of the wall?\n<image>"}, {"from": "gpt", "value": "yellow"}]
00000005303.jpg
5303_evjvqa
[{"from": "human", "value": "how many police soldiers are there?\n<image>"}, {"from": "gpt", "value": "there is one police soldier"}]
00000005303.jpg
5303_evjvqa
[{"from": "human", "value": "what color mask does police soldier wear?\n<image>"}, {"from": "gpt", "value": "white mask"}]
00000005303.jpg
5303_evjvqa
[{"from": "human", "value": "how many men wear dark blue vest?\n<image>"}, {"from": "gpt", "value": "one man"}]
00000005309.jpg
5309_evjvqa
[{"from": "human", "value": "how many girls wear green dresses?\n<image>"}, {"from": "gpt", "value": "one girl wears green dress"}]
00000005309.jpg
5309_evjvqa
[{"from": "human", "value": "how many boys wear black and white horizontal striped t-shirts?\n<image>"}, {"from": "gpt", "value": "one boy wears black and white horizontal striped t-shirt"}]
00000005310.jpg
5310_evjvqa
[{"from": "human", "value": "what is the boy in the red shirt doing on the cyclo?\n<image>"}, {"from": "gpt", "value": "surfing the phone"}]
00000005310.jpg
5310_evjvqa
[{"from": "human", "value": "what color shirt is the girl sitting on the cyclo wearing?\n<image>"}, {"from": "gpt", "value": "white shirt"}]
00000005313.jpg
5313_evjvqa
[{"from": "human", "value": "what is the boy doing?\n<image>"}, {"from": "gpt", "value": "repairing the signboard"}]
00000005313.jpg
5313_evjvqa
[{"from": "human", "value": "what color is the shirt of the signboard repairman?\n<image>"}, {"from": "gpt", "value": "blue"}]
00000005315.jpg
5315_evjvqa
[{"from": "human", "value": "how many motorbikes are parked in front of the house?\n<image>"}, {"from": "gpt", "value": "two motorbikes are parked in front of the house"}]
00000005315.jpg
5315_evjvqa
[{"from": "human", "value": "what color are the lanterns?\n<image>"}, {"from": "gpt", "value": "the lanterns are orange"}]
00000005315.jpg
5315_evjvqa
[{"from": "human", "value": "what is the main color of the house?\n<image>"}, {"from": "gpt", "value": "the main color of the house is yellow"}]
00000005317.jpg
5317_evjvqa
[{"from": "human", "value": "what color is the plate of spring rolls?\n<image>"}, {"from": "gpt", "value": "white"}]
00000005317.jpg
5317_evjvqa
[{"from": "human", "value": "how many shrimps are there in a spring roll?\n<image>"}, {"from": "gpt", "value": "there are two shrimps in a spring roll"}]
00000005317.jpg
5317_evjvqa
[{"from": "human", "value": "what pattern does the sauce bowl have?\n<image>"}, {"from": "gpt", "value": "purple flower pattern"}]
00000005319.jpg
5319_evjvqa
[{"from": "human", "value": "what is the main color of the floor?\n<image>"}, {"from": "gpt", "value": "grey"}]
00000005319.jpg
5319_evjvqa
[{"from": "human", "value": "what color are the pillows on the chair?\n<image>"}, {"from": "gpt", "value": "white"}]
00000005319.jpg
5319_evjvqa
[{"from": "human", "value": "what is the general color of the chairs?\n<image>"}, {"from": "gpt", "value": "brown"}]
00000005321.jpg
5321_evjvqa
[{"from": "human", "value": "what is the main color of the houses?\n<image>"}, {"from": "gpt", "value": "the main color is yellow"}]
00000005321.jpg
5321_evjvqa
[{"from": "human", "value": "how many people are on the boat?\n<image>"}, {"from": "gpt", "value": "two people are on the boat"}]
00000005321.jpg
5321_evjvqa
[{"from": "human", "value": "how many people are riding cyclos on the road?\n<image>"}, {"from": "gpt", "value": "one man is riding cyclo on the road"}]
00000005322.jpg
5322_evjvqa
[{"from": "human", "value": "what material is the house thatched with?\n<image>"}, {"from": "gpt", "value": "thatched with leaves"}]
00000005322.jpg
5322_evjvqa
[{"from": "human", "value": "how many doors does the house have?\n<image>"}, {"from": "gpt", "value": "three doors"}]
00000005323.jpg
5323_evjvqa
[{"from": "human", "value": "what are the woman in the red and black shirt and the woman in the pink shirt doing?\n<image>"}, {"from": "gpt", "value": "are trading"}]
00000005323.jpg
5323_evjvqa
[{"from": "human", "value": "what is the woman in white shirt with conical hat doing?\n<image>"}, {"from": "gpt", "value": "carrying the peddler"}]
00000005323.jpg
5323_evjvqa
[{"from": "human", "value": "how many people are carrying the peddler?\n<image>"}, {"from": "gpt", "value": "one person is carrying the peddler"}]
00000005324.jpg
5324_evjvqa
[{"from": "human", "value": "how many ships are moving on the sea?\n<image>"}, {"from": "gpt", "value": "three ships are moving on the sea"}]
00000005325.jpg
5325_evjvqa
[{"from": "human", "value": "how many bowls of pho are there?\n<image>"}, {"from": "gpt", "value": "there are three bowls of pho"}]
00000005325.jpg
5325_evjvqa
[{"from": "human", "value": "what is the tray that holds the bowls of pho?\n<image>"}, {"from": "gpt", "value": "stainless steel tray"}]
00000005325.jpg
5325_evjvqa
[{"from": "human", "value": "what color are the bowls of pho?\n<image>"}, {"from": "gpt", "value": "white"}]
00000005326.jpg
5326_evjvqa
[{"from": "human", "value": "what is the girl in the white shirt doing?\n<image>"}, {"from": "gpt", "value": "the girl is washing her hair"}]
00000005326.jpg
5326_evjvqa
[{"from": "human", "value": "how many green bottles of shampoo are there?\n<image>"}, {"from": "gpt", "value": "there are two bottles"}]
00000005326.jpg
5326_evjvqa
[{"from": "human", "value": "what color shirt is the girl washing her hair wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a white shirt"}]
00000005334.jpg
5334_evjvqa
[{"from": "human", "value": "what are the boy in the purple shirt and the girl in the white shirt doing?\n<image>"}, {"from": "gpt", "value": "the boy in the purple shirt is taking a picture of the girl in the white shirt"}]
00000005340.jpg
5340_evjvqa
[{"from": "human", "value": "what are the two young people doing?\n<image>"}, {"from": "gpt", "value": "kissing"}]
00000005340.jpg
5340_evjvqa
[{"from": "human", "value": "where are the two young people standing?\n<image>"}, {"from": "gpt", "value": "standing on the bridge"}]
00000005340.jpg
5340_evjvqa
[{"from": "human", "value": "what color shoes do the two people wear?\n<image>"}, {"from": "gpt", "value": "the two people wear white shoes"}]