image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000004661.jpg | 4661_evjvqa | [{"from": "human", "value": "where are people coming from?\n<image>"}, {"from": "gpt", "value": "going out from the market"}] |
00000004662.jpg | 4662_evjvqa | [{"from": "human", "value": "how many motorbikes are on the road?\n<image>"}, {"from": "gpt", "value": "there are two motorbikes running on the road"}] |
00000004662.jpg | 4662_evjvqa | [{"from": "human", "value": "what color shirt is the man jogging wearing?\n<image>"}, {"from": "gpt", "value": "wearing a yellow shirt"}] |
00000004662.jpg | 4662_evjvqa | [{"from": "human", "value": "how many people are standing on the sidewalk?\n<image>"}, {"from": "gpt", "value": "there are two people standing on the sidewalk"}] |
00000004664.jpg | 4664_evjvqa | [{"from": "human", "value": "where is everyone?\n<image>"}, {"from": "gpt", "value": "in the supermarket"}] |
00000004664.jpg | 4664_evjvqa | [{"from": "human", "value": "what are people doing?\n<image>"}, {"from": "gpt", "value": "shopping in the supermarket"}] |
00000004664.jpg | 4664_evjvqa | [{"from": "human", "value": "where is the child standing?\n<image>"}, {"from": "gpt", "value": "standing in a supermarket trolley"}] |
00000004666.jpg | 4666_evjvqa | [{"from": "human", "value": "how many flags are flied by the lakeside?\n<image>"}, {"from": "gpt", "value": "there are three flags flying by the lakeside"}] |
00000004667.jpg | 4667_evjvqa | [{"from": "human", "value": "where is everyone?\n<image>"}, {"from": "gpt", "value": "in a supermarket"}] |
00000004667.jpg | 4667_evjvqa | [{"from": "human", "value": "what color are supermarket staff uniforms?\n<image>"}, {"from": "gpt", "value": "uniforms are orange"}] |
00000004667.jpg | 4667_evjvqa | [{"from": "human", "value": "what shelf is the man in the white t-shirt standing next to?\n<image>"}, {"from": "gpt", "value": "standing next to the milk shelf"}] |
00000004671.jpg | 4671_evjvqa | [{"from": "human", "value": "what is the boy holding in his hand?\n<image>"}, {"from": "gpt", "value": "holding banknotes in hand"}] |
00000004671.jpg | 4671_evjvqa | [{"from": "human", "value": "what color is the car behind the guy?\n<image>"}, {"from": "gpt", "value": "the car is blue"}] |
00000004671.jpg | 4671_evjvqa | [{"from": "human", "value": "what is hanging on the rearview mirror of a motorcycle?\n<image>"}, {"from": "gpt", "value": "bags of food boxes are hung on the rearview mirror of the motorbike"}] |
00000004672.jpg | 4672_evjvqa | [{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a village by the river"}] |
00000004672.jpg | 4672_evjvqa | [{"from": "human", "value": "how many boats are running?\n<image>"}, {"from": "gpt", "value": "there is a boat running"}] |
00000004672.jpg | 4672_evjvqa | [{"from": "human", "value": "what time of day is the view of the village?\n<image>"}, {"from": "gpt", "value": "it is the afternoon"}] |
00000004674.jpg | 4674_evjvqa | [{"from": "human", "value": "how many people are in the group?\n<image>"}, {"from": "gpt", "value": "the group are four people"}] |
00000004674.jpg | 4674_evjvqa | [{"from": "human", "value": "what is behind the group of people?\n<image>"}, {"from": "gpt", "value": "behind is a house next to the lake"}] |
00000004674.jpg | 4674_evjvqa | [{"from": "human", "value": "how many people are wearing yellow shirts?\n<image>"}, {"from": "gpt", "value": "there is a man in a yellow t-shirt"}] |
00000004676.jpg | 4676_evjvqa | [{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a food shelf"}] |
00000004677.jpg | 4677_evjvqa | [{"from": "human", "value": "how many people are sitting by the roadside?\n<image>"}, {"from": "gpt", "value": "there are two people sitting on the side of the road"}] |
00000004677.jpg | 4677_evjvqa | [{"from": "human", "value": "how many motorbikes are parked next to the road?\n<image>"}, {"from": "gpt", "value": "there are two motorbikes parked next to the road"}] |
00000004678.jpg | 4678_evjvqa | [{"from": "human", "value": "how many girls are standing behind the product desk?\n<image>"}, {"from": "gpt", "value": "there are three girls"}] |
00000004678.jpg | 4678_evjvqa | [{"from": "human", "value": "how many products is the guy holding?\n<image>"}, {"from": "gpt", "value": "the guy is holding two products"}] |
00000004678.jpg | 4678_evjvqa | [{"from": "human", "value": "which side of girls does the boy stand?\n<image>"}, {"from": "gpt", "value": "standing to the left of the girls"}] |
00000004683.jpg | 4683_evjvqa | [{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a gallery of paintings"}] |
00000004683.jpg | 4683_evjvqa | [{"from": "human", "value": "how many chairs are placed in the room?\n<image>"}, {"from": "gpt", "value": "there is a chair placed in the room"}] |
00000004683.jpg | 4683_evjvqa | [{"from": "human", "value": "how many paintings are on display in the room?\n<image>"}, {"from": "gpt", "value": "there are three paintings on display in the room"}] |
00000004690.jpg | 4690_evjvqa | [{"from": "human", "value": "how many canoes are there on the river?\n<image>"}, {"from": "gpt", "value": "there are two canoes on the river"}] |
00000004690.jpg | 4690_evjvqa | [{"from": "human", "value": "how many people are driving the canoe?\n<image>"}, {"from": "gpt", "value": "there are four people driving the canoe"}] |
00000004690.jpg | 4690_evjvqa | [{"from": "human", "value": "how many people are on each boat?\n<image>"}, {"from": "gpt", "value": "there are two people on each boat"}] |
00000004691.jpg | 4691_evjvqa | [{"from": "human", "value": "how many people are rowing?\n<image>"}, {"from": "gpt", "value": "there are two people rowing"}] |
00000004691.jpg | 4691_evjvqa | [{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a river in the woods"}] |
00000004691.jpg | 4691_evjvqa | [{"from": "human", "value": "which side is the person behind the boat rowing towards?\n<image>"}, {"from": "gpt", "value": "swinging the paddle to the right"}] |
00000004694.jpg | 4694_evjvqa | [{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a market"}] |
00000004694.jpg | 4694_evjvqa | [{"from": "human", "value": "in which hand is the polka dot woman holding the bag?\n<image>"}, {"from": "gpt", "value": "holding the bag in her left hand"}] |
00000004694.jpg | 4694_evjvqa | [{"from": "human", "value": "how many children are walking in the market?\n<image>"}, {"from": "gpt", "value": "there are two children walking in the market"}] |
00000004695.jpg | 4695_evjvqa | [{"from": "human", "value": "how many people are standing at the sweet potato stand?\n<image>"}, {"from": "gpt", "value": "there are three people standing at the sweet potato stand"}] |
00000004695.jpg | 4695_evjvqa | [{"from": "human", "value": "what area are the three people standing in?\n<image>"}, {"from": "gpt", "value": "three people are standing at the supermarket sweet potato stand"}] |
00000004696.jpg | 4696_evjvqa | [{"from": "human", "value": "how many girls are looking at the phone?\n<image>"}, {"from": "gpt", "value": "there are two girls looking at the phone"}] |
00000004696.jpg | 4696_evjvqa | [{"from": "human", "value": "where are the two girls sitting?\n<image>"}, {"from": "gpt", "value": "two girls are sitting in a pub"}] |
00000004696.jpg | 4696_evjvqa | [{"from": "human", "value": "what color shirt does the person sitting on the left of the two girls wear?\n<image>"}, {"from": "gpt", "value": "wear a red shirt"}] |
00000004697.jpg | 4697_evjvqa | [{"from": "human", "value": "how many people are sitting in front of the house?\n<image>"}, {"from": "gpt", "value": "there is a person sitting in front of the house"}] |
00000004697.jpg | 4697_evjvqa | [{"from": "human", "value": "what color are the students' coat uniforms?\n<image>"}, {"from": "gpt", "value": "the coat uniform is blue"}] |
00000004697.jpg | 4697_evjvqa | [{"from": "human", "value": "how many lanterns are hanging behind the man?\n<image>"}, {"from": "gpt", "value": "there are three lanterns hanging"}] |
00000004699.jpg | 4699_evjvqa | [{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a market"}] |
00000004699.jpg | 4699_evjvqa | [{"from": "human", "value": "what color is the woman's apron?\n<image>"}, {"from": "gpt", "value": "the apron is yellow"}] |
00000004699.jpg | 4699_evjvqa | [{"from": "human", "value": "what color shirt is the child standing next to the clothes stall wearing?\n<image>"}, {"from": "gpt", "value": "wearing a white t-shirt"}] |
00000009601.jpg | 9601_evjvqa | [{"from": "human", "value": "how many tiers does the left shelf have?\n<image>"}, {"from": "gpt", "value": "the left shelf has four tiers"}] |
00000009603.jpg | 9603_evjvqa | [{"from": "human", "value": "where is this vegetable stall located?\n<image>"}, {"from": "gpt", "value": "at the corner of the street"}] |
00000009605.jpg | 9605_evjvqa | [{"from": "human", "value": "which side of the bus is the truck running?\n<image>"}, {"from": "gpt", "value": "the truck is running behind and on the left of the bus"}] |
00000009607.jpg | 9607_evjvqa | [{"from": "human", "value": "what color shirt is the woman without eyeglasses wearing?\n<image>"}, {"from": "gpt", "value": "the woman without eyeglasses is wearing a green shirt"}] |
00000009609.jpg | 9609_evjvqa | [{"from": "human", "value": "how does the announcement board look like?\n<image>"}, {"from": "gpt", "value": "the announcement board is rectangle, has red background and white text"}] |
00000009610.jpg | 9610_evjvqa | [{"from": "human", "value": "how many cups are held in the hands?\n<image>"}, {"from": "gpt", "value": "there are three cups held in the hands"}] |
00000009612.jpg | 9612_evjvqa | [{"from": "human", "value": "what color are the phoenix flowers on the branches?\n<image>"}, {"from": "gpt", "value": "the phoenix flowers on the branches are red"}] |
00000009613.jpg | 9613_evjvqa | [{"from": "human", "value": "what is the man wearing black shirt and glasses holding in his right hand?\n<image>"}, {"from": "gpt", "value": "he is holding a smartphone in his right hand"}] |
00000009614.jpg | 9614_evjvqa | [{"from": "human", "value": "what time of the day is it in this city area?\n<image>"}, {"from": "gpt", "value": "it is evening in this city area "}] |
00000009617.jpg | 9617_evjvqa | [{"from": "human", "value": "where is the man wearing red shirt standing?\n<image>"}, {"from": "gpt", "value": "the man wearing red shirt is standing in the leftmost position"}] |
00000009620.jpg | 9620_evjvqa | [{"from": "human", "value": "what shape is the yellow lantern?\n<image>"}, {"from": "gpt", "value": "the yellow lantern is a sphere"}] |
00000009622.jpg | 9622_evjvqa | [{"from": "human", "value": "how many joss sticks are sticked on the green ceremony table?\n<image>"}, {"from": "gpt", "value": "there are three joss sticks sticked on the green ceremony table"}] |
00000009631.jpg | 9631_evjvqa | [{"from": "human", "value": "where is the flag hanged compared to the entrance?\n<image>"}, {"from": "gpt", "value": "the flag is hanged to the left of the entrance"}] |
00000009632.jpg | 9632_evjvqa | [{"from": "human", "value": "what posture is the young man in white shirt in?\n<image>"}, {"from": "gpt", "value": "the young man in white shirt is in sitting posture"}] |
00000009633.jpg | 9633_evjvqa | [{"from": "human", "value": "what is the man wearing floppy hat doing?\n<image>"}, {"from": "gpt", "value": "he is carrying stuffs"}] |
00000009637.jpg | 9637_evjvqa | [{"from": "human", "value": "what do the roadside stalls sell?\n<image>"}, {"from": "gpt", "value": "clothing"}] |
00000009640.jpg | 9640_evjvqa | [{"from": "human", "value": "what is the woman in black shirt doing?\n<image>"}, {"from": "gpt", "value": "she is putting stuffs into the cart"}] |
00000009641.jpg | 9641_evjvqa | [{"from": "human", "value": "what is the sauce put in?\n<image>"}, {"from": "gpt", "value": "the sauce is put in a yellow bowl"}] |
00000009642.jpg | 9642_evjvqa | [{"from": "human", "value": "how many loafs of bread are on the plate?\n<image>"}, {"from": "gpt", "value": "there are two loaf of bread on the plate"}] |
00000009643.jpg | 9643_evjvqa | [{"from": "human", "value": "where can the red lantern be seen?\n<image>"}, {"from": "gpt", "value": "the red latern is hanged on the gate of the house on the left"}] |
00000009648.jpg | 9648_evjvqa | [{"from": "human", "value": "where is the inox chair?\n<image>"}, {"from": "gpt", "value": "the inox chair is behind the woman"}] |
00000009649.jpg | 9649_evjvqa | [{"from": "human", "value": "what dress is a person between two other people in black shirt wearing?\n<image>"}, {"from": "gpt", "value": "between two people in black shirt is a person in red ao dai"}] |
00000009650.jpg | 9650_evjvqa | [{"from": "human", "value": "how many men are on the stage?\n<image>"}, {"from": "gpt", "value": "there are five men on the stage"}] |
00000009651.jpg | 9651_evjvqa | [{"from": "human", "value": "where is the girl in black dress standing?\n<image>"}, {"from": "gpt", "value": "she is standing on the steps in front of the store"}] |
00000009654.jpg | 9654_evjvqa | [{"from": "human", "value": "what is this man doing?\n<image>"}, {"from": "gpt", "value": "he is pushing the cart"}] |
00000009656.jpg | 9656_evjvqa | [{"from": "human", "value": "what are these stalls selling?\n<image>"}, {"from": "gpt", "value": "selling clothing and fabric"}] |
00000009659.jpg | 9659_evjvqa | [{"from": "human", "value": "what color suit is the man wearing?\n<image>"}, {"from": "gpt", "value": "the man is wearing white suit"}] |
00000009660.jpg | 9660_evjvqa | [{"from": "human", "value": "what is the woman sitting in front of the black scooter selling?\n<image>"}, {"from": "gpt", "value": "she is selling fruit"}] |
00000009664.jpg | 9664_evjvqa | [{"from": "human", "value": "where is the kid in yellow shirt sitting?\n<image>"}, {"from": "gpt", "value": "the kid in yellow shirt is sitting in a cart"}] |
00000009667.jpg | 9667_evjvqa | [{"from": "human", "value": "how many motorbikes are parked on the sidewalk?\n<image>"}, {"from": "gpt", "value": "there are three motorbikes parked on the sidewalk"}] |
00000009669.jpg | 9669_evjvqa | [{"from": "human", "value": "how many christmas trees are displayed in this store?\n<image>"}, {"from": "gpt", "value": "there are three christmas trees displayed in this store"}] |
00000009670.jpg | 9670_evjvqa | [{"from": "human", "value": "what color is the scooter running next to the traffic light?\n<image>"}, {"from": "gpt", "value": "that scooter is brown"}] |
00000009671.jpg | 9671_evjvqa | [{"from": "human", "value": "what color handbag is the woman in yellow shirt wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing a red handbag"}] |
00000009672.jpg | 9672_evjvqa | [{"from": "human", "value": "what are the man wearing on his face?\n<image>"}, {"from": "gpt", "value": "a face mask and a mic"}] |
00000009673.jpg | 9673_evjvqa | [{"from": "human", "value": "what color shirt is the woman selecting vegetable wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing green shirt"}] |
00000009674.jpg | 9674_evjvqa | [{"from": "human", "value": "how does the flag hanged on the power pole look like?\n<image>"}, {"from": "gpt", "value": "the flag hanged on the power pole is rectangular, has red background and a yellow star"}] |
00000009677.jpg | 9677_evjvqa | [{"from": "human", "value": "what is this group of people coming here for?\n<image>"}, {"from": "gpt", "value": "for a tour"}] |
00000009683.jpg | 9683_evjvqa | [{"from": "human", "value": "how many men are wearing black vest?\n<image>"}, {"from": "gpt", "value": "there are three men wearing black vest"}] |
00000009685.jpg | 9685_evjvqa | [{"from": "human", "value": "where is this couple standing?\n<image>"}, {"from": "gpt", "value": "they are standing in front of the temple gate"}] |
00000009686.jpg | 9686_evjvqa | [{"from": "human", "value": "what color ao dai is the girl wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing red ao dai"}] |
00000009687.jpg | 9687_evjvqa | [{"from": "human", "value": "what color are the walls of this building painted?\n<image>"}, {"from": "gpt", "value": "the walls of this building are painted yellow"}] |
00000009689.jpg | 9689_evjvqa | [{"from": "human", "value": "what is this road covered by?\n<image>"}, {"from": "gpt", "value": "this road is covered by the canopy"}] |
00000009690.jpg | 9690_evjvqa | [{"from": "human", "value": "what are hanged on the balcony and on the entrance?\n<image>"}, {"from": "gpt", "value": "the red lanterns"}] |
00000009691.jpg | 9691_evjvqa | [{"from": "human", "value": "what color tiles is the floor of this grocery store tiled with?\n<image>"}, {"from": "gpt", "value": "the floor of this grocery store is tiled with white tiles"}] |
00000009694.jpg | 9694_evjvqa | [{"from": "human", "value": "how many pigeons are flapping their wings?\n<image>"}, {"from": "gpt", "value": "there are four pigeons flapping their wings"}] |
00000009695.jpg | 9695_evjvqa | [{"from": "human", "value": "what is the woman in red shirt doing while talking?\n<image>"}, {"from": "gpt", "value": "she is talking while listening to the cellphone"}] |
00000009697.jpg | 9697_evjvqa | [{"from": "human", "value": "what color cap is the man wearing?\n<image>"}, {"from": "gpt", "value": "he is wearing a pink cap"}] |
00000009700.jpg | 9700_evjvqa | [{"from": "human", "value": "what color is the scooter running last?\n<image>"}, {"from": "gpt", "value": "that scooter is blue"}] |
00000009702.jpg | 9702_evjvqa | [{"from": "human", "value": "who are standing by the street?\n<image>"}, {"from": "gpt", "value": "the firefighters are standing by the street"}] |
00000009705.jpg | 9705_evjvqa | [{"from": "human", "value": "what color are the letters on the signboard of the shop?\n<image>"}, {"from": "gpt", "value": "the letters on the signboard of the shop are white"}] |
00000009708.jpg | 9708_evjvqa | [{"from": "human", "value": "what is portrayed on the poster?\n<image>"}, {"from": "gpt", "value": "uncle ho holding a child"}] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.