image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000004329.jpg
|
4329_evjvqa
|
[{"from": "human", "value": "what is this large yard area used for?\n<image>"}, {"from": "gpt", "value": "to walk"}]
|
00000004332.jpg
|
4332_evjvqa
|
[{"from": "human", "value": "how many employees are going out of the store?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000004337.jpg
|
4337_evjvqa
|
[{"from": "human", "value": "what item is this man going to get?\n<image>"}, {"from": "gpt", "value": "the box of prepared food"}]
|
00000004337.jpg
|
4337_evjvqa
|
[{"from": "human", "value": "which stall is this man shopping at?\n<image>"}, {"from": "gpt", "value": "the ready-to-eat food stall"}]
|
00000004337.jpg
|
4337_evjvqa
|
[{"from": "human", "value": "where is the man going shopping?\n<image>"}, {"from": "gpt", "value": "in the supermarket"}]
|
00000004341.jpg
|
4341_evjvqa
|
[{"from": "human", "value": "where is the girl sitting?\n<image>"}, {"from": "gpt", "value": "on the pedestal of a wall"}]
|
00000004341.jpg
|
4341_evjvqa
|
[{"from": "human", "value": "what do people put on the wall?\n<image>"}, {"from": "gpt", "value": "there are three mirrors"}]
|
00000004342.jpg
|
4342_evjvqa
|
[{"from": "human", "value": "what is the man's bike riding on the right hand side selling?\n<image>"}, {"from": "gpt", "value": "selling all kinds of handicrafts made from bamboo"}]
|
00000004342.jpg
|
4342_evjvqa
|
[{"from": "human", "value": "what does the woman in a conical hat driving a bicycle on the left sell?\n<image>"}, {"from": "gpt", "value": "plastic utensils"}]
|
00000004342.jpg
|
4342_evjvqa
|
[{"from": "human", "value": "what hats do cyclists wear?\n<image>"}, {"from": "gpt", "value": "the conical hats"}]
|
00000004342.jpg
|
4342_evjvqa
|
[{"from": "human", "value": "how many people are riding bicycle?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000004346.jpg
|
4346_evjvqa
|
[{"from": "human", "value": "what store is the woman in the black hat standing at?\n<image>"}, {"from": "gpt", "value": "the clothes store"}]
|
00000004346.jpg
|
4346_evjvqa
|
[{"from": "human", "value": "how many people are visiting the clothes store?\n<image>"}, {"from": "gpt", "value": "there is the woman in black hat"}]
|
00000004346.jpg
|
4346_evjvqa
|
[{"from": "human", "value": "what item is the woman going to see?\n<image>"}, {"from": "gpt", "value": "clothes"}]
|
00000004347.jpg
|
4347_evjvqa
|
[{"from": "human", "value": "how is the street decorated?\n<image>"}, {"from": "gpt", "value": "the lanterns"}]
|
00000004347.jpg
|
4347_evjvqa
|
[{"from": "human", "value": "where are these lanterns decorated?\n<image>"}, {"from": "gpt", "value": "in the upper space of the pedestrian street"}]
|
00000004347.jpg
|
4347_evjvqa
|
[{"from": "human", "value": "how many people are using the conical hat?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000004349.jpg
|
4349_evjvqa
|
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "he is speaking at a conference"}]
|
00000004350.jpg
|
4350_evjvqa
|
[{"from": "human", "value": "what are children doing?\n<image>"}, {"from": "gpt", "value": "dancing and singing"}]
|
00000004351.jpg
|
4351_evjvqa
|
[{"from": "human", "value": "what is the woman in red doing?\n<image>"}, {"from": "gpt", "value": "choosing fruit"}]
|
00000004351.jpg
|
4351_evjvqa
|
[{"from": "human", "value": "how many people are standing at the fruit stall?\n<image>"}, {"from": "gpt", "value": "five"}]
|
00000004351.jpg
|
4351_evjvqa
|
[{"from": "human", "value": "who is the boy sitting in the trolley going in the supermarket with?\n<image>"}, {"from": "gpt", "value": "the woman in the white shirt"}]
|
00000004351.jpg
|
4351_evjvqa
|
[{"from": "human", "value": "how many people are standing nearby wearing masks?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000004354.jpg
|
4354_evjvqa
|
[{"from": "human", "value": "what is the person doing?\n<image>"}, {"from": "gpt", "value": "taking the medicine from the vaccine vial"}]
|
00000004356.jpg
|
4356_evjvqa
|
[{"from": "human", "value": "what are these two women doing?\n<image>"}, {"from": "gpt", "value": "they are peeling the fruit"}]
|
00000004356.jpg
|
4356_evjvqa
|
[{"from": "human", "value": "what do these two women do?\n<image>"}, {"from": "gpt", "value": "selling fruit"}]
|
00000004356.jpg
|
4356_evjvqa
|
[{"from": "human", "value": "where are these two women sitting?\n<image>"}, {"from": "gpt", "value": "at the fruit stall"}]
|
00000004358.jpg
|
4358_evjvqa
|
[{"from": "human", "value": "what is the woman in the green shirt doing?\n<image>"}, {"from": "gpt", "value": "checking the box of goods"}]
|
00000004358.jpg
|
4358_evjvqa
|
[{"from": "human", "value": "who is the man in the green shirt sitting with?\n<image>"}, {"from": "gpt", "value": "the woman in the green shirt"}]
|
00000004358.jpg
|
4358_evjvqa
|
[{"from": "human", "value": "how many people are sitting around the table?\n<image>"}, {"from": "gpt", "value": "three "}]
|
00000004360.jpg
|
4360_evjvqa
|
[{"from": "human", "value": "how many beds are there in this room?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000004362.jpg
|
4362_evjvqa
|
[{"from": "human", "value": "how many teachers take these children to visit historical monument?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000004363.jpg
|
4363_evjvqa
|
[{"from": "human", "value": "what is the woman have the son sitting in the stroller buying?\n<image>"}, {"from": "gpt", "value": "american beef"}]
|
00000004364.jpg
|
4364_evjvqa
|
[{"from": "human", "value": "what job does this man do?\n<image>"}, {"from": "gpt", "value": "he is a worker cleaning the streets"}]
|
00000004364.jpg
|
4364_evjvqa
|
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "he is pushing a garbage truck on the street"}]
|
00000004364.jpg
|
4364_evjvqa
|
[{"from": "human", "value": "where is the man going?\n<image>"}, {"from": "gpt", "value": "he is cleaning the streets"}]
|
00000004367.jpg
|
4367_evjvqa
|
[{"from": "human", "value": "how many people riding motorcycles wearing the helmets?\n<image>"}, {"from": "gpt", "value": "only the man"}]
|
00000004367.jpg
|
4367_evjvqa
|
[{"from": "human", "value": "what color shirt does the man riding a motorcycle wearing a helmet wear?\n<image>"}, {"from": "gpt", "value": "white"}]
|
00000004369.jpg
|
4369_evjvqa
|
[{"from": "human", "value": "what is the girl in pink ao dai holding in her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a bouquet of fresh flowers"}]
|
00000004369.jpg
|
4369_evjvqa
|
[{"from": "human", "value": "where is the girl holding a bouquet of flowers standing?\n<image>"}, {"from": "gpt", "value": "she is standing on a busy street"}]
|
00000004377.jpg
|
4377_evjvqa
|
[{"from": "human", "value": "where does this water stall sell?\n<image>"}, {"from": "gpt", "value": "on the railway"}]
|
00000004377.jpg
|
4377_evjvqa
|
[{"from": "human", "value": "where are these people sitting and drinking water?\n<image>"}, {"from": "gpt", "value": "on the railway "}]
|
00000004379.jpg
|
4379_evjvqa
|
[{"from": "human", "value": "where is the girl in black shirt standing?\n<image>"}, {"from": "gpt", "value": "on the book street"}]
|
00000004380.jpg
|
4380_evjvqa
|
[{"from": "human", "value": "what items does this stall sell?\n<image>"}, {"from": "gpt", "value": "types of scarve"}]
|
00000004380.jpg
|
4380_evjvqa
|
[{"from": "human", "value": "how many people come to buy scarves?\n<image>"}, {"from": "gpt", "value": "three"}]
|
00000004380.jpg
|
4380_evjvqa
|
[{"from": "human", "value": "what is the woman selling scarves doing?\n<image>"}, {"from": "gpt", "value": "returning excess money to customers"}]
|
00000004380.jpg
|
4380_evjvqa
|
[{"from": "human", "value": "what are these three women buying?\n<image>"}, {"from": "gpt", "value": "scarves"}]
|
00000004381.jpg
|
4381_evjvqa
|
[{"from": "human", "value": "what is the boy in orange doing?\n<image>"}, {"from": "gpt", "value": "he is walking the dog"}]
|
00000004381.jpg
|
4381_evjvqa
|
[{"from": "human", "value": "how many motorbikes are running behind the boy?\n<image>"}, {"from": "gpt", "value": "only one"}]
|
00000004381.jpg
|
4381_evjvqa
|
[{"from": "human", "value": "who is the boy in the orange shirt walking with?\n<image>"}, {"from": "gpt", "value": "the dog"}]
|
00000004381.jpg
|
4381_evjvqa
|
[{"from": "human", "value": "on which side of the road do people park their cars?\n<image>"}, {"from": "gpt", "value": "on the left"}]
|
00000004382.jpg
|
4382_evjvqa
|
[{"from": "human", "value": "what is this man carrying on his back?\n<image>"}, {"from": "gpt", "value": "the guitar"}]
|
00000004382.jpg
|
4382_evjvqa
|
[{"from": "human", "value": "what is the man doing? \n<image>"}, {"from": "gpt", "value": "he is burning incense and worshiping"}]
|
00000004382.jpg
|
4382_evjvqa
|
[{"from": "human", "value": "where is this man going sightseeing?\n<image>"}, {"from": "gpt", "value": "in the ancient pagoda"}]
|
00000004389.jpg
|
4389_evjvqa
|
[{"from": "human", "value": "how many ships are parked on the water surface?\n<image>"}, {"from": "gpt", "value": "two"}]
|
00000004395.jpg
|
4395_evjvqa
|
[{"from": "human", "value": "what do people do at these stalls?\n<image>"}, {"from": "gpt", "value": "to pay"}]
|
00000009800.jpg
|
9800_evjvqa
|
[{"from": "human", "value": "how many people are walking on the street?\n<image>"}, {"from": "gpt", "value": "there is one person walking on the street"}]
|
00000009801.jpg
|
9801_evjvqa
|
[{"from": "human", "value": "how many people are standing in front of the blue house?\n<image>"}, {"from": "gpt", "value": "there is one person standing in front of the blue house"}]
|
00000009807.jpg
|
9807_evjvqa
|
[{"from": "human", "value": "what is above the painting?\n<image>"}, {"from": "gpt", "value": "above the painting is an air conditioner"}]
|
00000009808.jpg
|
9808_evjvqa
|
[{"from": "human", "value": "what is the man doing?\n<image>"}, {"from": "gpt", "value": "the man is watching a medicine bottle on the hand of the girl"}]
|
00000009809.jpg
|
9809_evjvqa
|
[{"from": "human", "value": "what is the main building material of this house?\n<image>"}, {"from": "gpt", "value": "the main building material of this house is wood"}]
|
00000009811.jpg
|
9811_evjvqa
|
[{"from": "human", "value": "what is the girl sitting leaning against?\n<image>"}, {"from": "gpt", "value": "the girl is sitting leaning against the lamppost"}]
|
00000009812.jpg
|
9812_evjvqa
|
[{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a floating market"}]
|
00000009813.jpg
|
9813_evjvqa
|
[{"from": "human", "value": "where is the woman sitting?\n<image>"}, {"from": "gpt", "value": "the woman is sitting at a seafood stall"}]
|
00000009814.jpg
|
9814_evjvqa
|
[{"from": "human", "value": "how many buses are in front of the motorbike taxi driver?\n<image>"}, {"from": "gpt", "value": "there is one bus in front of the motorbike taxi driver"}]
|
00000009818.jpg
|
9818_evjvqa
|
[{"from": "human", "value": "what color are the light bulbs on the tower by the lake?\n<image>"}, {"from": "gpt", "value": "the light bulbs are yellow"}]
|
00000009819.jpg
|
9819_evjvqa
|
[{"from": "human", "value": "how many people are shopping in the supermarket?\n<image>"}, {"from": "gpt", "value": "there are three people shopping in the supermarket"}]
|
00000009820.jpg
|
9820_evjvqa
|
[{"from": "human", "value": "what is the man holding in his hand?\n<image>"}, {"from": "gpt", "value": "the man is holding an apple in his hand"}]
|
00000009822.jpg
|
9822_evjvqa
|
[{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a war museum"}]
|
00000009824.jpg
|
9824_evjvqa
|
[{"from": "human", "value": "what color is the taxi?\n<image>"}, {"from": "gpt", "value": "the taxi is brown"}]
|
00000009825.jpg
|
9825_evjvqa
|
[{"from": "human", "value": "what color shirt is the man wearing?\n<image>"}, {"from": "gpt", "value": "the man is wearing yellow shirt"}]
|
00000009826.jpg
|
9826_evjvqa
|
[{"from": "human", "value": "how many people are shopping?\n<image>"}, {"from": "gpt", "value": "there are two people shopping"}]
|
00000009827.jpg
|
9827_evjvqa
|
[{"from": "human", "value": "how many paintings are hanged on the wall?\n<image>"}, {"from": "gpt", "value": "there are two paintings hanged on the wall"}]
|
00000009828.jpg
|
9828_evjvqa
|
[{"from": "human", "value": "what color is the truck parked on the roadside?\n<image>"}, {"from": "gpt", "value": "the truck is red"}]
|
00000009832.jpg
|
9832_evjvqa
|
[{"from": "human", "value": "where is the man standing next to the cyclo looking at?\n<image>"}, {"from": "gpt", "value": "the man is looking at a building across the street"}]
|
00000009833.jpg
|
9833_evjvqa
|
[{"from": "human", "value": "what scenery does the painting portray?\n<image>"}, {"from": "gpt", "value": "the painting portrays the river countryside scenery"}]
|
00000009835.jpg
|
9835_evjvqa
|
[{"from": "human", "value": "what color shirt is the man wearing?\n<image>"}, {"from": "gpt", "value": "the man is wearing white shirt"}]
|
00000009837.jpg
|
9837_evjvqa
|
[{"from": "human", "value": "how many young girls are being awarded?\n<image>"}, {"from": "gpt", "value": "there are two young girls being awarded"}]
|
00000009839.jpg
|
9839_evjvqa
|
[{"from": "human", "value": "what color shirt is the man standing wearing?\n<image>"}, {"from": "gpt", "value": "the man standing is wearing white shirt"}]
|
00000009841.jpg
|
9841_evjvqa
|
[{"from": "human", "value": "what color is the roof of the pagoda?\n<image>"}, {"from": "gpt", "value": "the roof of the pagoda is red"}]
|
00000009845.jpg
|
9845_evjvqa
|
[{"from": "human", "value": "how many women are in the room?\n<image>"}, {"from": "gpt", "value": "there are two women in the room"}]
|
00000009849.jpg
|
9849_evjvqa
|
[{"from": "human", "value": "what is everyone doing?\n<image>"}, {"from": "gpt", "value": "everyone is visiting a temple"}]
|
00000009852.jpg
|
9852_evjvqa
|
[{"from": "human", "value": "what is the girl holding the cellphone doing?\n<image>"}, {"from": "gpt", "value": "the girl holding the cellphone is using the cellphone to scan the qr code"}]
|
00000009853.jpg
|
9853_evjvqa
|
[{"from": "human", "value": "what stall is the girl standing at?\n<image>"}, {"from": "gpt", "value": "the girl is standing at the dairy stall"}]
|
00000009854.jpg
|
9854_evjvqa
|
[{"from": "human", "value": "how many medical staffs are there?\n<image>"}, {"from": "gpt", "value": "there are two medical staffs"}]
|
00000009857.jpg
|
9857_evjvqa
|
[{"from": "human", "value": "what color is the light of the church?\n<image>"}, {"from": "gpt", "value": "the light of the church is yellow"}]
|
00000009859.jpg
|
9859_evjvqa
|
[{"from": "human", "value": "what stall is on the left of the woman in white shirt?\n<image>"}, {"from": "gpt", "value": "on the left of the woman is durian stall"}]
|
00000009860.jpg
|
9860_evjvqa
|
[{"from": "human", "value": "how many tables are decorated with plant pot?\n<image>"}, {"from": "gpt", "value": "there is one table decorated with plant pot"}]
|
00000009865.jpg
|
9865_evjvqa
|
[{"from": "human", "value": "what is this area selling?\n<image>"}, {"from": "gpt", "value": "this area is selling school supplies"}]
|
00000009869.jpg
|
9869_evjvqa
|
[{"from": "human", "value": "how many bikes are on the right of the young man in purple shirt?\n<image>"}, {"from": "gpt", "value": "there is one bike"}]
|
00000009870.jpg
|
9870_evjvqa
|
[{"from": "human", "value": "what color are the chairs?\n<image>"}, {"from": "gpt", "value": "the chairs are red"}]
|
00000009871.jpg
|
9871_evjvqa
|
[{"from": "human", "value": "how many mannequins are behind the woman?\n<image>"}, {"from": "gpt", "value": "there are four mannequins behind the woman"}]
|
00000009873.jpg
|
9873_evjvqa
|
[{"from": "human", "value": "what is the supermarket staff doing?\n<image>"}, {"from": "gpt", "value": "the supermarket staff is arranging the products"}]
|
00000009874.jpg
|
9874_evjvqa
|
[{"from": "human", "value": "what color shirt is the woman on the right of the little girl wearing?\n<image>"}, {"from": "gpt", "value": "the woman is wearing black shirt"}]
|
00000009875.jpg
|
9875_evjvqa
|
[{"from": "human", "value": "what color shirt is the man riding white motorbike wearing?\n<image>"}, {"from": "gpt", "value": "the man is wearing red shirt"}]
|
00000009876.jpg
|
9876_evjvqa
|
[{"from": "human", "value": "how many people are heading to the book fair?\n<image>"}, {"from": "gpt", "value": "there are two people heading to the book fair"}]
|
00000009877.jpg
|
9877_evjvqa
|
[{"from": "human", "value": "what area is this?\n<image>"}, {"from": "gpt", "value": "this is a market on the shore"}]
|
00000009880.jpg
|
9880_evjvqa
|
[{"from": "human", "value": "what is on the right of the man riding bicycle?\n<image>"}, {"from": "gpt", "value": "on the right of the man riding bicycle is a motorbike"}]
|
00000009881.jpg
|
9881_evjvqa
|
[{"from": "human", "value": "how many tea cups are there?\n<image>"}, {"from": "gpt", "value": "there are two tea cups"}]
|
00000009882.jpg
|
9882_evjvqa
|
[{"from": "human", "value": "what is everone playing?\n<image>"}, {"from": "gpt", "value": "everyone is going on the swing ride"}]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.