image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000001484.jpg | 1484_evjvqa | [{"from": "human", "value": "where does the straight road in the picture lead?\n<image>"}, {"from": "gpt", "value": "on the lake shore"}] |
00000001484.jpg | 1484_evjvqa | [{"from": "human", "value": "how many people are carrying the burden in the picture?\n<image>"}, {"from": "gpt", "value": "just one"}] |
00000001484.jpg | 1484_evjvqa | [{"from": "human", "value": "how many transports are on the road in the picture?\n<image>"}, {"from": "gpt", "value": "there is a cyclo"}] |
00000001485.jpg | 1485_evjvqa | [{"from": "human", "value": "what is supermarket staff doing?\n<image>"}, {"from": "gpt", "value": "sorting goods"}] |
00000001485.jpg | 1485_evjvqa | [{"from": "human", "value": "what color uniform do supermarket staff wear?\n<image>"}, {"from": "gpt", "value": "blue"}] |
00000001496.jpg | 1496_evjvqa | [{"from": "human", "value": "what is the woman wearing conical hat doing?\n<image>"}, {"from": "gpt", "value": "buying fruits"}] |
00000001496.jpg | 1496_evjvqa | [{"from": "human", "value": "what is the car behind the man wearing the black t-shirt and the black hat?\n<image>"}, {"from": "gpt", "value": "carrying fruit"}] |
00000001499.jpg | 1499_evjvqa | [{"from": "human", "value": "how many cows are walking in the middle of the rice fields?\n<image>"}, {"from": "gpt", "value": "two"}] |
00000001499.jpg | 1499_evjvqa | [{"from": "human", "value": "what is the farmer following the cow tending?\n<image>"}, {"from": "gpt", "value": "to plow rice"}] |
00000001499.jpg | 1499_evjvqa | [{"from": "human", "value": "what vehicle are the two women using?\n<image>"}, {"from": "gpt", "value": "bicycle"}] |
00000001499.jpg | 1499_evjvqa | [{"from": "human", "value": "how many people are going with two buffaloes?\n<image>"}, {"from": "gpt", "value": "three"}] |
00000001499.jpg | 1499_evjvqa | [{"from": "human", "value": "how many people are riding bicycle?\n<image>"}, {"from": "gpt", "value": "two"}] |
00000006201.jpg | 6201_evjvqa | [{"from": "human", "value": "what color are the police officers' uniforms?\n<image>"}, {"from": "gpt", "value": "they are green."}] |
00000006205.jpg | 6205_evjvqa | [{"from": "human", "value": "what is she holding?\n<image>"}, {"from": "gpt", "value": "she is holding the frame of her eyeglasses."}] |
00000006205.jpg | 6205_evjvqa | [{"from": "human", "value": "what color are her sneakers?\n<image>"}, {"from": "gpt", "value": "her sneakers are black."}] |
00000006205.jpg | 6205_evjvqa | [{"from": "human", "value": "what color are her pants?\n<image>"}, {"from": "gpt", "value": "her pants are red."}] |
00000006210.jpg | 6210_evjvqa | [{"from": "human", "value": "what is the expression on his face?\n<image>"}, {"from": "gpt", "value": "he is smiling slightly."}] |
00000006210.jpg | 6210_evjvqa | [{"from": "human", "value": "what is she doing?\n<image>"}, {"from": "gpt", "value": "she is floating the flower lanterns. "}] |
00000006210.jpg | 6210_evjvqa | [{"from": "human", "value": "what colors are the flower lanterns in his hands?\n<image>"}, {"from": "gpt", "value": "he is holding a red flower lantern and a green one. "}] |
00000006212.jpg | 6212_evjvqa | [{"from": "human", "value": "what color is the man with hands on hips' t-shirt?\n<image>"}, {"from": "gpt", "value": "the man with hands on hips is wearing a white t-shirt."}] |
00000006212.jpg | 6212_evjvqa | [{"from": "human", "value": "what color is the man in white's cap?\n<image>"}, {"from": "gpt", "value": "his cap is blue."}] |
00000006212.jpg | 6212_evjvqa | [{"from": "human", "value": "what color is the t-shirt worn by the man with the blue cap?\n<image>"}, {"from": "gpt", "value": "his t-shirt is white."}] |
00000006219.jpg | 6219_evjvqa | [{"from": "human", "value": "what color is the gas station attendant's uniform?\n<image>"}, {"from": "gpt", "value": "the uniform is blue."}] |
00000006219.jpg | 6219_evjvqa | [{"from": "human", "value": "how many people are standing at the gas station?\n<image>"}, {"from": "gpt", "value": "there are three people standing at the gas station."}] |
00000006222.jpg | 6222_evjvqa | [{"from": "human", "value": "what kind of animal is the couple photographed with?\n<image>"}, {"from": "gpt", "value": "goat. "}] |
00000006222.jpg | 6222_evjvqa | [{"from": "human", "value": "what are the expressions on the man and woman's faces?\n<image>"}, {"from": "gpt", "value": "they are smiling happily."}] |
00000006222.jpg | 6222_evjvqa | [{"from": "human", "value": "what color is the man' shirt?\n<image>"}, {"from": "gpt", "value": "it is brown. "}] |
00000006223.jpg | 6223_evjvqa | [{"from": "human", "value": "what animal is the girl in black holding?\n<image>"}, {"from": "gpt", "value": "she is holding an octopus."}] |
00000006223.jpg | 6223_evjvqa | [{"from": "human", "value": "what color is the man's shirt?\n<image>"}, {"from": "gpt", "value": "it is red."}] |
00000006223.jpg | 6223_evjvqa | [{"from": "human", "value": "what is the woman in green doing?\n<image>"}, {"from": "gpt", "value": "the woman in green is selling seafood."}] |
00000006227.jpg | 6227_evjvqa | [{"from": "human", "value": "what is the girl in plaid shirt doing?\n<image>"}, {"from": "gpt", "value": "she is selecting vegetables."}] |
00000006227.jpg | 6227_evjvqa | [{"from": "human", "value": "how many women in red are selecting vegetables?\n<image>"}, {"from": "gpt", "value": "there is one woman in red selecting vegetables."}] |
00000006232.jpg | 6232_evjvqa | [{"from": "human", "value": "what is she doing?\n<image>"}, {"from": "gpt", "value": "she is sitting on the motorbike."}] |
00000006232.jpg | 6232_evjvqa | [{"from": "human", "value": "what color is her ao dai?\n<image>"}, {"from": "gpt", "value": "her ao dai is white with green patterns."}] |
00000006232.jpg | 6232_evjvqa | [{"from": "human", "value": "what color is the bouquet of flowers that she is holding?\n<image>"}, {"from": "gpt", "value": "she is holding a bouquet of yellow flowers."}] |
00000006235.jpg | 6235_evjvqa | [{"from": "human", "value": "what color is his shirt?\n<image>"}, {"from": "gpt", "value": "his shirt is pink."}] |
00000006235.jpg | 6235_evjvqa | [{"from": "human", "value": "what is the color of the shirt that the man is wearing?\n<image>"}, {"from": "gpt", "value": "it is pink. "}] |
00000006235.jpg | 6235_evjvqa | [{"from": "human", "value": "what color are his jeans?\n<image>"}, {"from": "gpt", "value": "he is wearing dark grey jeans."}] |
00000006236.jpg | 6236_evjvqa | [{"from": "human", "value": "what is the man in white doing?\n<image>"}, {"from": "gpt", "value": "he is aligning the fish."}] |
00000006236.jpg | 6236_evjvqa | [{"from": "human", "value": "what color is the man aligning the fish's outfit?\n<image>"}, {"from": "gpt", "value": "it is white."}] |
00000006236.jpg | 6236_evjvqa | [{"from": "human", "value": "how many men in white outfit are there?\n<image>"}, {"from": "gpt", "value": "there is one man in white outfit."}] |
00000006237.jpg | 6237_evjvqa | [{"from": "human", "value": "what is the dominant color of the store?\n<image>"}, {"from": "gpt", "value": "it is black."}] |
00000006237.jpg | 6237_evjvqa | [{"from": "human", "value": "what color are the letters?\n<image>"}, {"from": "gpt", "value": "they are red. "}] |
00000006238.jpg | 6238_evjvqa | [{"from": "human", "value": "what color is the shirt worn by the man driving the blue motorbike on the street?\n<image>"}, {"from": "gpt", "value": "it is red. "}] |
00000006238.jpg | 6238_evjvqa | [{"from": "human", "value": "what color is the man in red's motorbike?\n<image>"}, {"from": "gpt", "value": "it is blue."}] |
00000006242.jpg | 6242_evjvqa | [{"from": "human", "value": "how many motorbikes are running on the road?\n<image>"}, {"from": "gpt", "value": "there are two motorbikes running on the road."}] |
00000006242.jpg | 6242_evjvqa | [{"from": "human", "value": "how many traffic signs are there on the side of the road?\n<image>"}, {"from": "gpt", "value": "there are two traffic signs."}] |
00000006242.jpg | 6242_evjvqa | [{"from": "human", "value": "what colors are the poles of the traffic signs?\n<image>"}, {"from": "gpt", "value": "the poles are red and white. "}] |
00000006245.jpg | 6245_evjvqa | [{"from": "human", "value": "what is the woman in white doing?\n<image>"}, {"from": "gpt", "value": "she is putting on the bracelet for the young girl. "}] |
00000006245.jpg | 6245_evjvqa | [{"from": "human", "value": "what color is the young girl with the bracelet dressed in? \n<image>"}, {"from": "gpt", "value": "she is dressed in orange. "}] |
00000006248.jpg | 6248_evjvqa | [{"from": "human", "value": "how many living statues are there?\n<image>"}, {"from": "gpt", "value": "there is one living statue. "}] |
00000006248.jpg | 6248_evjvqa | [{"from": "human", "value": "how many people are wearing the santa claus costume?\n<image>"}, {"from": "gpt", "value": "there is one person wearing santa claus costume."}] |
00000006248.jpg | 6248_evjvqa | [{"from": "human", "value": "what is santa claus making?\n<image>"}, {"from": "gpt", "value": "he is making balloon animals."}] |
00000006253.jpg | 6253_evjvqa | [{"from": "human", "value": "what are the two young people doing?\n<image>"}, {"from": "gpt", "value": "they are holding hands and looking at each other. "}] |
00000006253.jpg | 6253_evjvqa | [{"from": "human", "value": "what color are their shirts?\n<image>"}, {"from": "gpt", "value": "their shirts are white."}] |
00000006253.jpg | 6253_evjvqa | [{"from": "human", "value": "what color are the boy's pants?\n<image>"}, {"from": "gpt", "value": "his pants are black."}] |
00000006254.jpg | 6254_evjvqa | [{"from": "human", "value": "what color is the female customer dressed in?\n<image>"}, {"from": "gpt", "value": "she is dressed in black."}] |
00000006254.jpg | 6254_evjvqa | [{"from": "human", "value": "what costume is the shopkeeper wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing cao bang province's traditional costume."}] |
00000006255.jpg | 6255_evjvqa | [{"from": "human", "value": "how many ships are sailing on the sea?\n<image>"}, {"from": "gpt", "value": "there are three ships. "}] |
00000006255.jpg | 6255_evjvqa | [{"from": "human", "value": "what is the dominant color of the ships?\n<image>"}, {"from": "gpt", "value": "it is white."}] |
00000006255.jpg | 6255_evjvqa | [{"from": "human", "value": "how many white ships are there on the sea?\n<image>"}, {"from": "gpt", "value": "there are three white ships on the sea. "}] |
00000006257.jpg | 6257_evjvqa | [{"from": "human", "value": "what color is the shirt worn by the girl standing next to the bicycle?\n<image>"}, {"from": "gpt", "value": "it is white."}] |
00000006257.jpg | 6257_evjvqa | [{"from": "human", "value": "how many men in white are walking on the street?\n<image>"}, {"from": "gpt", "value": "there is one man in white walking on the street."}] |
00000006257.jpg | 6257_evjvqa | [{"from": "human", "value": "what is the woman in pink doing?\n<image>"}, {"from": "gpt", "value": "she is walking along her child whom is on a small bicycle. "}] |
00000006260.jpg | 6260_evjvqa | [{"from": "human", "value": "how many motorbikes are parked in front of the store?\n<image>"}, {"from": "gpt", "value": "there is one motorbike parked in front of the store."}] |
00000006262.jpg | 6262_evjvqa | [{"from": "human", "value": "where is she?\n<image>"}, {"from": "gpt", "value": "she is walking on the road under the trees with yellow leaves."}] |
00000006262.jpg | 6262_evjvqa | [{"from": "human", "value": "what color are her shoes?\n<image>"}, {"from": "gpt", "value": "she is wearing a pair of black high heels."}] |
00000006262.jpg | 6262_evjvqa | [{"from": "human", "value": "what color is her scarf?\n<image>"}, {"from": "gpt", "value": "it is white."}] |
00000006265.jpg | 6265_evjvqa | [{"from": "human", "value": "what is he doing?\n<image>"}, {"from": "gpt", "value": "he is fixing stuffs."}] |
00000006265.jpg | 6265_evjvqa | [{"from": "human", "value": "what color is his shirt?\n<image>"}, {"from": "gpt", "value": "it is white."}] |
00000006265.jpg | 6265_evjvqa | [{"from": "human", "value": "how many men who are seated in the picture?\n<image>"}, {"from": "gpt", "value": "there is one man seated."}] |
00000006273.jpg | 6273_evjvqa | [{"from": "human", "value": "what color is the logo of the delivery company?\n<image>"}, {"from": "gpt", "value": "it is red."}] |
00000006275.jpg | 6275_evjvqa | [{"from": "human", "value": "what kind of hat is the woman in purple wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing a conical hat."}] |
00000006275.jpg | 6275_evjvqa | [{"from": "human", "value": "what color is the woman with the conical hat's outfit?\n<image>"}, {"from": "gpt", "value": "it is purple."}] |
00000006278.jpg | 6278_evjvqa | [{"from": "human", "value": "where are they?\n<image>"}, {"from": "gpt", "value": "they are in a museum."}] |
00000006278.jpg | 6278_evjvqa | [{"from": "human", "value": "what is the woman in red doing?\n<image>"}, {"from": "gpt", "value": "she is sitting on a bench."}] |
00000006278.jpg | 6278_evjvqa | [{"from": "human", "value": "how many women in red that are sitting in the museum?\n<image>"}, {"from": "gpt", "value": "there is one woman in red sitting in the museum."}] |
00000006280.jpg | 6280_evjvqa | [{"from": "human", "value": "where is she sitting?\n<image>"}, {"from": "gpt", "value": "she is sitting on the bow of the boat."}] |
00000006280.jpg | 6280_evjvqa | [{"from": "human", "value": "what color is her top?\n<image>"}, {"from": "gpt", "value": "it is blue."}] |
00000006280.jpg | 6280_evjvqa | [{"from": "human", "value": "what color is her skirt?\n<image>"}, {"from": "gpt", "value": "it is pink."}] |
00000006281.jpg | 6281_evjvqa | [{"from": "human", "value": "what is she doing?\n<image>"}, {"from": "gpt", "value": "she is holding a bag of onions on her hands."}] |
00000006281.jpg | 6281_evjvqa | [{"from": "human", "value": "what color is her handbag?\n<image>"}, {"from": "gpt", "value": "her handbag is black."}] |
00000006281.jpg | 6281_evjvqa | [{"from": "human", "value": "what color is the boy' scarf?\n<image>"}, {"from": "gpt", "value": "it is red."}] |
00000006283.jpg | 6283_evjvqa | [{"from": "human", "value": "what character is she dressing as?\n<image>"}, {"from": "gpt", "value": "she is dressing as a witch with a blue cloak."}] |
00000006283.jpg | 6283_evjvqa | [{"from": "human", "value": "what is the man in white doing?\n<image>"}, {"from": "gpt", "value": "he is holding the green witch hat on his hand."}] |
00000006283.jpg | 6283_evjvqa | [{"from": "human", "value": "how many men in blue are there?\n<image>"}, {"from": "gpt", "value": "there is one man in blue."}] |
00000006290.jpg | 6290_evjvqa | [{"from": "human", "value": "what is she doing?\n<image>"}, {"from": "gpt", "value": "she is promoting products."}] |
00000006290.jpg | 6290_evjvqa | [{"from": "human", "value": "what is her facial expresison?\n<image>"}, {"from": "gpt", "value": "she is smiling."}] |
00000006290.jpg | 6290_evjvqa | [{"from": "human", "value": "what color is her shirt?\n<image>"}, {"from": "gpt", "value": "she is wearing a white shirt."}] |
00000006293.jpg | 6293_evjvqa | [{"from": "human", "value": "what color is the banner that they are hanging?\n<image>"}, {"from": "gpt", "value": "it is red."}] |
00000006293.jpg | 6293_evjvqa | [{"from": "human", "value": "how many men in red are there?\n<image>"}, {"from": "gpt", "value": "there is one man in red."}] |
00000006297.jpg | 6297_evjvqa | [{"from": "human", "value": "how many people are drinking at the tables?\n<image>"}, {"from": "gpt", "value": "there are five people drinking at the tables."}] |
00000006297.jpg | 6297_evjvqa | [{"from": "human", "value": "what is the young man in blue doing?\n<image>"}, {"from": "gpt", "value": "he is reading a book."}] |
00000006297.jpg | 6297_evjvqa | [{"from": "human", "value": "how many girls in red are there?\n<image>"}, {"from": "gpt", "value": "there is one girl in red."}] |
00000006298.jpg | 6298_evjvqa | [{"from": "human", "value": "what is he doing?\n<image>"}, {"from": "gpt", "value": "he is cleaning up."}] |
00000006298.jpg | 6298_evjvqa | [{"from": "human", "value": "what is the color of the jacket that he is wearing?\n<image>"}, {"from": "gpt", "value": "it is black."}] |
00000006298.jpg | 6298_evjvqa | [{"from": "human", "value": "what color is his jacket?\n<image>"}, {"from": "gpt", "value": "it is black. "}] |
00000001502.jpg | 1502_evjvqa | [{"from": "human", "value": "who is this girl standing next to?\n<image>"}, {"from": "gpt", "value": "nobody"}] |
00000001502.jpg | 1502_evjvqa | [{"from": "human", "value": "what is the girl standing against the wall wearing at her thigh level?\n<image>"}, {"from": "gpt", "value": "a handbag"}] |
00000001502.jpg | 1502_evjvqa | [{"from": "human", "value": "how is the girl standing against the wall dressed?\n<image>"}, {"from": "gpt", "value": "a light green shirt and a blue pair of jeans"}] |
Subsets and Splits