image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000001484.jpg
1484_evjvqa
[{"from": "human", "value": "where does the straight road in the picture lead?\n<image>"}, {"from": "gpt", "value": "on the lake shore"}]
00000001484.jpg
1484_evjvqa
[{"from": "human", "value": "how many people are carrying the burden in the picture?\n<image>"}, {"from": "gpt", "value": "just one"}]
00000001484.jpg
1484_evjvqa
[{"from": "human", "value": "how many transports are on the road in the picture?\n<image>"}, {"from": "gpt", "value": "there is a cyclo"}]
00000001485.jpg
1485_evjvqa
[{"from": "human", "value": "what is supermarket staff doing?\n<image>"}, {"from": "gpt", "value": "sorting goods"}]
00000001485.jpg
1485_evjvqa
[{"from": "human", "value": "what color uniform do supermarket staff wear?\n<image>"}, {"from": "gpt", "value": "blue"}]
00000001496.jpg
1496_evjvqa
[{"from": "human", "value": "what is the woman wearing conical hat doing?\n<image>"}, {"from": "gpt", "value": "buying fruits"}]
00000001496.jpg
1496_evjvqa
[{"from": "human", "value": "what is the car behind the man wearing the black t-shirt and the black hat?\n<image>"}, {"from": "gpt", "value": "carrying fruit"}]
00000001499.jpg
1499_evjvqa
[{"from": "human", "value": "how many cows are walking in the middle of the rice fields?\n<image>"}, {"from": "gpt", "value": "two"}]
00000001499.jpg
1499_evjvqa
[{"from": "human", "value": "what is the farmer following the cow tending?\n<image>"}, {"from": "gpt", "value": "to plow rice"}]
00000001499.jpg
1499_evjvqa
[{"from": "human", "value": "what vehicle are the two women using?\n<image>"}, {"from": "gpt", "value": "bicycle"}]
00000001499.jpg
1499_evjvqa
[{"from": "human", "value": "how many people are going with two buffaloes?\n<image>"}, {"from": "gpt", "value": "three"}]
00000001499.jpg
1499_evjvqa
[{"from": "human", "value": "how many people are riding bicycle?\n<image>"}, {"from": "gpt", "value": "two"}]
00000006201.jpg
6201_evjvqa
[{"from": "human", "value": "what color are the police officers' uniforms?\n<image>"}, {"from": "gpt", "value": "they are green."}]
00000006205.jpg
6205_evjvqa
[{"from": "human", "value": "what is she holding?\n<image>"}, {"from": "gpt", "value": "she is holding the frame of her eyeglasses."}]
00000006205.jpg
6205_evjvqa
[{"from": "human", "value": "what color are her sneakers?\n<image>"}, {"from": "gpt", "value": "her sneakers are black."}]
00000006205.jpg
6205_evjvqa
[{"from": "human", "value": "what color are her pants?\n<image>"}, {"from": "gpt", "value": "her pants are red."}]
00000006210.jpg
6210_evjvqa
[{"from": "human", "value": "what is the expression on his face?\n<image>"}, {"from": "gpt", "value": "he is smiling slightly."}]
00000006210.jpg
6210_evjvqa
[{"from": "human", "value": "what is she doing?\n<image>"}, {"from": "gpt", "value": "she is floating the flower lanterns. "}]
00000006210.jpg
6210_evjvqa
[{"from": "human", "value": "what colors are the flower lanterns in his hands?\n<image>"}, {"from": "gpt", "value": "he is holding a red flower lantern and a green one. "}]
00000006212.jpg
6212_evjvqa
[{"from": "human", "value": "what color is the man with hands on hips' t-shirt?\n<image>"}, {"from": "gpt", "value": "the man with hands on hips is wearing a white t-shirt."}]
00000006212.jpg
6212_evjvqa
[{"from": "human", "value": "what color is the man in white's cap?\n<image>"}, {"from": "gpt", "value": "his cap is blue."}]
00000006212.jpg
6212_evjvqa
[{"from": "human", "value": "what color is the t-shirt worn by the man with the blue cap?\n<image>"}, {"from": "gpt", "value": "his t-shirt is white."}]
00000006219.jpg
6219_evjvqa
[{"from": "human", "value": "what color is the gas station attendant's uniform?\n<image>"}, {"from": "gpt", "value": "the uniform is blue."}]
00000006219.jpg
6219_evjvqa
[{"from": "human", "value": "how many people are standing at the gas station?\n<image>"}, {"from": "gpt", "value": "there are three people standing at the gas station."}]
00000006222.jpg
6222_evjvqa
[{"from": "human", "value": "what kind of animal is the couple photographed with?\n<image>"}, {"from": "gpt", "value": "goat. "}]
00000006222.jpg
6222_evjvqa
[{"from": "human", "value": "what are the expressions on the man and woman's faces?\n<image>"}, {"from": "gpt", "value": "they are smiling happily."}]
00000006222.jpg
6222_evjvqa
[{"from": "human", "value": "what color is the man' shirt?\n<image>"}, {"from": "gpt", "value": "it is brown. "}]
00000006223.jpg
6223_evjvqa
[{"from": "human", "value": "what animal is the girl in black holding?\n<image>"}, {"from": "gpt", "value": "she is holding an octopus."}]
00000006223.jpg
6223_evjvqa
[{"from": "human", "value": "what color is the man's shirt?\n<image>"}, {"from": "gpt", "value": "it is red."}]
00000006223.jpg
6223_evjvqa
[{"from": "human", "value": "what is the woman in green doing?\n<image>"}, {"from": "gpt", "value": "the woman in green is selling seafood."}]
00000006227.jpg
6227_evjvqa
[{"from": "human", "value": "what is the girl in plaid shirt doing?\n<image>"}, {"from": "gpt", "value": "she is selecting vegetables."}]
00000006227.jpg
6227_evjvqa
[{"from": "human", "value": "how many women in red are selecting vegetables?\n<image>"}, {"from": "gpt", "value": "there is one woman in red selecting vegetables."}]
00000006232.jpg
6232_evjvqa
[{"from": "human", "value": "what is she doing?\n<image>"}, {"from": "gpt", "value": "she is sitting on the motorbike."}]
00000006232.jpg
6232_evjvqa
[{"from": "human", "value": "what color is her ao dai?\n<image>"}, {"from": "gpt", "value": "her ao dai is white with green patterns."}]
00000006232.jpg
6232_evjvqa
[{"from": "human", "value": "what color is the bouquet of flowers that she is holding?\n<image>"}, {"from": "gpt", "value": "she is holding a bouquet of yellow flowers."}]
00000006235.jpg
6235_evjvqa
[{"from": "human", "value": "what color is his shirt?\n<image>"}, {"from": "gpt", "value": "his shirt is pink."}]
00000006235.jpg
6235_evjvqa
[{"from": "human", "value": "what is the color of the shirt that the man is wearing?\n<image>"}, {"from": "gpt", "value": "it is pink. "}]
00000006235.jpg
6235_evjvqa
[{"from": "human", "value": "what color are his jeans?\n<image>"}, {"from": "gpt", "value": "he is wearing dark grey jeans."}]
00000006236.jpg
6236_evjvqa
[{"from": "human", "value": "what is the man in white doing?\n<image>"}, {"from": "gpt", "value": "he is aligning the fish."}]
00000006236.jpg
6236_evjvqa
[{"from": "human", "value": "what color is the man aligning the fish's outfit?\n<image>"}, {"from": "gpt", "value": "it is white."}]
00000006236.jpg
6236_evjvqa
[{"from": "human", "value": "how many men in white outfit are there?\n<image>"}, {"from": "gpt", "value": "there is one man in white outfit."}]
00000006237.jpg
6237_evjvqa
[{"from": "human", "value": "what is the dominant color of the store?\n<image>"}, {"from": "gpt", "value": "it is black."}]
00000006237.jpg
6237_evjvqa
[{"from": "human", "value": "what color are the letters?\n<image>"}, {"from": "gpt", "value": "they are red. "}]
00000006238.jpg
6238_evjvqa
[{"from": "human", "value": "what color is the shirt worn by the man driving the blue motorbike on the street?\n<image>"}, {"from": "gpt", "value": "it is red. "}]
00000006238.jpg
6238_evjvqa
[{"from": "human", "value": "what color is the man in red's motorbike?\n<image>"}, {"from": "gpt", "value": "it is blue."}]
00000006242.jpg
6242_evjvqa
[{"from": "human", "value": "how many motorbikes are running on the road?\n<image>"}, {"from": "gpt", "value": "there are two motorbikes running on the road."}]
00000006242.jpg
6242_evjvqa
[{"from": "human", "value": "how many traffic signs are there on the side of the road?\n<image>"}, {"from": "gpt", "value": "there are two traffic signs."}]
00000006242.jpg
6242_evjvqa
[{"from": "human", "value": "what colors are the poles of the traffic signs?\n<image>"}, {"from": "gpt", "value": "the poles are red and white. "}]
00000006245.jpg
6245_evjvqa
[{"from": "human", "value": "what is the woman in white doing?\n<image>"}, {"from": "gpt", "value": "she is putting on the bracelet for the young girl. "}]
00000006245.jpg
6245_evjvqa
[{"from": "human", "value": "what color is the young girl with the bracelet dressed in? \n<image>"}, {"from": "gpt", "value": "she is dressed in orange. "}]
00000006248.jpg
6248_evjvqa
[{"from": "human", "value": "how many living statues are there?\n<image>"}, {"from": "gpt", "value": "there is one living statue. "}]
00000006248.jpg
6248_evjvqa
[{"from": "human", "value": "how many people are wearing the santa claus costume?\n<image>"}, {"from": "gpt", "value": "there is one person wearing santa claus costume."}]
00000006248.jpg
6248_evjvqa
[{"from": "human", "value": "what is santa claus making?\n<image>"}, {"from": "gpt", "value": "he is making balloon animals."}]
00000006253.jpg
6253_evjvqa
[{"from": "human", "value": "what are the two young people doing?\n<image>"}, {"from": "gpt", "value": "they are holding hands and looking at each other. "}]
00000006253.jpg
6253_evjvqa
[{"from": "human", "value": "what color are their shirts?\n<image>"}, {"from": "gpt", "value": "their shirts are white."}]
00000006253.jpg
6253_evjvqa
[{"from": "human", "value": "what color are the boy's pants?\n<image>"}, {"from": "gpt", "value": "his pants are black."}]
00000006254.jpg
6254_evjvqa
[{"from": "human", "value": "what color is the female customer dressed in?\n<image>"}, {"from": "gpt", "value": "she is dressed in black."}]
00000006254.jpg
6254_evjvqa
[{"from": "human", "value": "what costume is the shopkeeper wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing cao bang province's traditional costume."}]
00000006255.jpg
6255_evjvqa
[{"from": "human", "value": "how many ships are sailing on the sea?\n<image>"}, {"from": "gpt", "value": "there are three ships. "}]
00000006255.jpg
6255_evjvqa
[{"from": "human", "value": "what is the dominant color of the ships?\n<image>"}, {"from": "gpt", "value": "it is white."}]
00000006255.jpg
6255_evjvqa
[{"from": "human", "value": "how many white ships are there on the sea?\n<image>"}, {"from": "gpt", "value": "there are three white ships on the sea. "}]
00000006257.jpg
6257_evjvqa
[{"from": "human", "value": "what color is the shirt worn by the girl standing next to the bicycle?\n<image>"}, {"from": "gpt", "value": "it is white."}]
00000006257.jpg
6257_evjvqa
[{"from": "human", "value": "how many men in white are walking on the street?\n<image>"}, {"from": "gpt", "value": "there is one man in white walking on the street."}]
00000006257.jpg
6257_evjvqa
[{"from": "human", "value": "what is the woman in pink doing?\n<image>"}, {"from": "gpt", "value": "she is walking along her child whom is on a small bicycle. "}]
00000006260.jpg
6260_evjvqa
[{"from": "human", "value": "how many motorbikes are parked in front of the store?\n<image>"}, {"from": "gpt", "value": "there is one motorbike parked in front of the store."}]
00000006262.jpg
6262_evjvqa
[{"from": "human", "value": "where is she?\n<image>"}, {"from": "gpt", "value": "she is walking on the road under the trees with yellow leaves."}]
00000006262.jpg
6262_evjvqa
[{"from": "human", "value": "what color are her shoes?\n<image>"}, {"from": "gpt", "value": "she is wearing a pair of black high heels."}]
00000006262.jpg
6262_evjvqa
[{"from": "human", "value": "what color is her scarf?\n<image>"}, {"from": "gpt", "value": "it is white."}]
00000006265.jpg
6265_evjvqa
[{"from": "human", "value": "what is he doing?\n<image>"}, {"from": "gpt", "value": "he is fixing stuffs."}]
00000006265.jpg
6265_evjvqa
[{"from": "human", "value": "what color is his shirt?\n<image>"}, {"from": "gpt", "value": "it is white."}]
00000006265.jpg
6265_evjvqa
[{"from": "human", "value": "how many men who are seated in the picture?\n<image>"}, {"from": "gpt", "value": "there is one man seated."}]
00000006273.jpg
6273_evjvqa
[{"from": "human", "value": "what color is the logo of the delivery company?\n<image>"}, {"from": "gpt", "value": "it is red."}]
00000006275.jpg
6275_evjvqa
[{"from": "human", "value": "what kind of hat is the woman in purple wearing?\n<image>"}, {"from": "gpt", "value": "she is wearing a conical hat."}]
00000006275.jpg
6275_evjvqa
[{"from": "human", "value": "what color is the woman with the conical hat's outfit?\n<image>"}, {"from": "gpt", "value": "it is purple."}]
00000006278.jpg
6278_evjvqa
[{"from": "human", "value": "where are they?\n<image>"}, {"from": "gpt", "value": "they are in a museum."}]
00000006278.jpg
6278_evjvqa
[{"from": "human", "value": "what is the woman in red doing?\n<image>"}, {"from": "gpt", "value": "she is sitting on a bench."}]
00000006278.jpg
6278_evjvqa
[{"from": "human", "value": "how many women in red that are sitting in the museum?\n<image>"}, {"from": "gpt", "value": "there is one woman in red sitting in the museum."}]
00000006280.jpg
6280_evjvqa
[{"from": "human", "value": "where is she sitting?\n<image>"}, {"from": "gpt", "value": "she is sitting on the bow of the boat."}]
00000006280.jpg
6280_evjvqa
[{"from": "human", "value": "what color is her top?\n<image>"}, {"from": "gpt", "value": "it is blue."}]
00000006280.jpg
6280_evjvqa
[{"from": "human", "value": "what color is her skirt?\n<image>"}, {"from": "gpt", "value": "it is pink."}]
00000006281.jpg
6281_evjvqa
[{"from": "human", "value": "what is she doing?\n<image>"}, {"from": "gpt", "value": "she is holding a bag of onions on her hands."}]
00000006281.jpg
6281_evjvqa
[{"from": "human", "value": "what color is her handbag?\n<image>"}, {"from": "gpt", "value": "her handbag is black."}]
00000006281.jpg
6281_evjvqa
[{"from": "human", "value": "what color is the boy' scarf?\n<image>"}, {"from": "gpt", "value": "it is red."}]
00000006283.jpg
6283_evjvqa
[{"from": "human", "value": "what character is she dressing as?\n<image>"}, {"from": "gpt", "value": "she is dressing as a witch with a blue cloak."}]
00000006283.jpg
6283_evjvqa
[{"from": "human", "value": "what is the man in white doing?\n<image>"}, {"from": "gpt", "value": "he is holding the green witch hat on his hand."}]
00000006283.jpg
6283_evjvqa
[{"from": "human", "value": "how many men in blue are there?\n<image>"}, {"from": "gpt", "value": "there is one man in blue."}]
00000006290.jpg
6290_evjvqa
[{"from": "human", "value": "what is she doing?\n<image>"}, {"from": "gpt", "value": "she is promoting products."}]
00000006290.jpg
6290_evjvqa
[{"from": "human", "value": "what is her facial expresison?\n<image>"}, {"from": "gpt", "value": "she is smiling."}]
00000006290.jpg
6290_evjvqa
[{"from": "human", "value": "what color is her shirt?\n<image>"}, {"from": "gpt", "value": "she is wearing a white shirt."}]
00000006293.jpg
6293_evjvqa
[{"from": "human", "value": "what color is the banner that they are hanging?\n<image>"}, {"from": "gpt", "value": "it is red."}]
00000006293.jpg
6293_evjvqa
[{"from": "human", "value": "how many men in red are there?\n<image>"}, {"from": "gpt", "value": "there is one man in red."}]
00000006297.jpg
6297_evjvqa
[{"from": "human", "value": "how many people are drinking at the tables?\n<image>"}, {"from": "gpt", "value": "there are five people drinking at the tables."}]
00000006297.jpg
6297_evjvqa
[{"from": "human", "value": "what is the young man in blue doing?\n<image>"}, {"from": "gpt", "value": "he is reading a book."}]
00000006297.jpg
6297_evjvqa
[{"from": "human", "value": "how many girls in red are there?\n<image>"}, {"from": "gpt", "value": "there is one girl in red."}]
00000006298.jpg
6298_evjvqa
[{"from": "human", "value": "what is he doing?\n<image>"}, {"from": "gpt", "value": "he is cleaning up."}]
00000006298.jpg
6298_evjvqa
[{"from": "human", "value": "what is the color of the jacket that he is wearing?\n<image>"}, {"from": "gpt", "value": "it is black."}]
00000006298.jpg
6298_evjvqa
[{"from": "human", "value": "what color is his jacket?\n<image>"}, {"from": "gpt", "value": "it is black. "}]
00000001502.jpg
1502_evjvqa
[{"from": "human", "value": "who is this girl standing next to?\n<image>"}, {"from": "gpt", "value": "nobody"}]
00000001502.jpg
1502_evjvqa
[{"from": "human", "value": "what is the girl standing against the wall wearing at her thigh level?\n<image>"}, {"from": "gpt", "value": "a handbag"}]
00000001502.jpg
1502_evjvqa
[{"from": "human", "value": "how is the girl standing against the wall dressed?\n<image>"}, {"from": "gpt", "value": "a light green shirt and a blue pair of jeans"}]