image
stringlengths 15
15
| id
stringlengths 8
12
| conversations
stringlengths 79
268
|
---|---|---|
00000003943.jpg | 3943_evjvqa | [{"from": "human", "value": "what is the woman in black shirt wearing on her waist?\n<image>"}, {"from": "gpt", "value": "she is wearing a handbag"}] |
00000003943.jpg | 3943_evjvqa | [{"from": "human", "value": "who stands on the right hand side of the woman in black shirt?\n<image>"}, {"from": "gpt", "value": "the man in white t-shirt stands on the right hand side of the woman in black shirt"}] |
00000003943.jpg | 3943_evjvqa | [{"from": "human", "value": "what are the man in white t-shirt and the woman next to him doing?\n<image>"}, {"from": "gpt", "value": "they are looking at the items on the shelves"}] |
00000003945.jpg | 3945_evjvqa | [{"from": "human", "value": "what is the woman in the pink coat doing?\n<image>"}, {"from": "gpt", "value": "the woman in the pink coat is calculating money by calculator"}] |
00000003945.jpg | 3945_evjvqa | [{"from": "human", "value": "what device is the woman in pink coat holding in her hand?\n<image>"}, {"from": "gpt", "value": "she is holding a calculator"}] |
00000003945.jpg | 3945_evjvqa | [{"from": "human", "value": "who is holding the calculator to calculate?\n<image>"}, {"from": "gpt", "value": "the woman in pink coat"}] |
00000003948.jpg | 3948_evjvqa | [{"from": "human", "value": "what pose is the man in the lime green shirt in?\n<image>"}, {"from": "gpt", "value": "he is in the sitting pose"}] |
00000003948.jpg | 3948_evjvqa | [{"from": "human", "value": "how many people wear green military uniforms?\n<image>"}, {"from": "gpt", "value": "two people wear green military uniforms"}] |
00000003948.jpg | 3948_evjvqa | [{"from": "human", "value": "what are these people doing?\n<image>"}, {"from": "gpt", "value": "they are raising the flag"}] |
00000003953.jpg | 3953_evjvqa | [{"from": "human", "value": "what does the man wear?\n<image>"}, {"from": "gpt", "value": "he wears a santa claus costume"}] |
00000003953.jpg | 3953_evjvqa | [{"from": "human", "value": "what store is the man in the santa claus costume standing at?\n<image>"}, {"from": "gpt", "value": "he is standing in the toy store"}] |
00000003953.jpg | 3953_evjvqa | [{"from": "human", "value": "what color is the backpack that the person standing near the green plastic chair wears?\n<image>"}, {"from": "gpt", "value": "that person wears a black backpack"}] |
00000003955.jpg | 3955_evjvqa | [{"from": "human", "value": "what vehicle is the man in purple shirt sitting on?\n<image>"}, {"from": "gpt", "value": "he is sitting on a motorbike"}] |
00000003957.jpg | 3957_evjvqa | [{"from": "human", "value": "what color is the outer surface of this ship painted?\n<image>"}, {"from": "gpt", "value": "the outer surface of this ship is painted white and orange"}] |
00000003957.jpg | 3957_evjvqa | [{"from": "human", "value": "what two colors do people paint on the outer surface of this ship?\n<image>"}, {"from": "gpt", "value": "orange and white"}] |
00000003957.jpg | 3957_evjvqa | [{"from": "human", "value": "what means of transportation are behind these people?\n<image>"}, {"from": "gpt", "value": "a ship"}] |
00000003960.jpg | 3960_evjvqa | [{"from": "human", "value": "what are piled up on the ground next to the person in red shirt?\n<image>"}, {"from": "gpt", "value": "the black baskets"}] |
00000003960.jpg | 3960_evjvqa | [{"from": "human", "value": "what is the person in the purple floral dress pushing?\n<image>"}, {"from": "gpt", "value": "that person is pushing a supermarket cart"}] |
00000003960.jpg | 3960_evjvqa | [{"from": "human", "value": "what is the person in the purple floral dress doing?\n<image>"}, {"from": "gpt", "value": "that person is going shopping"}] |
00000003961.jpg | 3961_evjvqa | [{"from": "human", "value": "what activity are these people participating in?\n<image>"}, {"from": "gpt", "value": "they are cleaning the beach"}] |
00000003961.jpg | 3961_evjvqa | [{"from": "human", "value": "what do some people carry sacks for?\n<image>"}, {"from": "gpt", "value": "to contain garbage collected on the beach"}] |
00000003961.jpg | 3961_evjvqa | [{"from": "human", "value": "whom is the man in gray shirt and black trousers carrying the sack with?\n<image>"}, {"from": "gpt", "value": "with a man in a military uniform wearing a white cap"}] |
00000003963.jpg | 3963_evjvqa | [{"from": "human", "value": "what is the woman in blue shirt doing?\n<image>"}, {"from": "gpt", "value": "she is taking inventory of the goods"}] |
00000003963.jpg | 3963_evjvqa | [{"from": "human", "value": "how many people are standing at this supermarket?\n<image>"}, {"from": "gpt", "value": "one person is standing at this supermarket"}] |
00000003963.jpg | 3963_evjvqa | [{"from": "human", "value": "how does the only person standing next to the shelf dress?\n<image>"}, {"from": "gpt", "value": "that person wears blue t-shirt and black trousers"}] |
00000003964.jpg | 3964_evjvqa | [{"from": "human", "value": "what is the woman in red apron doing?\n<image>"}, {"from": "gpt", "value": "she is handing the bag of bread to the woman wearing the handbag"}] |
00000003964.jpg | 3964_evjvqa | [{"from": "human", "value": "whom is the bakery employee giving the bag of bread to?\n<image>"}, {"from": "gpt", "value": "the bakery employee is giving the bag of bread to the woman wearing the handbag"}] |
00000003964.jpg | 3964_evjvqa | [{"from": "human", "value": "who is the employee of this bakery?\n<image>"}, {"from": "gpt", "value": "the woman wearing red apron"}] |
00000003970.jpg | 3970_evjvqa | [{"from": "human", "value": "how many bowls are there on the table?\n<image>"}, {"from": "gpt", "value": "there are three bowls on the table"}] |
00000003970.jpg | 3970_evjvqa | [{"from": "human", "value": "how many plates of dry pancakes are there on the table?\n<image>"}, {"from": "gpt", "value": "there are two plates of dry pancakes on the table"}] |
00000003970.jpg | 3970_evjvqa | [{"from": "human", "value": "what items are dry pancakes displayed on?\n<image>"}, {"from": "gpt", "value": "the plastic plates"}] |
00000003972.jpg | 3972_evjvqa | [{"from": "human", "value": "how many boats are there on the river?\n<image>"}, {"from": "gpt", "value": "there are two boats on the river"}] |
00000003972.jpg | 3972_evjvqa | [{"from": "human", "value": "how many people sit on each boat?\n<image>"}, {"from": "gpt", "value": "one people sits on each boat"}] |
00000003972.jpg | 3972_evjvqa | [{"from": "human", "value": "what is the weather like in this area?\n<image>"}, {"from": "gpt", "value": "it is sunny in this area"}] |
00000003973.jpg | 3973_evjvqa | [{"from": "human", "value": "what is the woman sitting on a chair holding?\n<image>"}, {"from": "gpt", "value": "she is holding a hand fan"}] |
00000003981.jpg | 3981_evjvqa | [{"from": "human", "value": "what vehicles are parked on the pavement?\n<image>"}, {"from": "gpt", "value": "the cars are parked on the pavement"}] |
00000003981.jpg | 3981_evjvqa | [{"from": "human", "value": "what is the weather like in this quarter of the town?\n<image>"}, {"from": "gpt", "value": "it is sunny in this quarter of the town"}] |
00000003981.jpg | 3981_evjvqa | [{"from": "human", "value": "what is the scooter rider in white shirt wearing on head?\n<image>"}, {"from": "gpt", "value": "a helmet"}] |
00000003985.jpg | 3985_evjvqa | [{"from": "human", "value": "where is the boy holding the balloon standing?\n<image>"}, {"from": "gpt", "value": "that boy is standing behind the man in black shirt wearing a black helmet"}] |
00000003985.jpg | 3985_evjvqa | [{"from": "human", "value": "what is floating above the head of the man wearing a black helmet with a white face mask?\n<image>"}, {"from": "gpt", "value": "a balloon"}] |
00000003985.jpg | 3985_evjvqa | [{"from": "human", "value": "whom is the woman in brown t-shirt carrying?\n<image>"}, {"from": "gpt", "value": "she is carrying a baby"}] |
00000003986.jpg | 3986_evjvqa | [{"from": "human", "value": "when are these people eating and drinking?\n<image>"}, {"from": "gpt", "value": "they are eating and drinking in the evening"}] |
00000003987.jpg | 3987_evjvqa | [{"from": "human", "value": "what is the woman wearing black shirt and white face mask doing?\n<image>"}, {"from": "gpt", "value": "she is selecting apples"}] |
00000003994.jpg | 3994_evjvqa | [{"from": "human", "value": "how many customers sit in the restaurant?\n<image>"}, {"from": "gpt", "value": "no customer sits in the restaurant"}] |
00000003995.jpg | 3995_evjvqa | [{"from": "human", "value": "what hat is the man on the motorbike wearing?\n<image>"}, {"from": "gpt", "value": "he is wearing a helmet"}] |
00000003995.jpg | 3995_evjvqa | [{"from": "human", "value": "what is the man in the red helmet doing?\n<image>"}, {"from": "gpt", "value": "he is driving a motorbike"}] |
00000003995.jpg | 3995_evjvqa | [{"from": "human", "value": "what is the man in the blue long-sleeve shirt doing?\n<image>"}, {"from": "gpt", "value": "the man in the blue long-sleeve shirt is driving a motorbike"}] |
00000003999.jpg | 3999_evjvqa | [{"from": "human", "value": "what are the students holding brooms for?\n<image>"}, {"from": "gpt", "value": "sweeping leaves on the yard"}] |
00000003999.jpg | 3999_evjvqa | [{"from": "human", "value": "what does the student wearing a hat use to sweep the yard?\n<image>"}, {"from": "gpt", "value": "that student uses a broom"}] |
00000005500.jpg | 5500_evjvqa | [{"from": "human", "value": "what is the girl in the black shirt holding in her right hand?\n<image>"}, {"from": "gpt", "value": "the girl in the black shirt is holding a cucumber"}] |
00000005500.jpg | 5500_evjvqa | [{"from": "human", "value": "what area is the woman in the red shirt standing in?\n<image>"}, {"from": "gpt", "value": "the woman in the red shirt is standing in the weighing area"}] |
00000005500.jpg | 5500_evjvqa | [{"from": "human", "value": "what color shirt does the staff in charge of weighing items wear?\n<image>"}, {"from": "gpt", "value": "the staff in charge of weighing items wears yellow shirt"}] |
00000005500.jpg | 5500_evjvqa | [{"from": "human", "value": "what color is the bag of the girl in the black shirt?\n<image>"}, {"from": "gpt", "value": "the bag is white"}] |
00000005501.jpg | 5501_evjvqa | [{"from": "human", "value": "how many pairs of chopsticks are on the table?\n<image>"}, {"from": "gpt", "value": "one pair of chopsticks is on the table"}] |
00000005501.jpg | 5501_evjvqa | [{"from": "human", "value": "what is the dish inside the porcelain bowl?\n<image>"}, {"from": "gpt", "value": "inside the porcelain bowl is the beef pho dish"}] |
00000005501.jpg | 5501_evjvqa | [{"from": "human", "value": "where are the bean sprouts put on the table?\n<image>"}, {"from": "gpt", "value": "bean sprouts are contained inside a stainless steel bowl"}] |
00000005502.jpg | 5502_evjvqa | [{"from": "human", "value": "how many people are walking in all?\n<image>"}, {"from": "gpt", "value": "three people are walking"}] |
00000005502.jpg | 5502_evjvqa | [{"from": "human", "value": "what is the object placed leaning against the fence?\n<image>"}, {"from": "gpt", "value": "the motorbike is placed leaning against the fence"}] |
00000005502.jpg | 5502_evjvqa | [{"from": "human", "value": "what color shirt are the two men walking together wearing?\n<image>"}, {"from": "gpt", "value": "both of them are wearing white shirts"}] |
00000005507.jpg | 5507_evjvqa | [{"from": "human", "value": "what color is the roof of the city gate?\n<image>"}, {"from": "gpt", "value": "the roof is orange"}] |
00000005507.jpg | 5507_evjvqa | [{"from": "human", "value": "where are the flagpoles placed?\n<image>"}, {"from": "gpt", "value": "the flagpoles are erected in front of the city gate"}] |
00000005508.jpg | 5508_evjvqa | [{"from": "human", "value": "what color is the ship?\n<image>"}, {"from": "gpt", "value": "the ship is white"}] |
00000005511.jpg | 5511_evjvqa | [{"from": "human", "value": "what is mounted on the sides of the tower?\n<image>"}, {"from": "gpt", "value": "the clock faces are mounted on the sides of the tower"}] |
00000005511.jpg | 5511_evjvqa | [{"from": "human", "value": "what color is the roof of the market?\n<image>"}, {"from": "gpt", "value": "the roof is red"}] |
00000005511.jpg | 5511_evjvqa | [{"from": "human", "value": "how many sides does the clock tower of the market have?\n<image>"}, {"from": "gpt", "value": "the tower has four sides"}] |
00000005512.jpg | 5512_evjvqa | [{"from": "human", "value": "what kind of flag is hung to the left of the entrance?\n<image>"}, {"from": "gpt", "value": "the festival flag is hung to the left of the entrance"}] |
00000005512.jpg | 5512_evjvqa | [{"from": "human", "value": "where is the trash can placed?\n<image>"}, {"from": "gpt", "value": "the trash can is placed below the festival flag"}] |
00000005512.jpg | 5512_evjvqa | [{"from": "human", "value": "what are the two sides of the leading stair decorated with?\n<image>"}, {"from": "gpt", "value": "two flower pots are placed on two side of the stair"}] |
00000005514.jpg | 5514_evjvqa | [{"from": "human", "value": "how many mobile policemen are standing in front of the store?\n<image>"}, {"from": "gpt", "value": "three mobile policemen are standing in front of the store"}] |
00000005514.jpg | 5514_evjvqa | [{"from": "human", "value": "what color is the store name board?\n<image>"}, {"from": "gpt", "value": "the store name board is blue"}] |
00000005514.jpg | 5514_evjvqa | [{"from": "human", "value": "how many mobile policemen do not wear masks?\n<image>"}, {"from": "gpt", "value": "one person do not wear a mask"}] |
00000005515.jpg | 5515_evjvqa | [{"from": "human", "value": "what color hat is the woman in the brown jacket wearing?\n<image>"}, {"from": "gpt", "value": "the woman is wearing a white hat"}] |
00000005515.jpg | 5515_evjvqa | [{"from": "human", "value": "how many baskets of carrots are on display?\n<image>"}, {"from": "gpt", "value": "two baskets of carrots"}] |
00000005515.jpg | 5515_evjvqa | [{"from": "human", "value": "which side are the two talking women standing compared to the woman in the white hat?\n<image>"}, {"from": "gpt", "value": "the two women are standing on the left of the woman in the red hat"}] |
00000005517.jpg | 5517_evjvqa | [{"from": "human", "value": "how many people are smiling?\n<image>"}, {"from": "gpt", "value": "two people are smiling"}] |
00000005517.jpg | 5517_evjvqa | [{"from": "human", "value": "what is the store selling?\n<image>"}, {"from": "gpt", "value": "the store is selling newspapers"}] |
00000005518.jpg | 5518_evjvqa | [{"from": "human", "value": "what is the man in the blue shirt doing?\n<image>"}, {"from": "gpt", "value": "the man is leading the motorbike"}] |
00000005519.jpg | 5519_evjvqa | [{"from": "human", "value": "what are the two kids doing?\n<image>"}, {"from": "gpt", "value": "watching the elephants"}] |
00000005519.jpg | 5519_evjvqa | [{"from": "human", "value": "what color hat is the boy wearing?\n<image>"}, {"from": "gpt", "value": "the boy is wearing a blue hat"}] |
00000005519.jpg | 5519_evjvqa | [{"from": "human", "value": "what kind of clothes is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a dress"}] |
00000005522.jpg | 5522_evjvqa | [{"from": "human", "value": "what is the food put in?\n<image>"}, {"from": "gpt", "value": "the foot is put in a porcelain bowl"}] |
00000005522.jpg | 5522_evjvqa | [{"from": "human", "value": "what item is placed inside the food bowl?\n<image>"}, {"from": "gpt", "value": "a spoon is placed inside the food bowl"}] |
00000005522.jpg | 5522_evjvqa | [{"from": "human", "value": "what is the food in the bowl?\n<image>"}, {"from": "gpt", "value": "in the bowl is pho"}] |
00000005523.jpg | 5523_evjvqa | [{"from": "human", "value": "what expression does the man have?\n<image>"}, {"from": "gpt", "value": "the man is smiling"}] |
00000005523.jpg | 5523_evjvqa | [{"from": "human", "value": "which hand is the man putting on his chest?\n<image>"}, {"from": "gpt", "value": "the man is putting his left hand on his chest"}] |
00000005523.jpg | 5523_evjvqa | [{"from": "human", "value": "what items does the store sell?\n<image>"}, {"from": "gpt", "value": "the store sells food and drinks"}] |
00000005523.jpg | 5523_evjvqa | [{"from": "human", "value": "what is the item in front of the man?\n<image>"}, {"from": "gpt", "value": "eggs trays are placed in front of the man"}] |
00000005525.jpg | 5525_evjvqa | [{"from": "human", "value": "where is the bike parked?\n<image>"}, {"from": "gpt", "value": "the bike is parked right in front of the pharmacy"}] |
00000005525.jpg | 5525_evjvqa | [{"from": "human", "value": "what color shirt is the girl standing in front of the pharmacy wearing?\n<image>"}, {"from": "gpt", "value": "the girl standing in front of the pharmacy is wearing a black shirt"}] |
00000005527.jpg | 5527_evjvqa | [{"from": "human", "value": "what is the man in the white shirt doing?\n<image>"}, {"from": "gpt", "value": "the man is walking"}] |
00000005527.jpg | 5527_evjvqa | [{"from": "human", "value": "what color are the benches painted?\n<image>"}, {"from": "gpt", "value": "the benches are painted black"}] |
00000005527.jpg | 5527_evjvqa | [{"from": "human", "value": "what is the main color of the exercise equipment painted?\n<image>"}, {"from": "gpt", "value": "the main color painted for the exercise equipment is orange"}] |
00000005528.jpg | 5528_evjvqa | [{"from": "human", "value": "what is the girl holding in her hand?\n<image>"}, {"from": "gpt", "value": "the girl is holding a milk pack and a shopping cart"}] |
00000005528.jpg | 5528_evjvqa | [{"from": "human", "value": "what counter is the girl standing at?\n<image>"}, {"from": "gpt", "value": "the girl is standing at the milk products counter"}] |
00000005528.jpg | 5528_evjvqa | [{"from": "human", "value": "what color is the shopping cart the girl is holding?\n<image>"}, {"from": "gpt", "value": "the shopping cart the girl is holding is blue"}] |
00000005532.jpg | 5532_evjvqa | [{"from": "human", "value": "what is the man in the red jacket doing?\n<image>"}, {"from": "gpt", "value": "the man in the red jacket is about to light a cigarette"}] |
00000005533.jpg | 5533_evjvqa | [{"from": "human", "value": "what is the man in the black and yellow shirt holding in his hands?\n<image>"}, {"from": "gpt", "value": "the man in the black and yellow shirt is holding a pair of pants in his hands"}] |
00000005533.jpg | 5533_evjvqa | [{"from": "human", "value": "what color dress is the woman reaching out to get the clothes wearing?\n<image>"}, {"from": "gpt", "value": "the woman reaching out to get the clothes is wearing the outfit with black and white stripe pattern"}] |
00000005533.jpg | 5533_evjvqa | [{"from": "human", "value": "what is the woman in black doing?\n<image>"}, {"from": "gpt", "value": "the woman in black is about to take something out of her bag"}] |
00000005534.jpg | 5534_evjvqa | [{"from": "human", "value": "what is on the back of the man's bicycle?\n<image>"}, {"from": "gpt", "value": "bundles of vegetables are being placed on the back of the man's bicycle"}] |