image
stringlengths
15
15
id
stringlengths
8
12
conversations
stringlengths
79
268
00000005534.jpg
5534_evjvqa
[{"from": "human", "value": "what color is the helmet of the man riding the yellow motorbike?\n<image>"}, {"from": "gpt", "value": "the helmet of the man is yellow"}]
00000005534.jpg
5534_evjvqa
[{"from": "human", "value": "what kind of hat is placed by the man inside the bicycle basket?\n<image>"}, {"from": "gpt", "value": "inside the bicycle basket is a pith helmet"}]
00000005537.jpg
5537_evjvqa
[{"from": "human", "value": "how many glasses are placed on the table?\n<image>"}, {"from": "gpt", "value": "two glasses are placed on the table"}]
00000005537.jpg
5537_evjvqa
[{"from": "human", "value": "how many bowls of rice are on the table?\n<image>"}, {"from": "gpt", "value": "one bowl of rice is on the table"}]
00000005537.jpg
5537_evjvqa
[{"from": "human", "value": "where is the sushi placed?\n<image>"}, {"from": "gpt", "value": "the sushi is placed in a large rectangular plate"}]
00000005540.jpg
5540_evjvqa
[{"from": "human", "value": "what is placed on the swing?\n<image>"}, {"from": "gpt", "value": "on the swing is a hat"}]
00000005541.jpg
5541_evjvqa
[{"from": "human", "value": "what posture is the man sitting?\n<image>"}, {"from": "gpt", "value": "the man is sitting with left leg resting on right leg"}]
00000005541.jpg
5541_evjvqa
[{"from": "human", "value": "what color is the shirt the man is wearing?\n<image>"}, {"from": "gpt", "value": "the man is wearing a white shirt"}]
00000005541.jpg
5541_evjvqa
[{"from": "human", "value": "what color is the shadow of the text on the photo?\n<image>"}, {"from": "gpt", "value": "the shadow of the text on the photo is red"}]
00000005542.jpg
5542_evjvqa
[{"from": "human", "value": "what is the man in the blue cap holding in his hand?\n<image>"}, {"from": "gpt", "value": "thê man in blue cap is holding a tablet"}]
00000005542.jpg
5542_evjvqa
[{"from": "human", "value": "what is in the arm of the child who is putting his arm around his friend?\n<image>"}, {"from": "gpt", "value": "the arm of the child has a watch"}]
00000005542.jpg
5542_evjvqa
[{"from": "human", "value": "what is the man presenting to the students by?\n<image>"}, {"from": "gpt", "value": "the man is presenting to the students by a micro"}]
00000005543.jpg
5543_evjvqa
[{"from": "human", "value": "what color shirt is the mc wearing?\n<image>"}, {"from": "gpt", "value": "the mc is wearing a black shirt"}]
00000005543.jpg
5543_evjvqa
[{"from": "human", "value": "how many female staff are in charge of holding the ribbon?\n<image>"}, {"from": "gpt", "value": "four female staff are in charge of holding the ribbon"}]
00000005545.jpg
5545_evjvqa
[{"from": "human", "value": "what is the man with the glasses doing?\n<image>"}, {"from": "gpt", "value": "the man with the glasses is reaching out to choose goods"}]
00000005545.jpg
5545_evjvqa
[{"from": "human", "value": "what color shirt is the woman standing next to the man with glasses wearing?\n<image>"}, {"from": "gpt", "value": "the woman standing next to the man with glasses is wearing a dark blue shirt"}]
00000005545.jpg
5545_evjvqa
[{"from": "human", "value": "what color shirt is the man wearing glasses wearing?\n<image>"}, {"from": "gpt", "value": "the man wearing glasses is wearing a blue shirt"}]
00000005545.jpg
5545_evjvqa
[{"from": "human", "value": "what is the counter that the woman and the man are choosing?\n<image>"}, {"from": "gpt", "value": "those two are choosing at the fruit counter"}]
00000005549.jpg
5549_evjvqa
[{"from": "human", "value": "what is the signboard of the beach resort made of?\n<image>"}, {"from": "gpt", "value": "the signboard of the beach resort is made of a rock"}]
00000005552.jpg
5552_evjvqa
[{"from": "human", "value": "what color shirt is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a blue shirt"}]
00000005552.jpg
5552_evjvqa
[{"from": "human", "value": "what is the girl holding in her hand?\n<image>"}, {"from": "gpt", "value": "the girl is holding a conical hat in her hand"}]
00000005552.jpg
5552_evjvqa
[{"from": "human", "value": "what kind of clothes is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a long sleeved outfit"}]
00000005554.jpg
5554_evjvqa
[{"from": "human", "value": "where is the woman in the red shirt sitting compared to the boy in the blue shirt?\n<image>"}, {"from": "gpt", "value": "the woman in the red shirt is sitting opposite the boy in the blue shirt"}]
00000005554.jpg
5554_evjvqa
[{"from": "human", "value": "what is the boy in the blue shirt doing?\n<image>"}, {"from": "gpt", "value": "the boy in the blue shirt is on tiptoe to drink"}]
00000005557.jpg
5557_evjvqa
[{"from": "human", "value": "what shape is the supermarket's chandelier designed with?\n<image>"}, {"from": "gpt", "value": "the supermarket's chandelier is designed with the same shape as a drop of water"}]
00000005557.jpg
5557_evjvqa
[{"from": "human", "value": "what is the animal standing in the middle of the supermarket lobby?\n<image>"}, {"from": "gpt", "value": "the dog is standing in the middle of the supermarket lobby"}]
00000005557.jpg
5557_evjvqa
[{"from": "human", "value": "what color is the supermarket's chandelier?\n<image>"}, {"from": "gpt", "value": "the supermarket's chandelier has two colors metallic gold and silver"}]
00000005560.jpg
5560_evjvqa
[{"from": "human", "value": "what hairstyle does the girl have?\n<image>"}, {"from": "gpt", "value": "the girl has short hairstyle"}]
00000005560.jpg
5560_evjvqa
[{"from": "human", "value": "what is the girl's face leaning on?\n<image>"}, {"from": "gpt", "value": "the girl's face is leaning on her right hand"}]
00000005560.jpg
5560_evjvqa
[{"from": "human", "value": "what color is the girl's hair?\n<image>"}, {"from": "gpt", "value": "the girl's hair is black"}]
00000005561.jpg
5561_evjvqa
[{"from": "human", "value": "where is the girl standing compared to the building?\n<image>"}, {"from": "gpt", "value": "the girl is standing in front of the stairs leading to the building"}]
00000005561.jpg
5561_evjvqa
[{"from": "human", "value": "what is the girl doing?\n<image>"}, {"from": "gpt", "value": "the girl is posing for a photo"}]
00000005561.jpg
5561_evjvqa
[{"from": "human", "value": "how many unopened doors does the building have?\n<image>"}, {"from": "gpt", "value": "one door is closed"}]
00000005563.jpg
5563_evjvqa
[{"from": "human", "value": "what is the girl in the orange shirt doing?\n<image>"}, {"from": "gpt", "value": "the girl in the orange shirt is introducing the product to the girl in the blue shirt"}]
00000005563.jpg
5563_evjvqa
[{"from": "human", "value": "who is introducing the product to the girl in the blue shirt?\n<image>"}, {"from": "gpt", "value": "the girl in the orange shirt is introducing the product to the girl in the blue shirt"}]
00000005564.jpg
5564_evjvqa
[{"from": "human", "value": "what outfit is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing an ao dai"}]
00000005564.jpg
5564_evjvqa
[{"from": "human", "value": "what is the boy wearing on his left hand?\n<image>"}, {"from": "gpt", "value": "the boy is wearing a watch on his left hand"}]
00000005564.jpg
5564_evjvqa
[{"from": "human", "value": "what pattern does the shirt the boy is wearing have on the front?\n<image>"}, {"from": "gpt", "value": "the shirt the boy is wearing has a tree branch pattern on the front"}]
00000005566.jpg
5566_evjvqa
[{"from": "human", "value": "how many trees are there in the painting?\n<image>"}, {"from": "gpt", "value": "there are three trees in the painting"}]
00000005566.jpg
5566_evjvqa
[{"from": "human", "value": "how many people are cycling in the painting?\n<image>"}, {"from": "gpt", "value": "one person is cycling"}]
00000005566.jpg
5566_evjvqa
[{"from": "human", "value": "what is the shape of the painting frame?\n<image>"}, {"from": "gpt", "value": "the painting frame has a rectangular shape"}]
00000005567.jpg
5567_evjvqa
[{"from": "human", "value": "what color shirt is the woman leading the little girl wearing?\n<image>"}, {"from": "gpt", "value": "the woman leading the little girl is wearing a purple shirt"}]
00000005567.jpg
5567_evjvqa
[{"from": "human", "value": "how many people are being consulted?\n<image>"}, {"from": "gpt", "value": "three people are being consulted by staff"}]
00000005567.jpg
5567_evjvqa
[{"from": "human", "value": "which side is the little girl standing compared to the woman in the purple shirt?\n<image>"}, {"from": "gpt", "value": "the little girl is standing on the left of the woman in the purple shirt"}]
00000005568.jpg
5568_evjvqa
[{"from": "human", "value": "what is the woman doing?\n<image>"}, {"from": "gpt", "value": "the woman is taking notes"}]
00000005568.jpg
5568_evjvqa
[{"from": "human", "value": "what color is the shirt the woman is wearing?\n<image>"}, {"from": "gpt", "value": "the woman is wearing a white shirt"}]
00000005568.jpg
5568_evjvqa
[{"from": "human", "value": "where is the woman standing?\n<image>"}, {"from": "gpt", "value": "the woman is standing in front of a food counter"}]
00000005569.jpg
5569_evjvqa
[{"from": "human", "value": "what is the girl doing?\n<image>"}, {"from": "gpt", "value": "the girl is drinking a water bottle"}]
00000005569.jpg
5569_evjvqa
[{"from": "human", "value": "what is the symbol on the left sleeve of the girl's shirt?\n<image>"}, {"from": "gpt", "value": "the left sleeve of the girl's shirt has a fish bone symbol"}]
00000005569.jpg
5569_evjvqa
[{"from": "human", "value": "which hand is the girl holding the water bottle with?\n<image>"}, {"from": "gpt", "value": "the girl is holding the water bottle with her left hand"}]
00000005571.jpg
5571_evjvqa
[{"from": "human", "value": "where is the man standing?\n<image>"}, {"from": "gpt", "value": "the man is standing in front of a picture"}]
00000005571.jpg
5571_evjvqa
[{"from": "human", "value": "what color is the shirt the man is wearing?\n<image>"}, {"from": "gpt", "value": "the shirt the man is wearing is black"}]
00000005571.jpg
5571_evjvqa
[{"from": "human", "value": "how many people appear in the picture behind the man?\n<image>"}, {"from": "gpt", "value": "one person is in the picture behind the man"}]
00000005572.jpg
5572_evjvqa
[{"from": "human", "value": "how many people are using the escalator?\n<image>"}, {"from": "gpt", "value": "two people are using the escalator"}]
00000005572.jpg
5572_evjvqa
[{"from": "human", "value": "what is the woman in the red shirt standing on the escalator holding?\n<image>"}, {"from": "gpt", "value": "the woman in the red shirt standing on the escalator is holding a drink cup"}]
00000005572.jpg
5572_evjvqa
[{"from": "human", "value": "which way are the two people on the escalator going?\n<image>"}, {"from": "gpt", "value": "the two people on the escalator are going upstairs"}]
00000005573.jpg
5573_evjvqa
[{"from": "human", "value": "how many santa claus keychains are there?\n<image>"}, {"from": "gpt", "value": "there is a santa claus keychain"}]
00000005573.jpg
5573_evjvqa
[{"from": "human", "value": "what color is the glove keychain?\n<image>"}, {"from": "gpt", "value": "the glove keychain is red"}]
00000005573.jpg
5573_evjvqa
[{"from": "human", "value": "how many stars are on the pine tree keychain?\n<image>"}, {"from": "gpt", "value": "there are four stars on the pine tree keychain"}]
00000005575.jpg
5575_evjvqa
[{"from": "human", "value": "how many entrances does the temple have?\n<image>"}, {"from": "gpt", "value": "the temple has three entrances"}]
00000005575.jpg
5575_evjvqa
[{"from": "human", "value": "how many human statues are placed in front of the temple gate?\n<image>"}, {"from": "gpt", "value": "two statues are placed in front of the temple gate"}]
00000005576.jpg
5576_evjvqa
[{"from": "human", "value": "what does the painting depict?\n<image>"}, {"from": "gpt", "value": "the painting depicts the scene of a road fork"}]
00000005580.jpg
5580_evjvqa
[{"from": "human", "value": "what is the girl wearing the mask doing?\n<image>"}, {"from": "gpt", "value": "the girl is standing and manipulating an electronic device"}]
00000005581.jpg
5581_evjvqa
[{"from": "human", "value": "which hand is the boy putting behind?\n<image>"}, {"from": "gpt", "value": "the boy is putting his left arm behind"}]
00000005581.jpg
5581_evjvqa
[{"from": "human", "value": "what color is the backpack the boy is carrying?\n<image>"}, {"from": "gpt", "value": "the backpack the boy is carrying is blue"}]
00000005581.jpg
5581_evjvqa
[{"from": "human", "value": "where is the boy in the orange shirt standing?\n<image>"}, {"from": "gpt", "value": "the boy in the orange shirt is standing in front of a temple gate"}]
00000005584.jpg
5584_evjvqa
[{"from": "human", "value": "what is the policeman doing?\n<image>"}, {"from": "gpt", "value": "the policeman is helping with the box"}]
00000005584.jpg
5584_evjvqa
[{"from": "human", "value": "what color shirt is the person the police officer helping wearing?\n<image>"}, {"from": "gpt", "value": "the helped man is wearing a blue shirt"}]
00000005587.jpg
5587_evjvqa
[{"from": "human", "value": "what is the girl in the black shirt doing?\n<image>"}, {"from": "gpt", "value": "the girl in the black shirt is using a mobile phone"}]
00000005588.jpg
5588_evjvqa
[{"from": "human", "value": "what stall is the girl standing next to?\n<image>"}, {"from": "gpt", "value": "the girl is standing next to the rice stall"}]
00000005588.jpg
5588_evjvqa
[{"from": "human", "value": "what color shirt is the girl wearing?\n<image>"}, {"from": "gpt", "value": "the girl is wearing a green shirt"}]
00000005589.jpg
5589_evjvqa
[{"from": "human", "value": "what booth are people standing in front of?\n<image>"}, {"from": "gpt", "value": "people are standing in front of the book booth"}]
00000005589.jpg
5589_evjvqa
[{"from": "human", "value": "what is the man in the orange shirt doing?\n<image>"}, {"from": "gpt", "value": "the man in the orange shirt is introducing books to everyone"}]
00000005589.jpg
5589_evjvqa
[{"from": "human", "value": "how many books is the man in the white shirt on the left the man in the orange shirt is holding?\n<image>"}, {"from": "gpt", "value": "the man on the left the man in the orange shirt is holding one book"}]
00000005594.jpg
5594_evjvqa
[{"from": "human", "value": "what is the main background color of the billboard?\n<image>"}, {"from": "gpt", "value": "the main background color of the billboard is red"}]
00000005597.jpg
5597_evjvqa
[{"from": "human", "value": "how many people appear in the photo?\n<image>"}, {"from": "gpt", "value": "two people appear in the photo"}]
00000001401.jpg
1401_evjvqa
[{"from": "human", "value": "where is the woman with the red scarf sitting?\n<image>"}, {"from": "gpt", "value": "she is sitting at the vegetable stall "}]
00000001401.jpg
1401_evjvqa
[{"from": "human", "value": "what is the woman wearing the red scarf doing?\n<image>"}, {"from": "gpt", "value": "selling vegetables"}]
00000001401.jpg
1401_evjvqa
[{"from": "human", "value": "how many people come to buy vegetables where the woman with the red scarf is?\n<image>"}, {"from": "gpt", "value": "nobody"}]
00000001401.jpg
1401_evjvqa
[{"from": "human", "value": "what items does the woman wearing a red scarf sell?\n<image>"}, {"from": "gpt", "value": "raw vegetables"}]
00000001402.jpg
1402_evjvqa
[{"from": "human", "value": "how many motorbikes are being built in front of the houses?\n<image>"}, {"from": "gpt", "value": "four"}]
00000001402.jpg
1402_evjvqa
[{"from": "human", "value": "how many women are standing at the end of the road?\n<image>"}, {"from": "gpt", "value": "two"}]
00000001402.jpg
1402_evjvqa
[{"from": "human", "value": "how many children are playing at the end of the road?\n<image>"}, {"from": "gpt", "value": "only one"}]
00000001403.jpg
1403_evjvqa
[{"from": "human", "value": "where is the group of people going to visit?\n<image>"}, {"from": "gpt", "value": "the museum"}]
00000001403.jpg
1403_evjvqa
[{"from": "human", "value": "which side of the stairs is more crowded?\n<image>"}, {"from": "gpt", "value": "on the right"}]
00000001403.jpg
1403_evjvqa
[{"from": "human", "value": "which side of the stairs is more sparsely?\n<image>"}, {"from": "gpt", "value": "on the left"}]
00000001405.jpg
1405_evjvqa
[{"from": "human", "value": "how many people are holding the camera?\n<image>"}, {"from": "gpt", "value": "two"}]
00000001405.jpg
1405_evjvqa
[{"from": "human", "value": "what is the girt in green doing?\n<image>"}, {"from": "gpt", "value": "photographing"}]
00000001405.jpg
1405_evjvqa
[{"from": "human", "value": "what is the boy in orange doing?\n<image>"}, {"from": "gpt", "value": "taking a photo"}]
00000001405.jpg
1405_evjvqa
[{"from": "human", "value": "what color shirt is the man holding the camera wearing?\n<image>"}, {"from": "gpt", "value": "orange"}]
00000001405.jpg
1405_evjvqa
[{"from": "human", "value": "what color shirt is the girl holding the camera wearing?\n<image>"}, {"from": "gpt", "value": "green"}]
00000001405.jpg
1405_evjvqa
[{"from": "human", "value": "where are the visitors going to visit?\n<image>"}, {"from": "gpt", "value": "the gallery of photos"}]
00000001405.jpg
1405_evjvqa
[{"from": "human", "value": "how many people did the male friend in the orange shirt take pictures with?\n<image>"}, {"from": "gpt", "value": "only one"}]
00000001405.jpg
1405_evjvqa
[{"from": "human", "value": "where is the girl in the green shirt taking pictures?\n<image>"}, {"from": "gpt", "value": "the gallery of photos"}]
00000001405.jpg
1405_evjvqa
[{"from": "human", "value": "how many people are standing in the photo gallery?\n<image>"}, {"from": "gpt", "value": "two"}]
00000001407.jpg
1407_evjvqa
[{"from": "human", "value": "who are drawed in the poster?\n<image>"}, {"from": "gpt", "value": "the woman, the childr and the medical staff"}]
00000001407.jpg
1407_evjvqa
[{"from": "human", "value": "what is the boy in the poster doing?\n<image>"}, {"from": "gpt", "value": "he is washing his hands"}]
00000001407.jpg
1407_evjvqa
[{"from": "human", "value": "what is the woman in the poster doing?\n<image>"}, {"from": "gpt", "value": "she is wearing a medical mask"}]
00000001408.jpg
1408_evjvqa
[{"from": "human", "value": "what is the woman wearing glasses doing?\n<image>"}, {"from": "gpt", "value": "buying filled fish"}]
00000001408.jpg
1408_evjvqa
[{"from": "human", "value": "how many filled fishes is the woman wearing glasses holding in her hand?\n<image>"}, {"from": "gpt", "value": "four"}]